Jie Zhang

ORCID: 0009-0004-8250-2000
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Advanced Graph Neural Networks
  • Complex Network Analysis Techniques
  • Topic Modeling
  • Graph Theory and Algorithms
  • Domain Adaptation and Few-Shot Learning
  • Recommender Systems and Techniques
  • Traffic Prediction and Management Techniques
  • Semantic Web and Ontologies
  • Bayesian Modeling and Causal Inference
  • Mental Health via Writing
  • Bioinformatics and Genomic Networks
  • Machine Learning and ELM
  • Human Mobility and Location-Based Analysis
  • Spam and Phishing Detection
  • Epigenetics and DNA Methylation
  • Data Stream Mining Techniques
  • Adversarial Robustness in Machine Learning
  • Machine Learning in Materials Science
  • Ferroelectric and Negative Capacitance Devices
  • Functional Brain Connectivity Studies
  • Network Security and Intrusion Detection
  • Caching and Content Delivery
  • Identification and Quantification in Food
  • Natural Language Processing Techniques
  • Geographic Information Systems Studies

Alibaba Group (China)
2025

East China University of Science and Technology
2024

Nanyang Technological University
2018-2023

Zhengzhou University
2022

Shanghai Center for Brain Science and Brain-Inspired Technology
2022

Fudan University
2020-2022

National Central University
2021-2022

The University of Sydney
2022

Cloud Computing Center
2022

Space Engineering University
2021

Recent advances in network embedding has revolutionized the field of graph and mining. However, (pre-)training embeddings for very large-scale networks is computationally challenging most existing methods. In this work, we present ProNE---a fast, scalable, effective model, whose single-thread version 10--400x faster than efficient benchmarks with 20 threads, including LINE, DeepWalk, node2vec, GraRep, HOPE. As a concrete example, single-version ProNE requires only 29 hours to embed hundreds...

10.24963/ijcai.2019/594 article EN 2019-07-28

We investigate how generative adversarial nets (GANs) can help semi-supervised learning on graphs. first provide insights working principles of over graphs and then present GraphSGAN, a novel approach to In generator classifier networks play competitive game. At equilibrium, generates fake samples in low-density areas between subgraphs. order discriminate from the real, implicitly takes density property subgraph into consideration. An efficient algorithm has been developed improve...

10.1145/3269206.3271768 preprint EN 2018-10-17

Linking entities from different sources is a fundamental task in building open knowledge graphs. Despite much research conducted related fields, the challenges of linkinglarge-scale heterogeneous entity graphs are far resolved. Employing two billion-scale academic (Microsoft Academic Graph and AMiner) as for our study, we propose unified framework --- LinKG to address problem large-scale linked graph. coupled with three linking modules, each which addresses one category entities. To link...

10.1145/3292500.3330785 article EN 2019-07-25

We study the problem of semi-supervised learning on graphs, for which graph neural networks (GNNs) have been extensively explored. However, most existing GNNs inherently suffer from limitations over-smoothing, non-robustness, and weak-generalization when labeled nodes are scarce. In this paper, we propose a simple yet effective framework -- GRAPH RANDOM NEURAL NETWORKS (GRAND) to address these issues. GRAND, first design random propagation strategy perform data augmentation. Then leverage...

10.48550/arxiv.2005.11079 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Existing network embedding algorithms based on generative adversarial networks (GANs) improve the robustness of node embeddings by selecting high-quality negative samples with generator to play against discriminator. Since most can be easily discriminated from positive in graphs, their poor competitiveness weakens function generator. Inspired sales skills market, this article, we present tripartite training for (TriATNE), a novel learning framework stable and robust embeddings. TriATNE...

10.1109/tcyb.2021.3061771 article EN IEEE Transactions on Cybernetics 2021-03-17

Link prediction in signed social networks is challenging because of the existence and imbalance three kinds status (positive, negative no-relation). Furthermore, there are a variety types no-relation reality, e.g., strangers frenemies, which cannot be well distinguished from other linked by existing approaches. In this paper, we propose novel Framework Integrating both Latent Explicit features (FILE), to better deal with improve overall link performance networks. particular, design two...

10.1609/aaai.v32i1.11262 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2018-04-25

Graph data mining has largely benefited from the recent developments of graph representation learning. Most attempts to improve representations have thus far focused on designing new network embedding or neural (GNN) architectures. Inspired by SGC and ProNE models, we instead focus enhancing any existing learned further smoothing them via filters. In this paper, introduce an automated framework AutoProNE achieve this. Specifically, automatically searches for a unique optimal set filters...

10.1109/tkde.2021.3115017 article EN publisher-specific-oa IEEE Transactions on Knowledge and Data Engineering 2021-01-01

Past research has examined how well these models grasp code syntax, yet their understanding of semantics still needs to be explored. We extensively analyze seven investigate represent syntax and semantics. This includes four prominent pre-trained (CodeBERT, GraphCodeBERT, CodeT5, UnixCoder) three large language (StarCoder, CodeLlama, CodeT5+). have developed probing tasks evaluate the models' abilities learn These focus on reconstructing semantic structures-such as AST, CFG, CDG, DDG -...

10.48550/arxiv.2212.10017 preprint EN other-oa arXiv (Cornell University) 2022-01-01

As the popularity of graph data increases, there is a growing need to count occurrences subgraph patterns interest, for variety applications. Many graphs are massive in scale and also fully dynamic (with insertions deletions edges), rendering exact computation these counts be infeasible. Common practice is, instead, use small set edges as sample estimate counts. Existing sampling algorithms with uniform probability. In this paper, we show that can do much better if based on their individual...

10.1109/icde55515.2023.00088 article EN 2022 IEEE 38th International Conference on Data Engineering (ICDE) 2023-04-01

Training graph neural networks (GNNs) on large-scale data holds immense promise for numerous real-world applications but remains a great challenge. Several disk-based GNN systems have been built to train graphs in single machine. However, they often fall short terms of performance, especially when training terabyte-scale graphs. This is because existing either overly focus minimizing the number SSD accesses or do not fully overlap with training, thus resulting substantial unnecessary...

10.48550/arxiv.2310.00837 preprint EN cc-by-nc-sa arXiv (Cornell University) 2023-01-01

Graph neural networks are a promising architecture for learning and inference with graph-structured data. Yet, how to generate informative, fixed dimensional features graphs varying size topology can still be challenging. Typically, this is achieved through graph-pooling, which summarizes graph by compressing all its nodes into single vector. Is such “collapsing-style” graph-pooling the only choice classification? From complex system’s point of view, properties system arise largely from...

10.1609/aaai.v36i8.20912 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2022-06-28

Network embedding aims to learn low-dimensional representations of nodes in a network, while the network structure and inherent properties are preserved. It has attracted tremendous attention recently due significant progress downstream learning tasks, such as node classification, link prediction, visualization. However, most existing methods suffer from expensive computations large volume networks. In this paper, we propose $10\times \sim 100\times$ faster method, called Progle, by...

10.48550/arxiv.1806.02623 preprint EN other-oa arXiv (Cornell University) 2018-01-01

Graph embedding has been widely used to process various downstream tasks on large-scale graphs, i.e. node classification, community detection and link prediction. Among methods, Deep Infomax (DGI) is a newly proposed method which achieves excellent performance in classification. However, such outstanding achievement may cause the over-mining issue of user privacy robustness this still unexplored. In paper, we investigate how disturb classification accuracy DGI from an attacker's perspective....

10.1109/iscas45731.2020.9180817 article EN 2022 IEEE International Symposium on Circuits and Systems (ISCAS) 2020-09-29

Recent advancements in Graph Neural Networks have led to state-of-the-art performance on graph representation learning. However, the majority of existing works process directed graphs by symmetrization, which causes loss directional information. To address this issue, we introduce magnetic Laplacian, a discrete Schr\"odinger operator with field, preserves edge directionality encoding it into complex phase an electric charge parameter. By adopting truncated variant PageRank named Linear-...

10.48550/arxiv.2110.07570 preprint EN other-oa arXiv (Cornell University) 2021-01-01

<div>Graph Convolutional Networks (GCNs) derive in spiration from recent advances computer vision, by stacking layers of first-order filters followed a nonlinear activation function to learn graph representations. Although GCNs have been shown boost the performance many network analysis tasks, they still face tremendous challenges learning Heterogeneous Information (HINs), where relations play decisive role knowledge reasoning. In addition, there may exist multi-aspect representations...

10.36227/techrxiv.19311104 preprint EN cc-by 2022-03-11

Graph Convolutional Networks (GCNs) derive inspiration from recent advances in computer vision, by stacking layers of first-order filters followed a nonlinear activation function to learn entity or graph embeddings. Although GCNs have been shown boost the performance many network analysis tasks, they still face tremendous challenges learning Heterogeneous Information (HINs), where relations play decisive role knowledge reasoning. What's more, there are multiaspect representations entities...

10.1109/tnnls.2022.3213799 article EN IEEE Transactions on Neural Networks and Learning Systems 2022-11-03
Coming Soon ...