Haifeng Chen

ORCID: 0000-0002-9363-738X
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Topic Modeling
  • Advanced Graph Neural Networks
  • Time Series Analysis and Forecasting
  • Natural Language Processing Techniques
  • Machine Learning in Healthcare
  • Explainable Artificial Intelligence (XAI)
  • Domain Adaptation and Few-Shot Learning
  • Advanced Text Analysis Techniques
  • Anomaly Detection Techniques and Applications
  • Adversarial Robustness in Machine Learning
  • Recommender Systems and Techniques
  • Music and Audio Processing
  • Text and Document Classification Technologies
  • Neural Networks and Applications
  • Sentiment Analysis and Opinion Mining
  • Multimodal Machine Learning Applications
  • Privacy-Preserving Technologies in Data
  • Complex Network Analysis Techniques
  • Software System Performance and Reliability
  • Speech and Audio Processing
  • Face and Expression Recognition
  • Human Mobility and Location-Based Analysis
  • Data Stream Mining Techniques
  • Artificial Intelligence in Healthcare and Education
  • Traffic Prediction and Management Techniques

NEC (United States)
2020-2025

Princeton University
2007-2024

Tsinghua University
2022

NEC (Japan)
2020-2021

Georgia Institute of Technology
2019

Rutgers, The State University of New Jersey
2004

The Nonlinear autoregressive exogenous (NARX) model, which predicts the current value of a time series based upon its previous values as well and past multiple driving (exogenous) series, has been studied for decades. Despite fact that various NARX models have developed, few them can capture long-term temporal dependencies appropriately select relevant to make predictions. In this paper, we propose dual-stage attention-based recurrent neural network (DA-RNN) address these two issues. first...

10.48550/arxiv.1704.02971 preprint EN other-oa arXiv (Cornell University) 2017-01-01

Despite recent progress in Graph Neural Networks (GNNs), explaining predictions made by GNNs remains a challenging open problem. The leading method independently addresses the local explanations (i.e., important subgraph structure and node features) to interpret why GNN model makes prediction for single instance, e.g. or graph. As result, explanation generated is painstakingly customized each instance. unique interpreting instance not sufficient provide global understanding of learned model,...

10.48550/arxiv.2011.04573 preprint EN cc-by arXiv (Cornell University) 2020-01-01

Text summarization has been a crucial problem in natural language processing (NLP) for several decades. It aims to condense lengthy documents into shorter versions while retaining the most critical information. Various methods have proposed text summarization, including extractive and abstractive summarization. The emergence of large models (LLMs) like GPT3 ChatGPT recently created significant interest using these tasks. Recent studies \cite{goyal2022news, zhang2023benchmarking} shown that...

10.48550/arxiv.2302.08081 preprint EN cc-by arXiv (Cornell University) 2023-01-01

To subvert recent advances in perimeter and host security, the attacker community has developed employed various attack vectors to make a malware much stealthier than before penetrate target system prolong its presence.Such advanced or "stealthy malware" makes use of techniques impersonate abuse benign applications legitimate tools minimize footprints system.It is thus difficult for traditional detection tools, such as scanners, detect it, normally does not expose malicious payload file...

10.14722/ndss.2020.24167 article EN 2020-01-01

Click-through rate (CTR) prediction is a critical task in online advertising and marketing. For this problem, existing approaches, with shallow or deep architectures, have three major drawbacks. First, they typically lack persuasive rationales to explain the outcomes of models. Unexplainable predictions recommendations may be difficult validate thus unreliable untrustworthy. In many applications, inappropriate suggestions even bring severe consequences. Second, approaches poor efficiency...

10.1145/3336191.3371785 article EN 2020-01-20

Various graph contrastive learning models have been proposed to improve the performance of tasks on datasets in recent years. While effective and prevalent, these are usually carefully customized. In particular, although all researches create two views, they differ greatly view augmentations, architectures, objectives. It remains an open question how build your model from scratch for particular datasets. this work, we aim fill gap by studying information is transformed transferred during...

10.48550/arxiv.2110.15438 preprint EN cc-by-nc-nd arXiv (Cornell University) 2021-01-01

Large language models (LLMs) have significantly advanced the field of natural processing (NLP), providing a highly useful, task-agnostic foundation for wide range applications. However, directly applying LLMs to solve sophisticated problems in specific domains meets many hurdles, caused by heterogeneity domain data, sophistication knowledge, uniqueness objectives, and diversity constraints (e.g., various social norms, cultural conformity, religious beliefs, ethical standards applications)....

10.48550/arxiv.2305.18703 preprint EN cc-by arXiv (Cornell University) 2023-01-01

Various contrastive learning approaches have been proposed in recent years and achieve significant empirical success. While effective prevalent, has less explored for time series data. A key component of is to select appropriate augmentations imposing some priors construct feasible positive samples, such that an encoder can be trained learn robust discriminative representations. Unlike image language domains where "desired'' augmented samples generated with the rule thumb guided by...

10.1609/aaai.v37i4.25575 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2023-06-26

Time series data is essential in various applications, including climate modeling, healthcare monitoring, and financial analytics. Understanding the contextual information associated with real-world time often for accurate reliable event predictions. In this paper, we introduce TimeCAP, a time-series processing framework that creatively employs Large Language Models (LLMs) as contextualizers of data, extending their typical usage predictors. TimeCAP incorporates two independent LLM agents:...

10.1609/aaai.v39i17.33989 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2025-04-11

Graph Neural Networks (GNNs) resurge as a trending research subject owing to their impressive ability capture representations from graph-structured data. However, the black-box nature of GNNs presents significant challenge in terms comprehending and trusting these models, thereby limiting practical applications mission-critical scenarios. Although there has been substantial progress field explaining recent years, majority studies are centered on static graphs, leaving explanation dynamic...

10.1145/3729173 article EN ACM Transactions on Knowledge Discovery from Data 2025-04-12

Multivariate time series data are becoming increasingly common in numerous real world applications, e.g., power plant monitoring, health care, wearable devices, automobile, etc. As a result, multivariate retrieval, i.e., given the current segment, how to obtain its relevant segments historical (or database), attracts significant amount of interest many fields. Building such system, however, is challenging since it requires compact representation raw which can explicitly encode temporal...

10.1145/3219819.3220108 article EN 2018-07-19

Multi-modal fusion overcomes the inherent limitations of single-sensor perception in 3D object detection autonomous driving. The 4D Radar and LiDAR can boost range more robust. Nevertheless, different data characteristics noise distributions between two sensors hinder performance improvement when directly integrating them. Therefore, we are first to propose a novel method termed <inline-formula xmlns:mml="http://www.w3.org/1998/Math/MathML"...

10.1109/tvt.2022.3230265 article EN IEEE Transactions on Vehicular Technology 2022-12-19

The problem of learning and forecasting underlying trends in time series data arises a variety applications, such as traffic management, energy optimization, etc. In literature, trend is characterized by the slope duration, its prediction then to forecast two values subsequent given historical series. For this problem, existing approaches mainly deal with case univariate However, many real-world there are multiple variables at play, handling all them same crucial for an accurate prediction....

10.1609/aaai.v34i02.5496 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2020-04-03

Forecasting on sparse multivariate time series (MTS) aims to model the predictors of future values given their incomplete past, which is important for many emerging applications. However, most existing methods process MTS’s individually, and do not leverage dynamic distributions underlying MTS’s, leading sub-optimal results when sparsity high. To address this challenge, we propose a novel generative model, tracks transition latent clusters, instead isolated feature representations, achieve...

10.1609/aaai.v35i1.16145 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2021-05-18

The goal of root cause analysis is to identify the underlying causes system problems by discovering and analyzing causal structure from monitoring data. It indispensable for maintaining stability robustness large-scale complex systems. Existing methods mainly focus on construction a single effective isolated network, whereas many real-world systems are exhibit interdependent structures (i.e., multiple networks interconnected cross-network links). In networks, malfunctioning effects...

10.1145/3580305.3599849 article EN Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2023-08-04

Deep learning systems on the cloud are increasingly targeted by attacks that attempt to steal sensitive data. Intel SGX has been proven effective protect confidentiality and integrity of such data during computation. However, state-of-the-art still suffer from substantial performance overhead induced limited physical memory SGX. This limitation significantly undermines usability deep due their memory-intensive characteristics.

10.1145/3419111.3421282 article EN 2020-10-12

The task of root cause analysis (RCA) is to identify the causes system faults/failures by analyzing monitoring data. Efficient RCA can greatly accelerate failure recovery and mitigate damages or financial losses. However, previous research has mostly focused on developing offline algorithms, which often require manually initiating process, a significant amount time data train robust model, then being retrained from scratch for new fault.

10.1145/3580305.3599392 article EN Proceedings of the 28th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2023-08-04

Time series analysis has witnessed the inspiring development from traditional autoregressive models, deep learning to recent Transformers and Large Language Models (LLMs). Efforts in leveraging vision models for time have also been made along way but are less visible community due predominant research on sequence modeling this domain. However, discrepancy between continuous discrete token space of LLMs, challenges explicitly correlations variates multivariate shifted some attentions equally...

10.48550/arxiv.2502.08869 preprint EN arXiv (Cornell University) 2025-02-12

Time series data is essential in various applications, including climate modeling, healthcare monitoring, and financial analytics. Understanding the contextual information associated with real-world time often for accurate reliable event predictions. In this paper, we introduce TimeCAP, a time-series processing framework that creatively employs Large Language Models (LLMs) as contextualizers of data, extending their typical usage predictors. TimeCAP incorporates two independent LLM agents:...

10.48550/arxiv.2502.11418 preprint EN arXiv (Cornell University) 2025-02-16

Large Language Models (LLMs) exhibit potential artificial generic intelligence recently, however, their usage is costly with high response latency. Given mixed LLMs own strengths and weaknesses, LLM routing aims to identify the most suitable model for each query in stream maximize quality minimize cost However, challenges involve: (1) dynamic trade-offs among quality, cost, latency; (2) enabling continual learning deployed systems; (3) navigating a varying (e.g., new addition or old removal)...

10.32388/ns4gu3 preprint EN 2025-04-03

Graph neural network (GNN), as a powerful representation learning model on graph data, attracts much attention across various disciplines. However, recent studies show that GNN is vulnerable to adversarial attacks. How make more robust? What are the key vulnerabilities in GNN? address and defense against attacks? In this paper, we propose DefNet, an effective framework for GNNs. particular, first investigate latent every layer of GNNs corresponding strategies including dual-stage aggregation...

10.48550/arxiv.1905.03679 preprint EN other-oa arXiv (Cornell University) 2019-01-01
Coming Soon ...