Chengyu Wang

ORCID: 0000-0003-1010-9678
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Topic Modeling
  • Natural Language Processing Techniques
  • Multimodal Machine Learning Applications
  • Speech Recognition and Synthesis
  • Domain Adaptation and Few-Shot Learning
  • Generative Adversarial Networks and Image Synthesis
  • Advanced Graph Neural Networks
  • Image Retrieval and Classification Techniques
  • Text and Document Classification Technologies
  • Biomedical Text Mining and Ontologies
  • Semantic Web and Ontologies
  • Advanced Image and Video Retrieval Techniques
  • Privacy-Preserving Technologies in Data
  • Text Readability and Simplification
  • Adversarial Robustness in Machine Learning
  • Video Analysis and Summarization
  • Sentiment Analysis and Opinion Mining
  • Web Data Mining and Analysis
  • Advanced Vision and Imaging
  • Anomaly Detection Techniques and Applications
  • Advanced Text Analysis Techniques
  • Speech and dialogue systems
  • Music and Audio Processing
  • Advanced Neural Network Applications
  • Computer Graphics and Visualization Techniques

Linyi University
2025

Alibaba Group (United States)
2020-2024

PLA Information Engineering University
2024

Alibaba Group (China)
2021-2024

Xi'an Shiyou University
2024

East China Normal University
2013-2023

Hefei University of Technology
2023

National University of Singapore
2023

National Cheng Kung University
2023

National University of Defense Technology
2023

10.1007/s11704-016-5228-9 article EN Frontiers of Computer Science 2016-09-26

This paper proposes the novel task of video generation conditioned on a SINGLE semantic label map, which provides good balance between flexibility and quality in process. Different from typical end-to-end approaches, model both scene content dynamics single step, we propose to decompose this difficult into two sub-problems. As current image methods do better than terms detail, synthesize high by only generating first frame. Then animate based its meaning obtain temporally coherent video,...

10.1109/cvpr.2019.00385 preprint EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2019-06-01

A taxonomy is a semantic hierarchy, consisting of concepts linked by is-a relations. While large number taxonomies have been constructed from human-compiled resources (e.g., Wikipedia), learning text corpora has received growing interest and essential for long-tailed domain-specific knowledge acquisition. In this paper, we overview recent advances on construction free texts, reorganizing relevant subtasks into complete framework. We also evaluation discuss challenges future research.

10.18653/v1/d17-1123 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2017-01-01

Taolin Zhang, Zerui Cai, Chengyu Wang, Minghui Qiu, Bite Yang, Xiaofeng He. Proceedings of the 59th Annual Meeting Association for Computational Linguistics and 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2021.

10.18653/v1/2021.acl-long.457 article EN cc-by 2021-01-01

Distant supervision assumes that any sentence containing the same entity pairs reflects identical relationships. Previous works of distantly supervised relation extraction (DSRE) task generally focus on sentence-level or bag-level de-noising techniques independently, neglecting explicit interaction with cross levels. In this paper, we propose a hierarchical contrastive learning Framework for Distantly Supervised (HiCLRE) to reduce noisy sentences, which integrate global structural...

10.18653/v1/2022.findings-acl.202 article EN cc-by Findings of the Association for Computational Linguistics: ACL 2022 2022-01-01

Pre-trained Language Models (PLMs) have achieved remarkable performance for various language understanding tasks in IR systems, which require the fine-tuning process based on labeled training data. For low-resource scenarios, prompt-based learning PLMs exploits prompts as task guidance and turns downstream into masked problems effective few-shot fine-tuning. In most existing approaches, high of heavily relies handcrafted verbalizers, may limit application such approaches real-world...

10.1145/3539597.3570398 article EN 2023-02-22

Recently, researchers have applied the word-character lattice framework to integrated word information, which has become very popular for Chinese named entity recognition (NER). However, prior approaches fuse information by different variants of encoders such as Lattice LSTM or Flat-Lattice Transformer, but are still not data-efficient indeed fully grasp depth interaction cross-granularity and important from lexicon. In this paper, we go beyond typical structure propose a novel...

10.1609/aaai.v37i11.26640 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2023-06-26

Fine-tuning large pre-trained language models on various downstream tasks with whole parameters is prohibitively expensive. Hence, Parameter-efficient fine-tuning has attracted attention that only optimizes a few task-specific the frozen model. In this work, we focus prefix tuning, which continuous vectors (i.e. pseudo tokens) inserted into Transformer layers. Based observation learned syntax and semantics representation varies lot at different layers, argue adaptive will be further tailored...

10.18653/v1/2023.acl-short.107 article EN cc-by 2023-01-01

Two facile and diversified protocols for the chem‐selective synthesis of 1,2‐dihydropyridine‐fused coumarins pyridocoumarins have been developed. Readily accessible substrates, mild reaction conditions good functional group tolerance make these procedures attractive construction various dihydropyridine‐fused pyridocoumarins, which may enrich develop heterocyclic chemistry medicinal chemistry.

10.1002/ejoc.202401244 article EN European Journal of Organic Chemistry 2025-01-13

Recent studies have shown that prompts improve the performance of large pre-trained language models for few-shot text classification. Yet, it is unclear how prompting knowledge can be transferred across similar NLP tasks purpose mutual reinforcement. Based on continuous prompt embeddings, we propose TransPrompt, a transferable framework learning tasks. In employ multi-task meta-knowledge acquisition procedure to train meta-learner captures cross-task knowledge. Two de-biasing techniques are...

10.18653/v1/2021.emnlp-main.221 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021-01-01

Haojie Pan, Chengyu Wang, Minghui Qiu, Yichang Zhang, Yaliang Li, Jun Huang. Proceedings of the 59th Annual Meeting Association for Computational Linguistics and 11th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2021.

10.18653/v1/2021.acl-long.236 article EN cc-by 2021-01-01

Few-shot Named Entity Recognition (NER) aims to identify named entities with very little annotated data. Previous methods solve this problem based on token-wise classification, which ignores the information of entity boundaries, and inevitably performance is affected by massive non-entity tokens. To end, we propose a seminal span-based prototypical network (SpanProto) that tackles few-shot NER via two-stage approach, including span extraction mention classification. In stage, transform...

10.18653/v1/2022.emnlp-main.227 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2022-01-01

Spoken Language Understanding (SLU) aims to interpret the meanings of human speeches in order support various human-machine interaction systems. A key technique for SLU is Automatic Speech Recognition (ASR), which transcribes speech signals into text contents. As output texts modern ASR systems unavoidably contain errors, mainstream models either trained or tested on transcribed by would not be sufficiently error robust. We present ARoBERT, an Robust BERT model, can fine-tuned solve a...

10.1109/taslp.2022.3153268 article EN IEEE/ACM Transactions on Audio Speech and Language Processing 2022-01-01

Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models (PLMs) on few-shot text classification by employing task-specific prompts. Yet, PLMs are unfamiliar with prompt-style expressions during pre-training, which limits learning downstream tasks.It would be desirable if models can acquire some prompting knowledge before adapting to specific NLP tasks. We present Unified Prompt Tuning (UPT) framework, leading better for BERT-style explicitly capturing semantics...

10.18653/v1/2022.findings-emnlp.37 article EN cc-by 2022-01-01

Knowledge-Enhanced Pre-trained Language Models (KEPLMs) are pre-trained models with relation triples injecting from knowledge graphs to improve language understanding abilities.Experiments show that our model outperforms other KEPLMs significantly over zero-shot probing tasks and multiple knowledge-aware tasks. To guarantee effective injection, previous studies integrate encoders for representing retrieved graphs. The operations retrieval encoding bring significant computational burdens,...

10.1609/aaai.v36i10.21425 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2022-06-28

10.1109/cvpr52733.2024.00747 article EN 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) 2024-06-16

The automatic construction of large-scale knowledge graphs has received much attention from both academia and industry in the past few years. Notable graph systems include Google Knowledge Graph, DBPedia, YAGO, NELL, Probase many others. organizes information a structured way by explicitly describing relations among entities. Since entity identification relation extraction are highly depending on language itself, data sources largely determine processed, extracted, ultimately how formed,...

10.1109/icdew.2015.7129545 article EN 2015-04-01

Chengyu Wang, Minghui Qiu, Taolin Zhang, Tingting Liu, Lei Li, Jianing Ming Jun Huang, Wei Lin. Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing: System Demonstrations. 2022.

10.18653/v1/2022.emnlp-demos.3 article EN cc-by 2022-01-01

Image-text retrieval is a challenging cross-modal task that arouses much attention. While the traditional methods cannot break down barriers between different modalities, Vision-Language Pre-trained (VLP) models greatly improve image-text performance based on massive pairs. Nonetheless, VLP-based are still prone to produce results be aligned with entities. Recent efforts try fix this problem at pre-training stage, which not only expensive but also unpractical due unavailable of full...

10.1145/3539597.3570481 article EN 2023-02-22
Coming Soon ...