Cheng Wen

ORCID: 0000-0002-5585-4957
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Speech Recognition and Synthesis
  • Advanced Text Analysis Techniques
  • Natural Language Processing Techniques
  • Topic Modeling
  • Sentiment Analysis and Opinion Mining
  • Underwater Vehicles and Communication Systems
  • Advanced Data Compression Techniques
  • Speech and Audio Processing
  • Text and Document Classification Technologies
  • Indoor and Outdoor Localization Technologies
  • Sparse and Compressive Sensing Techniques

Nanjing University
2023-2024

Didi Chuxing (China)
2021

Xi'an University of Technology
2021

This paper introduces a new open-sourced Mandarin speech corpus, called DiDiSpeech. It consists of about 800 hours data at 48kHz sampling rate from 6000 speakers and the corresponding texts. All in corpus is recorded quiet environment suitable for various processing tasks, such as voice conversion, multi-speaker text-to-speech automatic recognition. We conduct experiments with multiple tasks evaluate performance, showing that it promising to use both academic research practical application....

10.1109/icassp39728.2021.9414423 article EN ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2021-05-13

Aiming at the problem that Aspect-based sentiment analysis in Chinese has low recognition rate due to many steps, this paper proposes an improved BiLSTM-CRF model based on combine character vector and words position feature, which can extract attribute jointly simultaneously, while extracting Polarity judges of words. Experiments show improves precision by 9.2% 13.32%, recall 0.48% 21.29%, F-measure 7.33% 15.74% compared with Conditional Random Fields (CRF) Long Short Term Memory (LSTM)...

10.3233/jifs-192078 article EN Journal of Intelligent & Fuzzy Systems 2021-02-16

In this paper, we describe our speech generation system for the first Audio Deep Synthesis Detection Challenge (ADD 2022). Firstly, build an any-to-many voice conversion (VC) to convert source with arbitrary language content into target speaker's fake speech. Then converted generated from VC is post-processed in time-domain improve deception ability. The experimental results show that has adversarial ability against anti-spoofing detectors a little compromise audio quality and speaker...

10.1109/icassp43922.2022.9746164 article EN ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2022-04-27

One of the key challenges for wireless sensing systems is how to efficiently enable capabilities multiple devices while leveraging existing communication resources. In this paper, we propose DEWS, a distributed channel measurement scheme that allows transmitters perform tasks simultaneously, which considers three issues in tasks: multi-device resolution, reliability, and accuracy. First, use carefully designed Resource Unit (dRU) allocation based on OFDMA ensure simultaneously with entire...

10.1145/3699728 article EN Proceedings of the ACM on Interactive Mobile Wearable and Ubiquitous Technologies 2024-11-21

This paper introduces a new open-sourced Mandarin speech corpus, called DiDiSpeech. It consists of about 800 hours data at 48kHz sampling rate from 6000 speakers and the corresponding texts. All in corpus is recorded quiet environment suitable for various processing tasks, such as voice conversion, multi-speaker text-to-speech automatic recognition. We conduct experiments with multiple tasks evaluate performance, showing that it promising to use both academic research practical application....

10.48550/arxiv.2010.09275 preprint EN other-oa arXiv (Cornell University) 2020-01-01

This paper describes our submission to ICASSP 2023 MUG Challenge Track 4, Keyphrase Extraction, which aims extract keyphrases most relevant the conference theme from materials. We model challenge as a single-class Named Entity Recognition task and developed techniques for better performance on challenge: For data preprocessing, we encode split after word segmentation. In addition, increase amount of input information that can accept at one time by fusing multiple preprocessed sentences into...

10.1109/icassp49357.2023.10096850 article EN ICASSP 2022 - 2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2023-05-05
Coming Soon ...