Mingxuan Wang

ORCID: 0009-0000-9738-4263
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Natural Language Processing Techniques
  • Topic Modeling
  • Speech Recognition and Synthesis
  • Multimodal Machine Learning Applications
  • Toxic Organic Pollutants Impact
  • Speech and dialogue systems
  • Photorefractive and Nonlinear Optics
  • Catalytic Processes in Materials Science
  • Photonic and Optical Devices
  • Advanced Optical Imaging Technologies
  • Text Readability and Simplification
  • Air Quality and Health Impacts
  • Advancements in Battery Materials
  • Atmospheric chemistry and aerosols
  • Biosensors and Analytical Detection
  • Semiconductor materials and devices
  • SARS-CoV-2 detection and testing
  • Phase Equilibria and Thermodynamics
  • Electrochemical Analysis and Applications
  • Speech and Audio Processing
  • Microbial bioremediation and biosurfactants
  • Music and Audio Processing
  • Catalysis and Oxidation Reactions
  • Molecular Communication and Nanonetworks
  • Advanced oxidation water treatment

Research Center for Eco-Environmental Sciences
2022-2025

University of Chinese Academy of Sciences
2022-2025

Minzu University of China
2025

Huazhong University of Science and Technology
2025

Laboratoire de Synthèse Organique
2025

Chinese Academy of Sciences
2015-2024

Shanghai University of Medicine and Health Sciences
2024

Technical Institute of Physics and Chemistry
2023-2024

Shanghai University of Engineering Science
2023

Zhejiang University of Science and Technology
2023

How to learn a better speech representation for end-to-end speech-to-text translation (ST) with limited labeled data? Existing techniques often attempt transfer powerful machine (MT) capabilities ST, but neglect the discrepancy across modalities. In this paper, we propose Speech-TExt Manifold Mixup (STEMM) method calibrate such discrepancy. Specifically, mix up sequences of different modalities, and take both unimodal multimodal mixed as input model in parallel, regularize their output...

10.18653/v1/2022.acl-long.486 article EN cc-by Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2022-01-01

How can we learn unified representations for spoken utterances and their written text? Learning similar semantically speech text is important translation. To this end, propose ConST, a cross-modal contrastive learning method end-to-end speech-to-text We evaluate ConST variety of previous baselines on popular benchmark MuST-C. Experiments show that the proposed consistently outperforms methods, achieves an average BLEU 29.4. The analysis further verifies indeed closes representation gap...

10.18653/v1/2022.naacl-main.376 article EN cc-by Proceedings of the 2022 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies 2022-01-01

Although sequence-to-sequence attentional neural machine translation (NMT) has achieved great progress recently, it is confronted with two challenges: learning optimal model parameters for long parallel sentences and well exploiting different scopes of contexts. In this paper, partially inspired by the idea segmenting a sentence into short clauses, each which can be easily translated NMT, we propose hierarchy-to-sequence NMT to handle these challenges. Our encoder takes segmented clause...

10.1109/taslp.2018.2789721 article EN IEEE/ACM Transactions on Audio Speech and Language Processing 2018-01-04

Fandong Meng, Zhengdong Lu, Mingxuan Wang, Hang Li, Wenbin Jiang, Qun Liu. Proceedings of the 53rd Annual Meeting Association for Computational Linguistics and 7th International Joint Conference on Natural Language Processing (Volume 1: Long Papers). 2015.

10.3115/v1/p15-1003 article EN cc-by 2015-01-01

We propose to enhance the RNN decoder in a neural machine translator (NMT) with external memory, as natural but powerful extension state decoding RNN.This memory-enhanced is called MEMDEC.At each time during decoding, MEMDEC will read from this memory and write once, both content-based addressing.Unlike unbounded previous work (Bahdanau et al., 2014) store representation of source sentence, matrix predetermined size designed better capture information important for process at step.Our...

10.18653/v1/d16-1027 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2016-01-01

Photopolymers hold great promise for the preparation of transparent volume holographic gratings (VHG), which are core optical elements in many application fields. To improve recording property a two-stage photopolymer, four new (meth)acrylate monomers (CTA, CTMA, CTBA, CTBMA) with high refractive indices (1.59–1.63) designed and synthesized this study. Using them as one writing monomer, series photopolymer samples different formulations thicknesses fabricated recording. Among them,...

10.1021/acsami.3c01446 article EN ACS Applied Materials & Interfaces 2023-05-11

Deep Neural Networks (DNNs) have provably enhanced the state-of-the-art Machine Translation (NMT) with its capability in modeling complex functions and capturing linguistic structures. However NMT deep architecture encoder or decoder RNNs often suffer from severe gradient diffusion due to non-linear recurrent activations, which makes optimization much more difficult. To address this problem we propose a novel linear associative units (LAU) reduce propagation path inside unit. Different...

10.18653/v1/p17-1013 article EN cc-by Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2017-01-01

The recently proposed neural network joint model (NNJM) (Devlin et al., 2014) augments the n-gram target language with a heuristically chosen source context window, achieving state-of-the-art performance in SMT. In this paper, we give more systematic treatment by summarizing relevant information through convolutional architecture guided information. With different guiding signals during decoding, our specifically designed convolution+gating architectures can pinpoint parts of sentence that...

10.48550/arxiv.1503.01838 preprint EN other-oa arXiv (Cornell University) 2015-01-01

This paper proposes three distortion models to explicitly incorporate the word reordering knowledge into attention-based Neural Machine Translation (NMT) for further improving translation performance. Our proposed enable attention mechanism attend source words regarding both semantic requirement and penalty. Experiments on Chinese-English show that approaches can improve alignment quality achieve significant improvements over a basic NMT by large margins. Compared with previous works...

10.18653/v1/p17-1140 article EN cc-by Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers) 2017-01-01

In this study, we first investigate a novel capsule network with dynamic routing for linear time Neural Machine Translation (NMT), referred as \textsc{CapsNMT}. \textsc{CapsNMT} uses an aggregation mechanism to map the source sentence into matrix pre-determined size, and then applys deep LSTM decode target sequence from representation. Unlike previous work \cite{sutskever2014sequence} store passive bottom-up way, policy encodes iterative process decide credit attribution between nodes lower...

10.18653/v1/d19-1074 preprint EN cc-by 2019-01-01

Metal compounds play important roles in the formation of organic pollutants during thermal-related processes. However, metal-catalyzed predominant have not previously been characterized nor any detailed catalytic mechanisms clarified. Here, we preciously distinguished multiple free radical intermediates on metal catalyst surfaces pollutant through laboratory and theoretical studies. Differences between intermediate species, concentrations, under catalysis different were investigated. The...

10.1021/acs.est.2c05892 article EN Environmental Science & Technology 2022-09-28

How can speech-to-text translation (ST) perform as well machine (MT)? The key point is to bridge the modality gap between speech and text so that useful MT techniques be applied ST.Recently, approach of representing with unsupervised discrete units yields a new way ease problem. This motivates us propose Discrete Unit Back-translation(DUB) answer two questions (1) Is it better represent than continuous features in direct ST? (2) much benefit bring With DUB, back-translation technique...

10.18653/v1/2023.findings-acl.447 article EN cc-by Findings of the Association for Computational Linguistics: ACL 2022 2023-01-01

Many tasks in natural language processing, ranging from machine translation to question answering, can be reduced the problem of matching two sentences or more generally short texts. We propose a new approach problem, called Deep Match Tree (DeepMatch$_{tree}$), under general setting. The consists components, 1) mining algorithm discover patterns for short-texts, defined product space dependency trees, and 2) deep neural network texts using mined patterns, as well learning build having...

10.48550/arxiv.1503.02427 preprint EN other-oa arXiv (Cornell University) 2015-01-01

Chlorinated organic chemicals are produced and used extensively worldwide, their risks to the biology environment of increasing concern. However, chlorinated byproducts [e.g., polychlorinated dibenzo-p-dioxins dibenzofurans (PCDD/Fs)] formed during commercial manufacturing processes present in organochlorine products rarely reported. The knowledge on occurrences fate unintentional persistent chemical is necessary for accurate assessment production. Here, PCDD/Fs were tracked throughout...

10.1021/acs.est.2c07322 article EN Environmental Science & Technology 2023-01-22
Coming Soon ...