Liangyou Li

ORCID: 0000-0002-0279-003X
Publications
Citations
Views
---
Saved
---
About
Contact & Profiles
Research Areas
  • Natural Language Processing Techniques
  • Topic Modeling
  • Multimodal Machine Learning Applications
  • Text Readability and Simplification
  • Speech Recognition and Synthesis
  • Speech and dialogue systems
  • Adversarial Robustness in Machine Learning
  • Music and Audio Processing
  • Biomedical Text Mining and Ontologies
  • Laser and Thermal Forming Techniques
  • Criminal Justice and Corrections Analysis
  • Hepatitis B Virus Studies
  • Metallurgy and Material Forming
  • Algorithms and Data Compression
  • Industrial Automation and Control Systems
  • Industrial Vision Systems and Defect Detection
  • Handwritten Text Recognition Techniques
  • Text and Document Classification Technologies
  • Artificial Intelligence in Law
  • Optical measurement and interference techniques
  • Pharmacological Effects and Assays
  • Metal Forming Simulation Techniques
  • Domain Adaptation and Few-Shot Learning
  • Vibration and Dynamic Analysis
  • Evaluation Methods in Various Fields

Huzhou University
2024

Changzhou Academy of Intelli-Ag Equipment (China)
2022-2024

Huawei Technologies (China)
2019-2024

Huawei Technologies (Sweden)
2021-2023

Huaibei Normal University
2022

Monash University
2021

University of Macau
2021

Dublin City University
2014-2017

Large Language Models (LLMs) trained on extensive textual corpora have emerged as leading solutions for a broad array of Natural Processing (NLP) tasks. Despite their notable performance, these models are prone to certain limitations such misunderstanding human instructions, generating potentially biased content, or factually incorrect (hallucinated) information. Hence, aligning LLMs with expectations has become an active area interest within the research community. This survey presents...

10.48550/arxiv.2307.12966 preprint EN cc-by arXiv (Cornell University) 2023-01-01

LLM-as-a-Judge, which generates chain-of-thought (CoT) judgments, has become a widely adopted auto-evaluation method. However, its reliability is compromised by the CoT reasoning's inability to capture comprehensive and deeper details, often leading incomplete outcomes. Existing methods mainly rely on majority voting or criteria expansion, insufficient address limitation in CoT. We propose Crowd-based Comparative Evaluation, introduces additional crowd responses compare with candidate...

10.48550/arxiv.2502.12501 preprint EN arXiv (Cornell University) 2025-02-17

Simultaneous translation (ST) starts translations synchronously while reading source sentences, and is used in many online scenarios. The previous wait-k policy concise achieved good results ST. However, faces two weaknesses: low training speed caused by the recalculation of hidden states lack future information to guide training. For speed, we propose an incremental Transformer with average embedding layer (AEL) accelerate calculation during future-guided training, a conventional as teacher...

10.1609/aaai.v35i16.17696 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2021-05-18

End-to-end simultaneous speech translation (SST), which directly translates in one language into text another real-time, is useful many scenarios but has not been fully investigated. In this work, we propose RealTranS, an end-to-end model for SST. To bridge the modality gap between and text, RealTranS gradually downsamples input with interleaved convolution unidirectional Transformer layers acoustic modeling, then maps features space a weighted-shrinking operation semantic encoder. Besides,...

10.18653/v1/2021.findings-acl.218 article EN cc-by 2021-01-01

10.18653/v1/2024.emnlp-main.1124 article EN Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2024-01-01

Previous works have shown that contextual information can improve the performance of neural machine translation (NMT). However, most existing document-level NMT methods failed to leverage contexts beyond a few set previous sentences. How make use whole document as global is still challenge. To address this issue, we hypothesize be represented graph connects relevant regardless their distances. We employ several types relations, including adjacency, syntactic dependency, lexical consistency,...

10.18653/v1/2021.emnlp-main.663 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021-01-01

Previous work on document-level NMT usually focuses limited contexts because of degraded performance larger contexts. In this paper, we investigate using large with three main contributions: (1) Different from previous which pertrained models large-scale sentence-level parallel corpora, use pretrained language models, specifically BERT, are trained monolingual documents; (2) We propose context manipulation methods to control the influence contexts, lead comparable results systems small and...

10.48550/arxiv.1911.03110 preprint EN other-oa arXiv (Cornell University) 2019-01-01

We argue that the vulnerability of model parameters is crucial value to study robustness and generalization but little research has been devoted understanding this matter. In work, we propose an indicator measure neural network by exploiting their via parameter corruption. The proposed describes maximum loss variation in non-trivial worst-case scenario under For practical purposes, give a gradient-based estimation, which far more effective than random corruption trials can hardly induce...

10.1609/aaai.v35i13.17385 article EN Proceedings of the AAAI Conference on Artificial Intelligence 2021-05-18

To alleviate the data scarcity problem in End-to-end speech translation (ST), pre-training on for recognition and machine is considered as an important technique. However, modality gap between text prevents ST model from efficiently inheriting knowledge pre-trained models. In this work, we propose AdaTranS end-to-end ST. It adapts features with a new shrinking mechanism to mitigate length mismatch by predicting word boundaries. Experiments MUST-C dataset demonstrate that achieves better...

10.18653/v1/2023.findings-emnlp.154 article EN cc-by 2023-01-01

This paper describes Huawei’s neural machine translation systems for the WMT 2019 biomedical shared task. We trained and fine-tuned our on a combination of out-of-domain in-domain parallel corpora six directions covering English–Chinese, English–French English–German language pairs. Our submitted achieve best BLEU scores pairs according to official evaluation results. In English–Chinese task, are in second place. The enhanced performance is attributed more training sophisticated models...

10.18653/v1/w19-5420 article EN cc-by 2019-01-01

Learning multilingual and multi-domain translation model is challenging as the heterogeneous imbalanced data make converge inconsistently over different corpora in real world. One common practice to adjust share of each corpus training, so that learning process balanced low-resource cases can benefit from high resource ones. However, automatic balancing methods usually depend on intra- inter-dataset characteristics, which agnostic or requires human priors. In this work, we propose an...

10.18653/v1/2021.emnlp-main.580 article EN cc-by Proceedings of the 2021 Conference on Empirical Methods in Natural Language Processing 2021-01-01

This paper describes the DCU submission to WMT 2014 on German-English translation task. Our system uses phrasebased model with several popular techniques, including Lexicalized Reordering Model, Operation Sequence Model and Language interpolation. final is result of combination systems which have different pre-processing alignments.

10.3115/v1/w14-3314 article EN cc-by 2014-01-01

We argue that the vulnerability of model parameters is crucial value to study robustness and generalization but little research has been devoted understanding this matter. In work, we propose an indicator measure neural network by exploiting their via parameter corruption. The proposed describes maximum loss variation in non-trivial worst-case scenario under For practical purposes, give a gradient-based estimation, which far more effective than random corruption trials can hardly induce...

10.48550/arxiv.2006.05620 preprint EN other-oa arXiv (Cornell University) 2020-01-01

Dependency structure provides grammatical relations between words, which have shown to be effective in Statistical Machine Translation (SMT).In this paper, we present an open source module Moses implements a dependency-to-string model.We propose method transform the input dependency tree into corresponding constituent for reusing tree-based decoder Moses.In our experiments, achieves comparable results with standard model.Furthermore, enrich model via decomposition of structure, including...

10.3115/v1/w14-4014 article EN cc-by 2014-01-01

This paper describes our work in the WAT 2020 Indic Multilingual Translation Task. We participated all 7 language pairs (En Bn/Hi/Gu/Ml/Mr/Ta/Te) both directions under constrained condition—using only officially provided data. Using transformer as a baseline, Multi->En and En->Multi translation systems achieve best performances. Detailed data filtering domain selection are keys to performance enhancement experiment, with an average improvement of 2.6 BLEU scores for each pair system 4.6...

10.18653/v1/2020.wat-1.8 article EN 2020-01-01

This paper describes our work in participation of the IWSLT-2021 offline speech translation task. Our system was built a cascade form, including speaker diarization module, an Automatic Speech Recognition (ASR) module and Machine Translation (MT) module. We directly use LIUM SpkDiarization tool as The ASR is trained with three datasets from different sources, by multi-source training, using modified Transformer encoder. MT pretrained on large-scale WMT news dataset fine-tuned TED corpus....

10.48550/arxiv.2108.03845 preprint EN cc-by arXiv (Cornell University) 2021-01-01
Coming Soon ...