Alternating Language Modeling for Cross-Lingual Pre-Training
Concatenation (mathematics)
Code (set theory)
DOI:
10.1609/aaai.v34i05.6480
Publication Date:
2020-06-18T08:13:53Z
AUTHORS (6)
ABSTRACT
Language model pre-training has achieved success in many natural language processing tasks. Existing methods for cross-lingual adopt Translation Model to predict masked words with the concatenation of source sentence and its target equivalent. In this work, we introduce a novel method, called Alternating Modeling (ALM). It code-switches sentences different languages rather than simple concatenation, hoping capture rich context phrases. More specifically, randomly substitute phrases translations create code-switched sentences. Then, use these data train ALM learn languages. We evaluate our on downstream tasks machine translation classification. Experiments show that can outperform previous three benchmarks.1
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (0)
CITATIONS (21)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....