AMPLIFY: attention-based mixup for performance improvement and label smoothing in transformer
FOS: Computer and information sciences
Computer Science - Machine Learning
Mixup
Computer Science - Computation and Language
Data augmentation
Artificial Intelligence
Electronic computers. Computer science
Attention mechanism
QA75.5-76.95
Label smoothing
Model robustness
Computation and Language (cs.CL)
Machine Learning (cs.LG)
DOI:
10.7717/peerj-cs.2011
Publication Date:
2024-04-30T08:34:13Z
AUTHORS (2)
ABSTRACT
Mixup is an effective data augmentation method that generates new augmented samples by aggregating linear combinations of different original samples. However, if there are noises or aberrant features in the original samples, mixup may propagate them to the augmented samples, leading to over-sensitivity of the model to these outliers. To solve this problem, this paper proposes a new mixup method called AMPLIFY. This method uses the attention mechanism of Transformer itself to reduce the influence of noises and aberrant values in the original samples on the prediction results, without increasing additional trainable parameters, and the computational cost is very low, thereby avoiding the problem of high resource consumption in common mixup methods such as Sentence Mixup. The experimental results show that, under a smaller computational resource cost, AMPLIFY outperforms other mixup methods in text classification tasks on seven benchmark datasets, providing new ideas and new ways to further improve the performance of pre-trained models based on the attention mechanism, such as BERT, ALBERT, RoBERTa, and GPT. Our code can be obtained at https://github.com/kiwi-lilo/AMPLIFY.
SUPPLEMENTAL MATERIAL
Coming soon ....
REFERENCES (32)
CITATIONS (0)
EXTERNAL LINKS
PlumX Metrics
RECOMMENDATIONS
FAIR ASSESSMENT
Coming soon ....
JUPYTER LAB
Coming soon ....