Chinese Journal of Network and Information Security ›› 2023, Vol. 9 ›› Issue (5): 138-149.doi: 10.11959/j.issn.2096-109x.2023078

• Papers • Previous Articles    

Research on the robustness of neural machine translation systems in word order perturbation

Yuran ZHAO, Tang XUE, Gongshen LIU   

  1. School of Cyber Science and Engineering, Shanghai Jiao Tong University, Shanghai 200240, China
  • Revised:2023-03-02 Online:2023-10-01 Published:2023-10-01
  • Supported by:
    The National Natural Science Foundation of China(U21B2020);Shanghai Science and Technology Plan(22511104400)

Abstract:

Pre-trained language model is one of the most important models in the natural language processing field, as pre-train-finetune has become the paradigm in various NLP downstream tasks.Previous studies have proved integrating pre-trained language models (e.g., BERT) into neural machine translation (NMT) models can improve translation performance.However, it is still unclear whether these improvements stem from enhanced semantic or syntactic modeling capabilities, as well as how pre-trained knowledge impacts the robustness of the models.To address these questions, a systematic study was conducted to examine the syntactic ability of BERT-enhanced NMT models using probing tasks.The study revealed that the enhanced models showed proficiency in modeling word order, highlighting their syntactic modeling capabilities.In addition, an attacking method was proposed to evaluate the robustness of NMT models in handling word order.BERT-enhanced NMT models yielded better translation performance in most of the tasks, indicating that BERT can improve the robustness of NMT models.It was observed that BERT-enhanced NMT model generated poorer translations than vanilla NMT model after attacking in the English-German translation task, which meant that English BERT worsened model robustness in such a scenario.Further analyses revealed that English BERT failed to bridge the semantic gap between the original and perturbed sources, leading to more copying errors and errors in translating low-frequency words.These findings suggest that the benefits of pre-training may not always be consistent in downstream tasks, and careful consideration should be given to its usage.

Key words: neural machine translation, pre-training model, robustness, word order

CLC Number: 

No Suggested Reading articles found!