Journal of Computer Applications ›› 2022, Vol. 42 ›› Issue (12): 3679-3685.DOI: 10.11772/j.issn.1001-9081.2021101805

• Artificial intelligence • Previous Articles    

Neural machine translation integrating bidirectional-dependency self-attention mechanism

Zhijin LI1,2, Hua LAI1,2(), Yonghua WEN1,2, Shengxiang GAO1,2   

  1. 1.Faculty of Information Engineering and Automation,Kunming University of Science and Technology,Kunming Yunnan 650504,China
    2.Yunnan Key Laboratory of Artificial Intelligence (Kunming University of Science and Technology),Kunming Yunnan 650504,China
  • Received:2021-10-22 Revised:2022-01-06 Accepted:2022-01-24 Online:2022-04-26 Published:2022-12-10
  • Contact: Hua LAI
  • About author:LI Zhijin, born in 1997, M. S. candidate. His research interests include machine translation, natural language processing.
    WEN Yonghua, born in 1979, Ph. D. candidate. His research interests include machine translation.
    GAO Shengxiang,born in 1977, Ph. D., associate professor. Her research interests include machine translation, natural language processing, information retrieval.
  • Supported by:
    National Natural Science Foundation of China(61732005);Yunnan Province Major Science and Technology Special Project(202002AD080001-5);Yunnan Province High-tech Industry Special Project(201606)

融合双向依存自注意力机制的神经机器翻译

李治瑾1,2, 赖华1,2(), 文永华1,2, 高盛祥1,2   

  1. 1.昆明理工大学 信息工程与自动化学院,昆明 650504
    2.云南省人工智能重点实验室(昆明理工大学),昆明 650504
  • 通讯作者: 赖华
  • 作者简介:李治瑾(1997—),男,辽宁大连人,硕士研究生,主要研究方向:机器翻译、自然语言处理
    文永华(1979—),男(白族),云南大理人,博士研究生,主要研究方向:机器翻译
    文永华(1979—),男(白族),云南大理人,博士研究生,主要研究方向:机器翻译
    高盛祥(1977—),女,云南大理人,副教授,博士,主要研究方向:机器翻译、自然语言处理、信息检索。
  • 基金资助:
    国家自然科学基金资助项目(61732005);云南省重大科技专项(202002AD080001?5);云南省高新技术产业专项(201606)

Abstract:

Aiming at the problem of resource scarcity in neural machine translation, a method for fusion of dependency syntactic knowledge based on a Bidirectional-Dependency self-attention mechanism (Bi-Dependency) was proposed. Firstly, an external parser was used to parse the source sentence to obtain dependency parsing data. Then, the dependency parsing data was transformed into the position vector of the parent word and the weight matrix of the child word. Finally, the dependency knowledge was integrated into the multi-head attention mechanism of the Transformer encoder. By using Bi-Dependency, the translation model was able to simultaneously pay attention to the dependency information in both directions: the parent word to the child word and the child word to the parent word. Experimental results of bi-directional translation show that compared with the Transformer model, in the case of rich resources, the proposed method has the BLEU (BiLingual Evaluation Understudy) value on Chinese-Thai translation improved by 1.07 and 0.86 respectively, and the BLEU value on Chinese-English translation improved by 0.79 and 0.68 respectively; in the case of low resources, the proposed model has the BLEU value increased by 0.51 and 1.06 respectively on Chinese-Thai translation, and the BLEU value increased by 1.04 and 0.40 respectively on Chinese-English translation. It can be seen that Bi-Dependency provides the model with richer dependence information, which can effectively improve the translation performance.

Key words: neural machine translation, bidirectional-dependency attention, multi-head attention, parent word, child word

摘要:

针对神经机器翻译中资源稀缺的问题,提出了一种基于双向依存自注意力机制(Bi-Dependency)的依存句法知识融合方法。首先,利用外部解析器对源句子解析得到依存解析数据;然后,将依存解析数据转化为父词位置向量和子词权重矩阵;最后,将依存知识融合到Transformer编码器的多头注意力机制上。利用Bi-Dependency,翻译模型可以同时对父词到子词、子词到父词两个方向的依存信息进行关注。双向翻译的实验结果表明,与Transformer模型相比,在富资源情况下,所提方法在汉-泰翻译上的BLEU值分别提升了1.07和0.86,在汉-英翻译上的BLEU值分别提升了0.79和0.68;在低资源情况下,所提方法在汉-泰翻译上的BLEU值分别提升了0.51和1.06,在汉-英翻译上的BLEU值分别提升了1.04和0.40。可见Bi-Dependency为模型提供了更丰富的依存信息,能够有效提升翻译性能。

关键词: 神经机器翻译, 双向依存注意力, 多头注意力, 父词, 子词

CLC Number: