Journal of Computer Applications ›› 2024, Vol. 44 ›› Issue (3): 690-695.DOI: 10.11772/j.issn.1001-9081.2023040443

• Artificial intelligence • Previous Articles     Next Articles

Joint approach of intent detection and slot filling based on multi-task learning

Aiguo SHANG1,2, Xinjuan ZHU1,2()   

  1. 1.Shaanxi Key Laboratory of Clothing Intelligence (Xi’an Polytechnic University),Xi’an Shaanxi 710600,China
    2.School of Computer Science,Xi’an Polytechnic University,Xi’an Shaanxi 710600,China
  • Received:2023-04-18 Revised:2023-06-08 Accepted:2023-06-09 Online:2023-12-04 Published:2024-03-10
  • Contact: Xinjuan ZHU
  • About author:SHANG Aiguo, born in 1999,M. S. candidate. His research interests include spoken language understanding, natural language processing.
  • Supported by:
    National Key Research and Development Program of China(2019YFC1521405);Graduate Scientific Innovation Fund for Xi’an Polytechnic University(chx2022023)

基于多任务学习的意图检测和槽位填充联合方法

尚爱国1,2, 朱欣娟1,2()   

  1. 1.陕西省服装设计智能化重点实验室(西安工程大学),西安 710600
    2.西安工程大学 计算机科学学院,西安 710600
  • 通讯作者: 朱欣娟
  • 作者简介:尚爱国(1999—),男,陕西西安人,硕士研究生,主要研究方向:口语理解、自然语言处理
  • 基金资助:
    国家重点研发计划项目(2019YFC1521405);西安工程大学研究生创新基金资助项目(chx2022023)

Abstract:

With the application of pre-trained language models in Natural Language Processing (NLP) tasks, joint modeling of Intent Detection (ID) and Slot Filling (SF) has improved the performance of Spoken Language Understanding (SLU). Existing methods mostly focus on the interaction between intents and slots, neglecting the influence of modeling differential text sequences on SLU tasks. A joint method for Intent Detection and Slot Filling based on Multi-task Learning (IDSFML) was proposed. Firstly, differential texts were constructed using random mask strategy, and a neural network structure combining AutoEncoder and Attention mechanism (AEA) was designed to incorporate the features of differential text sequences into the SLU task. Secondly, a similarity distribution task was designed to make the representations of differential texts and original texts similar. Finally, three tasks of ID, SF and differential text sequence similarity distribution were jointly trained. Experimental results on Airline Travel Information Systems (ATIS) and SNIPS datasets show that, compared with the suboptimal baseline method SASGBC (Self-Attention and Slot-Gated on top of BERT with CRF), IDSFML improves the F1 scores of slot filling by 1.9 and 1.6 percentage points respectively, and improves the accuracy of intent detection by 0.2 and 0.4 percentage points respectively, enhancing the accuracy of spoken language understanding tasks.

Key words: Intent Detection (ID), Slot Filling (SF), multi-task learning, Spoken Language Understanding (SLU), attention mechanism

摘要:

随着预训练语言模型在自然语言处理(NLP)任务上的应用,意图检测(ID)和槽位填充(SF)联合建模提高了口语理解的性能。现有方法大多关注意图和槽位的相互作用,忽略了差异文本序列建模对口语理解(SLU)任务的影响。因此,提出一种基于多任务学习的意图检测和槽位填充联合方法(IDSFML)。首先,使用随机掩盖mask策略构造差异文本,设计结合自编码器和注意力机制的神经网络(AEA)结构,为口语理解任务融入差异文本序列的特征;其次,设计相似性分布任务,使差异文本和原始文本的表征相似;最后,联合训练ID、SF和差异文本序列相似性分布三个任务。在航班旅行信息系统(ATIS)和SNIPS数据集上的实验结果表明,IDSFML与表现次优的基线方法SASGBC(Self-Attention and Slot-Gated on top of BERT with CRF)相比,槽位填充F1值分别提升了1.9和1.6个百分点,意图检测准确率分别提升了0.2和0.4个百分点,提高了口语理解任务的准确率。

关键词: 意图检测, 槽位填充, 多任务学习, 口语理解, 注意力机制

CLC Number: