《计算机应用》唯一官方网站 ›› 2023, Vol. 43 ›› Issue (2): 365-373.DOI: 10.11772/j.issn.1001-9081.2021122167

• 人工智能 • 上一篇    

基于池化和特征组合增强BERT的答案选择模型

胡婕1,2(), 陈晓茜1, 张龑1,2   

  1. 1.湖北大学 计算机与信息工程学院,武汉 430062
    2.湖北省教育信息化工程技术研究中心(湖北大学),武汉 430062
  • 收稿日期:2021-12-29 修回日期:2022-06-04 接受日期:2022-06-10 发布日期:2022-06-30 出版日期:2023-02-10
  • 通讯作者: 胡婕
  • 作者简介:陈晓茜(1997—),女,河南平顶山人,硕士研究生,主要研究方向:自然语言处理
    张龑(1974—),男,湖北宜昌人,教授,博士,CCF会员,主要研究方向:软件工程、信息安全。
  • 基金资助:
    国家自然科学基金资助项目(61977021)

Answer selection model based on pooling and feature combination enhanced BERT

Jie HU1,2(), Xiaoxi CHEN1, Yan ZHANG1,2   

  1. 1.School of Computer Science and Information Engineering,Hubei University,Wuhan Hubei 430062,China
    2.Hubei Engineering Technology Research Center for Educational Informatization (Hubei University),Wuhan Hubei 430062,China
  • Received:2021-12-29 Revised:2022-06-04 Accepted:2022-06-10 Online:2022-06-30 Published:2023-02-10
  • Contact: Jie HU
  • About author:CHEN Xiaoxi, born in 1997, M. S. candidate. Her research interests include natural language processing.
    ZHANG Yan, born in 1974, Ph. D., professor. His research interests include software engineering, information security.
  • Supported by:
    National Natural Science Foundation of China(61977021)

摘要:

当前主流模型无法充分地表示问答对的语义,未充分考虑问答对主题信息间的联系并且激活函数存在软饱和的问题,而这些会影响模型的整体性能。针对这些问题,提出了一种基于池化和特征组合增强BERT的答案选择模型。首先,在预训练模型BERT的基础上增加对抗样本并引入池化操作来表示问答对的语义;其次,引入主题信息特征组合来加强问答对主题信息间的联系;最后,改进隐藏层的激活函数,并用拼接向量通过隐藏层和分类器完成答案选择任务。在SemEval-2016CQA和SemEval-2017CQA数据集上进行的验证结果表明,所提模型与tBERT模型相比,准确率分别提高了3.1个百分点和2.2个百分点;F1值分别提高了2.0个百分点和3.1个百分点。可见,所提模型在答案选择任务上的综合效果得到了有效提升,准确率和F1值均优于对比模型。

关键词: 答案选择, 预训练模型, 池化, 特征组合, 激活函数

Abstract:

Current main stream models cannot fully express the semantics of question and answer pairs, do not fully consider the relationships between the topic information of question and answer pairs, and the activation function has the problem of soft saturation, which affect the overall performance of the model. To solve these problems, an answer selection model based on pooling and feature combination enhanced BERT (Bi-directional Encoder Representations from Transformers) was proposed. Firstly, adversarial samples and pooling operation were introduced to represent the semantics of question and answer pairs based on the pre-training model BERT. Secondly, the relationships between topic information of question and answer pairs were strengthened by the feature combination of topic information. Finally, the activation function in the hidden layer was improved, and the splicing vector was used to complete the answer selection task through the hidden layer and classifier. Model validation was performed on datasets SemEval-2016CQA and SemEval-2017CQA. The results show that compared with tBERT model, the proposed model has the accuracy increased by 3.1 percentage points and 2.2 percentage points respectively, F1 score increased by 2.0 percentage points and 3.1 percentage points respectively. It can be seen that the comprehensive effect of the proposed model on the answer selection task is effectively improved, and both of the accuracy and F1 score of the model are better than those of the model for comparison.

Key words: answer selection, pre-training model, pooling, feature combination, activation function

中图分类号: