《计算机应用》唯一官方网站

• •    下一篇

基于改进的提示学习方法的双通道情感分析模型

沈君凤,周星辰,汤灿   

  1. 湖北大学
  • 收稿日期:2023-06-07 修回日期:2023-08-13 发布日期:2023-09-14 出版日期:2023-09-14
  • 通讯作者: 沈君凤

Dual-channel sentiment analysis model based on improved learning method

  • Received:2023-06-07 Revised:2023-08-13 Online:2023-09-14 Published:2023-09-14

摘要: 针对先前提示学习方法中存在的模板迭代更新周期长,泛化能力下降等问题,该文基于改进的提示学习方法提出一种双通道的情感分析模型。该模型首先在提示学习方法中将序列化后的提示模板同输入词向量一起引入到注意力机制结构中,随着输入词向量在多层注意力机制中更新的同时迭代更新提示模板。然后在另一通道采用ALBERT(A Lite BERT)模型来提取语义信息,最后采用集成的方式将提取出来的语义特征输出,以提升整体模型的泛化能力。该文模型在SemEval2014的Laptop和Restaurants数据集、ACL(Association for Computational Linguistics)的Twitter数据集以及斯坦福大学创建的SST-2数据集上进行实验,分类准确率达到80.88%,91.78%,76.78%,95.53%与基线模型Bert Large相比分别有 0.99%,1.13%,3.39%,2.84% 的提升。与改进前方法的模型相比在Restaurants数据集、Twitter数据集以及SST-2数据集上分别有2.88%,3.60%,2.06%的提升,且比原方法更早达到收敛状态。

Abstract: Aiming at the problems of long template iterative update cycle and reduced generalization ability in the previous learning method, A dual-channel sentiment analysis model was proposed based on the improved based on the improved learning method. In the learning method, the serialized templates were firstly introduced into the attention mechanism structure, and meanwhile, the templates were iteratively updated as the input word vectors were updated in the multi-layer attention mechanism. Then the semantic information was extracted by the ALBERT (A Lite BERT) model in another channel, and the extracted semantic features were proposed in an integrated way to improve the generalization ability of the overall model. The model was tested on the SemEval2014 Laptop and Restaurants datasets, the ACL’s (Association for Computational Linguistics) Twitter dataset, and the SST-2 dataset created by Stanford University. The classification accuracy reached 80.88%, 91.78%, 76.78%, 95.53% and compared with the baseline model Bert Large, there are 0.99%, 1.13%, 3.39%, and 2.84% improvements respectively. Compared with the model of the method before improvement, it has 2.88%, 3.60%, and 2.06% improvements on the Restaurants dataset, Twitter dataset, and SST-2 dataset respectively, and reaches the convergence state earlier than the original method.

中图分类号: