《计算机应用》唯一官方网站

• •    下一篇

基于掩码提示与门控记忆网络校准的关系抽取方法

魏超1,陈艳平1,王凯1,秦永彬2,黄瑞章2   

  1. 1. 贵州大学计算机科学与技术学院
    2. 贵州大学
  • 收稿日期:2023-06-26 修回日期:2023-08-16 发布日期:2023-08-30 出版日期:2023-08-30
  • 通讯作者: 魏超
  • 基金资助:
    国家自然科学基金资助项目

Relation extraction method based on mask and gated memory network calibration

  • Received:2023-06-26 Revised:2023-08-16 Online:2023-08-30 Published:2023-08-30

摘要: 摘 要: 针对关系抽取任务中实体关系语义挖掘困难和预测关系有偏差等问题,提出一种基于掩码提示与门控神经网络校准的关系抽取方法。该方法利用提示中的掩码学习实体之间在预训练语言模型语义空间中的潜在语义,通过构造掩码注意力权重矩阵,将离散的掩码语义空间相互关联,并采用门控校准网络将含有实体及关系语义的掩码表示融入到句子的全局语义中,将其作为关系提示对关系信息进行校准,随后将句子表示的最终表示映射到相应的关系类别上。该方法更好地利用提示中掩码的作用,并结合传统微调方法的学习句子全局语义的优势,充分激发了预训练语言模型的潜力。实验结果表明,该方法在SemEval(SemEval-2010 Task 8)数据集的F1值达到91.4%,相较于RELA生成式方法提高了1个百分点,相较于RELA(Relation Extraction with Label Augmentation)少了大约10%的参数量;在SciERC(Entities, Relations, and Coreference for Scientific)和CLTC(Chinese Literature Text Corpus)数据集上的抽取性能分别达到91.0%和82.8%。该方法在上述三个数据集上均明显优于对比的方法,并且实现了当前最先进的性能,证明了该方法的有效性。相比基于生成式的方法,该方法以较少的参数量实现了更优的抽取性能。

关键词: 关系抽取, 掩码, 门控神经网络, 预训练语言模型, 提示学习

Abstract: Abstract: To tackle the difficulty of semantic mining of entity relations and biased relation prediction in relation extraction tasks, a relation extraction method based on mask and gated memory network calibration was proposed. First of all, the latent semantics between entities within the pre-training language model semantic space were effectively learned through the utilization of masks in s. By constructing a mask attention weight matrix, the discrete masked semantic spaces were interconnected. And then, the gate-controlled calibration networks were incorporated to integrate the masked representations, which contain entity and relation semantics. Besides, these calibrated representations served as s to adjust the relation information, and the final representation of the calibrated sentence was mapped to the corresponding relation class. Finally, the potential of pre-training language models was fully exploited by the proposed approach through harnessing the power of masks in s and combining it with the advantages of traditional fine-tuning methods.The experimental results highlight the effectiveness of the proposed method. On the SemEval (SemEval-2010 Task 8) dataset, the F1 score reaches an impressive 91.4%, outperforming the RELA (Relation Extraction with Label Augmentation) generative method by 1 percentage point and reducing the parameter count by approximately 10% compared to RELA. Additionally, the extraction performances on the SciERC (Entities, Relations, and Coreference for Scientific) and CLTC (Chinese Literature Text Corpus) datasets are remarkable, achieving 91.0% and 82.8% respectively. The effectiveness of the proposed method is evident as it consistently outperforms the comparative methods on all three datasets mentioned above, establishing itself as the current state-of-the-art approach. Furthermore, superior extraction performance is achieved by this method compared to generative methods, while utilizing fewer parameters.

Key words: relation extraction, mask, gated neural network, pre-trained language model, tuning

中图分类号: