《计算机应用》唯一官方网站 ›› 2025, Vol. 45 ›› Issue (5): 1511-1519.DOI: 10.11772/j.issn.1001-9081.2024050675

• 人工智能 • 上一篇    

基于全域信息融合和多维关系感知的命名实体识别模型

胡婕1,2,3, 武帅星1, 曹芝兰1,2,3(), 张龑1,2,3   

  1. 1.湖北大学 计算机学院,武汉 430062
    2.大数据智能分析与行业应用湖北省重点实验室(湖北大学),武汉 430062
    3.智慧政务与人工智能应用湖北省工程研究中心(湖北大学),武汉 430062
  • 收稿日期:2024-05-27 修回日期:2024-08-09 接受日期:2024-08-30 发布日期:2024-09-05 出版日期:2025-05-10
  • 通讯作者: 曹芝兰
  • 作者简介:胡婕(1977—),女,湖北汉川人,教授,博士,主要研究方向:复杂语义大数据管理、自然语言处理
    武帅星(2000—),男,河南安阳人,硕士研究生,主要研究方向:自然语言处理
    曹芝兰(1971—),女,湖北麻城人,讲师,硕士,主要研究方向:自然语言处理
    张龑(1974—),男,湖北宜昌人,教授,博士,CCF会员,主要研究方向:软件工程、信息安全。
  • 基金资助:
    国家自然科学基金资助项目(61977021)

Named entity recognition model based on global information fusion and multi-dimensional relation perception

Jie HU1,2,3, Shuaixing WU1, Zhilan CAO1,2,3(), Yan ZHANG1,2,3   

  1. 1.School of Computer Science,Hubei University,Wuhan Hubei 430062,China
    2.Hubei Key Laboratory of Big Data Intelligent Analysis and Application (Hubei University),Wuhan Hubei 430062,China
    3.Engineering Research Center of Hubei Province in Intelligent Government Affairs and Application of Artificial Intelligence (Hubei University),Wuhan Hubei 430062,China
  • Received:2024-05-27 Revised:2024-08-09 Accepted:2024-08-30 Online:2024-09-05 Published:2025-05-10
  • Contact: Zhilan CAO
  • About author:HU Jie, born in 1977, Ph. D., professor. Her research interests include complex semantic big data management, natural language processing.
    WU Shuaixing, born in 2000, M. S. candidate. His research interests include natural language processing.
    CAO Zhilan, born in 1971, M. S., lecturer. Her research interests include natural language processing.
    ZHANG Yan, born in 1974, Ph. D., professor. His research interests include software engineering, information security.
  • Supported by:
    National Natural Science Foundation of China(61977021)

摘要:

现有的基于双向长短时记忆(BiLSTM)网络的命名实体识别(NER)模型难以全面理解文本的整体语义以及捕捉复杂的实体关系。因此,提出一种基于全域信息融合和多维关系感知的NER模型。首先,通过BERT (Bidirectional Encoder Representations from Transformers)获取输入序列的向量表示,并结合BiLSTM进一步学习输入序列的上下文信息。其次,提出由梯度稳定层和特征融合模块组成的全域信息融合机制:前者使模型保持稳定的梯度传播并更新优化输入序列的表示,后者则融合BiLSTM的前后向表示获取更全面的特征表示。接着,构建多维关系感知结构学习不同子空间单词的关联性,以捕获文档中复杂的实体关系。此外,使用自适应焦点损失函数动态调整不同类别实体的权重,提高模型对少数类实体的识别性能。最后,在7个公开数据集上将所提模型和11个基线模型进行对比,实验结果表明所提模型的F1值均优于对比模型,可见该模型的综合性较优。

关键词: 命名实体识别, 全域信息融合机制, 梯度稳定层, 多维关系感知, 自适应焦点损失

Abstract:

The existing Named Entity Recognition (NER) models based on Bidirectional Long Short-Term Memory (BiLSTM) network are difficult to fully understand the global semantics of text and capture the complex relationships between entities. Therefore, an NER model based on global information fusion and multi-dimensional relation perception was proposed. Firstly, BERT (Bidirectional Encoder Representations from Transformers) was used to obtain vector representation of the input sequence, and BiLSTM was combined to further learn context information of the input sequence. Secondly, a global information fusion mechanism composed of gradient stabilization layer and feature fusion module was proposed. With the former one, the model was able to maintain stable gradient propagation and update as well as optimize representation of the input sequence. In the latter one, the forward and backward representations of BiLSTM were integrated to obtain more comprehensive feature representation. Thirdly, a multi-dimensional relation perception structure was constructed to learn correlations between words in different subspaces in order to capture complex entity relationships in documents. In addition, the adaptive focus loss function was used to adjust the weights of different entity types dynamically to improve the recognition performance of the model for minority entities. Finally, experiments were conducted on 7 public datasets for the proposed model and 11 baseline models. The results show that all of the F1 values of the proposed model are higher than those of the comparison models, validating the comprehensive performance of the proposed model.

Key words: Named Entity Recognition (NER), global information fusion mechanism, gradient stabilization layer, multi-dimensional relation perception, adaptive focus loss

中图分类号: