Journal of Computer Applications ›› 2024, Vol. 44 ›› Issue (10): 3067-3073.DOI: 10.11772/j.issn.1001-9081.2023101407

• Artificial intelligence • Previous Articles     Next Articles

Large language model-driven stance-aware fact-checking

Yushan JIANG, Yangsen ZHANG()   

  1. Institute of Intelligent Information Processing,Beijing Information Science and Technology University,Beijing 100101,China
  • Received:2023-10-20 Revised:2024-02-19 Accepted:2024-02-21 Online:2024-10-15 Published:2024-10-10
  • Contact: Yangsen ZHANG
  • About author:JIANG Yushan, born in 1999, M. S. candidate. Her research interests include natural language processing, fact-checking.
  • Supported by:
    National Natural Science Foundation of China(62176023)

大语言模型驱动的立场感知事实核查

姜雨杉, 张仰森()   

  1. 北京信息科技大学 智能信息处理研究所,北京 100101
  • 通讯作者: 张仰森
  • 作者简介:姜雨杉(1999—),女,黑龙江黑河人,硕士研究生,CCF会员,主要研究方向:自然语言处理、事实核查
    张仰森(1962—),男,山西临猗人,教授,博士,CCF杰出会员,主要研究方向:自然语言处理、人工智能、Web内容安全 zhangyangsen@163.com
  • 基金资助:
    国家自然科学基金资助项目(62176023)

Abstract:

To address the issues of evidence stance imbalance and neglect of stance information in the field of Fact-Checking (FC), a Large Language Model-driven Stance-Aware fact-checking (LLM-SA) method was proposed. Firstly, a series of dialectical claims that differed from the original claim were generated by using a large language model, to capture different perspectives for fact-checking. Secondly, through semantic similarity calculations, the relevances of each piece of evidence sentence to the original claim and the dialectical claim were separately assessed, and the top k sentences with the highest semantic similarity to each of them were selected as the evidence to either support or oppose the original claim, which obtained evidences representing different stances, and helped the fact-checking model integrate information from multiple perspectives and evaluate the veracity of the claim more accurately. Finally, the BERT-StuSE (BERT-based Stance-infused Semantic Encoding network) model was introduced to fully incorporate the semantic and stance information of the evidence through the multi-head attention mechanism and make a more comprehensive and objective judgment on the relationship between the claim and the evidence. The experimental results on the CHEF dataset show that, compared to the BERT method, the Micro F1 value and Macro F1 value of the proposed method on the test set were improved by 3.52 and 3.90 percentage points, respectively, achieving a good level of performance. The experimental results demonstrate the effectiveness of the proposed method, and the significance of considering evidence from different stances and leveraging the stance information of the evidence for enhancing fact-checking performance.

Key words: Fact-Checking (FC), Natural Language Processing (NLP), Large Language Model (LLM), prompt engineering, stance awareness, multi-head attention mechanism

摘要:

为解决事实核查领域的证据立场不平衡和忽略立场信息的问题,提出一种大语言模型(LLM)驱动的立场感知事实核查(LLM-SA)方法。首先,使用LLM推理并生成一系列与原始声明立场不同的辩证声明,使事实核查任务能够获取不同立场的视角;其次,通过语义相似度计算衡量每个证据句子与原始声明及辩证声明之间的相关性,并从证据句子中分别选择与两者语义上最相近的k个句子,作为支持或反对原始声明的证据,从而获得代表不同立场的证据,有助于事实核查模型综合多角度的信息,更准确地评估声明的真实性;最后,引入BERT-StuSE(Bidirectional Encoder Representations from Transformers-based Stance-infused Semantic Encoding network)模型,利用多头注意力机制充分融合证据的语义和立场信息,并更全面、客观地判断声明和证据的关系。在CHEF数据集上的实验结果表明,与BERT方法相比,所提方法在测试集上的Micro F1值和Macro F1值分别提高了3.52、3.90个百分点,达到较好的水平。验证了所提方法的有效性,以及考虑不同立场的证据和充分利用证据的立场信息对事实核查的性能提升具有重要意义。

关键词: 事实核查, 自然语言处理, 大语言模型, 提示工程, 立场感知, 多头注意力机制

CLC Number: