《计算机应用》唯一官方网站 ›› 2023, Vol. 43 ›› Issue (S1): 67-74.DOI: 10.11772/j.issn.1001-9081.2022050710

• 人工智能 • 上一篇    下一篇

面向法律判决文书的长文档抽取式文摘方法——BIGDCNN

赵嘉昕1,2, 崔喆1()   

  1. 1.中国科学院 成都计算机应用研究所,成都 610041
    2.中国科学院大学 计算机科学与技术学院,北京 100049
  • 收稿日期:2022-05-18 修回日期:2022-07-21 接受日期:2022-07-21 发布日期:2023-07-04 出版日期:2023-06-30
  • 通讯作者: 崔喆
  • 作者简介:赵嘉昕(1998—),男,四川成都人,硕士研究生,CCF会员,主要研究方向:自然语言处理
    崔喆(1970—),男,四川巴中人,研究员,博士,主要研究方向:可信计算、信息安全。cuizhe@casit.com.cn
  • 基金资助:
    四川省科技计划项目(2020YFG0009);四川省重大科技专项(2019ZDZX0005)

BIGDCNN: Extractive summarization method for long legal judgment documents

Jiaxin ZHAO1,2, Zhe CUI1()   

  1. 1.Chengdu Institute of Computer Application,Chinese Academy of Sciences,Chengdu Sichuan 610041,China
    2.School of Computer Science and Technology,University of Chinese Academy of Sciences,Beijing 100049,China
  • Received:2022-05-18 Revised:2022-07-21 Accepted:2022-07-21 Online:2023-07-04 Published:2023-06-30
  • Contact: Zhe CUI

摘要:

针对法律判决文书信息点较多、结构化程度较高,传统的抽取式文摘方法容易产生冗余句子且无法覆盖全部关键信息的问题,提出BIGDCNN(BERT based Improved Gate Dilated Convolutional Neural Network)模型。首先将原始数据进行语料转换获取序列标注数据,再通过预训练语言模型BERT(Bidirectional Encoder Representations from Transformers)得到从词粒度到句子粒度的长文本表示;最后使用融合了改进门机制的膨胀卷积神经网络(DCNN)以及单模型融合方法,实现低冗余度提取原文关键信息的同时增强抗干扰性,并减小了梯度消失的风险。在法律判决文书自动文摘实验中,本模型的ROUGE-1、ROUGE-2、ROUGE-L评分为62.85%、46.56%、59.25%,较主流模型BERT+Transformer分别提升了15.10、15.75、12.97个百分点。BIGDCNN模型解决了传统抽取式文摘方法的问题,可以高效地运用在法律判决文书的自动文摘场景中。

关键词: 判决文书, 抽取式文摘, 预训练语言模型, 门机制, 单模型融合

Abstract:

In view of the large number of information points and the high degree of structure of legal judgment documents, the traditional extractive summarization method is prone to produce redundant sentences and cannot cover all the key information. A BERT (Bidirectional Encoder Representations from Transformers) based Improved Gate Dilated Convolutional Neural Network (BIGDCNN) model was proposed to deal with the above problems. Firstly, the original data was converted to sequence annotation data, then long text representation from word granularity to sentence granularity was obtained by pre-trained language model BERT. Finally, the Dilated Convolutional Neural Network (DCNN) combined with the improved gate mechanism and a single model fusion method were used to achieve low-redundancy extraction of key information in the original document while enhancing anti-interference and reducing the risk of gradient disappearance. In the automatic summarization experiment of legal judgment documents, the ROUGE-1, ROUGE-2, and ROUGE-L scores of this model were 62.85%, 46.56%, and 59.25%, which were 15.10, 15.75, and 12.97 percentage points higher than those of the mainstream model BERT+Transformer. The BIGDCNN model solves the problem of the traditional extractive summarization method, and can be efficiently used in the automatic summarization scene of legal judgment documents.

Key words: judgment documents, extractive summarization, pre-trained language model, gate mechanism, single model fusion

中图分类号: