Journal of Computer Applications

    Next Articles

Multimodal harmful content detection method based on weakly supervised modality semantic enhancement

  

  • Received:2024-10-12 Revised:2024-12-05 Accepted:2024-12-09 Online:2024-12-23 Published:2024-12-23
  • Contact: lei initialwang

基于弱监督模态语义增强的多模态有害信息检测方法

刘晋文1,2,3, 王磊1,2,3*, 马博1,2,3, 董瑞1,2,3, 杨雅婷1,2,3, 艾合塔木江·艾合麦提1,2,3, 王欣乐4   

  1. 1.中国科学院 新疆理化技术研究所,乌鲁木齐 830011; 2.中国科学院大学,北京 100049; 3.中国科学院 新疆民族语音语言信息处理实验室,乌鲁木齐 830011; 4.河海大学,南京 210000

  • 通讯作者: 王磊
  • 基金资助:
    新疆维吾尔自治区自然科学基金重点项目;“天山英才”科技创新领军人才项目;新疆维吾尔自治区重点研发计划项目;新疆维吾尔自治区重点研发计划项目;中国科学院青年创新促进会资助;新疆维吾尔自治区“天山英才”培养计划

Abstract: The proliferation of multimodal harmful information on social media not only undermines public interests but also severely disrupts social order, highlighting the urgent need for effective detection methods. Existing approaches have predominantly relied on pre-trained models to extract and integrate multimodal features, often neglecting the limitations of general semantics in harmful information detection and the complex, dynamic combinations of harmful content. To address these issues, a multimodal harmful content detection framework based on weakly Supervised modality semantic enhancement (weak-S) was introduced. Weakly supervised modality information was utilized to facilitate the harmful semantic alignment of multimodal features, and a low-rank bilinear pooling-based multimodal gated integration mechanism was designed to differentiate the contributions of various information sources. Experimental results show that the proposed method achieves F1-score improvements of 2.2 and 3.2 percentage points on the HarmP and MultiOFF benchmark datasets, respectively, outperforming SOTA (State-Of-The-Art) models and validating the significance of weakly supervised modality semantics in multimodal harmful information detection. Additionally, the method delivers a 1-percentage-point improvement in generalization performance for multimodal exaggeration detection tasks.

Key words: unimodal weak supervision, contrastive learning, gated integration, multimodal, harmful content detection

摘要: 社交媒体上多模态有害信息的泛滥,不仅侵害公众利益,还严重扰乱社会秩序,亟需有效的检测方法。现有研究依赖预训练模型提取与融合多模态特征,忽视了通用语义在有害信息检测任务中的局限性,且未能充分考虑有害信息复杂多变的组合形式。为此,提出一种基于弱监督模态语义增强的多模态有害信息检测方法(weak-S),所提方法通过引入弱监督模态信息辅助多模态特征的有害语义对齐,并设计一种低秩双线性池化的多模态门控集成机制,以区分不同信息的贡献度。实验结果表明,所提方法在HarmP和MuitiOFF等公开基准数据集上的F1值相较于SOTA(State-Of-The-Art)模型分别提高了2.2和3.2个百分点,验证了弱监督模态语义在多模态有害信息检测中的重要性。此外,所提方法还在多模态夸张检测任务上取得了1个百分点的泛化性能提升。

关键词: 单模态弱监督, 对比学习, 门控集成, 多模态, 有害信息检测

CLC Number: