期刊文献+

基于多尺度跨模态特征融合的图文情感分类模型

Image-Text Sentiment Classification Model Based on Multi-scale Cross-modal Feature Fusion
下载PDF
导出
摘要 图文情感分类任务常用早期融合和Transformer模型相结合的跨模态特征融合策略进行图文特征融合,但该策略更倾向于关注模态内部的独有信息,而忽略了模态间的相互联系和共有信息,导致跨模态特征融合效果不理想。针对此问题,提出一种基于多尺度跨模态特征融合的图文情感分类方法。局部尺度方面,基于跨模态注意力机制进行局部特征融合,使模型不仅关注图像和文本的独有信息,而且可以发现图像和文本之间的联系和共有信息。全局尺度方面,基于MLM损失进行全局特征融合,使模型对图像和文本数据进行全局建模,进一步挖掘图像和文本之间的联系,从而促进图像和文本特征的深度融合。在两个公开数据集MVSA-Single和MVSA-Multiple上与10个基线模型进行对比实验,结果表明所提方法在精度、F1值和模型参数量方面均具有明显优势,验证了其有效性。 For the image-text sentiment classification task,the cross-modal feature fusion strategy which combines early fusion and Transformer model is usually used for image-text feature fusion.However,this strategy prefers to focus on the unique information within a single modality,while ignoring the interconnections and common information among multiple modalities,resulting in unsatisfactory effect of cross-modal feature fusion.To solve this problem,a method of image-text classification based on multi-scale cross-modal feature fusion is proposed.On the one hand,for the local scale,local feature fusion is carried out based on the cross-modal attention mechanism,so that the model not only focuses onthe unique information of the image and text,but also explores the connection and common information between the image and text.On the other hand,for the global scale,global feature fusion based on MLM loss enables the model to conduct global modeling of image and text data,further mine the relationship between them,and thus promote the deep fusion of image and text features.Compared with ten baseline models on two public datasets,MVSA-Single and MVSA-Multiple,the proposed method shows distinct advantages in accuracy,F1 score,and model para-meter quantity,verifying its effectiveness.
作者 刘倩 白志豪 程春玲 归耀城 LIU Qian;BAI Zhihao;CHENG Chunling;GUI Yaocheng(School of Computer Science,Nanjing University of Posts and Telecommunications,Nanjing 210023,China;School of Modern Posts,Nanjing University of Posts and Telecommunications,Nanjing 210023,China)
出处 《计算机科学》 CSCD 北大核心 2024年第9期258-264,共7页 Computer Science
基金 江苏省双创博士项目(JSSCBS20210507)。
关键词 图文情感分类 跨模态特征融合 Transformer模型 注意力机制 MLM损失 Image-Text sentiment classification Cross-modal feature fusion Transformer model Attention mechanism MLM loss
  • 相关文献

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部