期刊文献+

基于多模态表示学习的情感分析框架 被引量:2

Sentiment Analysis Framework Based on Multimodal Representation Learning
下载PDF
导出
摘要 在多模态表示对整体损失的学习过程中,重构损失对模型的依赖性相对较小,导致隐含表示无法有效捕捉它们各自模态的细节。文中提出了一个基于多模态表示学习的多子空间情感分析框架。首先将每个模态投射到模态不变和模态特定两种不同的话语表示中,在模态不变表示中构建主共享子空间以及帮助该子空间减少模态差距的辅助共享子空间,在模态特定表示中构建私有子空间以捕获每个模态独有的特征,将所有子空间中的隐藏向量作为解码函数的输入并重构模态向量,以实现对重构损失的优化。然后,在融合阶段对每个模态表示执行基于Transformer的自注意力,使每个表示能从对整体情感取向具有协同作用的其他跨模态表示中获取潜在信息。最后,通过串联生成联合向量并利用全连接层生成任务预测。在两个公开数据集MOSI和MOSEI上的实验结果表明,该框架在大多数评价指标上都优于基线模型。 In the process of learning the overall loss of multimodal representations,the dependence of reconstruction loss on the model is relatively less,resulting in hidden representations that cannot effectively capture the details of their respective modalities.This paper proposes a multi-subspace sentiment analysis framework.Firstly,the framework projects each modality to two distinct utterance representations:modality-invariant and modality-specific.We construct the main shared subspace and the auxiliary shared subspace that helps the main subspace to reduce the modality gap in the modality-invariant representation.Also,construct the private subspaces in the modality-specific representation to capture the characteristic features of each modality.We take the hidden vectors in all subspaces as the input of the decoder function and reconstruct the modal vector to achieve optimization of reconstruction loss.Secondly,in the fusion procedure,we perform a multi-headed self-attention based on Transformer on these representations,so that each cross-modal representation can induce potential information from fellow representations that have a synergistic effect on the overall emotional orientation.Finally,we construct a joint-vector by using concatenation and use fully connected layers to generate task predictions.Experimental results on both MOSI and MOSEI datasets show that the proposed framework outperforms the baselines in most evaluation criteria.
作者 胡新荣 陈志恒 刘军平 彭涛 叶鹏 朱强 HU Xin-rong;CHEN Zhi-heng;LIU Jun-ping;PENG Tao;YE Peng;ZHU Qiang(Hubei Provincial Engineering Research Center for Intelligent Textile and Fashion,Wuhan Textile University,Wuhan 430200,China;Engineering Research Center of Hubei Province for Clothing Information,Wuhan Textile University,Wuhan 430200,China;School of Computer Science and Artificial Intelligence,Wuhan Textile University,Wuhan 430200,China)
出处 《计算机科学》 CSCD 北大核心 2022年第S02期631-636,共6页 Computer Science
基金 国家自然科学基金(61103085) 湖北省高等学校优秀中青年科技创新团队计划项目(T201807) 湖北省高校知识产权推进工程项目(GXYS2018009) 湖北省教育厅科学研究计划重点项目(D20191708)
关键词 多模态表示 情感分析 TRANSFORMER 自注意力 跨模态 Multimodal representation Sentiment analysis Transformer Self-attention Cross-modality
  • 相关文献

同被引文献28

引证文献2

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部