摘要
图像描述任务旨在针对一张给出的图像产生其对应描述。针对现有算法中语义信息理解不够全面的问题,提出了一个针对图像描述领域的多模态Transformer模型。该模型在注意模块中同时捕捉模态内和模态间的相互作用;更进一步使用ELMo获得包含上下文信息的文本特征,使模型获得更加丰富的语义描述输入。该模型可以对复杂的多模态信息进行更好地理解与推断并且生成更为准确的自然语言描述。该模型在Microsoft COCO数据集上进行了广泛的实验,实验结果表明,相比于使用bottom-up注意力机制以及LSTM进行图像描述的基线模型具有较大的效果提升,模型在BLEU-1、BLEU-2、BLEU-3、BLEU-4、ROUGE-L、CIDEr-D上分别有0.7、0.4、0.9、1.3、0.6、4.9个百分点的提高。
The task of image caption is aim to generate the corresponding description of a given image.In order to solve the problem of incomplete understanding of semantic information in existing algorithms,a multimodal Transformer model for image description is proposed.In the attention module,the model captures the interaction within and between modes simultaneously,and further uses ELMo to obtain word embeddings which containing context information,so that the model can obtain more rich semantic description as input.This model can better understand and infer complex multimodal information and generate more accurate natural language description.The model has been widely tested on Microsoft COCO dataset,and the experimental results show that it has a great improvement compared with the baseline model using bottom-up attention and LSTM.The model has an improvement of 0.7,0.4,0.9,1.3,0.6,4.9 percentage points on BLEU-1,BLEU-2,BLEU-3,BLEU-4,ROUGE-L,CIDEr-D respectively.
作者
杨文瑞
沈韬
朱艳
曾凯
刘英莉
YANG Wenrui;SHEN Tao;ZHU Yan;ZENG Kai;LIU Yingli(Faculty of Information Engineering and Automation,Kunming University of Science and Technology,Kunming 650500,China;Yunnan Key Laboratory of Computer Technologies Application,Kunming University of Science and Technology,Kunming 650500,China)
出处
《计算机工程与应用》
CSCD
北大核心
2022年第21期223-231,共9页
Computer Engineering and Applications
基金
国家自然科学基金(61971208,61671225,52061020,61702128)
云南省应用基础研究计划项目重点项目(2018FA034)
昆明理工大学人才培养项目(KKSY201703016)。