期刊文献+

VLP:A Survey on Vision-language Pre-training 被引量:3

原文传递
导出
摘要 In the past few years,the emergence of pre-training models has brought uni-modal fields such as computer vision(CV)and natural language processing(NLP)to a new era.Substantial works have shown that they are beneficial for downstream uni-modal tasks and avoid training a new model from scratch.So can such pre-trained models be applied to multi-modal tasks?Researchers have ex-plored this problem and made significant progress.This paper surveys recent advances and new frontiers in vision-language pre-training(VLP),including image-text and video-text pre-training.To give readers a better overall grasp of VLP,we first review its recent ad-vances in five aspects:feature extraction,model architecture,pre-training objectives,pre-training datasets,and downstream tasks.Then,we summarize the specific VLP models in detail.Finally,we discuss the new frontiers in VLP.To the best of our knowledge,this is the first survey focused on VLP.We hope that this survey can shed light on future research in the VLP field.
出处 《Machine Intelligence Research》 EI CSCD 2023年第1期38-56,共19页 机器智能研究(英文版)
基金 supported by the Key Research Program of the Chinese Academy of Sciences(No.ZDBSSSW-JSC006) the Strategic Priority Research Program of the Chinese Academy of Sciences(No.XDA 27030300).
  • 相关文献

同被引文献10

引证文献3

二级引证文献3

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部