期刊文献+

基于集成分类的暴恐图像自动标注方法 被引量:1

Violent image annotation using ensemble learning
下载PDF
导出
摘要 为减少暴恐图像对社会发展和青少年成长造成的不利影响,本文提出一种基于集成分类的暴恐图像自动标注方法,辅助筛除网页中的暴恐信息。该方法将暴恐图像的标注视作多标签分类问题,利用迁移学习训练多个子网络,然后通过集成学习对子网络的输出进行融合,同时在融合过程中针对各个标签在不同网络上的准确率进行权重分配,最后经过一系列矩阵运算得到图像的标注结果。实验结果表明,与传统机器学习算法相比,本文方法在准确率和召回率上都有较大提升,并改善了样本不均衡所造成的不同标签类别上模型标注精确度差异较大的问题。 In order to reduce the negative impact of the horror image on social development and adolescent growth, a violent image annotation algorithm based on ensemble learning is proposed, assisting in screening out the horror information in the webpage. The annotation of violent image is considered as a multi-label classification problem in this method. Multiple sub-networks are trained through transfer learning, and then the ensemble learning is introduced to fuse the outputs of sub-networks. In the process of fusion, weights are allocated according to the precision of each label on different networks, thus the annotation result is obtained through a series of matrix operations. The experimental results show that the proposed method achieves a great improvement in precision and recall than traditional machine learning algorithm, and also improves the problem that the precision of model annotation on different labels varies greatly due to the label category imbalance.
作者 严靓 周欣 何小海 熊淑华 卿粼波 YAN Liang;ZHOU Xin;HE Xiaohai;XIONG Shuhua;QING Linbo(School of Electronic Information,Sichuan University,Chengdu Sichuan 610065,China)
出处 《太赫兹科学与电子信息学报》 北大核心 2020年第2期306-312,共7页 Journal of Terahertz Science and Electronic Information Technology
基金 国家自然科学基金资助项目(61871278) 四川省成都市产业集群协同创新资助项目(2016-XT00-00015-GX) 四川省科技计划资助项目(2018HH0143)。
关键词 图像标注 多标签分类 集成学习 权重分配 样本不均衡 image annotation multi-label classification ensemble learning weight allocation label category imbalance
  • 相关文献

参考文献3

二级参考文献58

  • 1路晶,金奕江,马少平,茹立云.使用基于SVM的否定概率和法的图像标注[J].智能系统学报,2006,1(1):62-66. 被引量:2
  • 2赵敏,陈恩红,宋睿.基于集成学习的Adaboost演化决策树算法[J].计算机应用与软件,2007,24(3):1-2. 被引量:4
  • 3芮晓光,袁平波,何芳,俞能海.一种新的基于语义聚类和图算法的自动图像标注方法[J].中国图象图形学报,2007,12(2):239-244. 被引量:9
  • 4LI Q, GU Y, QIAN X. LCMKL: latent-community and multl-kernel learning based image annotation [C] //Proceedings of the 22nd ACM International Confer- ence on Information & Knowledge Management. New York, USA: ACM, 2013: 1469-1472.
  • 5QIAN X, HUA X S, HOU X. Tag filtering based on similar compatible principle [C]//Proceedings of IEEE International Conference on Image Processing. Piscat- away, NJ, USA: IEEE, 2012: 2349-2352.
  • 6QIAN X, HUA X S, TANG Y Y, et al. Social image tagging with diverse semantics [J]. IEEE Transactions on Cybernetics, 2014, 44(12): 2493-2508.
  • 7NGIAM J, KHOSLA A, KIM M, et al. Multimodal deep learning [C] // Proceedings of the 28th Interna- tional Conference on Machine Learning. New York, USA: ACM, 2011: 689-696.
  • 8OUYANG W, CHU X, WANG X. Multi-source deep learning for human pose estimation [C] // Proceedings of IEEE Conference on Computer Vision and Pattern Recognition. Piscataway, NJ, USA: IEEE, 2014: 2337-2344.
  • 9KIROS R, ZEMEL R, SALAKHUTDINOV R. Mul- timodal neural language models [J]. Journal of Ma- chine Learning Research, 2014, 32(1): 595-603.
  • 10SRIVASTAVA N, SALAKHUTDINOV R. Multi- modal learning with deep Boltzmann machines [C]// Proceedings of Advances in Neural Information Pro- cessing Systems. Cambridge, MA, USA: MIT, 2012: 2222-2230.

共引文献20

同被引文献6

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部