期刊文献+

一种机器学习中防止过拟合的Dropout优化算法 被引量:9

A Dropout Optimization Algorithm for Preventing Overfitting in Machine Learning
下载PDF
导出
摘要 针对机器学习中深度神经网络训练时常见的过拟合问题,提出了一种防止过拟合的Dropout优化算法.Dropout算法是在每批次的神经网络训练中,忽略掉一定概率的特征检测器,让某些神经元暂时停止工作,减少神经元之间的相互作用,隐式去除网络中的神经元、阻止某些特征的协同作用来缓解过拟合.算法中选择被暂时丢弃的神经元是随机概率,而优化算法在神经网络中应用伊辛模型来识别链接能量较低的神经元,并在训练和推理中把这些神经元暂时丢弃掉,算法使模型泛化性更强,有效缓解网络训练过拟合问题. Aiming at the common over-fitting problem of deep neural network training in machine learning,a Dropout optimization algorithm to prevent over-fitting is proposed.The Dropout algorithm neglects some probabilistic feature detectors in each batch of neural network training,makes some neurons stop working temporarily,reduces the interaction between neurons,implicitly removes the neurons in the network,and prevents the synergy of some features to alleviate over-fitting.The selection of temporarily discarded neurons in the algorithm is random probability,while the optimization algorithm applies Ising model to identify neurons with lower link energy in the neural network,and discards these neurons temporarily in training and reasoning.The algorithm makes the model more generalized and effectively alleviates the problem of over-fitting in network training.
作者 张云 李岚 王浩东 ZHANG Yun;LI Lan;WANG Hao-dong(School of Digital Medial,Lanzhou University of Arts and Science,Lanzhou 730000,China;Nanjing Post Distribution Center of China Post Express Logistics,Nanjing 210003,China)
出处 《兰州文理学院学报(自然科学版)》 2019年第6期84-87,共4页 Journal of Lanzhou University of Arts and Science(Natural Sciences)
基金 甘肃省高等学校科研项目(2016A-105) 甘肃省大学生创新创业训练计划项目(201611562017) 2017年教育部高教司第二批产学研协同育人项目“基于项目驱动的Java系列课程实践教学改革研究”(201702163024) 2019年甘肃省创新创业项目“基于‘多维度创新’的数字媒体技术专业教学体系研究与构建”
关键词 深度神经网络 过拟合 神经元 机器学习 伊辛模型 deep neural network over-fitting neurons machine learning Ising model
  • 相关文献

参考文献6

二级参考文献63

  • 1Rumelhart D E, Hinton G E,Williams R J.Learning rep- resentations by back-propagating errors [ J ]. Nature, 1986,323 (6088) : 533-536.
  • 2Bengio Y.Deep Learning of Representations:looking for- ward [ J]. Lecture Notes in Computer Science, 2013, 7978:1-37.
  • 3Bottou L, Bengio Y, Cun Y L.Global training of document processing systems using graph transformer networks [C3//Proceedings of the IEEE Computer Society Con- ference on Computer Vision and Pattern Recognition.San Juan, Puerto Rico : IEEE, 1997:489-494.
  • 4Hinton G E, Salakhutdinov R R. Reducing the dimension- ality of data with neural networks [ J ].Science, 2006,313 (5786) :504-507.
  • 5Bengio Y.Learning deep architectures for AI[ M ]. Hano- ver: The Association for Computing Machinery ,2009.
  • 6Lecun Y, Bottou L, Bengio Y, et al. Gradient based learning applied to document recognition [ J ].Proceedings of the IEEE, 1998,86 ( 11 ) : 2278-2324.
  • 7Hinton G E, Osindero S,Yw T.A fast learning algorithm for deep belief nets [ J ]. Neural Computation, 2006, 18 (7) : 1527-1554.
  • 8Deng L, Seltzer M, Yu D, et al. Binary coding of speech spectrograms using a deep auto-encoder [ C ]// Proceedings of the llth Annual Conference on Interna- tional Speech Communication Association. Chiba, Japan : Makuhari, 2010:1692-1695.
  • 9He K M, Zhang X, Ren S, et al. Deep residual learning for image recognition [ C ]//Proceedings of the International Conference on Computer Vision and Pattern Recognition. Las Vegas, Nevada.IEEE, 2016 : 770-778.
  • 10Krizhevsky A, Sutskever I, Hinton G E.Imagenet classi- fication with deep convolutional neural networks [ C ]// Proceeding of 26th Annual Conference on Neural Infor- mation Processing System. Lake Tahoe, USA: MIT Press, 2012 : 1097-1105.

共引文献1659

同被引文献93

引证文献9

二级引证文献30

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部