期刊文献+

卷积神经网络中激活函数的一种改进 被引量:18

An Improvement of the Activation Function in Convolutional Neural Networks
下载PDF
导出
摘要 卷积神经网络中激活函数的作用是激活神经元的特征,然后保留并映射出来,这是人工神经网络能模拟人脑机制,解决非线性问题的关键.针对传统卷积神经网络出现的震荡、不收敛甚至过拟合的情况,对激活ReLU函数进行优化.提出一种新型矫正激活函数,称其为ReLU阈值函数.通过对数据集caltech101和caltech256进行训练,证明其在图片分类上的性能要优于ReLU函数.其中用Alexnet网络模型对caltech101数据集进行训练时的分类准确率由之前的97.7%提高到99.3%,对caltech256数据集进行训练时的分类准确率由之前的65.4%提高到92.3%. The purpose of the activation function in the convolutional neural network is to activate the characteristics of neurons and then retain and map them. It is the key point that the artificial eural network can simulate human brain mechanisms and solve nonlinear problems. In this paper, to solve problems such as oscillation, non-convergence and even over-fitting of traditional convolutional neural networks, the activation of ReLU function is optimised. A new type of corrective activation function is proposed, which is called the ReLU threshold function. By training the data sets caltech101 and caltech256, it is proved that its performance in image classification is better than the ReLU function. The classification-accuracy rate of the caltech101 dataset was improved from 97.7% to 99.3% when using the Alexnet network model, and when training the caltech256 dataset. It was improved from 65.4% to 92.3%.
作者 刘小文 郭大波 李聪 LIU Xiaowen;GUO Dabo;LI Cong(College of Physics and Electronic Engineering, Shanxi University, Taiyuan 030006, China)
出处 《测试技术学报》 2019年第2期121-125,共5页 Journal of Test and Measurement Technology
基金 山西省基础研究资助项目(201601D102033 201801D121118)
关键词 卷积神经网络 激活函数 ReLU PReLU convolutional neural network activation function ReLU PReLU
  • 相关文献

参考文献6

二级参考文献80

  • 1Polana R, Nelson R C. Detection and recognition of periodic, nonrigid motion[J]. International Journal of Computer Vision(IJCV), 1997, 23(3): 261-282.
  • 2Laptev I, Marszalek M. Learning realistic human actions from movies[C]. Proceedings of IEEE Conference on Com- puter Vision and Pattern Recognition (CVPR08), 2008.. 1-8.
  • 3Scovanner P, Ali S, Shah M. A 3-dimensional SIFT descriptor and its application to action recognition[C]. Proceed- ings of the International Conference on Multimedia (MultiMedia'07). Augsburg, Germany, 2007.. 357-360.
  • 4LeCun Y, Boser B, Denker J S, et al. Backpropagation applied to handwritten zip code recognition[J]. Neural compu tation, 1989, 1(4)= 541-551.
  • 5Bengio Y, Lamblin P, Popovici D, et al. Greedy layer-wise training of deep networks[C]. Advances in Neural Infor mation Processing Systems. Cambridge: MIT Press, 2007: 153-160.
  • 6Hinton G E. A practical guide to training restricted Boltzmann machines[M]. Neural Networks= Tricks of the Trade Berlin= Springer-Verlag, 2013= 599-619.
  • 7Fukushima K. Neocognitron: A hierarchical neural network capable of visual pattern Recognition[J]. Neural net works, 1988, 1(2)= 119-130.
  • 8LeCun Y, Bottou L, Bengio Y, et al. Gradient-based learning applied to document recognition[J]. Proceedings of the IEEE, 1998, 86(11).. 2278-2324.
  • 9Ji Shuiwang, Xu Wei, Yang Ming, et al. 3D convolutional neural networks for human action recognition[J]. Pattern Analysis and Machine Intelligence. IEEE Transactions on, 2013, 35(1): 221-232.
  • 10Glorot X, Bordes A, Bengio Y. Deep sparse rectifier neural networks[J]. Aistats, 2011(15) : 315-323.

共引文献2143

同被引文献131

引证文献18

二级引证文献60

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部