期刊文献+

基于Dopout与ADAM优化器的改进CNN算法 被引量:104

Modified CNN algorithm based on Dropout and ADAM optimizer
原文传递
导出
摘要 在分析当前卷积神经网络模型特征提取过程中存在问题的基础上,提出了基于Dropout与ADAM优化器的改进卷积神经网络算法(MCNN-DA).设计了二次卷积神经网络结构,通过引入基于Re LU的激活函数以避免梯度消失问题,提高收敛速度;通过在全连接层和输出层之间加入Dropout层解决过拟合问题,并设计了ADAM优化器的最小化交叉熵.以MNIST和HCL2000数据集为测试数据,测试分析了ADAM优化器的不同学习率对算法性能的影响,得出当学习率处于0.04~0.08时,算法具有较好的识别性能.与三种算法的实验比较结果表明:本文算法的平均识别率最高可达99.21%;对于HCL2000测试集,本文算法的平均识别率比基于支持向量机优化的极速学习机算法提高了3.98%. A modified convolution neural network(CNN) algorithm was proposed based on Dropout and adaptive moment estimation(ADAM) optimizer(MCNN-DA) by analyzing the problems of CNN in extracting the convolution feature.A quadratic convolution neural network structure was designed for MCNN-DA,and Re LU was adopted as the activation function to avoid the vanishing gradient problem and accelerate the convergence.Focusing on the over-fitting problem,the algorithm employed an ADAM optimizer to minimize the cross entropy,which was implemented by inserting a Dropout layer into the all-connected layer and the output layer.Datasets MNIST and HCL2000 were used as the benchmark data,and the performance of ADAM optimizer was analyzed under different learning parameters,which shows that the proposed algorithm has better recognition performance when the learning rate is set to 0.04~0.08.Statistic results compared with three kinds of algorithms show that for the benchmark MNIST,the MCNN-DA exhibits high recognition rate of 99.21%;compared with reduced extreme learning machine algorithm optimized with support vector machine,the proposed algorithm's average increase of recognition rate is 3.98% for the benchmark HCL2000.
作者 杨观赐 杨静 李少波 胡建军 Yang Guanci;Yang Jing;Li Shaobo;Hu Jianjun(Key Laboratory of Advanced Manufacturing Technology of Ministry of Education,Guizhou University,Guiyang 550003,China;Department of Computer Science and Engineering,University of South Carolina,Columbia 29208,USA)
出处 《华中科技大学学报(自然科学版)》 EI CAS CSCD 北大核心 2018年第7期122-127,共6页 Journal of Huazhong University of Science and Technology(Natural Science Edition)
基金 国家自然科学基金资助项目(61640209) 贵州省科技计划资助项目(黔科合人字(2015)13号,黔科合LH字[2016]7433号) 贵州省科技厅基础平台计划资助项目(黔科合平台人才[2018]5702)
关键词 卷积神经网络 激活函数 梯度消失 ADAM优化器 梯度饱和问题 convolution neural network activate function gradient disappearence ADAM optimizer vanishing gradient problem
  • 相关文献

参考文献6

二级参考文献50

  • 1梁化楼,戴贵亮.人工神经网络与遗传算法的结合:进展及展望[J].电子学报,1995,23(10):194-200. 被引量:71
  • 2李飞,赵生妹,郑宝玉.量子神经网络及其在CDMA多用户检测中的应用[J].信号处理,2005,21(6):555-559. 被引量:6
  • 3BISHOP C M. Neural networks for pattern recognition [ M ]. New York : Oxford University Press, 1995.
  • 4张青贵.人工神经网络导论[M].北京:高等教育出版社,1992.
  • 5HUANG Guang-bing, ZHU Yu-qin, SIEW C K. Extreme learning machine : theory and applications [ J ]. Neurocomputing, 2006,70 ( 1-3 ) :489-501.
  • 6张雨浓,杨逸文,李巍.神经网络网络权值直接确定法[M].广州:中山大学出版社,2010.
  • 7YU Hao, WILAMOWSK B M. Levenberg-Marquardt training[ J ]. Intelligent Systems,2010 ( 12 ) : 1 - 6.
  • 8HUANG Guang-bing, CHEN Lei, SIEW C K. Universal approximation using incremental constructive feedforward networks with random hidden nodes[J]. [EEE Trans on Neural Networks,2006,17(4) : 879- 883,.
  • 9Femdndez-Delgado M, Cemadas E,Barro S, et al. Dowe Need Hundreds of Classifiers to Solve Real WorldClassification Problems? [J]. Journal of MachineLearning Research (SI 532-4435), 2014, 15(1):3133-3181.
  • 10Joachims T. Making Large-scale Support VectorMachine Learning Practical [C]// Advances in kernelmethods. USA: MIT Press, 1999: 169-184.

共引文献1745

同被引文献857

引证文献104

二级引证文献535

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部