期刊文献+

基于自适应学习率优化的AdaNet改进 被引量:5

Improved AdaNet based on adaptive learning rate optimization
下载PDF
导出
摘要 人工神经网络的自适应结构学习(AdaNet)是基于Boosting集成学习的神经结构搜索框架,可通过集成子网创建高质量的模型。现有的AdaNet所产生的子网之间的差异性不显著,因而限制了集成学习中泛化误差的降低。在AdaNet设置子网网络权重和集成子网的两个步骤中,使用Adagrad、RMSProp、Adam、RAdam等自适应学习率方法来改进现有AdaNet中的优化算法。改进后的优化算法能够为不同维度参数提供不同程度的学习率缩放,得到更分散的权重分布,以增加AdaNet产生子网的多样性,从而降低集成学习的泛化误差。实验结果表明,在MNIST(Mixed National Institute of Standards and Technology database)、Fashion-MNIST、带高斯噪声的Fashion-MNIST这三个数据集上,改进后的优化算法能提升AdaNet的搜索速度,而且该方法产生的更加多样性的子网能提升集成模型的性能。在F1值这一评估模型性能的指标上,改进后的方法相较于原方法,在三种数据集上的最大提升幅度分别为0.28%、1.05%和1.10%。 AdaNet(Adaptive structural learning of artificial neural Networks)is a neural architecture search framework based on Boosting ensemble learning,which can create high-quality models through integrated subnets.The difference between subnets generated by the existing AdaNet is not significant,which limits the reduction of generalization error in ensemble learning.In the two steps of AdaNet:setting subnet network weights and integrating subnets,Adagrad,RMSProp(Root Mean Square Prop),Adam,RAdam(Rectified Adam)and other adaptive learning rate methods were used to improve the existing optimization algorithms in AdaNet.The improved optimization algorithms were able to provide different degrees of learning rate scaling for different dimensional parameters,resulting in a more dispersed weight distribution,so as to increase the diversity of subnets generated by AdaNet,thereby reducing the generalization error of ensemble learning.The experimental results show that on the three datasets:MNIST(Mixed National Institute of Standards and Technology database),Fashion-MNIST and Fashion-MNIST with Gaussian noise,the improved optimization algorithms can improve the search speed of AdaNet,and more diverse subnets generated by the method can improve the performance of the ensemble model.For the F1 value,which is an index to evaluate the model performance,compared with the original method,the improved methods have the largest improvement of 0.28%,1.05%and 1.10%on the three datasets.
作者 刘然 刘宇 顾进广 LIU Ran;LIU Yu;GU Jinguang(College of Computer Science and Technology,Wuhan University of Science and Technology,Wuhan Hubei 430065,China;Key Laboratory of Intelligent Information Processing and Real-time Industrial System in Hubei Province(Wuhan University of Science and Technology),Wuhan Hubei 430065,China;Institute of Big Data Science and Engineering,Wuhan University of Science and Technology,Wuhan Hubei 430065,China;Key Laboratory of Rich-media Knowledge Organization and Service of Digital Publishing Content,National Press and Publication Administration(Wuhan University of Science and Technology),Beijing 100038,China)
出处 《计算机应用》 CSCD 北大核心 2020年第10期2804-2810,共7页 journal of Computer Applications
基金 国家自然科学基金资助项目(U1836118,61673004) 教育部新一代信息技术创新项目(2018A03025) 国家社会科学基金重大计划项目(11&ZD189)。
关键词 AdaNet 神经架构搜索 集成学习 自适应学习率方法 自动机器学习 AdaNet Neural Architecture Search(NAS) ensemble learning adaptive learning rate method Automated Machine Learning(AutoML)
  • 相关文献

参考文献3

二级参考文献13

共引文献66

同被引文献51

引证文献5

二级引证文献30

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部