期刊文献+

基于随机L-BFGS的二阶非凸稀疏优化算法 被引量:1

Second-Order Nonconvex Sparse Optimization Method Based on Stochastic L-BFGS
下载PDF
导出
摘要 在稀疏模型中普遍采用一阶优化算法进行学习,这些算法的普遍思路都是将迭代硬阈值算法与优化算法进行结合。相较于一阶优化算法,二阶优化算法很少被应用到稀疏优化问题中,因为Hessian矩阵及其逆矩阵的计算需要消耗极大的算力资源。所以为了有效且高效地利用二阶信息,提出一种新的随机L-BFGS硬阈值优化算法用于解决非凸的稀疏学习问题,算法将迭代硬阈值方法(IHT)引入随机L-BFGS算法,在保持模型性能的同时显著提升了算法的收敛速度,并在线性回归和逻辑回归上的实验结果上证明了新算法的优越性。 First-order optimization methods have been widely applied in the learning of sparse models. The fundamental idea of these methods is to incorporate the Iterative Hard Thresholding(IHT) algorithm into traditional optimization methods. Compared with the first-order methods, very few second-order methods have been applied to sparse optimization problems because of the tremendous computation that obtains the Hessian matrix and its inverse. In order to make use of the second-order information effectively and efficiently, this paper proposes a novel optimization method to solve the nonconvex sparse learning problems. The core idea of the proposed Stochastic L-BFGS Hard Thresholding approach is to incorporate the Stochastic L-BFGS into the Iterative Hard Thresholding(IHT) method, which significantly accelerates the convergence rate and maintains the effectiveness of the model. The experimental results on linear regression and logistic regression demonstrate the superiority of the proposed approach.
作者 刘光宇 张令威 杭仁龙 LIU Guang-yu;ZHANG Ling-wei;HANG Ren-long(Jiangsu Key Laboratory of Big Data Analysis Technology,Nanjing University of Information Science and Technology,Nanjing Jiangsu 210044,China)
出处 《计算机仿真》 北大核心 2022年第10期359-363,共5页 Computer Simulation
基金 江苏省青年基金项目(BK20180786) 国家自然科学基金(61906096)。
关键词 一阶优化算法 二阶信息 迭代硬阈值 稀疏学习 First-order optimization methods Second-order information IHT Sparse learning
  • 相关文献

参考文献1

二级参考文献42

  • 1Vapnik VN. Statistical Learning Theory. New York: Wiley-Interscience, 1998.
  • 2Zhang T. Statistical behavior and consistency of classification methods based on convex risk minimization. Annals of Statistics, 2004,32(l):56-85. [doi: 10.1214/aos/1079120130].
  • 3Zhang T. Statistical analysis of some multi-category large margin classification methods. Journal of Machine Learning Research, 2004,5:1225-1251.
  • 4Wang J, Tao Q. Machine learning: The state of the art. IEEE Intelligent Systems, 2008,23(6):49-55. [doi: 10.1109/MIS.2008.107].
  • 5Bennett KP, Parrado-Hemandez E. The interplay of optimization and machine learning research. Journal of Machine Learning Research, 2006,7:1265-1281.
  • 6Tibshirani R. Regression shrinkage and selection via the lasso. Journal of Royal Statistical Society (Series B), 1996,58(l):267-288.
  • 7Nesterov Y. Primal-Dual subgradient methods for convex problems. Mathematical Programming, 2009,120(l):221-259. [doi: 10. 1007/sl0107-007-0149-x].
  • 8Bertsekas DP, Nedic A, Ozdaglar AE. Convex Analysis and Optimization. Belmont: Athena Scientific, 2003.
  • 9Zinkevich M. Online convex programming and generalized infinitesimal gradient ascent. In: Proc. of the Int’l Conf. on Machine Learning. 2003. 928-936.
  • 10Shalev-Shwartz S, Singer Y, Srebro N. Pegasos: Primal estimated sub-gradient solver for SVM. In: Proc. of the Int’l Conf. on Machine Learning. 2007. 807-814. [doi: 10.1145/1273496.1273598].

共引文献22

同被引文献14

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部