We prove,under mild conditions,the convergence of a Riemannian gradient descent method for a hyperbolic neural network regression model,both in batch gradient descent and stochastic gradient descent.We also discuss a ...We prove,under mild conditions,the convergence of a Riemannian gradient descent method for a hyperbolic neural network regression model,both in batch gradient descent and stochastic gradient descent.We also discuss a Riemannian version of the Adam algorithm.We show numerical simulations of these algorithms on various benchmarks.展开更多
基金partially supported by NSF Grants DMS-1854434,DMS-1952644,and DMS-2151235 at UC Irvinesupported by NSF Grants DMS-1924935,DMS-1952339,DMS-2110145,DMS-2152762,and DMS-2208361,and DOE Grants DE-SC0021142 and DE-SC0002722.
文摘We prove,under mild conditions,the convergence of a Riemannian gradient descent method for a hyperbolic neural network regression model,both in batch gradient descent and stochastic gradient descent.We also discuss a Riemannian version of the Adam algorithm.We show numerical simulations of these algorithms on various benchmarks.