期刊文献+
共找到1篇文章
< 1 >
每页显示 20 50 100
Hyperparameter Tuning for Deep Neural Networks Based Optimization Algorithm 被引量:3
1
作者 D.Vidyabharathi V.Mohanraj 《Intelligent Automation & Soft Computing》 SCIE 2023年第6期2559-2573,共15页
For training the present Neural Network(NN)models,the standard technique is to utilize decaying Learning Rates(LR).While the majority of these techniques commence with a large LR,they will decay multiple times over ti... For training the present Neural Network(NN)models,the standard technique is to utilize decaying Learning Rates(LR).While the majority of these techniques commence with a large LR,they will decay multiple times over time.Decaying has been proved to enhance generalization as well as optimization.Other parameters,such as the network’s size,the number of hidden layers,drop-outs to avoid overfitting,batch size,and so on,are solely based on heuristics.This work has proposed Adaptive Teaching Learning Based(ATLB)Heuristic to identify the optimal hyperparameters for diverse networks.Here we consider three architec-tures Recurrent Neural Networks(RNN),Long Short Term Memory(LSTM),Bidirectional Long Short Term Memory(BiLSTM)of Deep Neural Networks for classification.The evaluation of the proposed ATLB is done through the various learning rate schedulers Cyclical Learning Rate(CLR),Hyperbolic Tangent Decay(HTD),and Toggle between Hyperbolic Tangent Decay and Triangular mode with Restarts(T-HTR)techniques.Experimental results have shown the performance improvement on the 20Newsgroup,Reuters Newswire and IMDB dataset. 展开更多
关键词 Deep learning deep neural network(DNN) learning rates(LR) recurrent neural network(RNN) cyclical learning rate(CLR) hyperbolic tangent decay(HTD) toggle between hyperbolic tangent decay and triangular mode with restarts(T-HTR) teaching learning based optimization(TLBO)
下载PDF
上一页 1 下一页 到第
使用帮助 返回顶部