The optimization of network topologies to retain the generalization ability by deciding when to stop overtraining an artificial neural network(ANN)is an existing vital challenge in ANN prediction works.The larger the ...The optimization of network topologies to retain the generalization ability by deciding when to stop overtraining an artificial neural network(ANN)is an existing vital challenge in ANN prediction works.The larger the dataset the ANN is trained with,the better generalization the prediction can give.In this paper,a large dataset of atmospheric corrosion data of carbon steel compiled from several resources is used to train and test a multilayer backpropagation ANN model as well as two conventional corrosion prediction models(linear and Klinesmith models).Unlike previous related works,a grid search-based hyperparameter tuning is performed to develop multiple hyperparameter combinations(network topologies)to train multiple ANNs with mini-batch stochastic gradient descent optimization algorithm to facilitate the training of a large dataset.After that,one selection strategy for the optimal hyperparameter combination is applied by an early stopping method to guarantee the generalization ability of the optimal network model.The correlation coefficients(R)of the ANN model can explain about 80%(more than 75%)of the variance of atmospheric corrosion of carbon steel,and the root mean square errors(RMSE)of three models show that the ANN model gives a better performance than the other two models with acceptable generalization.The influence of input parameters on the output is highlighted by using the fuzzy curve analysis method.The result reveals that TOW,Cl-and SO2 are the most important atmospheric chemical variables,which have a well-known nonlinear relationship with atmospheric corrosion.展开更多
Hierarchical Text Classification(HTC)aims to match text to hierarchical labels.Existing methods overlook two critical issues:first,some texts cannot be fully matched to leaf node labels and need to be classified to th...Hierarchical Text Classification(HTC)aims to match text to hierarchical labels.Existing methods overlook two critical issues:first,some texts cannot be fully matched to leaf node labels and need to be classified to the correct parent node instead of treating leaf nodes as the final classification target.Second,error propagation occurs when a misclassification at a parent node propagates down the hierarchy,ultimately leading to inaccurate predictions at the leaf nodes.To address these limitations,we propose an uncertainty-guided HTC depth-aware model called DepthMatch.Specifically,we design an early stopping strategy with uncertainty to identify incomplete matching between text and labels,classifying them into the corresponding parent node labels.This approach allows us to dynamically determine the classification depth by leveraging evidence to quantify and accumulate uncertainty.Experimental results show that the proposed DepthMatch outperforms recent strong baselines on four commonly used public datasets:WOS(Web of Science),RCV1-V2(Reuters Corpus Volume I),AAPD(Arxiv Academic Paper Dataset),and BGC.Notably,on the BGC dataset,it improvesMicro-F1 andMacro-F1 scores by at least 1.09%and 1.74%,respectively.展开更多
To detect uncorrectable frames and terminate the decoding procedure early, a probability stopping criterion for iterative analog decoding of low density parity check (LDPC) codes is proposed in this paper. By using ...To detect uncorrectable frames and terminate the decoding procedure early, a probability stopping criterion for iterative analog decoding of low density parity check (LDPC) codes is proposed in this paper. By using probabilities of satisfied checks to detect uncorrectable frames and terminate decoding, the proposed criterion could be applied to analog decoders without much structure modifications. Simulation results show that the proposed criterion can reduce the average number of iterations and achieve a better balance in bit error ratio (BER) performance and decoding complexity than other stopping criteria using extrinsic information.展开更多
This article presents two new kinds of artificial neural network (ANN) response surface methods (RSMs): the ANN RSM based on early stopping technique (ANNRSM-1), and the ANN RSM based on regularization theory ...This article presents two new kinds of artificial neural network (ANN) response surface methods (RSMs): the ANN RSM based on early stopping technique (ANNRSM-1), and the ANN RSM based on regularization theory (ANNRSM-2). The following improvements are made to the conventional ANN RSM (ANNRSM-0): 1) by monitoring the validation error during the training process, ANNRSM-1 determines the early stopping point and the training stopping point, and the weight vector at the early stopping point, which corresponds to the ANN model with the optimal generalization, is finally returned as the training result; 2) according to the regularization theory, ANNRSM-2 modifies the conventional training performance function by adding to it the sum of squares of the network weights, so the network weights are forced to have smaller values while the training error decreases. Tests show that the performance of ANN RSM becomes much better due to the above-mentioned improvements: first, ANNRSM-1 and ANNRSM-2 approximate to the limit state function (LSF) more accurately than ANNRSM-0; second, the estimated failure probabilities given by ANNRSM-1 and ANNRSM-2 have smaller errors than that obtained by ANNRSM-0; third, compared with ANNRSM-0, ANNRSM-1 and ANNRSM-2 require much fewer data samples to achieve stable failure probability results.展开更多
We consider a gradient iteration algorithm for prediction of functional linear regression under the framework of reproducing kernel Hilbert spaces.In the algorithm,we use an early stopping technique,instead of the cla...We consider a gradient iteration algorithm for prediction of functional linear regression under the framework of reproducing kernel Hilbert spaces.In the algorithm,we use an early stopping technique,instead of the classical Tikhonov regularization,to prevent the iteration from an overfitting function.Under mild conditions,we obtain upper bounds,essentially matching the known minimax lower bounds,for excess prediction risk.An almost sure convergence is also established for the proposed algorithm.展开更多
基金supported by National Key R&D Program of China[Grant Number 2017YFB0203703]111 Project[Grant Number B12012]Fundamental Research Funds for the Central Universities[Grant Number FRF-GF-19-029B].
文摘The optimization of network topologies to retain the generalization ability by deciding when to stop overtraining an artificial neural network(ANN)is an existing vital challenge in ANN prediction works.The larger the dataset the ANN is trained with,the better generalization the prediction can give.In this paper,a large dataset of atmospheric corrosion data of carbon steel compiled from several resources is used to train and test a multilayer backpropagation ANN model as well as two conventional corrosion prediction models(linear and Klinesmith models).Unlike previous related works,a grid search-based hyperparameter tuning is performed to develop multiple hyperparameter combinations(network topologies)to train multiple ANNs with mini-batch stochastic gradient descent optimization algorithm to facilitate the training of a large dataset.After that,one selection strategy for the optimal hyperparameter combination is applied by an early stopping method to guarantee the generalization ability of the optimal network model.The correlation coefficients(R)of the ANN model can explain about 80%(more than 75%)of the variance of atmospheric corrosion of carbon steel,and the root mean square errors(RMSE)of three models show that the ANN model gives a better performance than the other two models with acceptable generalization.The influence of input parameters on the output is highlighted by using the fuzzy curve analysis method.The result reveals that TOW,Cl-and SO2 are the most important atmospheric chemical variables,which have a well-known nonlinear relationship with atmospheric corrosion.
基金sponsored by the National Key Research and Development Program of China(No.2021YFF0704100)the National Natural Science Foundation of China(No.62136002)+1 种基金the Chongqing Natural Science Foundation(No.cstc2022ycjh-bgzxm0004)the Science and Technology Commission of Chongqing Municipality(CSTB2023NSCQ-LZX0006),respectively.
文摘Hierarchical Text Classification(HTC)aims to match text to hierarchical labels.Existing methods overlook two critical issues:first,some texts cannot be fully matched to leaf node labels and need to be classified to the correct parent node instead of treating leaf nodes as the final classification target.Second,error propagation occurs when a misclassification at a parent node propagates down the hierarchy,ultimately leading to inaccurate predictions at the leaf nodes.To address these limitations,we propose an uncertainty-guided HTC depth-aware model called DepthMatch.Specifically,we design an early stopping strategy with uncertainty to identify incomplete matching between text and labels,classifying them into the corresponding parent node labels.This approach allows us to dynamically determine the classification depth by leveraging evidence to quantify and accumulate uncertainty.Experimental results show that the proposed DepthMatch outperforms recent strong baselines on four commonly used public datasets:WOS(Web of Science),RCV1-V2(Reuters Corpus Volume I),AAPD(Arxiv Academic Paper Dataset),and BGC.Notably,on the BGC dataset,it improvesMicro-F1 andMacro-F1 scores by at least 1.09%and 1.74%,respectively.
基金supported by the National Natural Science Foundation of China(61601027)the Guangdong Provincial Science and Technology Project(2015B010101002)
文摘To detect uncorrectable frames and terminate the decoding procedure early, a probability stopping criterion for iterative analog decoding of low density parity check (LDPC) codes is proposed in this paper. By using probabilities of satisfied checks to detect uncorrectable frames and terminate decoding, the proposed criterion could be applied to analog decoders without much structure modifications. Simulation results show that the proposed criterion can reduce the average number of iterations and achieve a better balance in bit error ratio (BER) performance and decoding complexity than other stopping criteria using extrinsic information.
基金National High-tech Research and Development Program of China (2006AA04Z405)
文摘This article presents two new kinds of artificial neural network (ANN) response surface methods (RSMs): the ANN RSM based on early stopping technique (ANNRSM-1), and the ANN RSM based on regularization theory (ANNRSM-2). The following improvements are made to the conventional ANN RSM (ANNRSM-0): 1) by monitoring the validation error during the training process, ANNRSM-1 determines the early stopping point and the training stopping point, and the weight vector at the early stopping point, which corresponds to the ANN model with the optimal generalization, is finally returned as the training result; 2) according to the regularization theory, ANNRSM-2 modifies the conventional training performance function by adding to it the sum of squares of the network weights, so the network weights are forced to have smaller values while the training error decreases. Tests show that the performance of ANN RSM becomes much better due to the above-mentioned improvements: first, ANNRSM-1 and ANNRSM-2 approximate to the limit state function (LSF) more accurately than ANNRSM-0; second, the estimated failure probabilities given by ANNRSM-1 and ANNRSM-2 have smaller errors than that obtained by ANNRSM-0; third, compared with ANNRSM-0, ANNRSM-1 and ANNRSM-2 require much fewer data samples to achieve stable failure probability results.
基金supported in part by National Natural Science Foundation of China(Grant No.11871438)supported in part by the HKRGC GRF Nos.12300218,12300519,17201020,17300021,C1013-21GF,C7004-21GFJoint NSFC-RGC N-HKU76921。
文摘We consider a gradient iteration algorithm for prediction of functional linear regression under the framework of reproducing kernel Hilbert spaces.In the algorithm,we use an early stopping technique,instead of the classical Tikhonov regularization,to prevent the iteration from an overfitting function.Under mild conditions,we obtain upper bounds,essentially matching the known minimax lower bounds,for excess prediction risk.An almost sure convergence is also established for the proposed algorithm.