A theoretical methodology is suggested for finding the malaria parasites’presence with the help of an intelligent hyper-parameter tuned Deep Learning(DL)based malaria parasite detection and classification(HPTDL-MPDC)...A theoretical methodology is suggested for finding the malaria parasites’presence with the help of an intelligent hyper-parameter tuned Deep Learning(DL)based malaria parasite detection and classification(HPTDL-MPDC)in the smear images of human peripheral blood.Some existing approaches fail to predict the malaria parasitic features and reduce the prediction accuracy.The trained model initiated in the proposed system for classifying peripheral blood smear images into the non-parasite or parasite classes using the available online dataset.The Adagrad optimizer is stacked with the suggested pre-trained Deep Neural Network(DNN)with the help of the contrastive divergence method to pre-train.The features are extracted from the images in the proposed system to train the DNN for initializing the visible variables.The smear images show the concatenated feature to be utilized as the feature vector in the proposed system.Lastly,hyper-parameters are used to fine-tune DNN to calculate the class labels’probability.The suggested system outperforms more modern methodologies with an accuracy of 91%,precision of 89%,recall of 93%and F1-score of 91%.The HPTDL-MPDC has the primary application in detecting the parasite of malaria in the smear images of human peripheral blood.展开更多
Due to outstanding performance in cheminformatics,machine learning algorithms have been increasingly used to mine molecular properties and biomedical big data.The performance of machine learning models is known to cri...Due to outstanding performance in cheminformatics,machine learning algorithms have been increasingly used to mine molecular properties and biomedical big data.The performance of machine learning models is known to critically depend on the selection of the hyper-parameter configuration.However,many studies either explored the optimal hyper-parameters per the grid searching method or employed arbitrarily selected hyper-parameters,which can easily lead to achieving a suboptimal hyper-parameter configuration.In this study,Hyperopt library embedding with the Bayesian optimization is employed to find optimal hyper-parameters for different machine learning algorithms.Six drug discovery datasets,including solubility,probe-likeness,h ERG,Chagas disease,tuberculosis,and malaria,are used to compare different machine learning algorithms with ECFP6 fingerprints.This contribution aims to evaluate whether the Bernoulli Na?ve Bayes,logistic linear regression,Ada Boost decision tree,random forest,support vector machine,and deep neural networks algorithms with optimized hyper-parameters can offer any improvement in testing as compared with the referenced models assessed by an array of metrics including AUC,F1-score,Cohen’s kappa,Matthews correlation coefficient,recall,precision,and accuracy.Based on the rank normalized score approach,the Hyperopt models achieve better or comparable performance on 33 out 36 models for different drug discovery datasets,showing significant improvement achieved by employing the Hyperopt library.The open-source code of all the 6 machine learning frameworks employed in the Hyperopt python package is provided to make this approach accessible to more scientists,who are not familiar with writing code.展开更多
Regularized system identification has become the research frontier of system identification in the past decade.One related core subject is to study the convergence properties of various hyper-parameter estimators as t...Regularized system identification has become the research frontier of system identification in the past decade.One related core subject is to study the convergence properties of various hyper-parameter estimators as the sample size goes to infinity.In this paper,we consider one commonly used hyper-parameter estimator,the empirical Bayes(EB).Its convergence in distribution has been studied,and the explicit expression of the covariance matrix of its limiting distribution has been given.However,what we are truly interested in are factors contained in the covariance matrix of the EB hyper-parameter estimator,and then,the convergence of its covariance matrix to that of its limiting distribution is required.In general,the convergence in distribution of a sequence of random variables does not necessarily guarantee the convergence of its covariance matrix.Thus,the derivation of such convergence is a necessary complement to our theoretical analysis about factors that influence the convergence properties of the EB hyper-parameter estimator.In this paper,we consider the regularized finite impulse response(FIR)model estimation with deterministic inputs,and show that the covariance matrix of the EB hyper-parameter estimator converges to that of its limiting distribution.Moreover,we run numerical simulations to demonstrate the efficacy of ourtheoretical results.展开更多
In computer vision,convolutional neural networks have a wide range of uses.Images representmost of today’s data,so it’s important to know how to handle these large amounts of data efficiently.Convolutional neural ne...In computer vision,convolutional neural networks have a wide range of uses.Images representmost of today’s data,so it’s important to know how to handle these large amounts of data efficiently.Convolutional neural networks have been shown to solve image processing problems effectively.However,when designing the network structure for a particular problem,you need to adjust the hyperparameters for higher accuracy.This technique is time consuming and requires a lot of work and domain knowledge.Designing a convolutional neural network architecture is a classic NP-hard optimization challenge.On the other hand,different datasets require different combinations of models or hyperparameters,which can be time consuming and inconvenient.Various approaches have been proposed to overcome this problem,such as grid search limited to low-dimensional space and queuing by random selection.To address this issue,we propose an evolutionary algorithm-based approach that dynamically enhances the structure of Convolution Neural Networks(CNNs)using optimized hyperparameters.This study proposes a method using Non-dominated sorted genetic algorithms(NSGA)to improve the hyperparameters of the CNN model.In addition,different types and parameter ranges of existing genetic algorithms are used.Acomparative study was conducted with various state-of-the-art methodologies and algorithms.Experiments have shown that our proposed approach is superior to previous methods in terms of classification accuracy,and the results are published in modern computing literature.展开更多
Alzheimer’s disease(AD)is an intensifying disorder that causes brain cells to degenerate early and destruct.Mild cognitive impairment(MCI)is one of the early signs of AD that interferes with people’s regular functio...Alzheimer’s disease(AD)is an intensifying disorder that causes brain cells to degenerate early and destruct.Mild cognitive impairment(MCI)is one of the early signs of AD that interferes with people’s regular functioning and daily activities.The proposed work includes a deep learning approach with a multimodal recurrent neural network(RNN)to predict whether MCI leads to Alzheimer’s or not.The gated recurrent unit(GRU)RNN classifier is trained using individual and correlated features.Feature vectors are concate-nated based on their correlation strength to improve prediction results.The feature vectors generated are given as the input to multiple different classifiers,whose decision function is used to predict the final output,which determines whether MCI progresses onto AD or not.Our findings demonstrated that,compared to individual modalities,which provided an average accuracy of 75%,our prediction model for MCI conversion to AD yielded an improve-ment in accuracy up to 96%when used with multiple concatenated modalities.Comparing the accuracy of different decision functions,such as Support Vec-tor Machine(SVM),Decision tree,Random Forest,and Ensemble techniques,it was found that that the Ensemble approach provided the highest accuracy(96%)and Decision tree provided the lowest accuracy(86%).展开更多
The number of studies in the literature that diagnose cancer with machine learning using genome data is quite limited.These studies focus on the prediction performance,and the extraction of genomic factors that cause ...The number of studies in the literature that diagnose cancer with machine learning using genome data is quite limited.These studies focus on the prediction performance,and the extraction of genomic factors that cause disease is often overlooked.However,finding underlying genetic causes is very important in terms of early diagnosis,development of diagnostic kits,preventive medicine,etc.The motivation of our study was to diagnose bladder cancer(BCa)based on genetic data and to reveal underlying genetic factors by using machine-learning models.In addition,conducting hyper-parameter optimization to get the best performance from different models,which is overlooked in most studies,was another objective of the study.Within the framework of these motivations,C4.5,random forest(RF),artificial neural networks(ANN),and deep learning(DL)were used.In this way,the diagnostic performance of decision tree(DT)-based models and black box models on BCa was also compared.The most successful model,DL,yielded an area under the curve(AUC)of 0.985 and a mean square error(MSE)of 0.069.For each model,hyper-parameters were optimized by an evolutionary algorithm.On average,hyper-parameter optimization increased MSE,root mean square error(RMSE),LogLoss,and AUC by 30%,17.5%,13%,and 6.75%,respectively.The features causing BCa were extracted.For this purpose,entropy and Gini coefficients were used for DT-based methods,and the Gedeon variable importance was used for black box methods.The single nucleotide polymorphisms(SNPs)rs197412,rs2275928,rs12479919,rs798766 and rs2275928,whose BCa relations were proven in the literature,were found to be closely related to BCa.In addition,rs1994624 and rs2241766 susceptibility loci were proposed to be examined in future studies.展开更多
连续潮流(continuous power flow,CPF)是电力系统电压稳定分析的有效工具,也是解决常规潮流中病态潮流问题的方法之一。针对无平衡节点孤岛运行微电网系统的无平衡节点、且有下垂控制分布式电源装置的特性,提出一种无平衡节点孤岛运行...连续潮流(continuous power flow,CPF)是电力系统电压稳定分析的有效工具,也是解决常规潮流中病态潮流问题的方法之一。针对无平衡节点孤岛运行微电网系统的无平衡节点、且有下垂控制分布式电源装置的特性,提出一种无平衡节点孤岛运行微电网CPF计算方法。采用不要求雅可比矩阵非奇异,且具有全局收敛性的LM-TR方法求解初始点。预测环节采用结合局部参数化方法的切线法。校正环节提出新型的超球面参数化方法,并采用结合传统牛顿法和带Armijo型线性搜索牛顿法的组合牛顿法进行校正,以保证CPF校正计算成功,及实现整个CPF过程中在较高计算精度下一直采用较大定步长预测。对改造后的37节点和17节点无平衡节点孤岛运行微电网系统采用所提方法进行CPF计算,验证了其正确性和有效性。展开更多
Due to the fact that the vibration signal of the rotating machine is one-dimensional and the large-scale convolution kernel can obtain a better perception field, on the basis of the classical convolution neural networ...Due to the fact that the vibration signal of the rotating machine is one-dimensional and the large-scale convolution kernel can obtain a better perception field, on the basis of the classical convolution neural network model(LetNet-5), one-dimensional large-kernel convolution neural network(1 DLCNN) is designed. Since the hyper-parameters of 1 DLCNN have a greater impact on network performance, the genetic algorithm(GA) is used to optimize the hyper-parameters, and the method of optimizing the parameters of 1 DLCNN by the genetic algorithm is named GA-1 DLCNN. The experimental results show that the optimal network model based on the GA-1 DLCNN method can achieve 99.9% fault diagnosis accuracy, which is much higher than those of other traditional fault diagnosis methods. In addition, the 1 DLCNN is compared with one-dimencional small-kernel convolution neural network(1 DSCNN) and the classical two-dimensional convolution neural network model. The input sample lengths are set to be 128, 256, 512, 1 024, and 2 048, respectively, and the final diagnostic accuracy results and the visual scatter plot show that the effect of 1 DLCNN is optimal.展开更多
Machine learning models may outperform traditional statistical regression algorithms for predicting clinical outcomes.Proper validation of building such models and tuning their underlying algorithms is necessary to av...Machine learning models may outperform traditional statistical regression algorithms for predicting clinical outcomes.Proper validation of building such models and tuning their underlying algorithms is necessary to avoid over-fitting and poor generalizability,which smaller datasets can be more prone to.In an effort to educate readers interested in artificial intelligence and model-building based on machine-learning algorithms,we outline important details on crossvalidation techniques that can enhance the performance and generalizability of such models.展开更多
Geological structures often exhibit smooth characteristics away from sharp discontinuities. One aim of geophysical inversion is to recover information about the smooth structures as well as about the sharp discontinui...Geological structures often exhibit smooth characteristics away from sharp discontinuities. One aim of geophysical inversion is to recover information about the smooth structures as well as about the sharp discontinuities. Because no specific operator can provide a perfect sparse representation of complicated geological models, hyper-parameter regularization inversion based on the iterative split Bregman method was used to recover the features of both smooth and sharp geological structures. A novel preconditioned matrix was proposed, which counteracted the natural decay of the sensitivity matrix and its inverse matrix was calculated easily. Application of the algorithm to synthetic data produces density models that are good representations of the designed models. The results show that the algorithm proposed is feasible and effective.展开更多
Gaussian process(GP)has fewer parameters,simple model and output of probabilistic sense,when compared with the methods such as support vector machines.Selection of the hyper-parameters is critical to the performance o...Gaussian process(GP)has fewer parameters,simple model and output of probabilistic sense,when compared with the methods such as support vector machines.Selection of the hyper-parameters is critical to the performance of Gaussian process model.However,the common-used algorithm has the disadvantages of difficult determination of iteration steps,over-dependence of optimization effect on initial values,and easily falling into local optimum.To solve this problem,a method combining the Gaussian process with memetic algorithm was proposed.Based on this method,memetic algorithm was used to search the optimal hyper parameters of Gaussian process regression(GPR)model in the training process and form MA-GPR algorithms,and then the model was used to predict and test the results.When used in the marine long-range precision strike system(LPSS)battle effectiveness evaluation,the proposed MA-GPR model significantly improved the prediction accuracy,compared with the conjugate gradient method and the genetic algorithm optimization process.展开更多
A Bayesian estimator with informative prior distributions (a multi-normal and an inverted gamma distribution), adequate to displacement estimation at dam displacement monitoring networks, is presented. The hyper-par...A Bayesian estimator with informative prior distributions (a multi-normal and an inverted gamma distribution), adequate to displacement estimation at dam displacement monitoring networks, is presented. The hyper-parameters of the prior distributions are obtained by Bayesian empirical methods with non-informative meta-priors. The performances of the Bayes estimator and the classical generalized lest squares estimator are compared using two measurements of the horizontal monitoring network of a concrete gravity dam: the Penha Garcia dam (Portugal). In order to test the robustness of the two estimators, a gross error is added to one of the measured horizontal directions: the Bayes estimator proves to be significantly more robust than the classic maximum likelihood estimator.展开更多
In this paper we study the problem of model selection for a linear programming-based support vector machine for regression. We propose generalized method that is based on a quasi-Newton method that uses a globalizatio...In this paper we study the problem of model selection for a linear programming-based support vector machine for regression. We propose generalized method that is based on a quasi-Newton method that uses a globalization strategy and an inexact computation of first order information. We explore the case of two-class, multi-class, and regression problems. Simulation results among standard datasets suggest that the algorithm achieves insignificant variability when measuring residual statistical properties.展开更多
Domain shift is when the data used in training does not match the ones it will be applied to later on under similar conditions.Domain shift will reduce accuracy in results.To prevent this,domain adaptation is done,whi...Domain shift is when the data used in training does not match the ones it will be applied to later on under similar conditions.Domain shift will reduce accuracy in results.To prevent this,domain adaptation is done,which adapts the pre-trained model to the target domain.In real scenarios,the availability of labels for target data is rare thus resulting in unsupervised domain adaptation.Herein,we propose an innovative approach where source-free domain adaptation models and Generative Adversarial Networks(GANs)are integrated to improve the performance of computer vision or robotic vision-based systems in our study.Cosine Generative Adversarial Network(CosGAN)is developed as a GAN that uses cosine embedding loss to handle issues associated with unsupervised source-relax domain adaptations.For less complex architecture,the CosGAN training process has two steps that produce results almost comparable to other state-of-the-art techniques.The efficiency of CosGAN was compared by conducting experiments using benchmarked datasets.The approach was evaluated on different datasets and experimental results show superiority over existing state-of-the-art methods in terms of accuracy as well as generalization ability.This technique has numerous applications including wheeled robots,autonomous vehicles,warehouse automation,and all image-processing-based automation tasks so it can reshape the field of robotic vision with its ability to make robots adapt to new tasks and environments efficiently without requiring additional labeled data.It lays the groundwork for future expansions in robotic vision and applications.Although GAN provides a variety of outstanding features,it also increases the risk of instability and over-fitting of the training data thus making the data difficult to converge.展开更多
Accurate prediction of drilling efficiency is critical for developing the earth-rock excavation schedule.The single machine learning(ML)prediction models usually suffer from problems including parameter sensitivity an...Accurate prediction of drilling efficiency is critical for developing the earth-rock excavation schedule.The single machine learning(ML)prediction models usually suffer from problems including parameter sensitivity and overfitting.In addition,the influence of environmental and operational factors is often ignored.In response,a novel stacking-based ensemble learning method taking into account the combined effects of those factors is proposed.Through multiple comparison tests,four models,e Xtreme gradient boosting(XGBoost),random forest(RF),back propagation neural network(BPNN)as the base learners,and support vector regression(SVR)as the meta-learner,are selected for stacking.Furthermore,an improved cuckoo search optimization(ICSO)algorithm is developed for hyper-parameter optimization of the ensemble model.The application to a real-world project demonstrates that the proposed method outperforms the popular single ML method XGBoost and the ensemble model optimized by particle swarm optimization(PSO),with 16.43%and 4.88%improvements of mean absolute percentage error(MAPE),respectively.展开更多
文摘A theoretical methodology is suggested for finding the malaria parasites’presence with the help of an intelligent hyper-parameter tuned Deep Learning(DL)based malaria parasite detection and classification(HPTDL-MPDC)in the smear images of human peripheral blood.Some existing approaches fail to predict the malaria parasitic features and reduce the prediction accuracy.The trained model initiated in the proposed system for classifying peripheral blood smear images into the non-parasite or parasite classes using the available online dataset.The Adagrad optimizer is stacked with the suggested pre-trained Deep Neural Network(DNN)with the help of the contrastive divergence method to pre-train.The features are extracted from the images in the proposed system to train the DNN for initializing the visible variables.The smear images show the concatenated feature to be utilized as the feature vector in the proposed system.Lastly,hyper-parameters are used to fine-tune DNN to calculate the class labels’probability.The suggested system outperforms more modern methodologies with an accuracy of 91%,precision of 89%,recall of 93%and F1-score of 91%.The HPTDL-MPDC has the primary application in detecting the parasite of malaria in the smear images of human peripheral blood.
基金financial support provided by the National Key Research and Development Project(2019YFC0214403)Chongqing Joint Chinese Medicine Scientific Research Project(2021ZY023984)。
文摘Due to outstanding performance in cheminformatics,machine learning algorithms have been increasingly used to mine molecular properties and biomedical big data.The performance of machine learning models is known to critically depend on the selection of the hyper-parameter configuration.However,many studies either explored the optimal hyper-parameters per the grid searching method or employed arbitrarily selected hyper-parameters,which can easily lead to achieving a suboptimal hyper-parameter configuration.In this study,Hyperopt library embedding with the Bayesian optimization is employed to find optimal hyper-parameters for different machine learning algorithms.Six drug discovery datasets,including solubility,probe-likeness,h ERG,Chagas disease,tuberculosis,and malaria,are used to compare different machine learning algorithms with ECFP6 fingerprints.This contribution aims to evaluate whether the Bernoulli Na?ve Bayes,logistic linear regression,Ada Boost decision tree,random forest,support vector machine,and deep neural networks algorithms with optimized hyper-parameters can offer any improvement in testing as compared with the referenced models assessed by an array of metrics including AUC,F1-score,Cohen’s kappa,Matthews correlation coefficient,recall,precision,and accuracy.Based on the rank normalized score approach,the Hyperopt models achieve better or comparable performance on 33 out 36 models for different drug discovery datasets,showing significant improvement achieved by employing the Hyperopt library.The open-source code of all the 6 machine learning frameworks employed in the Hyperopt python package is provided to make this approach accessible to more scientists,who are not familiar with writing code.
基金supported in part by the National Natural Science Foundation of China(No.62273287)by the Shenzhen Science and Technology Innovation Council(Nos.JCYJ20220530143418040,JCY20170411102101881)the Thousand Youth Talents Plan funded by the central government of China.
文摘Regularized system identification has become the research frontier of system identification in the past decade.One related core subject is to study the convergence properties of various hyper-parameter estimators as the sample size goes to infinity.In this paper,we consider one commonly used hyper-parameter estimator,the empirical Bayes(EB).Its convergence in distribution has been studied,and the explicit expression of the covariance matrix of its limiting distribution has been given.However,what we are truly interested in are factors contained in the covariance matrix of the EB hyper-parameter estimator,and then,the convergence of its covariance matrix to that of its limiting distribution is required.In general,the convergence in distribution of a sequence of random variables does not necessarily guarantee the convergence of its covariance matrix.Thus,the derivation of such convergence is a necessary complement to our theoretical analysis about factors that influence the convergence properties of the EB hyper-parameter estimator.In this paper,we consider the regularized finite impulse response(FIR)model estimation with deterministic inputs,and show that the covariance matrix of the EB hyper-parameter estimator converges to that of its limiting distribution.Moreover,we run numerical simulations to demonstrate the efficacy of ourtheoretical results.
基金This research was supported by the Researchers Supporting Program(TUMAProject-2021-27)Almaarefa University,Riyadh,Saudi Arabia.
文摘In computer vision,convolutional neural networks have a wide range of uses.Images representmost of today’s data,so it’s important to know how to handle these large amounts of data efficiently.Convolutional neural networks have been shown to solve image processing problems effectively.However,when designing the network structure for a particular problem,you need to adjust the hyperparameters for higher accuracy.This technique is time consuming and requires a lot of work and domain knowledge.Designing a convolutional neural network architecture is a classic NP-hard optimization challenge.On the other hand,different datasets require different combinations of models or hyperparameters,which can be time consuming and inconvenient.Various approaches have been proposed to overcome this problem,such as grid search limited to low-dimensional space and queuing by random selection.To address this issue,we propose an evolutionary algorithm-based approach that dynamically enhances the structure of Convolution Neural Networks(CNNs)using optimized hyperparameters.This study proposes a method using Non-dominated sorted genetic algorithms(NSGA)to improve the hyperparameters of the CNN model.In addition,different types and parameter ranges of existing genetic algorithms are used.Acomparative study was conducted with various state-of-the-art methodologies and algorithms.Experiments have shown that our proposed approach is superior to previous methods in terms of classification accuracy,and the results are published in modern computing literature.
文摘Alzheimer’s disease(AD)is an intensifying disorder that causes brain cells to degenerate early and destruct.Mild cognitive impairment(MCI)is one of the early signs of AD that interferes with people’s regular functioning and daily activities.The proposed work includes a deep learning approach with a multimodal recurrent neural network(RNN)to predict whether MCI leads to Alzheimer’s or not.The gated recurrent unit(GRU)RNN classifier is trained using individual and correlated features.Feature vectors are concate-nated based on their correlation strength to improve prediction results.The feature vectors generated are given as the input to multiple different classifiers,whose decision function is used to predict the final output,which determines whether MCI progresses onto AD or not.Our findings demonstrated that,compared to individual modalities,which provided an average accuracy of 75%,our prediction model for MCI conversion to AD yielded an improve-ment in accuracy up to 96%when used with multiple concatenated modalities.Comparing the accuracy of different decision functions,such as Support Vec-tor Machine(SVM),Decision tree,Random Forest,and Ensemble techniques,it was found that that the Ensemble approach provided the highest accuracy(96%)and Decision tree provided the lowest accuracy(86%).
文摘The number of studies in the literature that diagnose cancer with machine learning using genome data is quite limited.These studies focus on the prediction performance,and the extraction of genomic factors that cause disease is often overlooked.However,finding underlying genetic causes is very important in terms of early diagnosis,development of diagnostic kits,preventive medicine,etc.The motivation of our study was to diagnose bladder cancer(BCa)based on genetic data and to reveal underlying genetic factors by using machine-learning models.In addition,conducting hyper-parameter optimization to get the best performance from different models,which is overlooked in most studies,was another objective of the study.Within the framework of these motivations,C4.5,random forest(RF),artificial neural networks(ANN),and deep learning(DL)were used.In this way,the diagnostic performance of decision tree(DT)-based models and black box models on BCa was also compared.The most successful model,DL,yielded an area under the curve(AUC)of 0.985 and a mean square error(MSE)of 0.069.For each model,hyper-parameters were optimized by an evolutionary algorithm.On average,hyper-parameter optimization increased MSE,root mean square error(RMSE),LogLoss,and AUC by 30%,17.5%,13%,and 6.75%,respectively.The features causing BCa were extracted.For this purpose,entropy and Gini coefficients were used for DT-based methods,and the Gedeon variable importance was used for black box methods.The single nucleotide polymorphisms(SNPs)rs197412,rs2275928,rs12479919,rs798766 and rs2275928,whose BCa relations were proven in the literature,were found to be closely related to BCa.In addition,rs1994624 and rs2241766 susceptibility loci were proposed to be examined in future studies.
文摘连续潮流(continuous power flow,CPF)是电力系统电压稳定分析的有效工具,也是解决常规潮流中病态潮流问题的方法之一。针对无平衡节点孤岛运行微电网系统的无平衡节点、且有下垂控制分布式电源装置的特性,提出一种无平衡节点孤岛运行微电网CPF计算方法。采用不要求雅可比矩阵非奇异,且具有全局收敛性的LM-TR方法求解初始点。预测环节采用结合局部参数化方法的切线法。校正环节提出新型的超球面参数化方法,并采用结合传统牛顿法和带Armijo型线性搜索牛顿法的组合牛顿法进行校正,以保证CPF校正计算成功,及实现整个CPF过程中在较高计算精度下一直采用较大定步长预测。对改造后的37节点和17节点无平衡节点孤岛运行微电网系统采用所提方法进行CPF计算,验证了其正确性和有效性。
基金The National Natural Science Foundation of China(No.51675098)
文摘Due to the fact that the vibration signal of the rotating machine is one-dimensional and the large-scale convolution kernel can obtain a better perception field, on the basis of the classical convolution neural network model(LetNet-5), one-dimensional large-kernel convolution neural network(1 DLCNN) is designed. Since the hyper-parameters of 1 DLCNN have a greater impact on network performance, the genetic algorithm(GA) is used to optimize the hyper-parameters, and the method of optimizing the parameters of 1 DLCNN by the genetic algorithm is named GA-1 DLCNN. The experimental results show that the optimal network model based on the GA-1 DLCNN method can achieve 99.9% fault diagnosis accuracy, which is much higher than those of other traditional fault diagnosis methods. In addition, the 1 DLCNN is compared with one-dimencional small-kernel convolution neural network(1 DSCNN) and the classical two-dimensional convolution neural network model. The input sample lengths are set to be 128, 256, 512, 1 024, and 2 048, respectively, and the final diagnostic accuracy results and the visual scatter plot show that the effect of 1 DLCNN is optimal.
文摘Machine learning models may outperform traditional statistical regression algorithms for predicting clinical outcomes.Proper validation of building such models and tuning their underlying algorithms is necessary to avoid over-fitting and poor generalizability,which smaller datasets can be more prone to.In an effort to educate readers interested in artificial intelligence and model-building based on machine-learning algorithms,we outline important details on crossvalidation techniques that can enhance the performance and generalizability of such models.
基金Projects(41174061,41374120)supported by the National Natural Science Foundation of China
文摘Geological structures often exhibit smooth characteristics away from sharp discontinuities. One aim of geophysical inversion is to recover information about the smooth structures as well as about the sharp discontinuities. Because no specific operator can provide a perfect sparse representation of complicated geological models, hyper-parameter regularization inversion based on the iterative split Bregman method was used to recover the features of both smooth and sharp geological structures. A novel preconditioned matrix was proposed, which counteracted the natural decay of the sensitivity matrix and its inverse matrix was calculated easily. Application of the algorithm to synthetic data produces density models that are good representations of the designed models. The results show that the algorithm proposed is feasible and effective.
基金Project(513300303)supported by the General Armament Department,China
文摘Gaussian process(GP)has fewer parameters,simple model and output of probabilistic sense,when compared with the methods such as support vector machines.Selection of the hyper-parameters is critical to the performance of Gaussian process model.However,the common-used algorithm has the disadvantages of difficult determination of iteration steps,over-dependence of optimization effect on initial values,and easily falling into local optimum.To solve this problem,a method combining the Gaussian process with memetic algorithm was proposed.Based on this method,memetic algorithm was used to search the optimal hyper parameters of Gaussian process regression(GPR)model in the training process and form MA-GPR algorithms,and then the model was used to predict and test the results.When used in the marine long-range precision strike system(LPSS)battle effectiveness evaluation,the proposed MA-GPR model significantly improved the prediction accuracy,compared with the conjugate gradient method and the genetic algorithm optimization process.
文摘A Bayesian estimator with informative prior distributions (a multi-normal and an inverted gamma distribution), adequate to displacement estimation at dam displacement monitoring networks, is presented. The hyper-parameters of the prior distributions are obtained by Bayesian empirical methods with non-informative meta-priors. The performances of the Bayes estimator and the classical generalized lest squares estimator are compared using two measurements of the horizontal monitoring network of a concrete gravity dam: the Penha Garcia dam (Portugal). In order to test the robustness of the two estimators, a gross error is added to one of the measured horizontal directions: the Bayes estimator proves to be significantly more robust than the classic maximum likelihood estimator.
文摘In this paper we study the problem of model selection for a linear programming-based support vector machine for regression. We propose generalized method that is based on a quasi-Newton method that uses a globalization strategy and an inexact computation of first order information. We explore the case of two-class, multi-class, and regression problems. Simulation results among standard datasets suggest that the algorithm achieves insignificant variability when measuring residual statistical properties.
文摘Domain shift is when the data used in training does not match the ones it will be applied to later on under similar conditions.Domain shift will reduce accuracy in results.To prevent this,domain adaptation is done,which adapts the pre-trained model to the target domain.In real scenarios,the availability of labels for target data is rare thus resulting in unsupervised domain adaptation.Herein,we propose an innovative approach where source-free domain adaptation models and Generative Adversarial Networks(GANs)are integrated to improve the performance of computer vision or robotic vision-based systems in our study.Cosine Generative Adversarial Network(CosGAN)is developed as a GAN that uses cosine embedding loss to handle issues associated with unsupervised source-relax domain adaptations.For less complex architecture,the CosGAN training process has two steps that produce results almost comparable to other state-of-the-art techniques.The efficiency of CosGAN was compared by conducting experiments using benchmarked datasets.The approach was evaluated on different datasets and experimental results show superiority over existing state-of-the-art methods in terms of accuracy as well as generalization ability.This technique has numerous applications including wheeled robots,autonomous vehicles,warehouse automation,and all image-processing-based automation tasks so it can reshape the field of robotic vision with its ability to make robots adapt to new tasks and environments efficiently without requiring additional labeled data.It lays the groundwork for future expansions in robotic vision and applications.Although GAN provides a variety of outstanding features,it also increases the risk of instability and over-fitting of the training data thus making the data difficult to converge.
基金supported by the Yalong River Joint Funds of the National Natural Science Foundation of China(No.U1965207)the National Natural Science Foundation of China(Nos.51839007,51779169,and 52009090)。
文摘Accurate prediction of drilling efficiency is critical for developing the earth-rock excavation schedule.The single machine learning(ML)prediction models usually suffer from problems including parameter sensitivity and overfitting.In addition,the influence of environmental and operational factors is often ignored.In response,a novel stacking-based ensemble learning method taking into account the combined effects of those factors is proposed.Through multiple comparison tests,four models,e Xtreme gradient boosting(XGBoost),random forest(RF),back propagation neural network(BPNN)as the base learners,and support vector regression(SVR)as the meta-learner,are selected for stacking.Furthermore,an improved cuckoo search optimization(ICSO)algorithm is developed for hyper-parameter optimization of the ensemble model.The application to a real-world project demonstrates that the proposed method outperforms the popular single ML method XGBoost and the ensemble model optimized by particle swarm optimization(PSO),with 16.43%and 4.88%improvements of mean absolute percentage error(MAPE),respectively.