This article delves into the analysis of performance and utilization of Support Vector Machines (SVMs) for the critical task of forest fire detection using image datasets. With the increasing threat of forest fires to...This article delves into the analysis of performance and utilization of Support Vector Machines (SVMs) for the critical task of forest fire detection using image datasets. With the increasing threat of forest fires to ecosystems and human settlements, the need for rapid and accurate detection systems is of utmost importance. SVMs, renowned for their strong classification capabilities, exhibit proficiency in recognizing patterns associated with fire within images. By training on labeled data, SVMs acquire the ability to identify distinctive attributes associated with fire, such as flames, smoke, or alterations in the visual characteristics of the forest area. The document thoroughly examines the use of SVMs, covering crucial elements like data preprocessing, feature extraction, and model training. It rigorously evaluates parameters such as accuracy, efficiency, and practical applicability. The knowledge gained from this study aids in the development of efficient forest fire detection systems, enabling prompt responses and improving disaster management. Moreover, the correlation between SVM accuracy and the difficulties presented by high-dimensional datasets is carefully investigated, demonstrated through a revealing case study. The relationship between accuracy scores and the different resolutions used for resizing the training datasets has also been discussed in this article. These comprehensive studies result in a definitive overview of the difficulties faced and the potential sectors requiring further improvement and focus.展开更多
The support vector machine(SVM)is a classical machine learning method.Both the hinge loss and least absolute shrinkage and selection operator(LASSO)penalty are usually used in traditional SVMs.However,the hinge loss i...The support vector machine(SVM)is a classical machine learning method.Both the hinge loss and least absolute shrinkage and selection operator(LASSO)penalty are usually used in traditional SVMs.However,the hinge loss is not differentiable,and the LASSO penalty does not have the Oracle property.In this paper,the huberized loss is combined with non-convex penalties to obtain a model that has the advantages of both the computational simplicity and the Oracle property,contributing to higher accuracy than traditional SVMs.It is experimentally demonstrated that the two non-convex huberized-SVM methods,smoothly clipped absolute deviation huberized-SVM(SCAD-HSVM)and minimax concave penalty huberized-SVM(MCP-HSVM),outperform the traditional SVM method in terms of the prediction accuracy and classifier performance.They are also superior in terms of variable selection,especially when there is a high linear correlation between the variables.When they are applied to the prediction of listed companies,the variables that can affect and predict financial distress are accurately filtered out.Among all the indicators,the indicators per share have the greatest influence while those of solvency have the weakest influence.Listed companies can assess the financial situation with the indicators screened by our algorithm and make an early warning of their possible financial distress in advance with higher precision.展开更多
Effective fault diagnosis and fault-tolerant control method for aeronautics electromechanical actuator is concerned in this paper.By borrowing the advantages of model-driven and data-driven methods,a fault tolerant no...Effective fault diagnosis and fault-tolerant control method for aeronautics electromechanical actuator is concerned in this paper.By borrowing the advantages of model-driven and data-driven methods,a fault tolerant nonsingular terminal sliding mode control method based on support vector machine(SVM)is proposed.A SVM is designed to estimate the fault by off-line learning from small sample data with solving convex quadratic programming method and is introduced into a high-gain observer,so as to improve the state estimation and fault detection accuracy when the fault occurs.The state estimation value of the observer is used for state reconfiguration.A novel nonsingular terminal sliding mode surface is designed,and Lyapunov theorem is used to derive a parameter adaptation law and a control law.It is guaranteed that the proposed controller can achieve asymptotical stability which is superior to many advanced fault-tolerant controllers.In addition,the parameter estimation also can help to diagnose the system faults because the faults can be reflected by the parameters variation.Extensive comparative simulation and experimental results illustrate the effectiveness and advancement of the proposed controller compared with several other main-stream controllers.展开更多
The distribution of data has a significant impact on the results of classification.When the distribution of one class is insignificant compared to the distribution of another class,data imbalance occurs.This will resu...The distribution of data has a significant impact on the results of classification.When the distribution of one class is insignificant compared to the distribution of another class,data imbalance occurs.This will result in rising outlier values and noise.Therefore,the speed and performance of classification could be greatly affected.Given the above problems,this paper starts with the motivation and mathematical representing of classification,puts forward a new classification method based on the relationship between different classification formulations.Combined with the vector characteristics of the actual problem and the choice of matrix characteristics,we firstly analyze the orderly regression to introduce slack variables to solve the constraint problem of the lone point.Then we introduce the fuzzy factors to solve the problem of the gap between the isolated points on the basis of the support vector machine.We introduce the cost control to solve the problem of sample skew.Finally,based on the bi-boundary support vector machine,a twostep weight setting twin classifier is constructed.This can help to identify multitasks with feature-selected patterns without the need for additional optimizers,which solves the problem of large-scale classification that can’t deal effectively with the very low category distribution gap.展开更多
With the widespread data collection and processing,privacy-preserving machine learning has become increasingly important in addressing privacy risks related to individuals.Support vector machine(SVM)is one of the most...With the widespread data collection and processing,privacy-preserving machine learning has become increasingly important in addressing privacy risks related to individuals.Support vector machine(SVM)is one of the most elementary learning models of machine learning.Privacy issues surrounding SVM classifier training have attracted increasing attention.In this paper,we investigate Differential Privacy-compliant Federated Machine Learning with Dimensionality Reduction,called FedDPDR-DPML,which greatly improves data utility while providing strong privacy guarantees.Considering in distributed learning scenarios,multiple participants usually hold unbalanced or small amounts of data.Therefore,FedDPDR-DPML enables multiple participants to collaboratively learn a global model based on weighted model averaging and knowledge aggregation and then the server distributes the global model to each participant to improve local data utility.Aiming at high-dimensional data,we adopt differential privacy in both the principal component analysis(PCA)-based dimensionality reduction phase and SVM classifiers training phase,which improves model accuracy while achieving strict differential privacy protection.Besides,we train Differential privacy(DP)-compliant SVM classifiers by adding noise to the objective function itself,thus leading to better data utility.Extensive experiments on three high-dimensional datasets demonstrate that FedDPDR-DPML can achieve high accuracy while ensuring strong privacy protection.展开更多
The selection of important factors in machine learning-based susceptibility assessments is crucial to obtain reliable susceptibility results.In this study,metaheuristic optimization and feature selection techniques we...The selection of important factors in machine learning-based susceptibility assessments is crucial to obtain reliable susceptibility results.In this study,metaheuristic optimization and feature selection techniques were applied to identify the most important input parameters for mapping debris flow susceptibility in the southern mountain area of Chengde City in Hebei Province,China,by using machine learning algorithms.In total,133 historical debris flow records and 16 related factors were selected.The support vector machine(SVM)was first used as the base classifier,and then a hybrid model was introduced by a two-step process.First,the particle swarm optimization(PSO)algorithm was employed to select the SVM model hyperparameters.Second,two feature selection algorithms,namely principal component analysis(PCA)and PSO,were integrated into the PSO-based SVM model,which generated the PCA-PSO-SVM and FS-PSO-SVM models,respectively.Three statistical metrics(accuracy,recall,and specificity)and the area under the receiver operating characteristic curve(AUC)were employed to evaluate and validate the performance of the models.The results indicated that the feature selection-based models exhibited the best performance,followed by the PSO-based SVM and SVM models.Moreover,the performance of the FS-PSO-SVM model was better than that of the PCA-PSO-SVM model,showing the highest AUC,accuracy,recall,and specificity values in both the training and testing processes.It was found that the selection of optimal features is crucial to improving the reliability of debris flow susceptibility assessment results.Moreover,the PSO algorithm was found to be not only an effective tool for hyperparameter optimization,but also a useful feature selection algorithm to improve prediction accuracies of debris flow susceptibility by using machine learning algorithms.The high and very high debris flow susceptibility zone appropriately covers 38.01%of the study area,where debris flow may occur under intensive human activities and heavy rainfall events.展开更多
Algorithms for steganography are methods of hiding data transfers in media files.Several machine learning architectures have been presented recently to improve stego image identification performance by using spatial i...Algorithms for steganography are methods of hiding data transfers in media files.Several machine learning architectures have been presented recently to improve stego image identification performance by using spatial information,and these methods have made it feasible to handle a wide range of problems associated with image analysis.Images with little information or low payload are used by information embedding methods,but the goal of all contemporary research is to employ high-payload images for classification.To address the need for both low-and high-payload images,this work provides a machine-learning approach to steganography image classification that uses Curvelet transformation to efficiently extract characteristics from both type of images.Support Vector Machine(SVM),a commonplace classification technique,has been employed to determine whether the image is a stego or cover.The Wavelet Obtained Weights(WOW),Spatial Universal Wavelet Relative Distortion(S-UNIWARD),Highly Undetectable Steganography(HUGO),and Minimizing the Power of Optimal Detector(MiPOD)steganography techniques are used in a variety of experimental scenarios to evaluate the performance of the proposedmethod.Using WOW at several payloads,the proposed approach proves its classification accuracy of 98.60%.It exhibits its superiority over SOTA methods.展开更多
AIM:To develop a classifier for traditional Chinese medicine(TCM)syndrome differentiation of diabetic retinopathy(DR),using optimized machine learning algorithms,which can provide the basis for TCM objective and intel...AIM:To develop a classifier for traditional Chinese medicine(TCM)syndrome differentiation of diabetic retinopathy(DR),using optimized machine learning algorithms,which can provide the basis for TCM objective and intelligent syndrome differentiation.METHODS:Collated data on real-world DR cases were collected.A variety of machine learning methods were used to construct TCM syndrome classification model,and the best performance was selected as the basic model.Genetic Algorithm(GA)was used for feature selection to obtain the optimal feature combination.Harris Hawk Optimization(HHO)was used for parameter optimization,and a classification model based on feature selection and parameter optimization was constructed.The performance of the model was compared with other optimization algorithms.The models were evaluated with accuracy,precision,recall,and F1 score as indicators.RESULTS:Data on 970 cases that met screening requirements were collected.Support Vector Machine(SVM)was the best basic classification model.The accuracy rate of the model was 82.05%,the precision rate was 82.34%,the recall rate was 81.81%,and the F1 value was 81.76%.After GA screening,the optimal feature combination contained 37 feature values,which was consistent with TCM clinical practice.The model based on optimal combination and SVM(GA_SVM)had an accuracy improvement of 1.92%compared to the basic classifier.SVM model based on HHO and GA optimization(HHO_GA_SVM)had the best performance and convergence speed compared with other optimization algorithms.Compared with the basic classification model,the accuracy was improved by 3.51%.CONCLUSION:HHO and GA optimization can improve the model performance of SVM in TCM syndrome differentiation of DR.It provides a new method and research idea for TCM intelligent assisted syndrome differentiation.展开更多
The turbidite channel of South China Sea has been highly concerned.Influenced by the complex fault and the rapid phase change of lithofacies,predicting the channel through conventional seismic attributes is not accura...The turbidite channel of South China Sea has been highly concerned.Influenced by the complex fault and the rapid phase change of lithofacies,predicting the channel through conventional seismic attributes is not accurate enough.In response to this disadvantage,this study used a method combining grey relational analysis(GRA)and support vectormachine(SVM)and established a set of prediction technical procedures suitable for reservoirs with complex geological conditions.In the case study of the Huangliu Formation in Qiongdongnan Basin,South China Sea,this study first dimensionalized the conventional seismic attributes of Gas Layer Group I and then used the GRA method to obtain the main relational factors.A higher relational degree indicates a higher probability of responding to the attributes of the turbidite channel.This study then accumulated the optimized attributes with the highest relational factors to obtain a first-order accumulated sequence,which was used as the input training sample of the SVM model,thus successfully constructing the SVM turbidite channel model.Drilling results prove that the GRA-SVMmethod has a high drilling coincidence rate.Utilizing the core and logging data and taking full use of the advantages of seismic inversion in predicting the sand boundary of water channels,this study divides the sedimentary microfacies of the Huangliu Formation in the Lingshui 17-2 Gas Field.This comprehensive study has shown that the GRA-SVM method has high accuracy for predicting turbidite channels and can be used as a superior turbidite channel prediction method under complex geological conditions.展开更多
The relationship among Mercer kernel, reproducing kernel and positive definite kernel in support vector machine (SVM) is proved and their roles in SVM are discussed. The quadratic form of the kernel matrix is used t...The relationship among Mercer kernel, reproducing kernel and positive definite kernel in support vector machine (SVM) is proved and their roles in SVM are discussed. The quadratic form of the kernel matrix is used to confirm the positive definiteness and their construction. Based on the Bochner theorem, some translation invariant kernels are checked in their Fourier domain. Some rotation invariant radial kernels are inspected according to the Schoenberg theorem. Finally, the construction of discrete scaling and wavelet kernels, the kernel selection and the kernel parameter learning are discussed.展开更多
针对浮选过程变量滞后、耦合特征及建模样本数量少所导致精矿品位难以准确预测的问题,提出了一种基于改进麻雀搜索算法(Improved Sparrow Search Algorithm,ISSA)优化混核最小二乘支持向量机(Hybrid Kernel Least Squares Support Vecto...针对浮选过程变量滞后、耦合特征及建模样本数量少所导致精矿品位难以准确预测的问题,提出了一种基于改进麻雀搜索算法(Improved Sparrow Search Algorithm,ISSA)优化混核最小二乘支持向量机(Hybrid Kernel Least Squares Support Vector Machine,HKLSSVM)的浮选过程精矿品位预测方法.首先采集浮选现场载流X荧光品位分析仪数据作为建模变量并进行预处理,建立基于最小二乘支持向量机(Least Squares Support Vector Machine,LSSVM)的预测模型,以此构建新型混合核函数,将输入空间映射至高维特征空间,再引入改进麻雀搜索算法对模型参数进行优化,提出基于ISSA-HKLSSVM方法实现精矿品位预测,最后开发基于LabVIEW的浮选精矿品位预测系统对本文提出方法实际验证.实验结果表明,本文提出方法对于浮选过程小样本建模具有良好拟合能力,相比现有方法提高了预测准确率,可实现精矿品位的准确在线预测,为浮选过程的智能调控提供实时可靠的精矿品位反馈信息.展开更多
Support vector machines (SVMs) are combined with the artificial immune network (aiNet), thus forming a new hybrid ai-SVM algorithm. The algorithm is used to reduce the number of samples and the training time of SV...Support vector machines (SVMs) are combined with the artificial immune network (aiNet), thus forming a new hybrid ai-SVM algorithm. The algorithm is used to reduce the number of samples and the training time of SVM on large datasets, aiNet is an artificial immune system (AIS) inspired method to perform the automatic data compression, extract the relevant information and retain the topology of the original sample distribution. The output of aiNet is a set of antibodies for representing the input dataset in a simplified way. Then the SVM model is built in the compressed antibody network instead of the original input data. Experimental results show that the ai-SVM algorithm is effective to reduce the computing time and simplify the SVM model, and the accuracy is not decreased.展开更多
文摘This article delves into the analysis of performance and utilization of Support Vector Machines (SVMs) for the critical task of forest fire detection using image datasets. With the increasing threat of forest fires to ecosystems and human settlements, the need for rapid and accurate detection systems is of utmost importance. SVMs, renowned for their strong classification capabilities, exhibit proficiency in recognizing patterns associated with fire within images. By training on labeled data, SVMs acquire the ability to identify distinctive attributes associated with fire, such as flames, smoke, or alterations in the visual characteristics of the forest area. The document thoroughly examines the use of SVMs, covering crucial elements like data preprocessing, feature extraction, and model training. It rigorously evaluates parameters such as accuracy, efficiency, and practical applicability. The knowledge gained from this study aids in the development of efficient forest fire detection systems, enabling prompt responses and improving disaster management. Moreover, the correlation between SVM accuracy and the difficulties presented by high-dimensional datasets is carefully investigated, demonstrated through a revealing case study. The relationship between accuracy scores and the different resolutions used for resizing the training datasets has also been discussed in this article. These comprehensive studies result in a definitive overview of the difficulties faced and the potential sectors requiring further improvement and focus.
文摘The support vector machine(SVM)is a classical machine learning method.Both the hinge loss and least absolute shrinkage and selection operator(LASSO)penalty are usually used in traditional SVMs.However,the hinge loss is not differentiable,and the LASSO penalty does not have the Oracle property.In this paper,the huberized loss is combined with non-convex penalties to obtain a model that has the advantages of both the computational simplicity and the Oracle property,contributing to higher accuracy than traditional SVMs.It is experimentally demonstrated that the two non-convex huberized-SVM methods,smoothly clipped absolute deviation huberized-SVM(SCAD-HSVM)and minimax concave penalty huberized-SVM(MCP-HSVM),outperform the traditional SVM method in terms of the prediction accuracy and classifier performance.They are also superior in terms of variable selection,especially when there is a high linear correlation between the variables.When they are applied to the prediction of listed companies,the variables that can affect and predict financial distress are accurately filtered out.Among all the indicators,the indicators per share have the greatest influence while those of solvency have the weakest influence.Listed companies can assess the financial situation with the indicators screened by our algorithm and make an early warning of their possible financial distress in advance with higher precision.
基金Supported by National Natural Science Foundation of China (Grant No.51975294)Fundamental Research Funds for the Central Universities of China (Grant No.30922010706)。
文摘Effective fault diagnosis and fault-tolerant control method for aeronautics electromechanical actuator is concerned in this paper.By borrowing the advantages of model-driven and data-driven methods,a fault tolerant nonsingular terminal sliding mode control method based on support vector machine(SVM)is proposed.A SVM is designed to estimate the fault by off-line learning from small sample data with solving convex quadratic programming method and is introduced into a high-gain observer,so as to improve the state estimation and fault detection accuracy when the fault occurs.The state estimation value of the observer is used for state reconfiguration.A novel nonsingular terminal sliding mode surface is designed,and Lyapunov theorem is used to derive a parameter adaptation law and a control law.It is guaranteed that the proposed controller can achieve asymptotical stability which is superior to many advanced fault-tolerant controllers.In addition,the parameter estimation also can help to diagnose the system faults because the faults can be reflected by the parameters variation.Extensive comparative simulation and experimental results illustrate the effectiveness and advancement of the proposed controller compared with several other main-stream controllers.
基金Hebei Province Key Research and Development Project(No.20313701D)Hebei Province Key Research and Development Project(No.19210404D)+13 种基金Mobile computing and universal equipment for the Beijing Key Laboratory Open Project,The National Social Science Fund of China(17AJL014)Beijing University of Posts and Telecommunications Construction of World-Class Disciplines and Characteristic Development Guidance Special Fund “Cultural Inheritance and Innovation”Project(No.505019221)National Natural Science Foundation of China(No.U1536112)National Natural Science Foundation of China(No.81673697)National Natural Science Foundation of China(61872046)The National Social Science Fund Key Project of China(No.17AJL014)“Blue Fire Project”(Huizhou)University of Technology Joint Innovation Project(CXZJHZ201729)Industry-University Cooperation Cooperative Education Project of the Ministry of Education(No.201902218004)Industry-University Cooperation Cooperative Education Project of the Ministry of Education(No.201902024006)Industry-University Cooperation Cooperative Education Project of the Ministry of Education(No.201901197007)Industry-University Cooperation Collaborative Education Project of the Ministry of Education(No.201901199005)The Ministry of Education Industry-University Cooperation Collaborative Education Project(No.201901197001)Shijiazhuang science and technology plan project(236240267A)Hebei Province key research and development plan project(20312701D)。
文摘The distribution of data has a significant impact on the results of classification.When the distribution of one class is insignificant compared to the distribution of another class,data imbalance occurs.This will result in rising outlier values and noise.Therefore,the speed and performance of classification could be greatly affected.Given the above problems,this paper starts with the motivation and mathematical representing of classification,puts forward a new classification method based on the relationship between different classification formulations.Combined with the vector characteristics of the actual problem and the choice of matrix characteristics,we firstly analyze the orderly regression to introduce slack variables to solve the constraint problem of the lone point.Then we introduce the fuzzy factors to solve the problem of the gap between the isolated points on the basis of the support vector machine.We introduce the cost control to solve the problem of sample skew.Finally,based on the bi-boundary support vector machine,a twostep weight setting twin classifier is constructed.This can help to identify multitasks with feature-selected patterns without the need for additional optimizers,which solves the problem of large-scale classification that can’t deal effectively with the very low category distribution gap.
基金supported in part by National Natural Science Foundation of China(Nos.62102311,62202377,62272385)in part by Natural Science Basic Research Program of Shaanxi(Nos.2022JQ-600,2022JM-353,2023-JC-QN-0327)+2 种基金in part by Shaanxi Distinguished Youth Project(No.2022JC-47)in part by Scientific Research Program Funded by Shaanxi Provincial Education Department(No.22JK0560)in part by Distinguished Youth Talents of Shaanxi Universities,and in part by Youth Innovation Team of Shaanxi Universities.
文摘With the widespread data collection and processing,privacy-preserving machine learning has become increasingly important in addressing privacy risks related to individuals.Support vector machine(SVM)is one of the most elementary learning models of machine learning.Privacy issues surrounding SVM classifier training have attracted increasing attention.In this paper,we investigate Differential Privacy-compliant Federated Machine Learning with Dimensionality Reduction,called FedDPDR-DPML,which greatly improves data utility while providing strong privacy guarantees.Considering in distributed learning scenarios,multiple participants usually hold unbalanced or small amounts of data.Therefore,FedDPDR-DPML enables multiple participants to collaboratively learn a global model based on weighted model averaging and knowledge aggregation and then the server distributes the global model to each participant to improve local data utility.Aiming at high-dimensional data,we adopt differential privacy in both the principal component analysis(PCA)-based dimensionality reduction phase and SVM classifiers training phase,which improves model accuracy while achieving strict differential privacy protection.Besides,we train Differential privacy(DP)-compliant SVM classifiers by adding noise to the objective function itself,thus leading to better data utility.Extensive experiments on three high-dimensional datasets demonstrate that FedDPDR-DPML can achieve high accuracy while ensuring strong privacy protection.
基金supported by the Second Tibetan Plateau Scientific Expedition and Research Program(Grant no.2019QZKK0904)Natural Science Foundation of Hebei Province(Grant no.D2022403032)S&T Program of Hebei(Grant no.E2021403001).
文摘The selection of important factors in machine learning-based susceptibility assessments is crucial to obtain reliable susceptibility results.In this study,metaheuristic optimization and feature selection techniques were applied to identify the most important input parameters for mapping debris flow susceptibility in the southern mountain area of Chengde City in Hebei Province,China,by using machine learning algorithms.In total,133 historical debris flow records and 16 related factors were selected.The support vector machine(SVM)was first used as the base classifier,and then a hybrid model was introduced by a two-step process.First,the particle swarm optimization(PSO)algorithm was employed to select the SVM model hyperparameters.Second,two feature selection algorithms,namely principal component analysis(PCA)and PSO,were integrated into the PSO-based SVM model,which generated the PCA-PSO-SVM and FS-PSO-SVM models,respectively.Three statistical metrics(accuracy,recall,and specificity)and the area under the receiver operating characteristic curve(AUC)were employed to evaluate and validate the performance of the models.The results indicated that the feature selection-based models exhibited the best performance,followed by the PSO-based SVM and SVM models.Moreover,the performance of the FS-PSO-SVM model was better than that of the PCA-PSO-SVM model,showing the highest AUC,accuracy,recall,and specificity values in both the training and testing processes.It was found that the selection of optimal features is crucial to improving the reliability of debris flow susceptibility assessment results.Moreover,the PSO algorithm was found to be not only an effective tool for hyperparameter optimization,but also a useful feature selection algorithm to improve prediction accuracies of debris flow susceptibility by using machine learning algorithms.The high and very high debris flow susceptibility zone appropriately covers 38.01%of the study area,where debris flow may occur under intensive human activities and heavy rainfall events.
基金financially supported by the Deanship of Scientific Research at King Khalid University under Research Grant Number(R.G.P.2/549/44).
文摘Algorithms for steganography are methods of hiding data transfers in media files.Several machine learning architectures have been presented recently to improve stego image identification performance by using spatial information,and these methods have made it feasible to handle a wide range of problems associated with image analysis.Images with little information or low payload are used by information embedding methods,but the goal of all contemporary research is to employ high-payload images for classification.To address the need for both low-and high-payload images,this work provides a machine-learning approach to steganography image classification that uses Curvelet transformation to efficiently extract characteristics from both type of images.Support Vector Machine(SVM),a commonplace classification technique,has been employed to determine whether the image is a stego or cover.The Wavelet Obtained Weights(WOW),Spatial Universal Wavelet Relative Distortion(S-UNIWARD),Highly Undetectable Steganography(HUGO),and Minimizing the Power of Optimal Detector(MiPOD)steganography techniques are used in a variety of experimental scenarios to evaluate the performance of the proposedmethod.Using WOW at several payloads,the proposed approach proves its classification accuracy of 98.60%.It exhibits its superiority over SOTA methods.
基金Supported by Hunan Province Traditional Chinese Medicine Research Project(No.B2023043)Hunan Provincial Department of Education Scientific Research Project(No.22B0386)Hunan University of Traditional Chinese Medicine Campus level Research Fund Project(No.2022XJZKC004).
文摘AIM:To develop a classifier for traditional Chinese medicine(TCM)syndrome differentiation of diabetic retinopathy(DR),using optimized machine learning algorithms,which can provide the basis for TCM objective and intelligent syndrome differentiation.METHODS:Collated data on real-world DR cases were collected.A variety of machine learning methods were used to construct TCM syndrome classification model,and the best performance was selected as the basic model.Genetic Algorithm(GA)was used for feature selection to obtain the optimal feature combination.Harris Hawk Optimization(HHO)was used for parameter optimization,and a classification model based on feature selection and parameter optimization was constructed.The performance of the model was compared with other optimization algorithms.The models were evaluated with accuracy,precision,recall,and F1 score as indicators.RESULTS:Data on 970 cases that met screening requirements were collected.Support Vector Machine(SVM)was the best basic classification model.The accuracy rate of the model was 82.05%,the precision rate was 82.34%,the recall rate was 81.81%,and the F1 value was 81.76%.After GA screening,the optimal feature combination contained 37 feature values,which was consistent with TCM clinical practice.The model based on optimal combination and SVM(GA_SVM)had an accuracy improvement of 1.92%compared to the basic classifier.SVM model based on HHO and GA optimization(HHO_GA_SVM)had the best performance and convergence speed compared with other optimization algorithms.Compared with the basic classification model,the accuracy was improved by 3.51%.CONCLUSION:HHO and GA optimization can improve the model performance of SVM in TCM syndrome differentiation of DR.It provides a new method and research idea for TCM intelligent assisted syndrome differentiation.
基金grateful for Science and Technology Innovation Ability Cultivation Project of Hebei Provincial Planning for College and Middle School Students(22E50590D)Priority Research Project of Langfang Education Sciences Planning(JCJY202130).
文摘The turbidite channel of South China Sea has been highly concerned.Influenced by the complex fault and the rapid phase change of lithofacies,predicting the channel through conventional seismic attributes is not accurate enough.In response to this disadvantage,this study used a method combining grey relational analysis(GRA)and support vectormachine(SVM)and established a set of prediction technical procedures suitable for reservoirs with complex geological conditions.In the case study of the Huangliu Formation in Qiongdongnan Basin,South China Sea,this study first dimensionalized the conventional seismic attributes of Gas Layer Group I and then used the GRA method to obtain the main relational factors.A higher relational degree indicates a higher probability of responding to the attributes of the turbidite channel.This study then accumulated the optimized attributes with the highest relational factors to obtain a first-order accumulated sequence,which was used as the input training sample of the SVM model,thus successfully constructing the SVM turbidite channel model.Drilling results prove that the GRA-SVMmethod has a high drilling coincidence rate.Utilizing the core and logging data and taking full use of the advantages of seismic inversion in predicting the sand boundary of water channels,this study divides the sedimentary microfacies of the Huangliu Formation in the Lingshui 17-2 Gas Field.This comprehensive study has shown that the GRA-SVM method has high accuracy for predicting turbidite channels and can be used as a superior turbidite channel prediction method under complex geological conditions.
基金Supported by the National Natural Science Foundation of China(60473035)~~
文摘The relationship among Mercer kernel, reproducing kernel and positive definite kernel in support vector machine (SVM) is proved and their roles in SVM are discussed. The quadratic form of the kernel matrix is used to confirm the positive definiteness and their construction. Based on the Bochner theorem, some translation invariant kernels are checked in their Fourier domain. Some rotation invariant radial kernels are inspected according to the Schoenberg theorem. Finally, the construction of discrete scaling and wavelet kernels, the kernel selection and the kernel parameter learning are discussed.
文摘针对浮选过程变量滞后、耦合特征及建模样本数量少所导致精矿品位难以准确预测的问题,提出了一种基于改进麻雀搜索算法(Improved Sparrow Search Algorithm,ISSA)优化混核最小二乘支持向量机(Hybrid Kernel Least Squares Support Vector Machine,HKLSSVM)的浮选过程精矿品位预测方法.首先采集浮选现场载流X荧光品位分析仪数据作为建模变量并进行预处理,建立基于最小二乘支持向量机(Least Squares Support Vector Machine,LSSVM)的预测模型,以此构建新型混合核函数,将输入空间映射至高维特征空间,再引入改进麻雀搜索算法对模型参数进行优化,提出基于ISSA-HKLSSVM方法实现精矿品位预测,最后开发基于LabVIEW的浮选精矿品位预测系统对本文提出方法实际验证.实验结果表明,本文提出方法对于浮选过程小样本建模具有良好拟合能力,相比现有方法提高了预测准确率,可实现精矿品位的准确在线预测,为浮选过程的智能调控提供实时可靠的精矿品位反馈信息.
文摘Support vector machines (SVMs) are combined with the artificial immune network (aiNet), thus forming a new hybrid ai-SVM algorithm. The algorithm is used to reduce the number of samples and the training time of SVM on large datasets, aiNet is an artificial immune system (AIS) inspired method to perform the automatic data compression, extract the relevant information and retain the topology of the original sample distribution. The output of aiNet is a set of antibodies for representing the input dataset in a simplified way. Then the SVM model is built in the compressed antibody network instead of the original input data. Experimental results show that the ai-SVM algorithm is effective to reduce the computing time and simplify the SVM model, and the accuracy is not decreased.