In this paper, sixty-eight research articles published between 2000 and 2017 as well as textbooks which employed four classification algorithms: K-Nearest-Neighbor (KNN), Support Vector Machines (SVM), Random Forest (...In this paper, sixty-eight research articles published between 2000 and 2017 as well as textbooks which employed four classification algorithms: K-Nearest-Neighbor (KNN), Support Vector Machines (SVM), Random Forest (RF) and Neural Network (NN) as the main statistical tools were reviewed. The aim was to examine and compare these nonparametric classification methods on the following attributes: robustness to training data, sensitivity to changes, data fitting, stability, ability to handle large data sizes, sensitivity to noise, time invested in parameter tuning, and accuracy. The performances, strengths and shortcomings of each of the algorithms were examined, and finally, a conclusion was arrived at on which one has higher performance. It was evident from the literature reviewed that RF is too sensitive to small changes in the training dataset and is occasionally unstable and tends to overfit in the model. KNN is easy to implement and understand but has a major drawback of becoming significantly slow as the size of the data in use grows, while the ideal value of K for the KNN classifier is difficult to set. SVM and RF are insensitive to noise or overtraining, which shows their ability in dealing with unbalanced data. Larger input datasets will lengthen classification times for NN and KNN more than for SVM and RF. Among these nonparametric classification methods, NN has the potential to become a more widely used classification algorithm, but because of their time-consuming parameter tuning procedure, high level of complexity in computational processing, the numerous types of NN architectures to choose from and the high number of algorithms used for training, most researchers recommend SVM and RF as easier and wieldy used methods which repeatedly achieve results with high accuracies and are often faster to implement.展开更多
We report a theoretical work on the properties of modulational instability and bright type nonlinear localized modes in one-dimensional easy-axis weak ferromagnetic spin lattices involving next-nearest-neighbor coupli...We report a theoretical work on the properties of modulational instability and bright type nonlinear localized modes in one-dimensional easy-axis weak ferromagnetic spin lattices involving next-nearest-neighbor couplings.With a linear stability analysis, we calculate the growth rates of the modulational instability, and plot the instability regions.When the strength of the next-nearest-neighbor coupling is large enough, two new asymmetric modulational instability regions appear near the boundary of the first Brillouin zone.Furthermore, analytical forms of the bright nonlinear localized modes are constructed by means of a quasi-discreteness approach.The influence of the next-nearest-neighbor coupling on the Brillouin zone center mode and boundary mode are discussed.In particular, we discover a reversal phenomenon of the propagation direction of the Brillouin zone boundary mode.展开更多
Missing values are prevalent in real-world datasets and they may reduce predictive performance of a learning algorithm. Dissolved Gas Analysis (DGA), one of the most deployable methods for detecting and predicting inc...Missing values are prevalent in real-world datasets and they may reduce predictive performance of a learning algorithm. Dissolved Gas Analysis (DGA), one of the most deployable methods for detecting and predicting incipient faults in power transformers is one of the casualties. Thus, this paper proposes filling-in the missing values found in a DGA dataset using the k-nearest neighbor imputation method with two different distance metrics: Euclidean and Cityblock. Thereafter, using these imputed datasets as inputs, this study applies Support Vector Machine (SVM) to built models which are used to classify transformer faults. Experimental results are provided to show the effectiveness of the proposed approach.展开更多
文摘In this paper, sixty-eight research articles published between 2000 and 2017 as well as textbooks which employed four classification algorithms: K-Nearest-Neighbor (KNN), Support Vector Machines (SVM), Random Forest (RF) and Neural Network (NN) as the main statistical tools were reviewed. The aim was to examine and compare these nonparametric classification methods on the following attributes: robustness to training data, sensitivity to changes, data fitting, stability, ability to handle large data sizes, sensitivity to noise, time invested in parameter tuning, and accuracy. The performances, strengths and shortcomings of each of the algorithms were examined, and finally, a conclusion was arrived at on which one has higher performance. It was evident from the literature reviewed that RF is too sensitive to small changes in the training dataset and is occasionally unstable and tends to overfit in the model. KNN is easy to implement and understand but has a major drawback of becoming significantly slow as the size of the data in use grows, while the ideal value of K for the KNN classifier is difficult to set. SVM and RF are insensitive to noise or overtraining, which shows their ability in dealing with unbalanced data. Larger input datasets will lengthen classification times for NN and KNN more than for SVM and RF. Among these nonparametric classification methods, NN has the potential to become a more widely used classification algorithm, but because of their time-consuming parameter tuning procedure, high level of complexity in computational processing, the numerous types of NN architectures to choose from and the high number of algorithms used for training, most researchers recommend SVM and RF as easier and wieldy used methods which repeatedly achieve results with high accuracies and are often faster to implement.
基金Project supported by the National Natural Science Foundation of China(Grant Nos.11604121 and 11875126)the Natural Science Fund Project of Hunan Province,China(Grant No.2017JJ3255)+1 种基金the National College Students’ Innovation Entrepreneurship Training Program,China(Grant No.201810531014)the Scientific Research Fund of Hunan Provincial Education Department,China(Grant No.17B212)
文摘We report a theoretical work on the properties of modulational instability and bright type nonlinear localized modes in one-dimensional easy-axis weak ferromagnetic spin lattices involving next-nearest-neighbor couplings.With a linear stability analysis, we calculate the growth rates of the modulational instability, and plot the instability regions.When the strength of the next-nearest-neighbor coupling is large enough, two new asymmetric modulational instability regions appear near the boundary of the first Brillouin zone.Furthermore, analytical forms of the bright nonlinear localized modes are constructed by means of a quasi-discreteness approach.The influence of the next-nearest-neighbor coupling on the Brillouin zone center mode and boundary mode are discussed.In particular, we discover a reversal phenomenon of the propagation direction of the Brillouin zone boundary mode.
文摘Missing values are prevalent in real-world datasets and they may reduce predictive performance of a learning algorithm. Dissolved Gas Analysis (DGA), one of the most deployable methods for detecting and predicting incipient faults in power transformers is one of the casualties. Thus, this paper proposes filling-in the missing values found in a DGA dataset using the k-nearest neighbor imputation method with two different distance metrics: Euclidean and Cityblock. Thereafter, using these imputed datasets as inputs, this study applies Support Vector Machine (SVM) to built models which are used to classify transformer faults. Experimental results are provided to show the effectiveness of the proposed approach.