In this paper, sixty-eight research articles published between 2000 and 2017 as well as textbooks which employed four classification algorithms: K-Nearest-Neighbor (KNN), Support Vector Machines (SVM), Random Forest (...In this paper, sixty-eight research articles published between 2000 and 2017 as well as textbooks which employed four classification algorithms: K-Nearest-Neighbor (KNN), Support Vector Machines (SVM), Random Forest (RF) and Neural Network (NN) as the main statistical tools were reviewed. The aim was to examine and compare these nonparametric classification methods on the following attributes: robustness to training data, sensitivity to changes, data fitting, stability, ability to handle large data sizes, sensitivity to noise, time invested in parameter tuning, and accuracy. The performances, strengths and shortcomings of each of the algorithms were examined, and finally, a conclusion was arrived at on which one has higher performance. It was evident from the literature reviewed that RF is too sensitive to small changes in the training dataset and is occasionally unstable and tends to overfit in the model. KNN is easy to implement and understand but has a major drawback of becoming significantly slow as the size of the data in use grows, while the ideal value of K for the KNN classifier is difficult to set. SVM and RF are insensitive to noise or overtraining, which shows their ability in dealing with unbalanced data. Larger input datasets will lengthen classification times for NN and KNN more than for SVM and RF. Among these nonparametric classification methods, NN has the potential to become a more widely used classification algorithm, but because of their time-consuming parameter tuning procedure, high level of complexity in computational processing, the numerous types of NN architectures to choose from and the high number of algorithms used for training, most researchers recommend SVM and RF as easier and wieldy used methods which repeatedly achieve results with high accuracies and are often faster to implement.展开更多
在处理雷达信号时,基于密度的空间聚类(Density-based spatial clustering of applications with noise,DBSCAN)分选算法依赖于参数或阈值的选取,影响分选的准确率。为此提出了一种改进的雷达信号脉冲分选算法,在DBSCAN聚类基础上结合了...在处理雷达信号时,基于密度的空间聚类(Density-based spatial clustering of applications with noise,DBSCAN)分选算法依赖于参数或阈值的选取,影响分选的准确率。为此提出了一种改进的雷达信号脉冲分选算法,在DBSCAN聚类基础上结合了K中位最近邻(K-median nearest neighbor,KMNN)算法,通过引入自衰减系数并设置阈值上限对参数值列表进行二次处理,可以自适应根据聚类结果与不同参数时的K值之间的关系确定最优的邻域半径和最少点个数,提高了分选的正确率。通过仿真实验验证了算法利用雷达脉冲描述字特征进行自适应分选的有效性。展开更多
文摘In this paper, sixty-eight research articles published between 2000 and 2017 as well as textbooks which employed four classification algorithms: K-Nearest-Neighbor (KNN), Support Vector Machines (SVM), Random Forest (RF) and Neural Network (NN) as the main statistical tools were reviewed. The aim was to examine and compare these nonparametric classification methods on the following attributes: robustness to training data, sensitivity to changes, data fitting, stability, ability to handle large data sizes, sensitivity to noise, time invested in parameter tuning, and accuracy. The performances, strengths and shortcomings of each of the algorithms were examined, and finally, a conclusion was arrived at on which one has higher performance. It was evident from the literature reviewed that RF is too sensitive to small changes in the training dataset and is occasionally unstable and tends to overfit in the model. KNN is easy to implement and understand but has a major drawback of becoming significantly slow as the size of the data in use grows, while the ideal value of K for the KNN classifier is difficult to set. SVM and RF are insensitive to noise or overtraining, which shows their ability in dealing with unbalanced data. Larger input datasets will lengthen classification times for NN and KNN more than for SVM and RF. Among these nonparametric classification methods, NN has the potential to become a more widely used classification algorithm, but because of their time-consuming parameter tuning procedure, high level of complexity in computational processing, the numerous types of NN architectures to choose from and the high number of algorithms used for training, most researchers recommend SVM and RF as easier and wieldy used methods which repeatedly achieve results with high accuracies and are often faster to implement.
文摘在处理雷达信号时,基于密度的空间聚类(Density-based spatial clustering of applications with noise,DBSCAN)分选算法依赖于参数或阈值的选取,影响分选的准确率。为此提出了一种改进的雷达信号脉冲分选算法,在DBSCAN聚类基础上结合了K中位最近邻(K-median nearest neighbor,KMNN)算法,通过引入自衰减系数并设置阈值上限对参数值列表进行二次处理,可以自适应根据聚类结果与不同参数时的K值之间的关系确定最优的邻域半径和最少点个数,提高了分选的正确率。通过仿真实验验证了算法利用雷达脉冲描述字特征进行自适应分选的有效性。