In clustering algorithms,the selection of neighbors significantly affects the quality of the final clustering results.While various neighbor relationships exist,such as K-nearest neighbors,natural neighbors,and shared...In clustering algorithms,the selection of neighbors significantly affects the quality of the final clustering results.While various neighbor relationships exist,such as K-nearest neighbors,natural neighbors,and shared neighbors,most neighbor relationships can only handle single structural relationships,and the identification accuracy is low for datasets with multiple structures.In life,people’s first instinct for complex things is to divide them into multiple parts to complete.Partitioning the dataset into more sub-graphs is a good idea approach to identifying complex structures.Taking inspiration from this,we propose a novel neighbor method:Shared Natural Neighbors(SNaN).To demonstrate the superiority of this neighbor method,we propose a shared natural neighbors-based hierarchical clustering algorithm for discovering arbitrary-shaped clusters(HC-SNaN).Our algorithm excels in identifying both spherical clusters and manifold clusters.Tested on synthetic datasets and real-world datasets,HC-SNaN demonstrates significant advantages over existing clustering algorithms,particularly when dealing with datasets containing arbitrary shapes.展开更多
Traditional clustering algorithms often struggle to produce satisfactory results when dealing with datasets withuneven density. Additionally, they incur substantial computational costs when applied to high-dimensional...Traditional clustering algorithms often struggle to produce satisfactory results when dealing with datasets withuneven density. Additionally, they incur substantial computational costs when applied to high-dimensional datadue to calculating similarity matrices. To alleviate these issues, we employ the KD-Tree to partition the dataset andcompute the K-nearest neighbors (KNN) density for each point, thereby avoiding the computation of similaritymatrices. Moreover, we apply the rules of voting elections, treating each data point as a voter and casting a votefor the point with the highest density among its KNN. By utilizing the vote counts of each point, we develop thestrategy for classifying noise points and potential cluster centers, allowing the algorithm to identify clusters withuneven density and complex shapes. Additionally, we define the concept of “adhesive points” between two clustersto merge adjacent clusters that have similar densities. This process helps us identify the optimal number of clustersautomatically. Experimental results indicate that our algorithm not only improves the efficiency of clustering butalso increases its accuracy.展开更多
The k-Nearest Neighbor method is one of the most popular techniques for both classification and regression purposes.Because of its operation,the application of this classification may be limited to problems with a cer...The k-Nearest Neighbor method is one of the most popular techniques for both classification and regression purposes.Because of its operation,the application of this classification may be limited to problems with a certain number of instances,particularly,when run time is a consideration.However,the classification of large amounts of data has become a fundamental task in many real-world applications.It is logical to scale the k-Nearest Neighbor method to large scale datasets.This paper proposes a new k-Nearest Neighbor classification method(KNN-CCL)which uses a parallel centroid-based and hierarchical clustering algorithm to separate the sample of training dataset into multiple parts.The introduced clustering algorithm uses four stages of successive refinements and generates high quality clusters.The k-Nearest Neighbor approach subsequently makes use of them to predict the test datasets.Finally,sets of experiments are conducted on the UCI datasets.The experimental results confirm that the proposed k-Nearest Neighbor classification method performs well with regard to classification accuracy and performance.展开更多
Intrusion detection aims to detect intrusion behavior and serves as a complement to firewalls.It can detect attack types of malicious network communications and computer usage that cannot be detected by idiomatic fire...Intrusion detection aims to detect intrusion behavior and serves as a complement to firewalls.It can detect attack types of malicious network communications and computer usage that cannot be detected by idiomatic firewalls.Many intrusion detection methods are processed through machine learning.Previous literature has shown that the performance of an intrusion detection method based on hybrid learning or integration approach is superior to that of single learning technology.However,almost no studies focus on how additional representative and concise features can be extracted to process effective intrusion detection among massive and complicated data.In this paper,a new hybrid learning method is proposed on the basis of features such as density,cluster centers,and nearest neighbors(DCNN).In this algorithm,data is represented by the local density of each sample point and the sum of distances from each sample point to cluster centers and to its nearest neighbor.k-NN classifier is adopted to classify the new feature vectors.Our experiment shows that DCNN,which combines K-means,clustering-based density,and k-NN classifier,is effective in intrusion detection.展开更多
Although k-nearest neighbors (KNN) is a popular fingerprint match algorithm for its simplicity and accuracy, because it is sensitive to the circumstances, a fuzzy c-means (FCM) clustering algorithm is applied to i...Although k-nearest neighbors (KNN) is a popular fingerprint match algorithm for its simplicity and accuracy, because it is sensitive to the circumstances, a fuzzy c-means (FCM) clustering algorithm is applied to improve it. Thus, a KNN-based two-step FCM weighted (KTFW) algorithm for indoor positioning in wireless local area networks (WLAN) is presented in this paper. In KTFW algorithm, k reference points (RPs) chosen by KNN are clustered through FCM based on received signal strength (RSS) and location coordinates. The right clusters are chosen according to rules, so three sets of RPs are formed including the set of k RPs chosen by KNN and are given different weights. RPs supposed to have better contribution to positioning accuracy are given larger weights to improve the positioning accuracy. Simulation results indicate that KTFW generally outperforms KNN and its complexity is greatly reduced through providing initial clustering centers for FCM.展开更多
Many classifiers and methods are proposed to deal with letter recognition problem. Among them, clustering is a widely used method. But only one time for clustering is not adequately. Here, we adopt data preprocessing ...Many classifiers and methods are proposed to deal with letter recognition problem. Among them, clustering is a widely used method. But only one time for clustering is not adequately. Here, we adopt data preprocessing and a re kernel clustering method to tackle the letter recognition problem. In order to validate effectiveness and efficiency of proposed method, we introduce re kernel clustering into Kernel Nearest Neighbor classification(KNN), Radial Basis Function Neural Network(RBFNN), and Support Vector Machine(SVM). Furthermore, we compare the difference between re kernel clustering and one time kernel clustering which is denoted as kernel clustering for short. Experimental results validate that re kernel clustering forms fewer and more feasible kernels and attain higher classification accuracy.展开更多
Interference signals recognition plays an important role in anti-jamming communication.With the development of deep learning,many supervised interference signals recognition algorithms based on deep learning have emer...Interference signals recognition plays an important role in anti-jamming communication.With the development of deep learning,many supervised interference signals recognition algorithms based on deep learning have emerged recently and show better performance than traditional recognition algorithms.However,there is no unsupervised interference signals recognition algorithm at present.In this paper,an unsupervised interference signals recognition method called double phases and double dimensions contrastive clustering(DDCC)is proposed.Specifically,in the first phase,four data augmentation strategies for interference signals are used in data-augmentation-based(DA-based)contrastive learning.In the second phase,the original dataset’s k-nearest neighbor set(KNNset)is designed in double dimensions contrastive learning.In addition,a dynamic entropy parameter strategy is proposed.The simulation experiments of 9 types of interference signals show that random cropping is the best one of the four data augmentation strategies;the feature dimensional contrastive learning in the second phase can improve the clustering purity;the dynamic entropy parameter strategy can improve the stability of DDCC effectively.The unsupervised interference signals recognition results of DDCC and five other deep clustering algorithms show that the clustering performance of DDCC is superior to other algorithms.In particular,the clustering purity of our method is above 92%,SCAN’s is 81%,and the other three methods’are below 71%when jammingnoise-ratio(JNR)is−5 dB.In addition,our method is close to the supervised learning algorithm.展开更多
Several typical supervised clustering methods such as Gaussian mixture model-based supervised clustering (GMM), k- nearest-neighbor (KNN), binary support vector machines (SVMs) and multiclass support vector mach...Several typical supervised clustering methods such as Gaussian mixture model-based supervised clustering (GMM), k- nearest-neighbor (KNN), binary support vector machines (SVMs) and multiclass support vector machines (MC-SVMs) were employed to classify the computer simulation data and two real microarray expression datasets. False positive, false negative, true positive, true negative, clustering accuracy and Matthews' correlation coefficient (MCC) were compared among these methods. The results are as follows: (1) In classifying thousands of gene expression data, the performances of two GMM methods have the maximal clustering accuracy and the least overall FP+FN error numbers on the basis of the assumption that the whole set of microarray data are a finite mixture of multivariate Gaussian distributions. Furthermore, when the number of training sample is very small, the clustering accuracy of GMM-Ⅱ method has superiority over GMM- Ⅰ method. (2) In general, the superior classification performance of the MC-SVMs are more robust and more practical, which are less sensitive to the curse of dimensionality, and not only next to GMM method in clustering accuracy to thousands of gene expression data, but also more robust to a small number of high-dimensional gene expression samples than other techniques. (3) Of the MC-SVMs, OVO and DAGSVM perform better on the large sample sizes, whereas five MC-SVMs methods have very similar performance on moderate sample sizes. In other cases, OVR, WW and CS yield better results when sample sizes are small. So, it is recommended that at least two candidate methods, choosing on the basis of the real data features and experimental conditions, should be performed and compared to obtain better clustering result.展开更多
The hypersonic interception in near space is a great challenge because of the target’s unpredictable trajectory, which demands the interceptors of trajectory cluster coverage of the predicted area and optimal traject...The hypersonic interception in near space is a great challenge because of the target’s unpredictable trajectory, which demands the interceptors of trajectory cluster coverage of the predicted area and optimal trajectory modification capability aiming at the consistently updating predicted impact point(PIP) in the midcourse phase. A novel midcourse optimal trajectory cluster generation and trajectory modification algorithm is proposed based on the neighboring optimal control theory. Firstly, the midcourse trajectory optimization problem is introduced; the necessary conditions for the optimal control and the transversality constraints are given.Secondly, with the description of the neighboring optimal trajectory existence theory(NOTET), the neighboring optimal control(NOC)algorithm is derived by taking the second order partial derivations with the necessary conditions and transversality conditions. The revised terminal constraints are reversely integrated to the initial time and the perturbations of the co-states are further expressed with the states deviations and terminal constraints modifications.Thirdly, the simulations of two different scenarios are carried out and the results prove the effectiveness and optimality of the proposed method.展开更多
基金This work was supported by Science and Technology Research Program of Chongqing Municipal Education Commission(KJZD-M202300502,KJQN201800539).
文摘In clustering algorithms,the selection of neighbors significantly affects the quality of the final clustering results.While various neighbor relationships exist,such as K-nearest neighbors,natural neighbors,and shared neighbors,most neighbor relationships can only handle single structural relationships,and the identification accuracy is low for datasets with multiple structures.In life,people’s first instinct for complex things is to divide them into multiple parts to complete.Partitioning the dataset into more sub-graphs is a good idea approach to identifying complex structures.Taking inspiration from this,we propose a novel neighbor method:Shared Natural Neighbors(SNaN).To demonstrate the superiority of this neighbor method,we propose a shared natural neighbors-based hierarchical clustering algorithm for discovering arbitrary-shaped clusters(HC-SNaN).Our algorithm excels in identifying both spherical clusters and manifold clusters.Tested on synthetic datasets and real-world datasets,HC-SNaN demonstrates significant advantages over existing clustering algorithms,particularly when dealing with datasets containing arbitrary shapes.
基金National Natural Science Foundation of China Nos.61962054 and 62372353.
文摘Traditional clustering algorithms often struggle to produce satisfactory results when dealing with datasets withuneven density. Additionally, they incur substantial computational costs when applied to high-dimensional datadue to calculating similarity matrices. To alleviate these issues, we employ the KD-Tree to partition the dataset andcompute the K-nearest neighbors (KNN) density for each point, thereby avoiding the computation of similaritymatrices. Moreover, we apply the rules of voting elections, treating each data point as a voter and casting a votefor the point with the highest density among its KNN. By utilizing the vote counts of each point, we develop thestrategy for classifying noise points and potential cluster centers, allowing the algorithm to identify clusters withuneven density and complex shapes. Additionally, we define the concept of “adhesive points” between two clustersto merge adjacent clusters that have similar densities. This process helps us identify the optimal number of clustersautomatically. Experimental results indicate that our algorithm not only improves the efficiency of clustering butalso increases its accuracy.
基金The authors received no specific funding for this work.
文摘The k-Nearest Neighbor method is one of the most popular techniques for both classification and regression purposes.Because of its operation,the application of this classification may be limited to problems with a certain number of instances,particularly,when run time is a consideration.However,the classification of large amounts of data has become a fundamental task in many real-world applications.It is logical to scale the k-Nearest Neighbor method to large scale datasets.This paper proposes a new k-Nearest Neighbor classification method(KNN-CCL)which uses a parallel centroid-based and hierarchical clustering algorithm to separate the sample of training dataset into multiple parts.The introduced clustering algorithm uses four stages of successive refinements and generates high quality clusters.The k-Nearest Neighbor approach subsequently makes use of them to predict the test datasets.Finally,sets of experiments are conducted on the UCI datasets.The experimental results confirm that the proposed k-Nearest Neighbor classification method performs well with regard to classification accuracy and performance.
文摘Intrusion detection aims to detect intrusion behavior and serves as a complement to firewalls.It can detect attack types of malicious network communications and computer usage that cannot be detected by idiomatic firewalls.Many intrusion detection methods are processed through machine learning.Previous literature has shown that the performance of an intrusion detection method based on hybrid learning or integration approach is superior to that of single learning technology.However,almost no studies focus on how additional representative and concise features can be extracted to process effective intrusion detection among massive and complicated data.In this paper,a new hybrid learning method is proposed on the basis of features such as density,cluster centers,and nearest neighbors(DCNN).In this algorithm,data is represented by the local density of each sample point and the sum of distances from each sample point to cluster centers and to its nearest neighbor.k-NN classifier is adopted to classify the new feature vectors.Our experiment shows that DCNN,which combines K-means,clustering-based density,and k-NN classifier,is effective in intrusion detection.
文摘Although k-nearest neighbors (KNN) is a popular fingerprint match algorithm for its simplicity and accuracy, because it is sensitive to the circumstances, a fuzzy c-means (FCM) clustering algorithm is applied to improve it. Thus, a KNN-based two-step FCM weighted (KTFW) algorithm for indoor positioning in wireless local area networks (WLAN) is presented in this paper. In KTFW algorithm, k reference points (RPs) chosen by KNN are clustered through FCM based on received signal strength (RSS) and location coordinates. The right clusters are chosen according to rules, so three sets of RPs are formed including the set of k RPs chosen by KNN and are given different weights. RPs supposed to have better contribution to positioning accuracy are given larger weights to improve the positioning accuracy. Simulation results indicate that KTFW generally outperforms KNN and its complexity is greatly reduced through providing initial clustering centers for FCM.
基金Supported by the National Science Foundation(No.IIS-9988642)the Multidisciplinary Research Program
文摘Many classifiers and methods are proposed to deal with letter recognition problem. Among them, clustering is a widely used method. But only one time for clustering is not adequately. Here, we adopt data preprocessing and a re kernel clustering method to tackle the letter recognition problem. In order to validate effectiveness and efficiency of proposed method, we introduce re kernel clustering into Kernel Nearest Neighbor classification(KNN), Radial Basis Function Neural Network(RBFNN), and Support Vector Machine(SVM). Furthermore, we compare the difference between re kernel clustering and one time kernel clustering which is denoted as kernel clustering for short. Experimental results validate that re kernel clustering forms fewer and more feasible kernels and attain higher classification accuracy.
基金This research was supported by the National Natural Science Foundation of China under Grant No.U19B2016.,and Zhejiang Provincial Key Lab of Data Storage and Transmission Technology,Hangzhou Dianzi University.
文摘Interference signals recognition plays an important role in anti-jamming communication.With the development of deep learning,many supervised interference signals recognition algorithms based on deep learning have emerged recently and show better performance than traditional recognition algorithms.However,there is no unsupervised interference signals recognition algorithm at present.In this paper,an unsupervised interference signals recognition method called double phases and double dimensions contrastive clustering(DDCC)is proposed.Specifically,in the first phase,four data augmentation strategies for interference signals are used in data-augmentation-based(DA-based)contrastive learning.In the second phase,the original dataset’s k-nearest neighbor set(KNNset)is designed in double dimensions contrastive learning.In addition,a dynamic entropy parameter strategy is proposed.The simulation experiments of 9 types of interference signals show that random cropping is the best one of the four data augmentation strategies;the feature dimensional contrastive learning in the second phase can improve the clustering purity;the dynamic entropy parameter strategy can improve the stability of DDCC effectively.The unsupervised interference signals recognition results of DDCC and five other deep clustering algorithms show that the clustering performance of DDCC is superior to other algorithms.In particular,the clustering purity of our method is above 92%,SCAN’s is 81%,and the other three methods’are below 71%when jammingnoise-ratio(JNR)is−5 dB.In addition,our method is close to the supervised learning algorithm.
基金This research was supported by the National Natural Science Foundation of China(30370758)Program for New Century Excellent Talents in Universities(NCET)of Ministry of Education to Dr.Xu Chenwu(NCET-05-0502).
文摘Several typical supervised clustering methods such as Gaussian mixture model-based supervised clustering (GMM), k- nearest-neighbor (KNN), binary support vector machines (SVMs) and multiclass support vector machines (MC-SVMs) were employed to classify the computer simulation data and two real microarray expression datasets. False positive, false negative, true positive, true negative, clustering accuracy and Matthews' correlation coefficient (MCC) were compared among these methods. The results are as follows: (1) In classifying thousands of gene expression data, the performances of two GMM methods have the maximal clustering accuracy and the least overall FP+FN error numbers on the basis of the assumption that the whole set of microarray data are a finite mixture of multivariate Gaussian distributions. Furthermore, when the number of training sample is very small, the clustering accuracy of GMM-Ⅱ method has superiority over GMM- Ⅰ method. (2) In general, the superior classification performance of the MC-SVMs are more robust and more practical, which are less sensitive to the curse of dimensionality, and not only next to GMM method in clustering accuracy to thousands of gene expression data, but also more robust to a small number of high-dimensional gene expression samples than other techniques. (3) Of the MC-SVMs, OVO and DAGSVM perform better on the large sample sizes, whereas five MC-SVMs methods have very similar performance on moderate sample sizes. In other cases, OVR, WW and CS yield better results when sample sizes are small. So, it is recommended that at least two candidate methods, choosing on the basis of the real data features and experimental conditions, should be performed and compared to obtain better clustering result.
基金supported by the National Natural Science Foundation of China(6150340861573374)
文摘The hypersonic interception in near space is a great challenge because of the target’s unpredictable trajectory, which demands the interceptors of trajectory cluster coverage of the predicted area and optimal trajectory modification capability aiming at the consistently updating predicted impact point(PIP) in the midcourse phase. A novel midcourse optimal trajectory cluster generation and trajectory modification algorithm is proposed based on the neighboring optimal control theory. Firstly, the midcourse trajectory optimization problem is introduced; the necessary conditions for the optimal control and the transversality constraints are given.Secondly, with the description of the neighboring optimal trajectory existence theory(NOTET), the neighboring optimal control(NOC)algorithm is derived by taking the second order partial derivations with the necessary conditions and transversality conditions. The revised terminal constraints are reversely integrated to the initial time and the perturbations of the co-states are further expressed with the states deviations and terminal constraints modifications.Thirdly, the simulations of two different scenarios are carried out and the results prove the effectiveness and optimality of the proposed method.