In clustering algorithms,the selection of neighbors significantly affects the quality of the final clustering results.While various neighbor relationships exist,such as K-nearest neighbors,natural neighbors,and shared...In clustering algorithms,the selection of neighbors significantly affects the quality of the final clustering results.While various neighbor relationships exist,such as K-nearest neighbors,natural neighbors,and shared neighbors,most neighbor relationships can only handle single structural relationships,and the identification accuracy is low for datasets with multiple structures.In life,people’s first instinct for complex things is to divide them into multiple parts to complete.Partitioning the dataset into more sub-graphs is a good idea approach to identifying complex structures.Taking inspiration from this,we propose a novel neighbor method:Shared Natural Neighbors(SNaN).To demonstrate the superiority of this neighbor method,we propose a shared natural neighbors-based hierarchical clustering algorithm for discovering arbitrary-shaped clusters(HC-SNaN).Our algorithm excels in identifying both spherical clusters and manifold clusters.Tested on synthetic datasets and real-world datasets,HC-SNaN demonstrates significant advantages over existing clustering algorithms,particularly when dealing with datasets containing arbitrary shapes.展开更多
The k-Nearest Neighbor method is one of the most popular techniques for both classification and regression purposes.Because of its operation,the application of this classification may be limited to problems with a cer...The k-Nearest Neighbor method is one of the most popular techniques for both classification and regression purposes.Because of its operation,the application of this classification may be limited to problems with a certain number of instances,particularly,when run time is a consideration.However,the classification of large amounts of data has become a fundamental task in many real-world applications.It is logical to scale the k-Nearest Neighbor method to large scale datasets.This paper proposes a new k-Nearest Neighbor classification method(KNN-CCL)which uses a parallel centroid-based and hierarchical clustering algorithm to separate the sample of training dataset into multiple parts.The introduced clustering algorithm uses four stages of successive refinements and generates high quality clusters.The k-Nearest Neighbor approach subsequently makes use of them to predict the test datasets.Finally,sets of experiments are conducted on the UCI datasets.The experimental results confirm that the proposed k-Nearest Neighbor classification method performs well with regard to classification accuracy and performance.展开更多
Traditional clustering algorithms often struggle to produce satisfactory results when dealing with datasets withuneven density. Additionally, they incur substantial computational costs when applied to high-dimensional...Traditional clustering algorithms often struggle to produce satisfactory results when dealing with datasets withuneven density. Additionally, they incur substantial computational costs when applied to high-dimensional datadue to calculating similarity matrices. To alleviate these issues, we employ the KD-Tree to partition the dataset andcompute the K-nearest neighbors (KNN) density for each point, thereby avoiding the computation of similaritymatrices. Moreover, we apply the rules of voting elections, treating each data point as a voter and casting a votefor the point with the highest density among its KNN. By utilizing the vote counts of each point, we develop thestrategy for classifying noise points and potential cluster centers, allowing the algorithm to identify clusters withuneven density and complex shapes. Additionally, we define the concept of “adhesive points” between two clustersto merge adjacent clusters that have similar densities. This process helps us identify the optimal number of clustersautomatically. Experimental results indicate that our algorithm not only improves the efficiency of clustering butalso increases its accuracy.展开更多
Although k-nearest neighbors (KNN) is a popular fingerprint match algorithm for its simplicity and accuracy, because it is sensitive to the circumstances, a fuzzy c-means (FCM) clustering algorithm is applied to i...Although k-nearest neighbors (KNN) is a popular fingerprint match algorithm for its simplicity and accuracy, because it is sensitive to the circumstances, a fuzzy c-means (FCM) clustering algorithm is applied to improve it. Thus, a KNN-based two-step FCM weighted (KTFW) algorithm for indoor positioning in wireless local area networks (WLAN) is presented in this paper. In KTFW algorithm, k reference points (RPs) chosen by KNN are clustered through FCM based on received signal strength (RSS) and location coordinates. The right clusters are chosen according to rules, so three sets of RPs are formed including the set of k RPs chosen by KNN and are given different weights. RPs supposed to have better contribution to positioning accuracy are given larger weights to improve the positioning accuracy. Simulation results indicate that KTFW generally outperforms KNN and its complexity is greatly reduced through providing initial clustering centers for FCM.展开更多
Interference signals recognition plays an important role in anti-jamming communication.With the development of deep learning,many supervised interference signals recognition algorithms based on deep learning have emer...Interference signals recognition plays an important role in anti-jamming communication.With the development of deep learning,many supervised interference signals recognition algorithms based on deep learning have emerged recently and show better performance than traditional recognition algorithms.However,there is no unsupervised interference signals recognition algorithm at present.In this paper,an unsupervised interference signals recognition method called double phases and double dimensions contrastive clustering(DDCC)is proposed.Specifically,in the first phase,four data augmentation strategies for interference signals are used in data-augmentation-based(DA-based)contrastive learning.In the second phase,the original dataset’s k-nearest neighbor set(KNNset)is designed in double dimensions contrastive learning.In addition,a dynamic entropy parameter strategy is proposed.The simulation experiments of 9 types of interference signals show that random cropping is the best one of the four data augmentation strategies;the feature dimensional contrastive learning in the second phase can improve the clustering purity;the dynamic entropy parameter strategy can improve the stability of DDCC effectively.The unsupervised interference signals recognition results of DDCC and five other deep clustering algorithms show that the clustering performance of DDCC is superior to other algorithms.In particular,the clustering purity of our method is above 92%,SCAN’s is 81%,and the other three methods’are below 71%when jammingnoise-ratio(JNR)is−5 dB.In addition,our method is close to the supervised learning algorithm.展开更多
针对基于接收信号强度的位置指纹室内定位算法定位精度不高的问题,提出了一种均值层次聚类和自适应加权K近邻(weighted K nearest neighbor,WKNN)的室内定位算法。算法首先在设置的参考点上采集蓝牙信号强度构建离线指纹数据库,然后采...针对基于接收信号强度的位置指纹室内定位算法定位精度不高的问题,提出了一种均值层次聚类和自适应加权K近邻(weighted K nearest neighbor,WKNN)的室内定位算法。算法首先在设置的参考点上采集蓝牙信号强度构建离线指纹数据库,然后采用均值层次聚类方法将所有参考点根据各自之间的相似度分为n个类,滤除掉相似度较小的参考点,最后根据待定位点和参考点间的信号距离的相似度,计算出距离差的标准差来自适应确定K值,并进行位置估算。实验结果表明,本文提出的算法在定位精度上比WKNN、动态加权K近邻(enhanced weighted K nearest neighbor,EWKNN)方法分别提升了30.0%和18.0%,在定位实时性上比WKNN和EWKNN方法分别提高了19.2%和28.4%。将该算法用于室内物体定位,可以同时提高定位精度和定位实时性。展开更多
基金This work was supported by Science and Technology Research Program of Chongqing Municipal Education Commission(KJZD-M202300502,KJQN201800539).
文摘In clustering algorithms,the selection of neighbors significantly affects the quality of the final clustering results.While various neighbor relationships exist,such as K-nearest neighbors,natural neighbors,and shared neighbors,most neighbor relationships can only handle single structural relationships,and the identification accuracy is low for datasets with multiple structures.In life,people’s first instinct for complex things is to divide them into multiple parts to complete.Partitioning the dataset into more sub-graphs is a good idea approach to identifying complex structures.Taking inspiration from this,we propose a novel neighbor method:Shared Natural Neighbors(SNaN).To demonstrate the superiority of this neighbor method,we propose a shared natural neighbors-based hierarchical clustering algorithm for discovering arbitrary-shaped clusters(HC-SNaN).Our algorithm excels in identifying both spherical clusters and manifold clusters.Tested on synthetic datasets and real-world datasets,HC-SNaN demonstrates significant advantages over existing clustering algorithms,particularly when dealing with datasets containing arbitrary shapes.
基金The authors received no specific funding for this work.
文摘The k-Nearest Neighbor method is one of the most popular techniques for both classification and regression purposes.Because of its operation,the application of this classification may be limited to problems with a certain number of instances,particularly,when run time is a consideration.However,the classification of large amounts of data has become a fundamental task in many real-world applications.It is logical to scale the k-Nearest Neighbor method to large scale datasets.This paper proposes a new k-Nearest Neighbor classification method(KNN-CCL)which uses a parallel centroid-based and hierarchical clustering algorithm to separate the sample of training dataset into multiple parts.The introduced clustering algorithm uses four stages of successive refinements and generates high quality clusters.The k-Nearest Neighbor approach subsequently makes use of them to predict the test datasets.Finally,sets of experiments are conducted on the UCI datasets.The experimental results confirm that the proposed k-Nearest Neighbor classification method performs well with regard to classification accuracy and performance.
基金National Natural Science Foundation of China Nos.61962054 and 62372353.
文摘Traditional clustering algorithms often struggle to produce satisfactory results when dealing with datasets withuneven density. Additionally, they incur substantial computational costs when applied to high-dimensional datadue to calculating similarity matrices. To alleviate these issues, we employ the KD-Tree to partition the dataset andcompute the K-nearest neighbors (KNN) density for each point, thereby avoiding the computation of similaritymatrices. Moreover, we apply the rules of voting elections, treating each data point as a voter and casting a votefor the point with the highest density among its KNN. By utilizing the vote counts of each point, we develop thestrategy for classifying noise points and potential cluster centers, allowing the algorithm to identify clusters withuneven density and complex shapes. Additionally, we define the concept of “adhesive points” between two clustersto merge adjacent clusters that have similar densities. This process helps us identify the optimal number of clustersautomatically. Experimental results indicate that our algorithm not only improves the efficiency of clustering butalso increases its accuracy.
文摘Although k-nearest neighbors (KNN) is a popular fingerprint match algorithm for its simplicity and accuracy, because it is sensitive to the circumstances, a fuzzy c-means (FCM) clustering algorithm is applied to improve it. Thus, a KNN-based two-step FCM weighted (KTFW) algorithm for indoor positioning in wireless local area networks (WLAN) is presented in this paper. In KTFW algorithm, k reference points (RPs) chosen by KNN are clustered through FCM based on received signal strength (RSS) and location coordinates. The right clusters are chosen according to rules, so three sets of RPs are formed including the set of k RPs chosen by KNN and are given different weights. RPs supposed to have better contribution to positioning accuracy are given larger weights to improve the positioning accuracy. Simulation results indicate that KTFW generally outperforms KNN and its complexity is greatly reduced through providing initial clustering centers for FCM.
基金This research was supported by the National Natural Science Foundation of China under Grant No.U19B2016.,and Zhejiang Provincial Key Lab of Data Storage and Transmission Technology,Hangzhou Dianzi University.
文摘Interference signals recognition plays an important role in anti-jamming communication.With the development of deep learning,many supervised interference signals recognition algorithms based on deep learning have emerged recently and show better performance than traditional recognition algorithms.However,there is no unsupervised interference signals recognition algorithm at present.In this paper,an unsupervised interference signals recognition method called double phases and double dimensions contrastive clustering(DDCC)is proposed.Specifically,in the first phase,four data augmentation strategies for interference signals are used in data-augmentation-based(DA-based)contrastive learning.In the second phase,the original dataset’s k-nearest neighbor set(KNNset)is designed in double dimensions contrastive learning.In addition,a dynamic entropy parameter strategy is proposed.The simulation experiments of 9 types of interference signals show that random cropping is the best one of the four data augmentation strategies;the feature dimensional contrastive learning in the second phase can improve the clustering purity;the dynamic entropy parameter strategy can improve the stability of DDCC effectively.The unsupervised interference signals recognition results of DDCC and five other deep clustering algorithms show that the clustering performance of DDCC is superior to other algorithms.In particular,the clustering purity of our method is above 92%,SCAN’s is 81%,and the other three methods’are below 71%when jammingnoise-ratio(JNR)is−5 dB.In addition,our method is close to the supervised learning algorithm.
文摘针对基于接收信号强度的位置指纹室内定位算法定位精度不高的问题,提出了一种均值层次聚类和自适应加权K近邻(weighted K nearest neighbor,WKNN)的室内定位算法。算法首先在设置的参考点上采集蓝牙信号强度构建离线指纹数据库,然后采用均值层次聚类方法将所有参考点根据各自之间的相似度分为n个类,滤除掉相似度较小的参考点,最后根据待定位点和参考点间的信号距离的相似度,计算出距离差的标准差来自适应确定K值,并进行位置估算。实验结果表明,本文提出的算法在定位精度上比WKNN、动态加权K近邻(enhanced weighted K nearest neighbor,EWKNN)方法分别提升了30.0%和18.0%,在定位实时性上比WKNN和EWKNN方法分别提高了19.2%和28.4%。将该算法用于室内物体定位,可以同时提高定位精度和定位实时性。