期刊文献+

基于GainRatio降维算法的流量聚类研究 被引量:2

Traffic Clustering Research Based on GainRatio Dimesionality Reduction Algorithm
下载PDF
导出
摘要 随着网络数据流量的快速增长,需要高效的流量分类技术来实现网络管理、流量控制和安全检测。传统基于端口和有效负载的流量分类方法准确率低,无监督学习方法往往仅采用单一的聚类算法对数据进行聚类分析,且较少研究对数据本身的处理。为了解决上述问题,提出了先运用GainRatio信息增益率方法对原始数据进行降维处理,再将降维后的数据进行聚类的方法。实验结果表明:提出的方法不仅有效地提高了运行效率,而且随着聚类个数的增加,也明显地提高了高准确率的收敛速度。 With the rapid growth of network data traffic, efficient traffic classification technologies are required to implement network management, flow control and security detection.The traditional port-based and payload-based classification methods have low accuracy, and the unsupervised learning method often adopts only a single clustering algorithm to cluster the data. To solve problems mentioned above, a method of reducing the dimensionality of the original data by using the GainRatio information gain rate method and then clustering the dimensionality-reduced data is proposed. The results show that the proposed method not only effectively improves the operating rate, but also accelerates the convergence rate of high accuracy with the increase of the number of clusters.
作者 高锐 刘北水 李丹 刘杰 尤博 GAO Rui;LIU Beishui;LI Dan;LIU Jie;YOU Bo(CEPREI,Guangzhou 510610,China)
出处 《电子产品可靠性与环境试验》 2020年第S02期51-55,共5页 Electronic Product Reliability and Environmental Testing
基金 2018年工业转型升级资金项目-信息编码核心算法检测评估能力建设 广州市科技计划一般项目(201804010316) 国家重点研发计划项目(2019YFC0118800) 国家重点研发计划项目(2018YFC1201104)资助。
关键词 机器学习 流量聚类 网络安全 维度下降 信息增益 machine learning traffic clustering network security dimensionality reduction information gain ratio
  • 相关文献

参考文献1

二级参考文献13

  • 1Azizi A. Efficient IRIS recognition through improvement of fea-ture extraction and subset selection [ J ]. International journalof computer science and information security ,2009,2( 1) :72-73.
  • 2Yu Lei, Liu Huan. Efficient feature selection via analysis ofrelevance and redundancy [ J ]. Machine learning research,2004,5(1):1207-1220.
  • 3Hall M A,Smith L A. Feature subset selection: A correlationbased filter approach [ C ] //Proc of international conference onneural information processing. [ s. 1. ] : [ s. n. ] ,1997 :2-4.
  • 4Hsu W H. Genetic wrappers for feature selection in decisiontree induction and variable ordering in Bayesian networkstructure learning [ J ]. Information sciences,2004, 163 ( 1):105-120.
  • 5Karegowda A G, Manjunath A S, Jayaram M A. Comparativestudy of attribute selection using gain ratio and correlationbased feature selection [ J ]. International journal of informationtechnology and knowledge management, 2010, 2(2): 271 -274.
  • 6Chena Y,Abrahama A’Yanga B. Feature selection and classi-fication using flexible neural tree [ J ]. Neurocomputing,2006,70(1):306-308.
  • 7Koller D, Sahami M. Toward optimal feature selection[ C ]//Proceedings of international conference on machine learning,[s.l. ].[s. n. ],1996:162-187.
  • 8Forman G. An extensive empirical study of feature selectionmetrics for text classification[J]. Machine research learning,2003(3) :1289-1305.
  • 9Sheikhi N, Rahmani A, Mohsenzadeh M. An unsupervised fea-ture selection method based on genetic algorithm[ J]. Interna-tional journal of computer science and information security,2011,24(3) ;117-120.
  • 10Yuan Jinhui,Li Jianmin, Zhang Bo. Learning concepts fromlarge scale imbalanced data sets using support cluster ma-chines [C ]//Proceedings of the 14th annual ACM internation-al conference on multimedia, [s. 1. ] :ACM,2006:441-450.

共引文献22

同被引文献20

引证文献2

二级引证文献3

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部