期刊文献+

基于邻域互信息最大相关性最小冗余度的特征选择 被引量:1

Neighborhood Mutual Information Based on Max Relevance and Min Redundancy Feature Selection
下载PDF
导出
摘要 特征选择是一种重要的数据预处理步骤,其中互信息是一类重要的信息度量方法.本文针对互信息不能很好地处理数值型的特征,介绍了邻域信息熵与邻域互信息.其次,设计了基于邻域互信息的最大相关性最小冗余度的特征排序算法.最后,用此算法选择前若干特征进行分类并与其它算法比较分类精度.实验结果表明本文提出算法在分类精度方面且优于或相当于其它流行特征选择算法. Feature selection is an important data preprocessing technique, where mutual information has been widely studied in information measure. However, mutual information cannot directly calculate relevancy among numeric features. In this paper, we first introduce neighborhood entropy and neighborhood mutual information. Then, we propose neighborhood mutual information based max relevance and min redundancy feature selection. Finally, experimental results show that the proposed method can effectively select a discriminative feature subset, and outperform or equal to other popular feature selection algorithms in classification performance.
作者 林培榕
出处 《漳州师范学院学报(自然科学版)》 2013年第4期13-18,共6页 Journal of ZhangZhou Teachers College(Natural Science)
基金 福建省自然科学基金项目(2013J01259) 漳州市科技计划项目(ZZ2013J04)
关键词 特征选择 邻域互信息 最大相关性 最小冗余度 feature selection neighborhood mutual information max relevance min redundancy
  • 相关文献

参考文献12

  • 1I. Guyon, A. Elisseeff. An introduction to variable and feature selection[J]. Journal of Machine Learning Research, 2003, (3): 1157-1182.
  • 2M. Dash, H. Liu. Feature selection for classification[J]. Intelligent Data Analysis, 1997, (1): 131-156.
  • 3W. Z. Zhu, G. Q. Si, Y. B. Zhang, J. C. Wang. Neighborhood effective information ratio for hybrid feature subset evaluation and selection[J]. Neuro computing, 2013, (99): 25-37.
  • 4M. A. Hall. Correlation-based Feature Selection for Discrete and Numeric Class Machine Learning[C]//Proceedings of the Seventh International Conference on Machine Learning, 2003: 359-366.
  • 5L. Yu, H. Liu. Feature selection for high-dimensional data: a fast correlationbased filter solution[C]//Proceedings of the Twentieth International Conference on Machine Learning, 2003: 856-863.
  • 6H. C. Peng, E H. Long, C. Ding. Feature selection based on mutual information: criteria of Max-Dependency, Max-Relevance, and Min-Redundancy[C]//IEEE Transactions on Pattern Analysis and Machine Intelligence, 2005, (27): 1226-1238.
  • 7Kononenko. Estimating attributes: analysis and extensions of RELIEF[C]//Proceedings European Conference Machine Learning, 1994: 171-182.
  • 8R. Gilad-Bachrach, A. Navot, N. Tishby. Margin based feature selection-theory and algorithms[C]//proceedings of the 21st International Conference on Machine Learning, 2004: 40-48.
  • 9R. Tibshirani. Regression shrinkage and selection via the lasso[J]. Journal of the Royal Statistical Society. Series B, 1996, (58): 267-288.
  • 10I. Guyon, J. Weston, S. Barnhill, V. Machine Learning, 2002, (46): 389-422.

同被引文献6

引证文献1

二级引证文献7

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部