期刊文献+

一种类内方差与相关度结合的特征选择算法 被引量:3

A feature selection algorithm combining within-class variance with correlation measure
下载PDF
导出
摘要 为了解决现有特征选择算法没有同时考虑特征之间以及特征与类别之间的相关性,且存在计算量大、适用范围窄等问题,从均方误差最小的分类训练准则出发,并借鉴线性鉴别分析的思想,提出了一种类内方差与相关度结合的特征选择算法,并使用核方法将其推广到可以解决非线性分类的特征选择问题.该算法不仅同时考虑了样本特征之间以及特征与分类标号之间的相关性,而且使得类内方差最小,有效地提高了分类器的性能.仿真实验表明:该算法适用于对特征数量多、特征相关性强的数据集进行特征选择,其选择的特征子集能够显著提高分类精度,具有较大的优越性. To solve the problems of large calculation and narrow application scope of present feature selection algorithms which do not take the correlation among sample features and that between sample features and classification labels into account, a feature selection algorithm combining within-class variance with correlation measure was proposed based on the principle of the minimum mean square error for classifier and the idea of linear discrimination analysis, then the algorithm was expanded to solve nonlinear feature selection problems by using kernel technology. The proposed algorithm can not only consider both the correlation among the features and that between the features and classification labels, but also minimize the within-class variance, effectively improving the generalization performance of classifier. The experimental results show that the proposed algorithm can be used to select important features from these data sets with numbers of correlation characteristics, and effectively improves the classification accuracy.
出处 《哈尔滨工业大学学报》 EI CAS CSCD 北大核心 2011年第3期132-136,共5页 Journal of Harbin Institute of Technology
基金 现代焊接生产技术国家重点实验室开放课题研究基金资助 江苏省高技术研究资助项目(BG2007013)
关键词 类内方差 相关度 特征选择 LDA 分类 within-class variance correlation measure feature selection LDA classification
  • 相关文献

参考文献1

二级参考文献2

  • 1Geoffrey I. Webb,Janice R. Boughton,Zhihai Wang. Not So Naive Bayes: Aggregating One-Dependence Estimators[J] 2005,Machine Learning(1):5~24
  • 2Nir Friedman,Dan Geiger,Moises Goldszmidt. Bayesian Network Classifiers[J] 1997,Machine Learning(2-3):131~163

共引文献1

同被引文献22

  • 1LIU Huan, YU Lei. Toward integrating feature selection algorithms for classification and clustering [ J]. IEEE Trans on Knowledge and Data Engineering,2005,17 (3) :491 - 502.
  • 2ZHANG Dao-qiang, CHEN Song-can, ZHOU Zhi-hua. Constraint score:a new filter method for feature selection with pair-wise con- straints [ J ]. Pattern Recognition,2008,41 (5) : 1440-1451.
  • 3Van DUCK G, Van HULLE M M. Speeding up the wrapper feature subset selection in regression by mutual information relevance and re- dundancy analysis [ C ]//Lecture Notes in Computer Science, vo14131. Berlin : Springer-Verlag, 2006 : 31 - 40.
  • 4YANG Cheng-huei, CHUANG Li-yeh, YANG Cheng-hong. IG-GA: a hybrid filter/wrapper method for feature selection of microarray data [J]. Journal of Medical and Biological Engineering, 2009,30 (1) :23-28.
  • 5COVER T M. The best two independent measurements are not the two best [ J]. IEEE Yrans on Systems, Man, and Cybernetics, 1974, 4(1) :116-117.
  • 6HSU H H, HSIEH C W, LU Ming-da. Hybrid feature selection by combining filters and wrappers [ J ]. Expert Systems with Applica- tions,2011,38(7) :8144-8150.
  • 7YU Lei, LIU Huan. Efficient feature selection via analysis of relevance and redundancy [ J ]. Journal of Machine Learning Research, 2004,5 ( 12 ) : 1205-1224.
  • 8VAFAIE H, De JONG K. Genetic algorithms as a tool for feature selec- tion in machine learning [ C ]//Proc of the 4th IEEE International Conference on Tools with AI. Washington DC:IEEE Computer Socie- tv. 1992:200- 203.
  • 9WITTEN I H, FRANK E. Data mining: practical machine learning tools and techniques[ M ]. 2nd ed. San Francisco : Morgan Kaufmann Publisher, 2005.
  • 10G.A. Rattá,J. Vega,A. Murari.Improved feature selection based on genetic algorithms for real time disruption prediction on JET[J].Fusion Engineering and Design.2012(9)

引证文献3

二级引证文献18

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部