期刊文献+

一种新的双重融合的半监督特征选择算法 被引量:1

Novel Dual Ensemble Based Semi-supervised Feature Selection Method
下载PDF
导出
摘要 针对特征选择算法的鲁棒性和稳定性问题以及现实应用领域中大量的廉价未标签数据的利用问题,提出一种基于双重融合策略的半监督特征选择算法.该方法综合利用弱分类器融合技术和未标签数据包含的数据集的簇的结构信息来扩充标签数据集,然后在得到的标签数据集上采用不同的特征选择算法,对不同的特征结果进行简单的融合操作,得到最终的特征子集.在一些公共数据集和有毒性预测数据集上的实验结果表明该方法在改善学习精度上有很好的应用前景. A Dual Ensemble based Semi-Supervised Feature Selection Method(ESSFS),where multiple feature selection methods are combined to yield more robust results,is proposed in this paper.It can deal with the problem of robustness or stability of feature selection techniques as well as the use of "cheaper" unlabeled data in several practical applications.At first,method of weak classifiers' ensemble combined with the information provided by clusters was used to enlarge the labeled dataset.Then,different feature selection methods were applied to this dataset,the final feature subset was obtained after the feature selection ensemble.Experimental results carried out on some public datasets collected from the UCI machine learning repository and predictive toxicology domain show that ESSFS has a promising performance on the improvement of the learning accuracy.
作者 陈红 郭躬德
出处 《小型微型计算机系统》 CSCD 北大核心 2010年第8期1604-1608,共5页 Journal of Chinese Computer Systems
基金 福建省自然科学基金项目(2007J0016)资助 教育部回国留学人员基金项目(教外司留[2008]890号)资助
关键词 特征选择 半监督 双重融合 稳定性 feature selection semi-supervised dual ensemble stability
  • 相关文献

参考文献18

  • 1ndl J,Knowles J. Feature subset selection in unsupervised learning via multiobjective optimization [ J ]. Int'l J. Computational Intelligence Research, 2006,2(3) :21%238.
  • 2Dash M, Liu H. Handling large unsupervised data via dimensionality reduction[C]. In: Proceedings of 1999 SIGMOD Research Issues in Data Mining and Knowledge Discovery. DMKD-99, Work-shop, 1999.
  • 3张丽新,王家钦,赵雁南,杨泽红.机器学习中的特征选择[J].计算机科学,2004,31(11):180-184. 被引量:18
  • 4Yvan Saeys, Thomas Abeel,Yves Van de Peer. Robust feature selection using ensemble feature selection techniques[M]. Machine Learning and Knowledge Discovery in Databases, vol 5212. Springer Berlin / Heidelberg, 2008,313-325.
  • 5Hand D, Marmila H, Smyth P. Principles of data mining [ M ]. The MIT Press, 2001.
  • 6Pierre-Emmanuel Jouve, Nicolas Nicoloyannis. A filter feature selection method for clustering[ M]. Foundations of Intelligent Systerns,vol 3488. Springer Berlin / Heidelberg, 2005,583-593.
  • 7Ren Jiang-tao, Qiu Zheng-yuan, Wei Fan,et al. Forward semi-supervised feature selection[ M]. Advances in Knowledge Discovery and Data Mining, voi 5012. Springer Berlin / Heidelberg, 2008, 970-976.
  • 8Marina Skuriehina, Robert P W Duin. Combining feature subsets in feature selection [ M]. Multiple Classifier Systems, vol 3541. Springer Berlin / Heidelberg, 2005,165-175.
  • 9Hugo Silva, Ana Fred. Feature subspace ensembles: a parallel classifter combination scheme using feature selection [ M ]. Multiple Classifier Systems, vol 4472. Springer Berlin / Heidelberg, 2007, 261-270.
  • 10Opitz D. Feature selection for ensembles[ C]. In: 16th National Conference on Artificial Intelligence. AAAI (1999) ,379-384.

二级参考文献53

  • 1Handl J,Knowles J.Feature subset selection in unsupervised learning via muhiobjective optimization[J].Int'l J Computational Intelligence Research,2006,2(3) :217-238.
  • 2Dash M,Liu H.Handling large unsupervised data via dimensionality reduction[C]//Proceedings of 1999 SIGMOD Research Issues in Data Mining and Knowledge Discovery(DMKD-99) Workshop, 1999.
  • 3Liu H,Yu L.Towards integrating feature selection algorithms for classification and clustering[J].IEEE Transactions on Knowledge and Data Engineering,2005,17(3 ) : 1-12.
  • 4Zhao Z,Liu H.Semi-supervised feature selection via spectral analysis[C]//SIAM International Conference on Data Mining,SDM 2007, 2007.
  • 5EU FP5 Quality of Life DEMETRA QLTR-2001-00691:Development of environmental modules for evaluation of toxicity of pesticide residues in agriculture.
  • 6Huang Y,Guo G,Neagu D.A partial coverage based approach to classification[C]//Proc of the 19th IEEE International Conference on Tools with Artificial Intelligence,ICTAI 2007,2007,1:275-280.
  • 7Nakatani Y,Zhu K,Uehara K.Semi-supendsed learning using feature selection based on maximum density subgraphs[J].Systems and Computers in Japan,2007,38(9):32-43.
  • 8Wang H,Duntsch I,Bell D.Data reduction based on hyper relations[C]//Proc of KDD, 1998 : 349-353.
  • 9Huang Y,Guo G,Chen H,et al.Dynamic partial coverage based feature selection method[C]//Proc of FSKD,2008.
  • 10UCI Repository of Machine Learning Databases[EB/OL].http://www. ics.uci.edu/-mlearn/MLRository.html.

共引文献18

同被引文献25

  • 1张莉,孙钢,郭军.基于K-均值聚类的无监督的特征选择方法[J].计算机应用研究,2005,22(3):23-24. 被引量:29
  • 2Boutcmedjet S, Bouguila N, Ziou D. A Hybrid Feature Extraction Selection Approach for High-Dimensional Non-Gaussian Data Cluste- ring. IEEE Trans on Pattern Analysis and Machine Intelligence, 2009, 31 (8) : 1429-1443.
  • 3Boutsidis C, Mahoney M W, Drineas P. Unsupervised Feature Selection for Principal Components Analysis // Proe of the 14th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Las Vegas, USA, 2008:61-69.
  • 4Cai D, He X F, Han J W. Sparse Projections over Graph//Proe of the 2lst AAAI Conference on Artifieial Inte|Iigence. Chicago, USA, 2008:610-615.
  • 5Xu Z L, King I, Lyu M R T, et al. Discriminative Semi-Supervised Feature Selection via Manifold Regularization. IEEE Trans on Neu- ral Networks, 2010, 21 (7) : 1033-1047.
  • 6He X F, Cai D, Niyogi P.' Laplacian Score for Feature Selection// Proc o: the Advances in Neural lnfonnat:,n Process:,r,g Systems lg. Vancouver, Canada, 2005:507-514.
  • 7Cai D, Zhang C Y, He X F. Unsupervised Feature Selection for Multi-cluster Data// Proc of the 16th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining. Washing- ton, USA, 2010:333-342.
  • 8Li Z C, Yang Y, Liu J, et al. Unsupervised Feature Selection Using Nonnegative Spectral Analysis // Proc of the 26th AAAI Conference on Artificial Intelligence. Toronto, Canada, 2012 : 1026-1032.
  • 9Zhu X F, Huang Z, Yang Y, et al. Self-Taught Dimensionality Reduction on the High-Dimensional Small-Sized Data. Pattern Recognition, 2013, 46(1 ): 215-229.
  • 10He X F, Niyogi P. D.:cality Preserving Projections// Proe of the Advances in Neural Information Processing Systems 16. Vancouver, Canada, 2003 : 153-160.

引证文献1

二级引证文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部