期刊文献+

基于随机子空间的多分类器集成 被引量:4

Multi-Classifier Ensemble Based on Random Feature Subspace
下载PDF
导出
摘要 提出了一种基于随机子空间的多分类器集成算法RFSEn.首先选择一个合适的子空间大小,然后随机选择特征子集并投影,并得到子空间上的基分类器,从而通过基分类器构成集成分类器,并由集成分类器来进行文本的分类.将该算法与单一分类器和基于重抽样技术的bagging算法进行了比较,在标准数据集上进行了实验.结果表明,该方法不仅优于单一分类器的分类性能,而且一定程度上优于bagging算法. In this paper,we propose an ensemble algorithm called RFSEn which is based on random feature subspaee. First, an appropriate feature subset size is selected, then subsets of features are randomly and projected on the training set, and the primary classifiers of subspace are obtained, and thus ensembled classifiers are formed with these primary classifiers. At last, we use the ensembled classifier to classify the text. We compare the algorithm with bagging algorithm which is based on re-sampling techniques and single classifier on the standard datasets. The results show that RFSEn al- gorithm is not only superior to single classifier in performance, but better than bagging algorithm in some degree.
作者 叶云龙 杨明
出处 《南京师范大学学报(工程技术版)》 CAS 2008年第4期87-90,共4页 Journal of Nanjing Normal University(Engineering and Technology Edition)
关键词 随机子空间 分类器集成 重抽样 random sub-space, classifier ensemble, re-sampling
  • 相关文献

参考文献11

  • 1[1]Dietterich T G.Machine learning research:four current directions[J].AI Magazine,1997,18(4):97-136.
  • 2[2]Freund Y,Schapire R E.Experiments with a new boosting algorithm[C]// Proceedings of the 13th International Conference on Machine Learning.San Francisco:Morgan Kaufmann,1996:148-156.
  • 3[3]Breiman L.Bagging predictors[J].Maching Learning,1996,24(2):123 -140.
  • 4[4]Zhou Z H,Wu J,Tang W.Ensembling neural networks:many could be better than all[J].Artificial Intelligence,2002,137(1/2):239-263.
  • 5[5]Weiss S M,Apte C,Damerau FJ.Maximizing text-mining performance[J].IEEE Intelligent Systems,1999,14 (4):63 -69.
  • 6[6]Schapire R E,Singer Y.Boostexter:a boosting-based system for text categorization[J].Machine Learning,2000,39(223):135-168.
  • 7[7]Tumer K,Ghosh J.Classifier combining:analytical results and implications[C]// Proceeding of the AAAI-96 Workshop on Integrating Multiple Learned Models for Improving and Scaling Machine Learing Algorithms.Portland:AAAI Press,1996.
  • 8[8]Wang Xiaogang,Tang Xiaoou.Using random subspace to combine multiple features for face recognition[C]// Proceeding of the 6th IEEE International Conference on Automatic Face and Gesture Recognition.Los Alamitos:IEEE Computer Society Press,2004:284-289.
  • 9[9]Bay S D.Combining nearest neighbor classifiers through multiple feature subsets[C]// Proceeding of the Proceedings of the 17th International Conference on Machine Learning.Madison,WI:Morgan Kaufmann,1998:37 -45.
  • 10[10]Robert Brylla.Attribute bagging:improving accuracy of classifier ensembles by using random feature subsets[J].Pattern Recognition,2003,36(6):1 291 -1 302.

同被引文献46

  • 1Sebastiani F. Machine learning in automated text categorization[J ]. ACM computing survey, 2002,34 ( 1 ) : 1 - 47.
  • 2Tumer K, Ghosh J. Classifier combining: analytical results and implications, working notes from the workshop 'integrating multiple learned models' [C]//Proc. of 13th National Conference on Artificial Intelligence. Portland, Oregon, 1996.
  • 3John G, Kohavi R, Pfleger K. Irrelevant features and the subset selection problem[ C]//The Eleventh International Conference on Machine Learning. San Fransico: Morgan Kaufmann, 1994.121 - 129.
  • 4Moore A W, Lee M S. Efficient algorithms for minimizing cross validation error[C]//Proceedings of the 17th International Conference on Machine Learning. San Fransico: Morgan Kaufmann, 1994 : 190 - 198.
  • 5Kohavi R, John G H. Wrappers for feature subset selection [ J ]. Artificial Intelligence, 1997,97 ( 1/2) : 273 - 324.
  • 6富春岩,葛茂松.一种能够适应概念漂移变化的数据流分类方法[J].智能系统学报,2007,2(4):86-91. 被引量:5
  • 7Widmer G Kubat M. Learning in the presence of concept drift and hidden contexts.Machine Learning,1996,23(l): 69-101.
  • 8Tsymbal A, Pechenizkiy M, Cunningham P, et al. Dynamic integration of classifiers for handling concept drift. Information Fusion, 2008,9(1):56-68.
  • 9Hanen LK, Salamon E Neutral network ensemble. IEEE Trans. on Pattern Anaylsis and Machine Intelligence, 1990, 12(10):993 -1001.
  • 10Street W, Kim Y, A streaming ensemble algorithm (SEA) for large-scale classification. Proc. of 7th ACM SIGKDD Int. Conf. on Knowledge Discovery and Data Mining KDD-2001. New York: ACM Press, 2001:77-382.

引证文献4

二级引证文献10

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部