期刊文献+

主动协同半监督粗糙集分类模型 被引量:7

A Semi-Supervised Rough Set Model for Classification Based on Active Learning and Co-Training
原文传递
导出
摘要 粗糙集理论是一种有监督学习模型,一般需要适量有标记的数据来训练分类器.但现实一些问题往往存在大量无标记的数据,而有标记数据由于标记代价过大较为稀少.文中结合主动学习和协同训练理论,提出一种可有效利用无标记数据提升分类性能的半监督粗糙集模型.该模型利用半监督属性约简算法提取两个差异性较大的约简构造基分类器,然后基于主动学习思想在无标记数据中选择两分类器分歧较大的样本进行人工标注,并将更新后的分类器交互协同学习.UCI数据集实验对比分析表明,该模型能明显提高分类学习性能,甚至能达到数据集的最优值. Rough set theory, as an effective supervised learning model, usually relies on the availability of an amount of labeled data to train the classifier. Howerer, in many practical problems, large amount of unlabeled data are readily available, and labeled ones are fairly expensive to obtain because of high cost. In this paper, a semi-supervised rough set model is proposed to deal with the partially labeled data. The proposed model firstly employs two diverse semi-supervised reducts to train its base classifiers on labeled data. The unlabeled ramified samples for two base classifiers are selected to be labeled based on the principle of active learning, and then the updated classifiers learn from each other by labeling confident unlabeled samples to its concomitant. The experimental results on selected UCI datasets show that the proposed model greatly improves the classification performance of partially labeled data, and even the bestperformance of dataset is obtained.
出处 《模式识别与人工智能》 EI CSCD 北大核心 2012年第5期745-754,共10页 Pattern Recognition and Artificial Intelligence
基金 国家自然科学基金项目(No.60970061 61075056 61103067) 中国博士后科学基金项目(No.2011M500626 2011M500815) 上海市重点学科建设项目(No.B004)资助
关键词 粗糙集 差别矩阵 半监督约简 主动学习 协同训练 Rough Set, Discernibility Matrix, Semi-Supervised Reduction, Active Learning, Co-Training
  • 相关文献

参考文献1

二级参考文献34

  • 1[1]D Cohn,Atlas R Ladner.Improving generalization with active learning.Machine Learning,1994,5(2):201-221
  • 2[2]Y Freund,H S Seung,E Shamir,et al.Selective sampling using the query by committee algorithm.Machine Learning,1997,28(2-3):133-168
  • 3[3]M Kaariainen.Active learning in the non-realizable case.In:Proc of the 17th Int'l Conf on Algorithmic Learning Theory.Berlin:Springer,2006.63-77
  • 4[4]M -F Balcan,A Beygelzimer,J Langford.Agnostic active learning.In:Proc of the 23rd Int'l Conf on Machine Learning.San Francisco,CA:Morgan Kaufmann,2006
  • 5[5]S Dasgupta.Coarse sample complexity bounds for active learning.In:Proc of Advances in Neural Information Processing Systems.Cambridge,MA:MIT Press,2005
  • 6[6]S Dasgupta,A T Kalai,C Monteleoni.Analysis of perceptron-based active learning.In:Proc of the 18th Annual Conf on Learning Theory.Berlin:Springer,2005
  • 7[7]I Dagon,S Engelson.Committee-based sampling for training probabilistic classifiers.In:Proc of the 12th Int'l Conf on Machine Learning.San Francisco,CA:Morgan Kaufmann,1995.150-157
  • 8[8]S Arganmon-Engelson,I Dagon.Committee-based sample selection for probabilistic classifiers.Journal of Artificial Intelligence research,1999,11:335-360
  • 9[9]D D Lewis,W A Gail.A sequential algorithm for training text classifiers.In:Proc of the 17th ACM Int'l Conf on Research and Development in Information Retrieval.Berlin:Springer,1994.3-12
  • 10[10]A K McCallum,K Nigam.Employing EM in pool-based active learning for text classification.In:Proc of the 15th Int'l Conf on Machine Learning.San Francisco,CA:Morgan Kaufmann,1998

共引文献30

同被引文献105

引证文献7

二级引证文献26

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部