期刊文献+

模型决策树:一种决策树加速算法 被引量:13

Model Decision Tree: An Accelerated Algorithm of Decision Tree
下载PDF
导出
摘要 决策树算法采用递归方法构建,训练效率较低,过度分类的决策树可能产生过拟合现象.因此,文中提出模型决策树算法.首先在训练数据集上采用基尼指数递归生成一棵不完全决策树,然后使用一个简单分类模型对其中的非纯伪叶结点(非叶结点且结点包含的样本不属于同一类)进行分类,生成最终的决策树.相比原始的决策树算法,这样产生的模型决策树能在算法精度不损失或损失很小的情况下,提高决策树的训练效率.在标准数据集上的实验表明,文中提出的模型决策树在速度上明显优于决策树算法,具备一定的抗过拟合能力. The decision tree algorithm is constructed in a recursive style. Therefore, the low training efficiency is yielded and the over-classification of decision tree may produce overfitting. An accelerated algorithm called model decision tree (MDT) is proposed in this paper. An incomplete classification decision tree is established via the Gini index on the training dataset firstly. Then a simple model is utilized to classify impure pseudo leaf nodes, which are neither leaf nodes nor in the same class. Consequently, the final MDT is generated. Compared with DT, MDT improves the training efficiency with smaller loss of classification accuracy or even no loss. The experimental results on benchmark datasets show that the proposed MDT is much faster than DT and it has a certain ability to avoid overfitting.
作者 尹儒 门昌骞 王文剑 刘澍泽 YIN Ru;MEN Changqian;WANG Wenjian;LIU Shuze(School of Computer and Information Technology,Shanxi University,Taiyuan 030006;Key Laboratory of Computational Intelligence and Chinese Information Processing of Ministry of Education,Shanxi University,Taiyuan 030006;Department of Computer Science,Rensselaer Polytechnic Institute,Troy,NY 12180)
出处 《模式识别与人工智能》 EI CSCD 北大核心 2018年第7期643-652,共10页 Pattern Recognition and Artificial Intelligence
基金 国家自然科学基金项目(No.61673249)、山西省回国留学人员科研基金项目(No.2016-004)、赛尔网络下一代互联网技术创新项目(No.NGIL20170601)
关键词 基尼指数 决策树(DT) 模型决策树 分类 Gini Index Decision Tree Model Decision Tree Classification
  • 相关文献

参考文献1

二级参考文献30

  • 1Stone P,Veloso M.A layered approach to learning client behaviors in the RoboCup soccer server[J].Applied Artificial Intelligence(AAI),1998,12(2-3):165-187.
  • 2Ankerst M,Elsen C,Ester M,et al.Visual classification:an interactive approach to decision tree construction[C].San Diego:In Proceedings of International Conference on Knowledge Discovery and Data Mining,1999.
  • 3Fournier D,Cremilleux B.A quality index for decision tree pruning[J].Knowledge-based Systems,2002,15(1):37-43.
  • 4Sebban M,Nock R,Chauchat J H,et al.Impact of learning set quality and size on decision tree performances[J].IJCSS,2000,1(1):85-105.
  • 5Oates T,Jensen D.The effects of training set size on decision tree complexity[C].Nashville,Tennessee:Proc of the 14th International Conference on Machine Learning,1997.
  • 6Brodley C E,Friedl M A.Identifying and eliminating mislabeled training instances[C].USA:Proceedings of the Thirteenth National Conference on Artificial Intelligence,1996.
  • 7Elouedi Z,Mellouli K,Smets Ph.Decision trees using the belief function theory[C].Madrid,Spain:Proceedings of the Eighth International Conference IPMU,2000.
  • 8Yildiz O T,Alpaydin E.Omnivariate decision trees[J].IEEE Transactions on Neural Networks,2001,12(6):1539-1546.
  • 9Swere E,Mulvaney D J.Robot navigation using decision trees[R].Loughborougk,UK:Electronic Systems and Control Division Research,2003.
  • 10Brezillon P,Pasquier L,Pomerol J.Context and decision graphs in incident management on a subway line[C].Trento,Italy:Proceedings of the 2nd International and Interdisciplinary Conference on Modeling and Using Context,1999.

共引文献61

同被引文献74

引证文献13

二级引证文献57

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部