期刊文献+

基于双属性节点部分匹配的决策树改进算法

Improved Decision Tree Based on Double-attribute and Part-match
下载PDF
导出
摘要 在决策树算法中,即使存在两个"最好"属性,也只是随机选择一个作为根或节点属性.因此,决策树算法产生的分类规则较少.此外,决策树算法采用全匹配测试实例,测试实例最多匹配一条分类规则甚至没有匹配,进而影响分类准确率.针对该问题,提出了基于双属性节点部分匹配的决策树改进算法(DAID3):首先,如果存在两个信息熵相等或相近的"最好"属性,DAID3算法选择两个属性构建节点,它们的属性值及组合作为分枝.因此,每个训练实例可能被多条分类规则覆盖.其次,判断新实例时,在分枝节点上可能匹配到多条路径,为了选择最好路径,为每个分枝节点设置了节点强度.最后,如果不存在一条从根节点到叶子节点的路径全匹配测试实例,则找出部分匹配该实例的路径,返回该路径的终节点强度最大的类标值.为了便于部分匹配时返回强度最大的类标值,为每一个分枝节点设置节点类标值.实验结果表明,与决策树算法相比,DAID3算法具有分类规则多且有更高的分类准确率. Decision tree algorithm only chooses one attribute as the root or the node attributes, even if there are two best attributes. So the decision tree algorithm produces few rules. Using full-match method, a new instance only matches one rule or none, which decrease the accuracy of the decision tree. Aiming at the problem, we improve decision tree based on double-attribute and part-match(DAID3): First, if two best attributes have same or approximate entropy, DAID3 algorithm selects the two attributes to build the node, and the two attribute values and their combination as the node’ branches. So a train instance can be covered one more times. Second, in the branch node, a new instance may be matched by one more branch path. In order to select the best one; we measure each branch node by the strong. Finally, for a new instance, if there is not a path from the root to a leaf to fully match the instance, DAID3 finds the longest part path, and returns the max-strong class label of the end note. So we labeled each branch node by the max-strong class label. Experimental results show that DAID3 algorithm has higher accuracy than the decision tree more.
出处 《闽南师范大学学报(自然科学版)》 2015年第4期41-46,共6页 Journal of Minnan Normal University:Natural Science
基金 国家自然科学基金资助项目(61170129) 福建省自然科学基金资助项目(2013J01259)
关键词 分类 不平衡数据 小类 选择度 data mining classification decision tree DAID3
  • 相关文献

参考文献14

  • 1Quinlan J R. Induction of decision trees[J]. Mach Learn, 1986, 1(1): 81-106.
  • 2丁春荣,李龙澍,杨宝华.基于粗糙集的决策树构造算法[J].计算机工程,2010,36(11):75-77. 被引量:23
  • 3Huang L, Huang M, Guo B, et al. A New Method for Constructing Decision Tree Based on Rough Set Theory[C]//2013 IEEE International Conference on Granular Computing (GrC). IEEE Computer Society, 2007: 241-244.
  • 4Davis J V, Ha J, Rossbach C J, et al. Cost-sensitive decision tree learning for forensic classification[M]. Machine Learning: Springer, 2006.
  • 5Deng L, Song J-Y. Decision tree classification algorithm based on cost and benefit dual-sensitive[C]//2014 IEEE International Conference on the Eleetro/Information Technology (EIT). IEEE, 2014: 320-323.
  • 6Min F, Zhu W. A competition strategy to cost-sensitive decision trees[M]. Rough Sets and Knowledge Technology: Springer, 2012.
  • 7An S, Shi H, Hu Q, et al. Fuzzy Rough Decision Trees[J]. Fundamenta Informaticae, 2014, 132(3): 381-399.
  • 8Zhai J-H. Fuzzy decision tree based on fuzzy-rough technique[J]. Soft Computing, 2011, 15(6): 1087-1096.
  • 9Hoens T R, Qian Q, Chawla N V, et al. Building decision trees for the multi-class imbalance problem[M]. Advances in Know- ledge Discovery and Data Mining: Springer, 2012.
  • 10Liu W, Chawla S, Cieslak D A, et al. A Robust Decision Tree Algorithm for Imbalanced Data Sets[C]//2010 Proceedings of the SIAM International Conference on Data Mining. SIAM 2010, 2010: 766-777.

二级参考文献8

  • 1乔梅,韩文秀.基于Rough集的决策树算法[J].天津大学学报(自然科学与工程技术版),2005,38(9):842-846. 被引量:9
  • 2Quinlan J R. Induction of Decision Tree[J]. Machine Learning, 1986, 1(1): 81-106.
  • 3Pawlak Z. Rough Sets[J]. International Journal of Computer Information Science, 1982, 11(5): 341-356.
  • 4Ziarko W. Probabilistic Decision Tables in the Variable Precision Rough Set Model[J]. Computational Intelligence, 2001, 17(3): 593-603.
  • 5Blake C M. UCI Machine Learning Repository[EB/OL]. (2005-10-21 ). http ://www.ics.uci.edu/-mlearn/MLRepository.html.
  • 6张文修 吴伟志 梁吉业.粗糙集理论与方法[M].北京:科学出版社,2003.107-112.
  • 7高静,徐章艳,宋威,杨炳儒.一种新的基于粗糙集模型的决策树算法[J].计算机工程,2008,34(3):9-11. 被引量:16
  • 8庞哈利,高政威,左军伟,卞玉倩.基于变精度粗糙集的分类决策树构造方法[J].系统工程与电子技术,2008,30(11):2160-2163. 被引量:9

共引文献22

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部