期刊文献+

一种懒惰式决策树和普通决策树结合的分类模型--半懒惰式决策树 被引量:1

A CLASSIFIER HYBRID MODEL BASED ON REGULAR DECISION TREE AND LAZY DECISION TREE—SEMI-LDTREE
下载PDF
导出
摘要 懒惰式决策树分类是一种非常有效的分类方法。它从概念上为每一个测试实例建立一棵"最优"的决策树。但是,大多数的研究是基于小的数据集合之上。在大的数据集合上,它的分类速度慢、内存消耗大、易被噪声误导等缺点,影响了其分类性能。通过分析懒惰式决策树和普通决策树的分类原则,提出了一种新的决策树分类模型,Semi-LDtree。它生成的决策树的节点,如普通决策树一样,包含单变量分裂,但是叶子节点相当于一个懒惰式决策树分类器。这种分类模型保留了普通决策树良好的可解释性,实验结果表明它提高了分类速度和分类精确度,在某些分类任务上它的分类性能经常性地胜过两者,特别是在大的数据集合上。 Lazy decision tree is a very effective classification method. It conceptually constructs the "best" decision tree for each test instance. However, most studies are done on small databases. In some larger databases, lazy decision tree shows its deficiencies in classification speed, memory consumption, and it is easily induced by noises, which affects its classification performance. On the basis of the analysis on the classification principles of regular decision tree classification model and Lazy decision tree classification model, a new decision tree classification model, Semi-LDtree is proposed. The decision tree nodes generated by the new model contain univariate splits like regular decision trees, but the leaves are egual to Lazy decision tree classifiers. This classification model retains the interpretability of regular decision tree. Experimental results show that this model has higher classification accuracy and faster classification speed, and it frequently outperforms the traditional two models especially in larger databases tested.
出处 《计算机应用与软件》 CSCD 北大核心 2008年第12期229-230,238,共3页 Computer Applications and Software
关键词 懒惰式决策树 朴素贝叶斯 半懒惰式决策树算机 Lazy decision tree Naive bayes Semi-LDtree
  • 相关文献

参考文献8

  • 1Han J, Kamber M. Data Mining Concepts and Techniques [ M ]. San Francisco: Morgan Kaufmann Publishers,2001:185 - 219.
  • 2Mitchell TM. Machine Learning[ M ]. McGraw Hill, 1997 : 112 - 140.
  • 3石洪波,王志海,黄厚宽,励晓健.一种限定性的双层贝叶斯分类模型[J].软件学报,2004,15(2):193-199. 被引量:44
  • 4Simovici Dan A, Jaroszewicz Szymon. A Metric Approach to Building Decision Trees Based on Goodman-Kruskal Association Index[ C]. PAKDD ,2004 : 181 - 190.
  • 5Witten IH, Frank E. Data Mining: Practical Machine Learning Tools and Techniques with Java Implementations [ M ]. Seattle : Morgan Kaufmann ,2000.
  • 6Friedman JH, Kohavi Ron, Yeogirl Yun. Lazy Decision Trees[ C]. AAAI-96 ,1996 :717 - 724.
  • 7Kohavi R. Scaling up the accuracy of Naive-Bayes classifiers:A decision-tree Hybrid[ C ]. In : Simoudis E, Han j, Fayyad UM, eds. Proc. of the 2 Int'l Conf. on Knowledge Discovery and Data Mining. Menlo Park : AAAI Press, 1996:202-207.
  • 8Newman D J, Hettich S, Blake C L, et al. UCI Repository of machine learning databases [ http://www, ics. uci. edu/- mlearn/MLRepository. html]. Irvine, CA: University of California, Department of Information and Computer Science, 1998.

二级参考文献15

  • 1Friedman N,Geiger D,Goldszmidt M.Bayesian network classifiers.Machine Learning,1997,29(2-3):131-163.
  • 2Langley P,Iba W,Thompson K.An analysis of Bayesian classifiers.In:Rosenbloom P,Szolovits P,eds.Proc.of the 10th National Conf.on Artificial Intelligence.Menlo Park:AAAI Press,1992.223-228.
  • 3Kononenko I.Seminaive Bayesian classifier.In:Kodratoff Y,ed.Proc.of the 6th European Working Session on Learning.New York:Springer-Verlag,1991.206-219.
  • 4Pazzani MJ.Searching for dependencies in Bayesian classifiers.In:Fisher D,Lenz HJ,eds.Learning from Data:Artificial Intelligence and Statistics V.New York:Springer-Verlag.1996.239-248.
  • 5Langley P,Sage S.Induction of selective Bayesian classifiers.In:Mantaras RL,Poole DL,eds.Proc.of the 10th Conf.on Uncertainty in Artificial Intelligence.San Francisco:Morgan Kaufmann Publishers,1994.399-406.
  • 6Webb GI,Pazzani MJ.Adjusted probability naive Bayesian induction.In:Antoniou G,Slaney JK,eds.Proc.of the 11th Australian Joint Conf.on Artificial Intelligence.Berlin:Springer-Verlag,1998.285-295.
  • 7Kohavi R.Scaling up the accuracy of Naive-Bayes classifiers:A decision-tree hybrid.In:Simoudis E,Han J,Fayyad UM,eds.Proc.of the 2nd Int'l Conf.on Knowledge Discovery and Data Mining.Menlo Park:AAAI Press,1996.202~207.
  • 8Keogh EJ,Pazzani MJ.Learning augmented Bayesian classifiers:A comparison of distribution-based and classification-based approaches.In:Heckerman DE,Whittaker J,eds.Proc.of the Uncertainty'99:The 7th Int'l Workshop on Artificial Intelligence and Statistics.
  • 9Cheng J,Greiner R.Comparing Bayesian network classifiers.In:Laskey KB,Prade H,eds.Proc.of the 15th Conf.on Uncertainty in Artificial Intelligence.San Francisco:Morgan Kaufmann Publishers,1999.101-108.
  • 10Chickering DM,Geiger D,Heckerman D.Learning Bayesian networks is NP-complete.In:Fisher DH,Lenz HJ,eds.Learning from Data:Artificial Intelligence and Statistics V.New York:Springer-Verlag,1996.121-130.

共引文献43

同被引文献14

  • 1李根国,桂亚东,刘欣.浅谈高性能计算的地位及应用[J].计算机应用与软件,2006,23(9):3-4. 被引量:15
  • 2付萍,薛定宇,林明秀,徐心和.基于多线索的车辆跟踪方法研究[J].系统仿真学报,2007,19(22):5299-5303. 被引量:1
  • 3Randal L S,Tom P.Learning Perl. . 2009
  • 4Wu Z,Xie M,Tian Yu.Optimization design of the X&S charts for monitoring process capability. Jour-nal of Manufacturing Systems . 2002
  • 5Yin Xiaoming,Xie M.Finger identification in handgesture based Human-Robot interaction. Roboticsand Autonomous Systems . 2001
  • 6Rinaldo Christian Tanumara,Xie M,Au Chi Kit.Learning Human-like color categorization through in-teraction. International Journal of ComputationalIntelligence . 2006
  • 7Guo Dong,Ming Xie.Color clustering and learning for image segmentation based on neural networks. IEEE Transactions on Neural Networks . 2005
  • 8Bertozzi M,Broggi A.GOLD: a parallel real-time stereo vision system for generic obstacle and lane detection. IEEE Transactions on Image Processing . 1998
  • 9Scott Meyers.More Effective C++. . 2003
  • 10Hudak,P.Conception, Evolution, and Application of Functional Programming Languages. ACM Computing Surveys . 1989

引证文献1

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部