摘要
决策树的优化是决策树学习算法中十分重要的分支.以ID3为基础,提出了改进的优化算法.每当选择一个新的属性时,算法不是仅仅考虑该属性带来的信息增益,而是考虑到选择该属性后继续选择的属性带来的信息增益,即同时考虑树的两层结点.提出的改进算法的时间复杂性与ID3相同,对于逻辑表达式的归纳,改进算法明显优于ID3.
Optimization of decision tree is a significant branch in decision tree learning algorithm. An optimized learning algorithm of ID3, a typical decision tree learning algorithm is presented in this paper. When the algorithm selects a new attribute, not only the information gain of the current attribute, but also the information gain of succeeding attributes of this attribute is taken into consideration. In other words, the information gain of attributes in two levels of the decision tree is used. The computational complexity of the modified ID3 (MID3) is the same as that of the ID3. When the two algorithms are applied to learning logic expressions, the performance of MID3 is better than that of ID3.
出处
《软件学报》
EI
CSCD
北大核心
1998年第10期797-800,共4页
Journal of Software
基金
国家863高科技项目
关键词
机器学习
决策树
分类
信息增益
熵
Machine learning, decision tree, classification, information gain, entropy.