期刊文献+

逻辑模型树算法性能分析与改进研究 被引量:6

The research on analysis and improvement of logistic model tree
下载PDF
导出
摘要 逻辑模型树(LMT)算法是基于树归纳和逻辑回归的一种分类算法。为验证LMT算法的优势,利用3个UCI标准数据集建模,将LMT算法与其他决策树方法进行对比分析。针对LMT算法在建立逻辑回归模型时会导致较高的计算复杂性的问题,研究利用赤池信息量准则改进LMT算法,提升算法时间性能,避免模型过度拟合。在UCI标准数据集和烟叶综合质量评价数据中应用改进的LMT算法进行建模验证,结果表明,该改进方法在模型精度和召回率方面基本优于其他决策树方法,时间性能比改进前提升50%左右,能较好地评价烟叶综合质量。 Logistic Model Trees (LMT) algorithm is a classification algorithm which is based on tree induction and logistic regression. To verify the advantage of LMT, compare and analyze LMT with other decision tree methods on three UCI data sets. Because in logistic model trees, logistic regression models can lead to the high computational complexity. This issue can be addressed by using the AIC criterion to improve LMT. It can improve time performance of algorithm and prevent over fitting models. The modification of LMT is used on UCI data sets and tobacco comprehensive quality evaluation data. And the result demonstrates that this method is superior to other decision tree methods in model precision and recall rate and time performance is about 50% faster than the unimproved. It can evaluate tobacco comprehensive quality well.
出处 《微型机与应用》 2014年第23期25-28,共4页 Microcomputer & Its Applications
基金 青岛市科技计划项目(12-4-1-9-JX) 国家科技支撑计划项目(013BAH17F01)
关键词 逻辑模型树 UCI标准数据集 烟叶综合质量评价数据 赤池信息量准则 模型精度 召回率 logistic model tree UCI data sets tobacco comprehensive quality evaluation data Akaike information criterion model precision recall rate
  • 相关文献

参考文献11

  • 1田苗苗.数据挖掘之决策树方法概述[J].长春大学学报,2004,14(6):48-51. 被引量:42
  • 2PERLICH C,PROVOST F,SIMONOFF J S.Tree induction vs.logistic regression:a learning-curve analysis[J].The Journal of Machine Learning Research,2003,(4):211-255.
  • 3LANDWEHR N,HALL M,FRANK E.Logistic model trees[J].Machine Learning,2005,59(1/2):161-205.
  • 4FRIEDMAN J,HASTIE T,TIBSHIRANI R.Special invited paper.additive logistic regression:A statistical view of boosting[J].Annals of statistics,2000,28(2):337-374.
  • 5BREIMAN L,FRIEDMAN H,OLSHEN J A.Classification and Regression Trees[M].New York:Wadsworth,1984.
  • 6SUMNER M,FRANK E,HALL M.Speeding up logistic model tree induction[M].Knowledge Discovery in Databases:PKDD 2005,Springer Berlin Heidelberg,2005.
  • 7富春枫,荀鹏程,赵杨,陈峰.logitboost及其在判别分析中的应用[J].中国卫生统计,2006,23(2):98-100. 被引量:11
  • 8季桂树,陈沛玲,宋航.决策树分类算法研究综述[J].科技广场,2007(1):9-12. 被引量:38
  • 9AKAIKE H.Information theory and an extension of the maximum likelihood principle[M].Breakthroughs in statistics,Springer New York,1992.
  • 10De Ridder F,PINTELON R,SCHOUKENS J,et al.Modified AIC and MDL model selection criteria for short data records[J].Instrumentation and Measurement,IEEE Transactions on,2005,54(1):144-150.

二级参考文献14

  • 1迟庆云.决策树分类算法及其应用[J].枣庄学院学报,2005,22(5):29-32. 被引量:9
  • 2乔向杰,陈功平.数据挖掘中分类算法的可扩展性研究[J].信阳师范学院学报(自然科学版),2006,19(2):239-242. 被引量:6
  • 3Schapire RE.The Boosting Approach to Machine Learning An Overview.MSRI Workshop on Nonlinear Estimation and Classification,2002.
  • 4Schapire RE.The strength of weak learnability.Machine Learning,1990,5(2):197-227.
  • 5Freund Y,Schapire RE.A decision-theoretic generalization of on-line learning and an application to boosting.Journal of Computer and System Sciences,1997,55(1):119-139.
  • 6Friedman JH,Trevor,Robert Tibshirani.Additive logistic regression:A statistical view of boosting.The Annals of Statistics,2000,38(2):337-374.
  • 7King RD.The Stalog databases.fttp.strath.ac.uk(130.159.248.24).
  • 8Quinlan JR.C4.5:Programs for Machine Learning.Morgan Kaufmann,1993.
  • 9Drucker H,Schapire R,Simard P.Boosting performance in neural networks.International Journal of Pattern Recognition and Artificial Intelligence,1993,7(4):705-719.
  • 10Dietterich TG.An experimental comparison of three methods for constructing ensembles of decision trees:Bagging,boosting,and randomization.Machine Learning,2000,40(2):139-158.

共引文献88

同被引文献49

引证文献6

二级引证文献24

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部