期刊文献+

基于深度森林的量表数据挖掘方法 被引量:4

Deep forest based inventory data mining method
下载PDF
导出
摘要 在数据挖掘领域中,量表是间接获取样本属性数据的重要工具。针对量表数据离散、稀疏、二值化的特点,导致其难以进行分析挖掘的问题。文中采用了基于深度森林的量表数据挖掘方法,实验分别对老年健康综合评估数据库中的两个量表进行对比分析。实验结果表明,在所选取的两个量表中,提取到的关键属性数量相比于原始量表属性数量分别下降了30%和40%,且比基线模型下降了16%和18%。同时,提出的方法可在保证分类性能基本不变的情况下,进一步降低提取到的关键属性数量。 In data mining,inventory is one of the important equipment for acquiring data.However,the crucial attribute analysis often fails on inventory data due to the discrete,sparse and binary charac teristics of inventory.In this paper,we propose a deep forest based inventory data mining method.The experiments are carried out on two inventory datasets from the elderly health comprehensive evaluating database.The results show that in the selected two datasets,the extracted crucial attributes are reduced by 30%and 40%respectively compared with the original ones.Moreover,the number of crucial attributes extracted by proposed model is reduced by 16%and 18%compared with baselines.Besides,it also indicates that the proposed model can further reduce the number of crucial attributes while maintaining the performance of classification.
作者 佟彤 罗森林 潘丽敏 张铁梅 TONG Tong;LUO Sen-lin;PAN Li-min;ZHANG Tie-mei(Information System&Security and Countermeasures Experiments Center,Beijing Institute of Technology,Beijing 100081,China;Key Laboratory of Geriatric Medicine of Ministry of Health,National Geriatric Center,Beijing Hospital,Beijing 100730,China)
出处 《电子设计工程》 2020年第13期88-91,96,共5页 Electronic Design Engineering
关键词 深度森林 量表数据 数据挖掘 属性提取 deep forest inventory data data mining attribute extraction
  • 相关文献

参考文献9

二级参考文献85

  • 1康重庆,夏清,张伯明.电力系统负荷预测研究综述与发展方向的探讨[J].电力系统自动化,2004,28(17):1-11. 被引量:496
  • 2贺铿.关于信息产业和信息产业投入产出表的编制方法[J].数量经济技术经济研究,1989(2):34-40. 被引量:22
  • 3GUYON I, ELISSEEFF A. An introduction to variable and feature selection[J]. J Mach Learn Res, 2003, 3: 1157-1182.
  • 4ABDI H, WILLIAMS. "Principal component analysis" Wiley interdisciplinary reviews[J]. Computational Statistics, 2010, 2: 433-459.
  • 5KOHAVI R, JOHN G H. Wrappers for feature subset selection[J]. Artiflntell, 1997, 97: 273-324.
  • 6JUHA R. Overfitting in making comparisons between variable selection method[J]. Journal of Machine Learning Research, 2003, 3: 1371-1382.
  • 7LIU Yi, ZHENG Yuan. FS_SFS: a novel feature selection method for support vector machines[J]. Pattern Recognit, 2006, 39: 1333-1345.
  • 8LIU Huan, SETIONO R. A probabilistic approach to feature selection: a filter solution[C]//Proceedings of the Thirteenth International Conference on Machine Learning. Bari: [s.n.], 1996, 319-327.
  • 9CHEN W, CHANG X, WANG H, et al. Automatic word clustering for text categorization using global information [C]//Asia Information Retrieval Syrup. Beijing: Springer- Verlag, 2004, 1-11.
  • 10XIONG M, FANG Z, ZHAO J. Biomarker identification by feature wrappers[J]. Genome Res, 2001, lh 1878-1187.

共引文献92

同被引文献37

引证文献4

二级引证文献10

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部