期刊文献+

构造样本k近邻数据的多标签分类算法 被引量:7

Multi-label classification algorithm of structure sample k-nearest neighbors data
下载PDF
导出
摘要 在多标签分类问题中,通过k近邻的分类思想,构造测试样本关于近邻样本类别标签的新数据,通过回归模型建立在新数据下的多标签分类算法。计算测试样本在每个标签上考虑距离的k近邻,构造出每个样本关于标签的新数据集。对新数据集采取线性回归和Logistic回归,给出基于样本k近邻数据的多标签分类算法。为了进一步利用原始数据的信息,考虑每个标签关于原始属性的Markov边界,结合新数据的特征建立新的回归模型,提出考虑Markov边界的多标签分类算法。实验结果表明所给出的方法性能优于常用的多标签学习算法。 In multi-label classification, this paper constructs the new dataset about the nearest neighbors sample class mark through the classification idea of the k-nearest neighbors. The multi-label classification algorithm are established on the new dataset through the regression model. Firstly, this paper calculates the k-nearest neighbors distance of the test samples in each label and constructs new dataset of each sample on the label set. Secondly, the multi label classification algorithm is given based on sample k-nearest neighbors dataset, using linear regression and Logistic regression. In order to further exploit the information of original dataset, considering the Markov boundary of the original property each label and combining the feature of the new dataset to establish a new regression model, a multi-label classification algorithm about Markov boundary is proposed. The experimental results show that the multi-label learning method is better than the common learning algorithm.
出处 《计算机工程与应用》 CSCD 北大核心 2018年第6期135-142,共8页 Computer Engineering and Applications
基金 国家自然科学基金(No.11501435) 西安市科技计划项目(No.CXY1441(2)) 西安工程大学研究生创新基金(No.CX201726)
关键词 多标签分类 LOGISTIC回归 K近邻 Markov边界 multi-label classification Logistic regression k-nearest neighbors Markov boundary
  • 相关文献

参考文献1

二级参考文献46

  • 1Schapire R E, Singer Y. BoosTcxter: A boostnlg bsed syslem for text categorizaion[J]. Machine Lcarning . 39(2/3): 135- 168.
  • 2McCallum A. Muhi-lahcl lext classification with a micture model trained by EM[C] //Proc of *he Working Nolcs of 11/ AAAI'99 Workshop on Text I.earning. Menlo Park, CA: AAAI Press, 1999.
  • 3Elissecff A, Weston J. A kcrtxel method for multi -labeclledclassification [C] //Advances in Neural Informalion Processing Systcms 14. Cambridge, MA: M1T Press, 2002: 681 -687.
  • 4QiGJ, HuaX S, Rui Y, et al. Corrclaativcmulti label vidco annotation [C] //Proc of the 15th ACM Int Conf on Muhimedia. New York: ACM, 2007:17- 26.
  • 5Aha D W. Specied A1 review issoe on lazy learning [J ]. Artificial Intelligcnce Review, 1997. 11(1/2/3/4/5): 7 -10.
  • 6Zhang M L,, Zhou Z H. ML-hNN: A lazy lcarning approach to multi label learning [J]. Paltern Recognition, 2007. 10 (7): 2038 -2048.
  • 7Freund Y, Sc:hapire R E. A dccision theoretic gcncralization of on-linc learning and an applocation to boosting[G]//Lecture Notcs in Computer Scicnce 904.Bcrlin:Springer.1995:23-37.
  • 8Dempstcr A P, 1.aird N M, Rubin D B+ Maxitnuntlikclihood from incomplete data via the EM algorithm[J].Journal of the Royal Statistics Socicty B, 1977, 39(1): 1-38.
  • 9Ueda N, Saito K. Parametric mixturc models for multi label text [C] //Advances in Neural Information Processing Systems 15. Cambridge, MA= MITPress, 2003:721-728.
  • 10Dumais S T, Platt J, Heckerman D, et al. Inductive learning algorithm and representation for text categorization [C]// Proc of the 7th ACM Int Conf on Information and Knowledge Management. New York: ACM, 1998= 148-155.

共引文献38

同被引文献74

引证文献7

二级引证文献34

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部