期刊文献+

多标记数据特征提取方法的核改进 被引量:4

Kernel improvement of multi-label feature extraction method
下载PDF
导出
摘要 针对多标记数据特征提取方法中输出核函数没有准确刻画标记间的相关性的问题,在充分度量标记间相关性的基础上,提出了两种新的输出核函数构造方法。第一种方法首先将多标记数据转化为单标记数据,并使用标记集合来刻画标记间的相关性;然后从损失函数的角度出发定义新的输出核函数。第二种方法是利用互信息来度量标记间的两两相关性,在此基础上进一步构造新的输出核函数。3个多标记数据集上2种分类器的实验结果表明,与原有核函数对应的多标记特征提取方法相比,基于损失函数的输出核函数对应的特征提取方法性能最好,5个评价指标的性能平均提高了10%左右,尤其在Yeast数据集上,Coverage指标下降幅度达到了30%左右;基于互信息的输出核函数次之,性能平均提高了5%左右。实验结果表明,基于新的输出核函数的特征提取方法能够更加有效地提取特征,并进一步简化分类器的学习过程,提高分类器的泛化性能。 Focusing on the issue that the label kernel functions do not take the correlation between labels into consideration in the multi-label feature extraction method, two construction methods of new label kernel functions were proposed. In the first method, the multi-label data were transformed into single-label data, and thus the correlation between labels could be characterized by the label set; then a new label kernel function was defined from the perspective of loss function of single-label data. In the second method, mutual information was used to characterize the correlation between labels, and a new label kernel function was proposed from the perspective of mutual information. Experiments on three real-life data sets using two multi-label classifiers demonstrated that the best method of all measures was feature extraction method with label kernel function based on loss function and the performance of five evaluation measures on average increased by 10% ; especially on the data set Yeast, the evaluation measure Coverage reached a decline of about 30%. Closely followed by feature extraction method with label kernel function based on mutual information and the performance of five evaluation measures on average increased by 5%. The theoretical analysis and simulation results show that the feature extraction methods based on new output kernel functions can effectively extract features, simplify learning process of multi-label classifiers and, moreover, improve the performance of multi-label classification.
出处 《计算机应用》 CSCD 北大核心 2015年第7期1939-1944,1954,共7页 journal of Computer Applications
基金 国家自然科学基金资助项目(61272095) 山西省回国留学人员科研项目(2013-014) 河北省教育厅项目(Z2014106)
关键词 多标记学习 特征提取 核函数 损失函数 互信息 multi-label learning feature extraction kernel function loss function mutual information
  • 相关文献

参考文献18

  • 1UEDA N, SAITO K. Parametric mixture models for multi-labeled text [C]// Advances in Neural Information Processing Systems 15. Cambridge: MIT Press, 2003: 721-728.
  • 2KAZAWA H, IZUMITANI T, TAIRA H, et al. Maximal margin labeling for multi-topic text categorization [C]// Advances in Neural information Processing Systems 17. Cambridge: MIT Press, 2005: 649-656.
  • 3SCHAPIRE R.R, SINGER Y. Boostexter: a boosting-based system for text categorization [J]. Machine Learning, 2000, 39(2/3): 135-168.
  • 4ZHANG M, ZHOU Z. Multi-label neural networks with applications to functional genomics and text categorization [J]. IEEE Transactions on Knowledge and Data Engineering, 2006, 18(10): 1338-1351.
  • 5BOUTELL M R, LUO J, SHEN X, et al. Learning multi-label scene classification [J]. Pattern Recognition, 2004, 37(9): 1757-1771.
  • 6KANG F, JIN R, SUKTHANKAR R. Correlated label propagation with application to multi-label learning [C]// Proceeding of the 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition. Piscataway: IEEE, 2006: 1719-1726.
  • 7QI G, HUA X, RUI Y, et al. Correlative multi-label video annotation [C]// Proceedings of the 15th ACM International Conference on Multimedia. New York: ACM, 2007: 17-26.
  • 8ELISSEEFF A, WESTON J. A kernel method for multi-labelled classification [C]// Advances in Neural Information Processing Systems 14. Cambridge: MIT Press, 2002: 681-687.
  • 9BARUTCUOGLU Z, SCHAPIRE R E, TROYANSKAYA O G. Hierarchical multi-label prediction of gene function [J]. Bioinformatics, 2006, 22(7): 830-836.
  • 10ZHANG Y, ZHOU Z. Multi-label dimensionality reduction via dependence maximization [J]. ACM Transactions on Knowledge Discovery from Data, 2010, 4(3):14.

同被引文献30

引证文献4

二级引证文献17

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部