期刊文献+

Dimensionality Reduction by Mutual Information for Text Classification 被引量:2

Dimensionality Reduction by Mutual Information for Text Classification
下载PDF
导出
摘要 The frame of text classification system was presented. The high dimensionality in feature space for text classification was studied. The mutual information is a widely used information theoretic measure, in a descriptive way, to measure the stochastic dependency of discrete random variables. The measure method was used as a criterion to reduce high dimensionality of feature vectors in text classification on Web. Feature selections or conversions were performed by using maximum mutual information including linear and non-linear feature conversions. Entropy was used and extended to find right features commendably in pattern recognition systems. Favorable foundation would be established for text classification mining. The frame of text classification system was presented. The high dimensionality in feature space for text classification was studied. The mutual information is a widely used information theoretic measure, in a descriptive way, to measure the stochastic dependency of discrete random variables. The measure method was used as a criterion to reduce high dimensionality of feature vectors in text classification on Web. Feature selections or conversions were performed by using maximum mutual information including linear and non-linear feature conversions. Entropy was used and extended to find right features commendably in pattern recognition systems. Favorable foundation would be established for text classification mining.
出处 《Journal of Beijing Institute of Technology》 EI CAS 2005年第1期32-36,共5页 北京理工大学学报(英文版)
基金 theNational"973"ProgramProjects(G1998030414)
关键词 text classification mutual information dimensionality reduction text classification mutual information dimensionality reduction
  • 相关文献

参考文献9

  • 1TorkkolaK,WilliamM.Campbellmutualinformationin learningfeaturetransformations[].thInternational ConfonMachineLearning.2000
  • 2TorkkolaK.Nonlinearfeaturetransformsusingmaxi mummutualinformation[].ConfoftheIJCNN.2001
  • 3KarypisG,HanEH.Conceptindexing:Afastdimen sionalityreductionalgorithmwithapplicationstodocu mentretrievalandclassification[].thInt’lConfIn formationandKnowledgeManagementCIKM.2000
  • 4HutterM.Distributionofmutualinformation[].th InternationalConferenceonNeuralInformationProcess ingSystems.2001
  • 5YangY,PedersenJO.Acomparativestudyonfeature selectionintextclassification[].KDD SixthACM SIGKDDInternationalConferenceonKnowledgeDiscov eryandDataMining.2000
  • 6TorkkolaK.Onfeatureextractionbymutualinformation maximization. http:∥citeseer.nj.nec.com/ torkkola02feature.htm .
  • 7JongzhiS.Knowledgediscovery[]..2002
  • 8GalavottiL,SebastianiF,SimiM.Featureselectionand negativeevidenceinautomatedtextclassification[].KDD SixthACMSIGKDDInternationConference onKnowledgeDiscoveryandDataMining.2000
  • 9ZaffalonM,HutterM.Robustfeatureselectionbymu tualinformationdistributions. http:∥www. idsia.ch/~marcus/ai/feature.htm .

同被引文献17

引证文献2

二级引证文献8

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部