期刊文献+

基于特征加权的朴素贝叶斯分类器 被引量:40

Naive Bayesian Classiflers Using Feature Weighting
下载PDF
导出
摘要 朴素贝叶斯分类器是一种广泛使用的分类算法,其计算效率和分类效果均十分理想。但是,由于其基础假设“朴素贝叶斯假设”与现实存在一定的差异,因此在某些数据上可能导致较差的分类结果。现在存在多种方法试图通过放松朴素贝叶斯假设来增强贝叶斯分类器的分类效果,但是通常会导致计算代价大幅提高。该文利用特征加权技术来增强朴素贝叶斯分类器。特征加权参数直接从数据导出,可以看作是计算某个类别的后验概率时,某个属性对于该计算的影响程度。数值实验表明,特征加权朴素贝叶斯分类器(FNVNB)的效果与其他的一些常用分类算法,例如树扩展朴素贝叶斯(TAN)和朴素贝叶斯树(NBTree)等的分类效果相当,其平均错误率都在17%左右;在计算速度上,FWNB接近于NB,比TAN和NBTree快至少一个数量级。 Naive Bayesian classifiers are widely used in machine learning due to their computational efficiency and competitive accuracy. However, their conditional attribute independence assumption can result in bad performance in real world problems. A number of techniques have explored the simple relaxations of the attribute independence assumption to increase accuracy, but always cost much more computing time. In this paper, we investigate enhancement of naive Bayes classifier using feature weighting technique. The feature weighting coefficients are directly induced from dataset, and can be regarded as the significance of each attribute when evaluating the posterior probability of the particular class value. Experiment results show that the new algorithm — Feature Weighting Naive Bayes(FWNB) can reach the same classification performance as state - of - the - art classifiers like TAN and NBTree, and all the mean error rate are around 18 percent, but the train time of FWNB is reduced at lear one quantitative level.
作者 程克非 张聪
出处 《计算机仿真》 CSCD 2006年第10期92-94,150,共4页 Computer Simulation
关键词 贝叶斯分类器 朴素贝叶斯分类器 特征加权 Bayesian classification Naive Bayesian classification Feature weighting
  • 相关文献

参考文献15

  • 1M Pankaj and W W Benjamin. Artificial neural networks:concepts and theory[M]. Los Alamitos, Calif. : IEEE Computer Society Press,1992.
  • 2J R Quinlan. C4.5: programs for machine learning[M]. San Mateo, Calif. : Morgan Kaufmann Publishers,1993.
  • 3R O Duda, P E Hart, D G Stork. Pattern Classification(2nd Edition) [ M]. Wiley - Interscience. 2000.
  • 4P Langley, W Iba and K Thompson. An analysis of bayesian classifiers[ C ]. In AAAI ' 90,1992. 223 - 228.
  • 5M Jaeger. Probabilistic Classifiers and the Concepts They Recognize[ C]. ICML 2003. 266 - 273
  • 6N Friedman, D Geiger and M Goldszmidt. Bayesian Network Classifiers[J]. Machine Learning, Kluwer Academic Publishers, Boston, 1997,29:131 - 163.
  • 7I Kononenko. Semi naive Bayesian classifier[C].Proceedings of European Conference on Artificial Intelligence,1991. 206 - 219.
  • 8P Langley and S Sage. Induction of selective Bayesian classifiers[C]. Proceedings of the Tenth Conference on Uncertainty in Artificial Intelligence. 1994, pp. 339 - 406.Seattle, WA: Morgan Kaufmann.
  • 9G I Webb & M J Pazzani. Adjusted probability naive Bayesian induction[C]. Proceedings of the Eleventh Australian Joint Conference on Artificial Intelligence. 1998, pp. 285 - 295.Berlin: Springer - Verlag.
  • 10R Kohavi. Scaling up the accuracy of naive - Bayes classifiers: A decision - tree hybrid[C]. Proceedings of the Second International Conference on Knowledge Discovery and Data Mining. 1996, pp. 202 -207. Menlo Park, CA: TheAAAI Press.

同被引文献264

引证文献40

二级引证文献204

相关作者

内容加载中请稍等...

相关机构

内容加载中请稍等...

相关主题

内容加载中请稍等...

浏览历史

内容加载中请稍等...
;
使用帮助 返回顶部