朴素贝叶斯分类器
特征(语言学)
加权
计算机科学
人工智能
Bayes错误率
贝叶斯定理
机器学习
水准点(测量)
模式识别(心理学)
条件独立性
数据挖掘
贝叶斯分类器
贝叶斯概率
支持向量机
医学
哲学
语言学
大地测量学
放射科
地理
作者
Liangxiao Jiang,Chaoqun Li,Shasha Wang,Lungan Zhang
标识
DOI:10.1016/j.engappai.2016.02.002
摘要
Naive Bayes (NB) continues to be one of the top 10 data mining algorithms due to its simplicity, efficiency and efficacy. Of numerous proposals to improve the accuracy of naive Bayes by weakening its feature independence assumption, the feature weighting approach has received less attention from researchers. Moreover, to our knowledge, all of the existing feature weighting approaches only incorporate the learned feature weights into the classification of formula of naive Bayes and do not incorporate the learned feature weights into its conditional probability estimates at all. In this paper, we propose a simple, efficient, and effective feature weighting approach, called deep feature weighting (DFW), which estimates the conditional probabilities of naive Bayes by deeply computing feature weighted frequencies from training data. Empirical studies on a collection of 36 benchmark datasets from the UCI repository show that naive Bayes with deep feature weighting rarely degrades the quality of the model compared to standard naive Bayes and, in many cases, improves it dramatically. Besides, we apply the proposed deep feature weighting to some state-of-the-art naive Bayes text classifiers and have achieved remarkable improvements.
科研通智能强力驱动
Strongly Powered by AbleSci AI