特征选择
计算机科学
特征(语言学)
人工智能
最小冗余特征选择
机器学习
光学(聚焦)
选择(遗传算法)
滤波器(信号处理)
集合(抽象数据类型)
数据挖掘
关系(数据库)
模式识别(心理学)
信息增益
特征学习
决策树
训练集
统计分类
特征提取
计算复杂性理论
作者
Ron Kohavi,George H. John
标识
DOI:10.1016/s0004-3702(97)00043-x
摘要
In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a feature subset selection method should consider how the algorithm and the training set interact. We explore the relation between optimal feature subset selection and relevance. Our wrapper method searches for an optimal feature subset tailored to a particular algorithm and a domain. We study the strengths and weaknesses of the wrapper approach and show a series of improved designs. We compare the wrapper approach to induction without feature subset selection and to Relief, a filter approach to feature subset selection. Significant improvement in accuracy is achieved for some datasets for the two families of induction algorithms used: decision trees and Naive-Bayes.
科研通智能强力驱动
Strongly Powered by AbleSci AI