特征选择
计算机科学
特征(语言学)
人工智能
最小冗余特征选择
机器学习
光学(聚焦)
相关性(法律)
选择(遗传算法)
滤波器(信号处理)
集合(抽象数据类型)
数据挖掘
关系(数据库)
模式识别(心理学)
哲学
语言学
物理
法学
政治学
光学
计算机视觉
程序设计语言
作者
Ron Kohavi,George H. John
标识
DOI:10.1016/s0004-3702(97)00043-x
摘要
In the feature subset selection problem, a learning algorithm is faced with the problem of selecting a relevant subset of features upon which to focus its attention, while ignoring the rest. To achieve the best possible performance with a particular learning algorithm on a particular training set, a feature subset selection method should consider how the algorithm and the training set interact. We explore the relation between optimal feature subset selection and relevance. Our wrapper method searches for an optimal feature subset tailored to a particular algorithm and a domain. We study the strengths and weaknesses of the wrapper approach and show a series of improved designs. We compare the wrapper approach to induction without feature subset selection and to Relief, a filter approach to feature subset selection. Significant improvement in accuracy is achieved for some datasets for the two families of induction algorithms used: decision trees and Naive-Bayes.
科研通智能强力驱动
Strongly Powered by AbleSci AI