特征选择
维数之咒
计算机科学
成对比较
最小冗余特征选择
预处理器
冗余(工程)
人工智能
数据预处理
降维
滤波器(信号处理)
相关性
数据挖掘
模式识别(心理学)
高维数据聚类
机器学习
特征(语言学)
选择(遗传算法)
数学
聚类分析
操作系统
哲学
语言学
计算机视觉
几何学
出处
期刊:International Conference on Machine Learning
日期:2003-08-21
卷期号:: 856-863
被引量:2065
摘要
Feature selection, as a preprocessing step to machine learning, is effective in reducing dimensionality, removing irrelevant data, increasing learning accuracy, and improving result comprehensibility. However, the recent increase of dimensionality of data poses a severe challenge to many existing feature selection methods with respect to efficiency and effectiveness. In this work, we introduce a novel concept, predominant correlation, and propose a fast filter method which can identify relevant features as well as redundancy among relevant features without pairwise correlation analysis. The efficiency and effectiveness of our method is demonstrated through extensive comparisons with other methods using real-world data of high dimensionality
科研通智能强力驱动
Strongly Powered by AbleSci AI