特征选择
可解释性
最小冗余特征选择
计算机科学
分类器(UML)
数据挖掘
特征(语言学)
人工智能
模式识别(心理学)
稳健性(进化)
理论(学习稳定性)
机器学习
算法
语言学
生物化学
基因
哲学
化学
作者
Utkarsh Mahadeo Khaire,R. Dhanalakshmi
标识
DOI:10.1016/j.jksuci.2019.06.012
摘要
Feature selection technique is a knowledge discovery tool which provides an understanding of the problem through the analysis of the most relevant features. Feature selection aims at building better classifier by listing significant features which also helps in reducing computational overload. Due to existing high throughput technologies and their recent advancements are resulting in high dimensional data due to which feature selection is being treated as handy and mandatory in such datasets. This actually questions the interpretability and stability of traditional feature selection algorithms. The high correlation in features frequently produces multiple equally optimal signatures, which makes traditional feature selection method unstable and thus leading to instability which reduces the confidence of selected features. Stability is the robustness of the feature preferences it produces to perturbation of training samples. Stability indicates the reproducibility power of the feature selection method. High stability of the feature selection algorithm is equally important as the high classification accuracy when evaluating feature selection performance. In this paper, we provide an overview of feature selection techniques and instability of the feature selection algorithm. We also present some of the solutions which can handle the different source of instability.
科研通智能强力驱动
Strongly Powered by AbleSci AI