冗余(工程)
判别式
特征选择
计算机科学
人工智能
回归
模式识别(心理学)
子空间拓扑
特征提取
降维
最小冗余特征选择
缩小
特征(语言学)
数据挖掘
机器学习
数学
哲学
操作系统
统计
语言学
程序设计语言
作者
Xueyuan Xu,Xia Wu,Fulin Wei,Wei Zhong,Feiping Nie
出处
期刊:IEEE Transactions on Knowledge and Data Engineering
[Institute of Electrical and Electronics Engineers]
日期:2021-02-17
卷期号:34 (11): 5056-5069
被引量:28
标识
DOI:10.1109/tkde.2021.3059523
摘要
Feature selection has attracted a lot of attention in obtaining discriminative and non-redundant features from high-dimension data. Compared with traditional filter and wrapper methods, embedded methods can obtain a more informative feature subset by fully considering the importance of features in the classification tasks. However, the existing embedded methods emphasize the above importance of features and mostly ignore the correlation between the features, which leads to retain the correlated and redundant features with similar scores in the feature subset. To solve the problem, we propose a novel supervised embedded feature selection framework, called feature selection under global redundancy minimization in orthogonal regression (GRMOR). The proposed framework can effectively recognize redundant features from a global view of redundancy among the features. We also incorporate the large margin constraint into GRMOR for robust multi-class classification. Compared with the traditional embedded methods based on least square regression, the proposed framework utilizes orthogonal regression to preserve more discriminative information in the subspace, which can help accurately rank the importance of features in the classification tasks. Experimental results on twelve public datasets demonstrate that the proposed framework can obtain superior classification performance and redundancy removal performance than twelve other feature selection methods.
科研通智能强力驱动
Strongly Powered by AbleSci AI