特征选择
人工智能
计算机科学
机器学习
模式识别(心理学)
特征(语言学)
降维
最小冗余特征选择
特征学习
粒度
维数之咒
稳健性(进化)
多标签分类
数据挖掘
基因
操作系统
哲学
生物化学
语言学
化学
作者
Wenbin Qian,Yinsong Xiong,Jun Yang,Wenhao Shu
标识
DOI:10.1016/j.ins.2021.08.076
摘要
Feature selection plays a crucial role in machine learning and data mining, and improves the performance of learning models by selecting a distinguishing feature subset and eliminating irrelevant features. Existing feature selection methods are mainly used for single-label learning and multi-label learning; however, there are only a few feature selection methods for label distribution learning. Label distribution learning has the “curse of dimensionality” problem, similar to that in multi-label learning. In label distribution learning, the related labels of each sample have different levels of importance. Therefore, multi-label feature selection algorithms can not be directly applied to label distribution data, and discretizing the label distribution data into multi-label data would result in the loss of some important supervised information. To solve this problem, a novel feature selection algorithm for label distribution learning is proposed in this paper. The proposed method utilizes neighborhood granularity to explore feature similarity, and it uses a correlation coefficient to generate the label correlations. In addition, sparse learning is used to improve the robustness and control complexity. Experimental results indicate that our proposed method is more effective than five state-of-art feature selection algorithms on twelve datasets, with respect to six representative evaluation measures.
科研通智能强力驱动
Strongly Powered by AbleSci AI