特征选择
特征(语言学)
人工智能
早熟收敛
计算机科学
人口
特征向量
约束(计算机辅助设计)
趋同(经济学)
最优化问题
模式识别(心理学)
进化算法
多目标优化
机器学习
选择(遗传算法)
数据挖掘
数学优化
数学
粒子群优化
几何学
哲学
社会学
人口学
经济
经济增长
语言学
作者
Ruwang Jiao,Bing Xue,Mengjie Zhang
标识
DOI:10.1109/tevc.2022.3215745
摘要
Reducing the number of selected features and improving the classification performance are two major objectives in feature selection, which can be viewed as a multi-objective optimization problem. Multi-objective feature selection in classification has its unique characteristics, such as it has a strong preference for the classification performance over the number of selected features. Besides, solution duplication often appears in both the search and the objective spaces, which degenerates the diversity and results in the premature convergence of the population. To deal with the above issues, in this paper, during the evolutionary training process, a multi-objective feature selection problem is reformulated and solved as a constrained multi-objective optimization problem, which adds a constraint on the classification performance for each solution (e.g., feature subset) according to the distribution of nondominated solutions, with the aim of selecting promising feature subsets that contain more informative and strongly relevant features, which are beneficial to improve the classification performance. Furthermore, based on the distribution of feature subsets in the objective space and their similarity in the search space, a duplication analysis and handling method is proposed to enhance the diversity of the population. Experimental results demonstrate that the proposed method outperforms six state-of-the-art algorithms and is computationally efficient on 18 classification datasets.
科研通智能强力驱动
Strongly Powered by AbleSci AI