特征选择
线性子空间
计算机科学
模式识别(心理学)
子空间拓扑
特征(语言学)
维数之咒
降维
特征向量
人工智能
预处理器
k-最近邻算法
重量
随机子空间法
数据挖掘
数学
语言学
李代数
哲学
纯数学
几何学
作者
Baoshuang Zhang,Yanying Li,Zheng Chai
标识
DOI:10.1016/j.knosys.2022.109400
摘要
Feature selection is an important preprocessing technology for dimensionality reduction, which reduces the dimension of the dataset by acquiring a subset of features with the largest amount of information, and improves the classification accuracy to the greatest extent at the same time. Although different types of feature selection algorithms have achieved remarkable success, most of them lack the ability to mine information in different subspaces, and ignore the useful information contained in the abundant samples. In this research, a novel random multi-subspace based ReliefF (RBEFF) is proposed for feature selection. In this method, firstly, multiple feature partitions containing a large number of random subspaces with the same size are generated. Secondly, the ReliefF algorithm is used in each random subspace to obtain the local weight of the feature. The local weight vectors of random subspaces in each feature partition are combined to obtain the full weight vector. Finally, the full weight vectors of multiple feature partitions are integrated into the final weight vector, which contains the final weight of each feature in the original feature space feature. The feature selection is carried out dynamically according to the final weight vector. We evaluated the performance of the RBEFF on 28 UCI datasets with different sizes and compare RBEFF with 6 feature selection algorithms using KNN and DT classifiers’ three evaluation indicators. The comparisons and experimental results demonstrate the effectiveness, competitiveness, and superiority of RBEFF in solving feature selection problems.
科研通智能强力驱动
Strongly Powered by AbleSci AI