计算机科学
特征选择
人工智能
机器学习
特征(语言学)
标记数据
半监督学习
模式识别(心理学)
约束(计算机辅助设计)
对偶(语法数字)
规范(哲学)
数学
艺术
法学
几何学
哲学
文学类
语言学
政治学
作者
Han Zhang,Maoguo Gong,Feiping Nie,Xuelong Li
标识
DOI:10.1016/j.neucom.2022.05.090
摘要
Semi-supervised feature selection alleviates the annotation burden of supervised feature learning by exploiting data under a handful of supervision information. The mainstream technique is to employ a linear regression framework that jointly learns labeled and unlabeled samples. However, existing approaches always encounter the deficiencies in two aspects: 1) the performance of models are severely degenerated once predicted labels are unreliable; 2) the balance of objectives in regards to two types of data are not well considered. In the article, we propose unified dual-label semi-supervised learning for top-k feature selection. The technique defines a soft label matrix to indicate the probability of samples belonging to each class. From the probability, the model could recognize unclassifiable samples that lay around the boundaries. Meanwhile, the label matrix is equipped with an exponent parameter γ. It endows the soft labels dual effects that the labeled and unlabeled data are tactfully discriminated. For the purpose of feature selection, we impose the ℓ2,0-norm constraint on the projection matrix, such that the exact top-k features are picked out. An iteration algorithm is designed to solve the given problem, by which large-scale data are facilely tackled. We conduct experiments that validate the superiority of the proposed method against the state-of-the-art competitors.
科研通智能强力驱动
Strongly Powered by AbleSci AI