惩罚法
弹性网正则化
估计员
Lasso(编程语言)
离群值
数学
数学优化
特征选择
背景(考古学)
回归
稳健回归
计算机科学
统计
人工智能
古生物学
万维网
生物
作者
Sophie Lambert‐Lacroix,Laurent Zwald
标识
DOI:10.1080/10485252.2016.1190359
摘要
We intend to combine Huber's loss with an adaptive reversed version as a penalty function. The purpose is twofold: first we would like to propose an estimator that is robust to data subject to heavy-tailed errors or outliers. Second we hope to overcome the variable selection problem in the presence of highly correlated predictors. For instance, in this framework, the adaptive least absolute shrinkage and selection operator (lasso) is not a very satisfactory variable selection method, although it is a popular technique for simultaneous estimation and variable selection. We call this new penalty the adaptive BerHu penalty. As for elastic net penalty, small coefficients contribute through their norm to this penalty while larger coefficients cause it to grow quadratically (as ridge regression). We will show that the estimator associated with Huber's loss combined with the adaptive BerHu penalty enjoys theoretical properties in the fixed design context. This approach is compared to existing regularisation methods such as adaptive elastic net and is illustrated via simulation studies and real data.
科研通智能强力驱动
Strongly Powered by AbleSci AI