贝叶斯概率
计算机科学
先验概率
协变量
特征选择
变量(数学)
后验概率
数学优化
数据挖掘
机器学习
数学
人工智能
数学分析
作者
Veronika Ročková,Edward I. George
标识
DOI:10.1080/01621459.2013.869223
摘要
AbstractDespite rapid developments in stochastic search algorithms, the practicality of Bayesian variable selection methods has continued to pose challenges. High-dimensional data are now routinely analyzed, typically with many more covariates than observations. To broaden the applicability of Bayesian variable selection for such high-dimensional linear regression contexts, we propose EMVS, a deterministic alternative to stochastic search based on an EM algorithm which exploits a conjugate mixture prior formulation to quickly find posterior modes. Combining a spike-and-slab regularization diagram for the discovery of active predictor sets with subsequent rigorous evaluation of posterior model probabilities, EMVS rapidly identifies promising sparse high posterior probability submodels. External structural information such as likely covariate groupings or network topologies is easily incorporated into the EMVS framework. Deterministic annealing variants are seen to improve the effectiveness of our algorithms by mitigating the posterior multimodality associated with variable selection priors. The usefulness of the EMVS approach is demonstrated on real high-dimensional data, where computational complexity renders stochastic search to be less practical.KEY WORDS: Dynamic posterior explorationHigh dimensionalityRegularization plotsSparsitySSVS
科研通智能强力驱动
Strongly Powered by AbleSci AI