TY - JOUR
T1 - Bayesian empirical likelihood for ridge and lasso regressions
AU - Bedoui, Adel
AU - Lazar, Nicole A.
N1 - Funding Information:
Nicole A. Lazar was supported in part by the National Science Foundation under grant NSF IIS-1607919 .
Publisher Copyright:
© 2020 Elsevier B.V.
PY - 2020/5
Y1 - 2020/5
N2 - Ridge and lasso regression models, which are also known as regularization methods, are widely used methods in machine learning and inverse problems that introduce additional information to solve ill-posed problems and/or perform feature selection. The ridge and lasso estimates for linear regression parameters can be interpreted as Bayesian posterior estimates when the regression parameters have Normal and independent Laplace (i.e., double-exponential) priors, respectively. A significant challenge in regularization problems is that these approaches assume that data are normally distributed, which makes them not robust to model misspecification. A Bayesian approach for ridge and lasso models based on empirical likelihood is proposed. This method is semiparametric because it combines a nonparametric model and a parametric model. Hence, problems with model misspecification are avoided. Under the Bayesian empirical likelihood approach, the resulting posterior distribution lacks a closed form and has a nonconvex support, which makes the implementation of traditional Markov chain Monte Carlo (MCMC) methods such as Gibbs sampling and Metropolis–Hastings very challenging. To solve the nonconvex optimization and nonconvergence problems, the tailored Metropolis–Hastings approach is implemented. The asymptotic Bayesian credible intervals are derived.
AB - Ridge and lasso regression models, which are also known as regularization methods, are widely used methods in machine learning and inverse problems that introduce additional information to solve ill-posed problems and/or perform feature selection. The ridge and lasso estimates for linear regression parameters can be interpreted as Bayesian posterior estimates when the regression parameters have Normal and independent Laplace (i.e., double-exponential) priors, respectively. A significant challenge in regularization problems is that these approaches assume that data are normally distributed, which makes them not robust to model misspecification. A Bayesian approach for ridge and lasso models based on empirical likelihood is proposed. This method is semiparametric because it combines a nonparametric model and a parametric model. Hence, problems with model misspecification are avoided. Under the Bayesian empirical likelihood approach, the resulting posterior distribution lacks a closed form and has a nonconvex support, which makes the implementation of traditional Markov chain Monte Carlo (MCMC) methods such as Gibbs sampling and Metropolis–Hastings very challenging. To solve the nonconvex optimization and nonconvergence problems, the tailored Metropolis–Hastings approach is implemented. The asymptotic Bayesian credible intervals are derived.
UR - http://www.scopus.com/inward/record.url?scp=85078116384&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85078116384&partnerID=8YFLogxK
U2 - 10.1016/j.csda.2020.106917
DO - 10.1016/j.csda.2020.106917
M3 - Article
AN - SCOPUS:85078116384
SN - 0167-9473
VL - 145
JO - Computational Statistics and Data Analysis
JF - Computational Statistics and Data Analysis
M1 - 106917
ER -