Estimating the error variance in a high-dimensional linear model

View Researcher's Other Codes

Disclaimer: The provided code links for this paper are external links. Science Nest has no responsibility for the accuracy, legality or content of these links. Also, by downloading this code(s), you agree to comply with the terms of use as set out by the author(s) of the code(s).

Authors Guo Yu, Jacob Bien
Journal/Conference Name Submitted
Paper Category
Paper Abstract The lasso has been studied extensively as a tool for estimating the coefficient vector in the high-dimensional linear model; however, considerably less is known about estimating the error variance. Indeed, most well-known theoretical properties of the lasso, including recent advances in selective inference with the lasso, are established under the assumption that the underlying error variance is known. Yet the error variance in practice is, of course, unknown. In this paper, we propose the natural lasso estimator for the error variance, which maximizes a penalized likelihood objective. A key aspect of the natural lasso is that the likelihood is expressed in terms of the natural parameterization of the multiparameter exponential family of a Gaussian with unknown mean and variance. The result is a remarkably simple estimator with provably good performance in terms of mean squared error. These theoretical results do not require placing any assumptions on the design matrix or the true regression coefficients. We also propose a companion estimator, called the organic lasso, which theoretically does not require tuning of the regularization parameter. Both estimators do well compared to preexisting methods, especially in settings where successful recovery of the true support of the coefficient vector is hard.
Date of publication 2017
Code Programming Language R

Copyright Researcher 2021