Abstract:
|
Consider the multiple linear regression model Y = ?1 + ?2x2 +···+ ?p x p + e = xT ?+e with sample size n. This paper compares the six shrinkage estimators: forward selection, lasso, partial least squares, principal components regression, lasso variable selection, and ridge regression, with large sample theory and two new prediction intervals that are asymptotically optimal if the estimator ?ˆ is a consistent estimator of ?. Few prediction intervals have been developed for p > n, and they are not asymptotically optimal. For p fixed, the large sample theory for variable selection estimators like forward selection is new, and the theory shows that lasso variable selection is ?n consistent under much milder conditions than lasso. This paper also simplifies the proofs of the large sample theory for lasso, ridge regression, and elastic net.
|