Skip to yearly menu bar Skip to main content


Poster

A Residual Bootstrap for High-Dimensional Regression with Near Low-Rank Designs

Miles Lopes

Level 2, room 210D

Abstract: We study the residual bootstrap (RB) method in the context of high-dimensional linear regression. Specifically, we analyze the distributional approximation of linear contrasts c(β^ρβ), where β^ρ is a ridge-regression estimator. When regression coefficients are estimated via least squares, classical results show that RB consistently approximates the laws of contrasts, provided that pn, where the design matrix is of size n×p. Up to now, relatively little work has considered how additional structure in the linear model may extend the validity of RB to the setting where p/n1. In this setting, we propose a version of RB that resamples residuals obtained from ridge regression. Our main structural assumption on the design matrix is that it is nearly low rank --- in the sense that its singular values decay according to a power-law profile. Under a few extra technical assumptions, we derive a simple criterion for ensuring that RB consistently approximates the law of a given contrast. We then specialize this result to study confidence intervals for mean response values Xiβ, where Xi is the ith row of the design. More precisely, we show that conditionally on a Gaussian design with near low-rank structure, RB \emph{simultaneously} approximates all of the laws Xi(β^ρβ), i=1,,n. This result is also notable as it imposes no sparsity assumptions on β. Furthermore, since our consistency results are formulated in terms of the Mallows (Kantorovich) metric, the existence of a limiting distribution is not required.

Live content is unavailable. Log in and register to view live content