325

Variance Breakdown of Huber (M)-estimators: n/pm(1,)n/p \rightarrow m \in (1,\infty)

Abstract

A half century ago, Huber evaluated the minimax asymptotic variance in scalar location estimation, $ \min_\psi \max_{F \in {\cal F}_\epsilon} V(\psi, F) = \frac{1}{I(F_\epsilon^*)} $, where V(ψ,F)V(\psi,F) denotes the asymptotic variance of the (M)(M)-estimator for location with score function ψ\psi, and I(Fϵ)I(F_\epsilon^*) is the minimal Fisher information $ \min_{{\cal F}_\epsilon} I(F)$ over the class of ϵ\epsilon-Contaminated Normal distributions. We consider the linear regression model Y=Xθ0+WY = X\theta_0 + W, Wii.i.d.FW_i\sim_{\text{i.i.d.}}F, and iid Normal predictors Xi,jX_{i,j}, working in the high-dimensional-limit asymptotic where the number nn of observations and pp of variables both grow large, while n/pm(1,)n/p \rightarrow m \in (1,\infty); hence mm plays the role of `asymptotic number of observations per parameter estimated'. Let Vm(ψ,F)V_m(\psi,F) denote the per-coordinate asymptotic variance of the (M)(M)-estimator of regression in the n/pmn/p \rightarrow m regime. Then VmVV_m \neq V; however VmVV_m \rightarrow V as mm \rightarrow \infty. In this paper we evaluate the minimax asymptotic variance of the Huber (M)(M)-estimate. The statistician minimizes over the family (ψλ)λ>0(\psi_\lambda)_{\lambda > 0} of all tunings of Huber (M)(M)-estimates of regression, and Nature maximizes over gross-error contaminations FFϵF \in {\cal F}_\epsilon. Suppose that I(Fϵ)m>1I(F_\epsilon^*) \cdot m > 1. Then $ \min_\lambda \max_{F \in {\cal F}_\epsilon} V_m(\psi_\lambda, F) = \frac{1}{I(F_\epsilon^*) - 1/m} $. Strikingly, if I(Fϵ)m1I(F_\epsilon^*) \cdot m \leq 1, then the minimax asymptotic variance is ++\infty. The breakdown point is where the Fisher information per parameter equals unity.

View on arXiv
Comments on this paper