23
0

Choosing the pp in LpL_p loss: rate adaptivity on the symmetric location problem

Abstract

Given univariate random variables Y1,,YnY_1, \ldots, Y_n with the Uniform(θ01,θ0+1)\text{Uniform}(\theta_0 - 1, \theta_0 + 1) distribution, the sample midrange Y(n)+Y(1)2\frac{Y_{(n)}+Y_{(1)}}{2} is the MLE for θ0\theta_0 and estimates θ0\theta_0 with error of order 1/n1/n, which is much smaller compared with the 1/n1/\sqrt{n} error rate of the usual sample mean estimator. However, the sample midrange performs poorly when the data has say the Gaussian N(θ0,1)N(\theta_0, 1) distribution, with an error rate of 1/logn1/\sqrt{\log n}. In this paper, we propose an estimator of the location θ0\theta_0 with a rate of convergence that can, in many settings, adapt to the underlying distribution which we assume to be symmetric around θ0\theta_0 but is otherwise unknown. When the underlying distribution is compactly supported, we show that our estimator attains a rate of convergence of n1αn^{-\frac{1}{\alpha}} up to polylog factors, where the rate parameter α\alpha can take on any value in (0,2](0, 2] and depends on the moments of the underlying distribution. Our estimator is formed by the γ\ell^\gamma-center of the data, for a γ2\gamma\geq2 chosen in a data-driven way -- by minimizing a criterion motivated by the asymptotic variance. Our approach can be directly applied to the regression setting where θ0\theta_0 is a function of observed features and motivates the use of γ\ell^\gamma loss function for γ>2\gamma > 2 in certain settings.

View on arXiv
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.