424

On Regression in Extreme Regions

Electronic Journal of Statistics (EJS), 2023
Main:13 Pages
1 Figures
Bibliography:1 Pages
Appendix:32 Pages
Abstract

In the classic regression problem, the value of a real-valued random variable YY is to be predicted based on the observation of a random vector XX, taking its values in Rd\mathbb{R}^d with d1d\geq 1 say. The statistical learning problem consists in building a predictive function f^:RdR\hat{f}:\mathbb{R}^d\to \mathbb{R} based on independent copies of the pair (X,Y)(X,Y) so that YY is approximated by f^(X)\hat{f}(X) with minimum error in the mean-squared sense. Motivated by various applications, ranging from environmental sciences to finance or insurance, special attention is paid here to the case of extreme (i.e. very large) observations XX. Because of their rarity, they contribute in a negligible manner to the (empirical) error and the predictive performance of empirical quadratic risk minimizers can be consequently very poor in extreme regions. In this paper, we develop a general framework for regression in the extremes. It is assumed that XX's conditional distribution given YY belongs to a non parametric class of heavy-tailed probability distributions. It is then shown that an asymptotic notion of risk can be tailored to summarize appropriately predictive performance in extreme regions of the input space. It is also proved that minimization of an empirical and non asymptotic version of this 'extreme risk', based on a fraction of the largest observations solely, yields regression functions with good generalization capacity. In addition, numerical results providing strong empirical evidence of the relevance of the approach proposed are displayed.

View on arXiv
Comments on this paper