ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.02716
7
13

Approximate Leave-One-Out for High-Dimensional Non-Differentiable Learning Problems

4 October 2018
Shuaiwen Wang
Wenda Zhou
A. Maleki
Haihao Lu
Vahab Mirrokni
ArXivPDFHTML
Abstract

Consider the following class of learning schemes: \begin{equation} \label{eq:main-problem1} \hat{\boldsymbol{\beta}} := \underset{\boldsymbol{\beta} \in \mathcal{C}}{\arg\min} \;\sum_{j=1}^n \ell(\boldsymbol{x}_j^\top\boldsymbol{\beta}; y_j) + \lambda R(\boldsymbol{\beta}), \qquad \qquad \qquad (1) \end{equation} where xi∈Rp\boldsymbol{x}_i \in \mathbb{R}^pxi​∈Rp and yi∈Ry_i \in \mathbb{R}yi​∈R denote the ithi^{\rm th}ith feature and response variable respectively. Let ℓ\ellℓ and RRR be the convex loss function and regularizer, β\boldsymbol{\beta}β denote the unknown weights, and λ\lambdaλ be a regularization parameter. C⊂Rp\mathcal{C} \subset \mathbb{R}^{p}C⊂Rp is a closed convex set. Finding the optimal choice of λ\lambdaλ is a challenging problem in high-dimensional regimes where both nnn and ppp are large. We propose three frameworks to obtain a computationally efficient approximation of the leave-one-out cross validation (LOOCV) risk for nonsmooth losses and regularizers. Our three frameworks are based on the primal, dual, and proximal formulations of (1). Each framework shows its strength in certain types of problems. We prove the equivalence of the three approaches under smoothness conditions. This equivalence enables us to justify the accuracy of the three methods under such conditions. We use our approaches to obtain a risk estimate for several standard problems, including generalized LASSO, nuclear norm regularization, and support vector machines. We empirically demonstrate the effectiveness of our results for non-differentiable cases.

View on arXiv
Comments on this paper