ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1906.02314
15
27

A Tunable Loss Function for Robust Classification: Calibration, Landscape, and Generalization

5 June 2019
Tyler Sypherd
Mario Díaz
J. Cava
Gautam Dasarathy
Peter Kairouz
Lalitha Sankar
ArXivPDFHTML
Abstract

We introduce a tunable loss function called α\alphaα-loss, parameterized by α∈(0,∞]\alpha \in (0,\infty]α∈(0,∞], which interpolates between the exponential loss (α=1/2\alpha = 1/2α=1/2), the log-loss (α=1\alpha = 1α=1), and the 0-1 loss (α=∞\alpha = \inftyα=∞), for the machine learning setting of classification. Theoretically, we illustrate a fundamental connection between α\alphaα-loss and Arimoto conditional entropy, verify the classification-calibration of α\alphaα-loss in order to demonstrate asymptotic optimality via Rademacher complexity generalization techniques, and build-upon a notion called strictly local quasi-convexity in order to quantitatively characterize the optimization landscape of α\alphaα-loss. Practically, we perform class imbalance, robustness, and classification experiments on benchmark image datasets using convolutional-neural-networks. Our main practical conclusion is that certain tasks may benefit from tuning α\alphaα-loss away from log-loss (α=1\alpha = 1α=1), and to this end we provide simple heuristics for the practitioner. In particular, navigating the α\alphaα hyperparameter can readily provide superior model robustness to label flips (α>1\alpha > 1α>1) and sensitivity to imbalanced classes (α<1\alpha < 1α<1).

View on arXiv
Comments on this paper