ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2301.13304
33
13

Understanding Self-Distillation in the Presence of Label Noise

30 January 2023
Rudrajit Das
Sujay Sanghavi
ArXivPDFHTML
Abstract

Self-distillation (SD) is the process of first training a \enquote{teacher} model and then using its predictions to train a \enquote{student} model with the \textit{same} architecture. Specifically, the student's objective function is (ξ∗ℓ(teacher’s predictions, student’s predictions)+(1−ξ)∗ℓ(given labels, student’s predictions))\big(\xi*\ell(\text{teacher's predictions}, \text{ student's predictions}) + (1-\xi)*\ell(\text{given labels}, \text{ student's predictions})\big)(ξ∗ℓ(teacher’s predictions, student’s predictions)+(1−ξ)∗ℓ(given labels, student’s predictions)), where ℓ\ellℓ is some loss function and ξ\xiξ is some parameter ∈[0,1]\in [0,1]∈[0,1]. Empirically, SD has been observed to provide performance gains in several settings. In this paper, we theoretically characterize the effect of SD in two supervised learning problems with \textit{noisy labels}. We first analyze SD for regularized linear regression and show that in the high label noise regime, the optimal value of ξ\xiξ that minimizes the expected error in estimating the ground truth parameter is surprisingly greater than 1. Empirically, we show that ξ>1\xi > 1ξ>1 works better than ξ≤1\xi \leq 1ξ≤1 even with the cross-entropy loss for several classification datasets when 50\% or 30\% of the labels are corrupted. Further, we quantify when optimal SD is better than optimal regularization. Next, we analyze SD in the case of logistic regression for binary classification with random label corruption and quantify the range of label corruption in which the student outperforms the teacher in terms of accuracy. To our knowledge, this is the first result of its kind for the cross-entropy loss.

View on arXiv
Comments on this paper