ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2402.09327
25
6

Information Complexity of Stochastic Convex Optimization: Applications to Generalization and Memorization

14 February 2024
Idan Attias
Gintare Karolina Dziugaite
Mahdi Haghifam
Roi Livni
Daniel M. Roy
ArXivPDFHTML
Abstract

In this work, we investigate the interplay between memorization and learning in the context of \emph{stochastic convex optimization} (SCO). We define memorization via the information a learning algorithm reveals about its training data points. We then quantify this information using the framework of conditional mutual information (CMI) proposed by Steinke and Zakynthinou (2020). Our main result is a precise characterization of the tradeoff between the accuracy of a learning algorithm and its CMI, answering an open question posed by Livni (2023). We show that, in the L2L^2L2 Lipschitz--bounded setting and under strong convexity, every learner with an excess error ε\varepsilonε has CMI bounded below by Ω(1/ε2)\Omega(1/\varepsilon^2)Ω(1/ε2) and Ω(1/ε)\Omega(1/\varepsilon)Ω(1/ε), respectively. We further demonstrate the essential role of memorization in learning problems in SCO by designing an adversary capable of accurately identifying a significant fraction of the training samples in specific SCO problems. Finally, we enumerate several implications of our results, such as a limitation of generalization bounds based on CMI and the incompressibility of samples in SCO problems.

View on arXiv
Comments on this paper