ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2306.01922
13
1

Agnostic Multi-Group Active Learning

2 June 2023
Nick Rittler
Kamalika Chaudhuri
ArXivPDFHTML
Abstract

Inspired by the problem of improving classification accuracy on rare or hard subsets of a population, there has been recent interest in models of learning where the goal is to generalize to a collection of distributions, each representing a ``group''. We consider a variant of this problem from the perspective of active learning, where the learner is endowed with the power to decide which examples are labeled from each distribution in the collection, and the goal is to minimize the number of label queries while maintaining PAC-learning guarantees. Our main challenge is that standard active learning techniques such as disagreement-based active learning do not directly apply to the multi-group learning objective. We modify existing algorithms to provide a consistent active learning algorithm for an agnostic formulation of multi-group learning, which given a collection of GGG distributions and a hypothesis class H\mathcal{H}H with VC-dimension ddd, outputs an ϵ\epsilonϵ-optimal hypothesis using O~((ν2/ϵ2+1)GdθG2log⁡2(1/ϵ)+Glog⁡(1/ϵ)/ϵ2)\tilde{O}\left( (\nu^2/\epsilon^2+1) G d \theta_{\mathcal{G}}^2 \log^2(1/\epsilon) + G\log(1/\epsilon)/\epsilon^2 \right)O~((ν2/ϵ2+1)GdθG2​log2(1/ϵ)+Glog(1/ϵ)/ϵ2) label queries, where θG\theta_{\mathcal{G}}θG​ is the worst-case disagreement coefficient over the collection. Roughly speaking, this guarantee improves upon the label complexity of standard multi-group learning in regimes where disagreement-based active learning algorithms may be expected to succeed, and the number of groups is not too large. We also consider the special case where each distribution in the collection is individually realizable with respect to H\mathcal{H}H, and demonstrate O~(GdθGlog⁡(1/ϵ))\tilde{O}\left( G d \theta_{\mathcal{G}} \log(1/\epsilon) \right)O~(GdθG​log(1/ϵ)) label queries are sufficient for learning in this case. We further give an approximation result for the full agnostic case inspired by the group realizable strategy.

View on arXiv
Comments on this paper