ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2205.10927
16
6

Fast ABC-Boost: A Unified Framework for Selecting the Base Class in Multi-Class Classification

22 May 2022
Ping Li
Weijie Zhao
ArXivPDFHTML
Abstract

The work in ICML'09 showed that the derivatives of the classical multi-class logistic regression loss function could be re-written in terms of a pre-chosen "base class" and applied the new derivatives in the popular boosting framework. In order to make use of the new derivatives, one must have a strategy to identify/choose the base class at each boosting iteration. The idea of "adaptive base class boost" (ABC-Boost) in ICML'09, adopted a computationally expensive "exhaustive search" strategy for the base class at each iteration. It has been well demonstrated that ABC-Boost, when integrated with trees, can achieve substantial improvements in many multi-class classification tasks. Furthermore, the work in UAI'10 derived the explicit second-order tree split gain formula which typically improved the classification accuracy considerably, compared with using only the fist-order information for tree-splitting, for both multi-class and binary-class classification tasks. In this paper, we develop a unified framework for effectively selecting the base class by introducing a series of ideas to improve the computational efficiency of ABC-Boost. Our framework has parameters (s,g,w)(s,g,w)(s,g,w). At each boosting iteration, we only search for the "sss-worst classes" (instead of all classes) to determine the base class. We also allow a "gap" ggg when conducting the search. That is, we only search for the base class at every g+1g+1g+1 iterations. We furthermore allow a "warm up" stage by only starting the search after www boosting iterations. The parameters sss, ggg, www, can be viewed as tunable parameters and certain combinations of (s,g,w)(s,g,w)(s,g,w) may even lead to better test accuracy than the "exhaustive search" strategy. Overall, our proposed framework provides a robust and reliable scheme for implementing ABC-Boost in practice.

View on arXiv
Comments on this paper