57
1

Classification Under Ambiguity: When Is Average-K Better Than Top-K?

Abstract

When many labels are possible, choosing a single one can lead to low precision. A common alternative, referred to as top-KK classification, is to choose some number KK (commonly around 5) and to return the KK labels with the highest scores. Unfortunately, for unambiguous cases, K>1K>1 is too many and, for very ambiguous cases, K5K \leq 5 (for example) can be too small. An alternative sensible strategy is to use an adaptive approach in which the number of labels returned varies as a function of the computed ambiguity, but must average to some particular KK over all the samples. We denote this alternative average-KK classification. This paper formally characterizes the ambiguity profile when average-KK classification can achieve a lower error rate than a fixed top-KK classification. Moreover, it provides natural estimation procedures for both the fixed-size and the adaptive classifier and proves their consistency. Finally, it reports experiments on real-world image data sets revealing the benefit of average-KK classification over top-KK in practice. Overall, when the ambiguity is known precisely, average-KK is never worse than top-KK, and, in our experiments, when it is estimated, this also holds.

View on arXiv
Comments on this paper