ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1810.10321
57
25
v1v2 (latest)

Active Ranking with Subset-wise Preferences

23 October 2018
Aadirupa Saha
Aditya Gopalan
ArXiv (abs)PDFHTML
Abstract

We consider the problem of probably approximately correct (PAC) ranking nnn items by adaptively eliciting subset-wise preference feedback. At each round, the learner chooses a subset of kkk items and observes stochastic feedback indicating preference information of the winner (most preferred) item of the chosen subset drawn according to a Plackett-Luce (PL) subset choice model unknown a priori. The objective is to identify an ϵ\epsilonϵ-optimal ranking of the nnn items with probability at least 1−δ1 - \delta1−δ. When the feedback in each subset round is a single Plackett-Luce-sampled item, we show (ϵ,δ)(\epsilon, \delta)(ϵ,δ)-PAC algorithms with a sample complexity of O(nϵ2ln⁡nδ)O\left(\frac{n}{\epsilon^2} \ln \frac{n}{\delta} \right)O(ϵ2n​lnδn​) rounds, which we establish as being order-optimal by exhibiting a matching sample complexity lower bound of Ω(nϵ2ln⁡nδ)\Omega\left(\frac{n}{\epsilon^2} \ln \frac{n}{\delta} \right)Ω(ϵ2n​lnδn​)---this shows that there is essentially no improvement possible from the pairwise comparisons setting (k=2k = 2k=2). When, however, it is possible to elicit top-mmm (≤k\leq k≤k) ranking feedback according to the PL model from each adaptively chosen subset of size kkk, we show that an (ϵ,δ)(\epsilon, \delta)(ϵ,δ)-PAC ranking sample complexity of O(nmϵ2ln⁡nδ)O\left(\frac{n}{m \epsilon^2} \ln \frac{n}{\delta} \right)O(mϵ2n​lnδn​) is achievable with explicit algorithms, which represents an mmm-wise reduction in sample complexity compared to the pairwise case. This again turns out to be order-wise unimprovable across the class of symmetric ranking algorithms. Our algorithms rely on a novel {pivot trick} to maintain only nnn itemwise score estimates, unlike O(n2)O(n^2)O(n2) pairwise score estimates that has been used in prior work. We report results of numerical experiments that corroborate our findings.

View on arXiv
Comments on this paper