ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.23409
39
0

LIRA: A Learning-based Query-aware Partition Framework for Large-scale ANN Search

30 March 2025
Ximu Zeng
Liwei Deng
Penghao Chen
Xu Chen
Han Su
Kai Zheng
ArXivPDFHTML
Abstract

Approximate nearest neighbor search is fundamental in information retrieval. Previous partition-based methods enhance search efficiency by probing partial partitions, yet they face two common issues. In the query phase, a common strategy is to probe partitions based on the distance ranks of a query to partition centroids, which inevitably probes irrelevant partitions as it ignores data distribution. In the partition construction phase, all partition-based methods face the boundary problem that separates a query's nearest neighbors to multiple partitions, resulting in a long-tailed kNN distribution and degrading the optimal nprobe (i.e., the number of probing partitions). To address this gap, we propose LIRA, a LearnIng-based queRy-aware pArtition framework. Specifically, we propose a probing model to directly probe the partitions containing the kNN of a query, which can reduce probing waste and allow for query-aware probing with nprobe individually. Moreover, we incorporate the probing model into a learning-based redundancy strategy to mitigate the adverse impact of the long-tailed kNN distribution on search efficiency. Extensive experiments on real-world vector datasets demonstrate the superiority of LIRA in the trade-off among accuracy, latency, and query fan-out. The codes are available atthis https URL.

View on arXiv
@article{zeng2025_2503.23409,
  title={ LIRA: A Learning-based Query-aware Partition Framework for Large-scale ANN Search },
  author={ Ximu Zeng and Liwei Deng and Penghao Chen and Xu Chen and Han Su and Kai Zheng },
  journal={arXiv preprint arXiv:2503.23409},
  year={ 2025 }
}
Comments on this paper