ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.09944
74
0

Query-Focused Retrieval Heads Improve Long-Context Reasoning and Re-ranking

11 June 2025
Wuwei Zhang
Fangcong Yin
Howard Yen
Danqi Chen
Xi Ye
    LRM
ArXiv (abs)PDFHTML
Main:5 Pages
4 Figures
Bibliography:1 Pages
6 Tables
Appendix:8 Pages
Abstract

Recent work has identified retrieval heads (Wu et al., 2025b), a subset of attention heads responsible for retrieving salient information in long-context language models (LMs), as measured by their copy-paste behavior in Needle-in-a-Haystack tasks. In this paper, we introduce QRHEAD (Query-Focused Retrieval Head), an improved set of attention heads that enhance retrieval from long context. We identify QRHEAD by aggregating attention scores with respect to the input query, using a handful of examples from real-world tasks (e.g., long-context QA). We further introduce QR- RETRIEVER, an efficient and effective retriever that uses the accumulated attention mass of QRHEAD as retrieval scores. We use QR- RETRIEVER for long-context reasoning by selecting the most relevant parts with the highest retrieval scores. On multi-hop reasoning tasks LongMemEval and CLIPPER, this yields over 10% performance gains over full context and outperforms strong dense retrievers. We also evaluate QRRETRIEVER as a re-ranker on the BEIR benchmark and find that it achieves strong zero-shot performance, outperforming other LLM-based re-rankers such as RankGPT. Further analysis shows that both the querycontext attention scoring and task selection are crucial for identifying QRHEAD with strong downstream utility. Overall, our work contributes a general-purpose retriever and offers interpretability insights into the long-context capabilities of LMs.

View on arXiv
@article{zhang2025_2506.09944,
  title={ Query-Focused Retrieval Heads Improve Long-Context Reasoning and Re-ranking },
  author={ Wuwei Zhang and Fangcong Yin and Howard Yen and Danqi Chen and Xi Ye },
  journal={arXiv preprint arXiv:2506.09944},
  year={ 2025 }
}
Comments on this paper