90
2
v1v2v3 (latest)

MA4DIV: Multi-Agent Reinforcement Learning for Search Result Diversification

Abstract

Search result diversification (SRD), which aims to ensure that documents in a ranking list cover a broad range of subtopics, is a significant and widely studied problem in Information Retrieval and Web Search. Existing methods primarily utilize a paradigm of "greedy selection", i.e., selecting one document with the highest diversity score at a time or optimize an approximation of the objective function. These approaches tend to be inefficient and are easily trapped in a suboptimal state. To address these challenges, we introduce Multi-Agent reinforcement learning (MARL) for search result DIVersity, which called MA4DIV. In this approach, each document is an agent and the search result diversification is modeled as a cooperative task among multiple agents. By modeling the SRD ranking problem as a cooperative MARL problem, this approach allows for directly optimizing the diversity metrics, such as α\alpha-NDCG, while achieving high training efficiency. We conducted experiments on public TREC datasets and a larger scale dataset in the industrial setting. The experiemnts show that MA4DIV achieves substantial improvements in both effectiveness and efficiency than existing baselines, especially on the industrial dataset. The code of MA4DIV can be seen onthis https URL.

View on arXiv
@article{chen2025_2403.17421,
  title={ MA4DIV: Multi-Agent Reinforcement Learning for Search Result Diversification },
  author={ Yiqun Chen and Jiaxin Mao and Yi Zhang and Dehong Ma and Long Xia and Jun Fan and Daiting Shi and Zhicong Cheng and Simiu Gu and Dawei Yin },
  journal={arXiv preprint arXiv:2403.17421},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.