ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.13410
60
0

Tell Me Why: Incentivizing Explanations

20 February 2025
Siddarth Srinivasan
Ezra Karger
Michiel Bakker
Yiling Chen
    FAtt
ArXivPDFHTML
Abstract

Common sense suggests that when individuals explain why they believe something, we can arrive at more accurate conclusions than when they simply state what they believe. Yet, there is no known mechanism that provides incentives to elicit explanations for beliefs from agents. This likely stems from the fact that standard Bayesian models make assumptions (like conditional independence of signals) that preempt the need for explanations, in order to show efficient information aggregation. A natural justification for the value of explanations is that agents' beliefs tend to be drawn from overlapping sources of information, so agents' belief reports do not reveal all that needs to be known. Indeed, this work argues that rationales-explanations of an agent's private information-lead to more efficient aggregation by allowing agents to efficiently identify what information they share and what information is new. Building on this model of rationales, we present a novel 'deliberation mechanism' to elicit rationales from agents in which truthful reporting of beliefs and rationales is a perfect Bayesian equilibrium.

View on arXiv
@article{srinivasan2025_2502.13410,
  title={ Tell Me Why: Incentivizing Explanations },
  author={ Siddarth Srinivasan and Ezra Karger and Michiel Bakker and Yiling Chen },
  journal={arXiv preprint arXiv:2502.13410},
  year={ 2025 }
}
Comments on this paper