Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2505.02072
Cited By
What do Language Model Probabilities Represent? From Distribution Estimation to Response Prediction
4 May 2025
Eitan Wagner
Omri Abend
Re-assign community
ArXiv (abs)
PDF
HTML
Papers citing
"What do Language Model Probabilities Represent? From Distribution Estimation to Response Prediction"
27 / 27 papers shown
Title
Enough Coin Flips Can Make LLMs Act Bayesian
Ritwik Gupta
Rodolfo Corona
Jiaxin Ge
Eric Wang
Dan Klein
Trevor Darrell
David M. Chan
BDL
LRM
101
3
0
06 Mar 2025
Mind Your Theory: Theory of Mind Goes Deeper Than Reasoning
Eitan Wagner
Nitay Alon
J. Barnby
Omri Abend
LRM
180
2
0
18 Dec 2024
Benchmarking Distributional Alignment of Large Language Models
Nicole Meister
Carlos Guestrin
Tatsunori Hashimoto
ALM
86
6
0
08 Nov 2024
Diverging Preferences: When do Annotators Disagree and do Models Know?
Michael J.Q. Zhang
Zhilin Wang
Jena D. Hwang
Yi Dong
Olivier Delalleau
Yejin Choi
Eunsol Choi
Xiang Ren
Valentina Pyatkin
103
13
0
18 Oct 2024
CONTESTS: a Framework for Consistency Testing of Span Probabilities in Language Models
Eitan Wagner
Yuli Slavutsky
Omri Abend
71
1
0
30 Sep 2024
From Distributional to Overton Pluralism: Investigating Large Language Model Alignment
Thom Lake
Eunsol Choi
Greg Durrett
102
14
0
25 Jun 2024
Can Large Language Models Faithfully Express Their Intrinsic Uncertainty in Words?
G. Yona
Roee Aharoni
Mor Geva
HILM
101
32
0
27 May 2024
RLHF Deciphered: A Critical Analysis of Reinforcement Learning from Human Feedback for LLMs
Shreyas Chaudhari
Pranjal Aggarwal
Vishvak Murahari
Tanmay Rajpurohit
Ashwin Kalyan
Karthik Narasimhan
Ameet Deshpande
Bruno Castro da Silva
91
38
0
12 Apr 2024
Calibrating the Confidence of Large Language Models by Eliciting Fidelity
Mozhi Zhang
Mianqiu Huang
Rundong Shi
Linsen Guo
Chong Peng
Peng Yan
Yaqian Zhou
Xipeng Qiu
86
13
0
03 Apr 2024
Cognitive Dissonance: Why Do Language Model Outputs Disagree with Internal Representations of Truthfulness?
Kevin Liu
Stephen Casper
Dylan Hadfield-Menell
Jacob Andreas
HILM
122
39
0
27 Nov 2023
On the Calibration of Large Language Models and Alignment
Chiwei Zhu
Benfeng Xu
Quan Wang
Yongdong Zhang
Zhendong Mao
155
44
0
22 Nov 2023
Large Language Models Are Not Robust Multiple Choice Selectors
Chujie Zheng
Hao Zhou
Fandong Meng
Jie Zhou
Minlie Huang
128
247
0
07 Sep 2023
Position: Key Claims in LLM Research Have a Long Tail of Footnotes
Anna Rogers
A. Luccioni
155
21
0
14 Aug 2023
Llama 2: Open Foundation and Fine-Tuned Chat Models
Hugo Touvron
Louis Martin
Kevin R. Stone
Peter Albert
Amjad Almahairi
...
Sharan Narang
Aurelien Rodriguez
Robert Stojnic
Sergey Edunov
Thomas Scialom
AI4MH
ALM
468
12,124
0
18 Jul 2023
On the Efficacy of Sampling Adapters
Clara Meister
Tiago Pimentel
Luca Malagutti
Ethan Gotlieb Wilcox
Ryan Cotterell
122
14
0
07 Jul 2023
How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources
Yizhong Wang
Hamish Ivison
Pradeep Dasigi
Jack Hessel
Tushar Khot
...
David Wadden
Kelsey MacMillan
Noah A. Smith
Iz Beltagy
Hannaneh Hajishirzi
ALM
ELM
120
393
0
07 Jun 2023
Self-Instruct: Aligning Language Models with Self-Generated Instructions
Yizhong Wang
Yeganeh Kordi
Swaroop Mishra
Alisa Liu
Noah A. Smith
Daniel Khashabi
Hannaneh Hajishirzi
ALM
SyDa
LRM
195
2,264
0
20 Dec 2022
A Measure-Theoretic Characterization of Tight Language Models
Li Du
Lucas Torroba Hennigen
Tiago Pimentel
Clara Meister
Jason Eisner
Ryan Cotterell
93
32
0
20 Dec 2022
Stop Measuring Calibration When Humans Disagree
Joris Baan
Wilker Aziz
Barbara Plank
Raquel Fernández
92
56
0
28 Oct 2022
Transformer Feed-Forward Layers Build Predictions by Promoting Concepts in the Vocabulary Space
Mor Geva
Avi Caciularu
Ke Wang
Yoav Goldberg
KELM
138
389
0
28 Mar 2022
Calibration of Machine Reading Systems at Scale
Shehzaad Dhuliawala
Leonard Adolphs
Rajarshi Das
Mrinmaya Sachan
72
12
0
20 Mar 2022
The World of an Octopus: How Reporting Bias Influences a Language Model's Perception of Color
Cory Paik
Stéphane Aroca-Ouellette
Alessandro Roncone
Katharina Kann
62
34
0
15 Oct 2021
How Can We Know When Language Models Know? On the Calibration of Language Models for Question Answering
Zhengbao Jiang
Jun Araki
Haibo Ding
Graham Neubig
UQCV
67
439
0
02 Dec 2020
Language Models are Few-Shot Learners
Tom B. Brown
Benjamin Mann
Nick Ryder
Melanie Subbiah
Jared Kaplan
...
Christopher Berner
Sam McCandlish
Alec Radford
Ilya Sutskever
Dario Amodei
BDL
991
42,651
0
28 May 2020
Fine-Tuning Language Models from Human Preferences
Daniel M. Ziegler
Nisan Stiennon
Jeff Wu
Tom B. Brown
Alec Radford
Dario Amodei
Paul Christiano
G. Irving
ALM
503
1,775
0
18 Sep 2019
The Curious Case of Neural Text Degeneration
Ari Holtzman
Jan Buys
Li Du
Maxwell Forbes
Yejin Choi
215
3,215
0
22 Apr 2019
On Calibration of Modern Neural Networks
Chuan Guo
Geoff Pleiss
Yu Sun
Kilian Q. Weinberger
UQCV
301
5,889
0
14 Jun 2017
1