ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2311.08877
  4. Cited By
Llamas Know What GPTs Don't Show: Surrogate Models for Confidence
  Estimation

Llamas Know What GPTs Don't Show: Surrogate Models for Confidence Estimation

15 November 2023
Vaishnavi Shrivastava
Percy Liang
Ananya Kumar
ArXivPDFHTML

Papers citing "Llamas Know What GPTs Don't Show: Surrogate Models for Confidence Estimation"

22 / 22 papers shown
Title
Comparing Uncertainty Measurement and Mitigation Methods for Large Language Models: A Systematic Review
Comparing Uncertainty Measurement and Mitigation Methods for Large Language Models: A Systematic Review
Toghrul Abbasli
Kentaroh Toyoda
Yuan Wang
Leon Witt
Muhammad Asif Ali
Yukai Miao
Dan Li
Qingsong Wei
UQCV
92
0
0
25 Apr 2025
Gauging Overprecision in LLMs: An Empirical Study
Gauging Overprecision in LLMs: An Empirical Study
Adil Bahaj
Hamed Rahimi
Mohamed Chetouani
Mounir Ghogho
72
0
0
16 Apr 2025
The challenge of uncertainty quantification of large language models in medicine
The challenge of uncertainty quantification of large language models in medicine
Zahra Atf
Seyed Amir Ahmad Safavi-Naini
Peter Lewis
Aref Mahjoubfar
Nariman Naderi
Thomas Savage
Ali Soroush
21
0
0
07 Apr 2025
Beyond Accuracy: The Role of Calibration in Self-Improving Large Language Models
Beyond Accuracy: The Role of Calibration in Self-Improving Large Language Models
Liangjie Huang
Dawei Li
Huan Liu
Lu Cheng
LRM
36
0
0
03 Apr 2025
Calibrating LLM Confidence with Semantic Steering: A Multi-Prompt Aggregation Framework
Ziang Zhou
Tianyuan Jin
Jieming Shi
Qing Li
LLMSV
68
0
0
04 Mar 2025
I Know What I Don't Know: Improving Model Cascades Through Confidence Tuning
I Know What I Don't Know: Improving Model Cascades Through Confidence Tuning
Stephan Rabanser
Nathalie Rauschmayr
Achin Kulshrestha
Petra Poklukar
Wittawat Jitkrittum
Sean Augenstein
Congchao Wang
Federico Tombari
42
0
0
26 Feb 2025
A Survey of Calibration Process for Black-Box LLMs
A Survey of Calibration Process for Black-Box LLMs
Liangru Xie
Hui Liu
Jingying Zeng
Xianfeng Tang
Yan Han
Chen Luo
Jing Huang
Zhen Li
Suhang Wang
Qi He
74
1
0
17 Dec 2024
Epistemic Integrity in Large Language Models
Epistemic Integrity in Large Language Models
Bijean Ghafouri
Shahrad Mohammadzadeh
James Zhou
Pratheeksha Nair
Jacob-Junqi Tian
Mayank Goel
Reihaneh Rabbany
Jean-Francois Godbout
Kellin Pelrine
HILM
48
0
0
10 Nov 2024
A Survey of Uncertainty Estimation in LLMs: Theory Meets Practice
A Survey of Uncertainty Estimation in LLMs: Theory Meets Practice
Hsiu-Yuan Huang
Yutong Yang
Zhaoxi Zhang
Sanwoo Lee
Yunfang Wu
44
9
0
20 Oct 2024
Enhancing Healthcare LLM Trust with Atypical Presentations Recalibration
Enhancing Healthcare LLM Trust with Atypical Presentations Recalibration
Jeremy Qin
Bang Liu
Quoc Dinh Nguyen
35
2
0
05 Sep 2024
Crowd-Calibrator: Can Annotator Disagreement Inform Calibration in
  Subjective Tasks?
Crowd-Calibrator: Can Annotator Disagreement Inform Calibration in Subjective Tasks?
Urja Khurana
Eric T. Nalisnick
Antske Fokkens
Swabha Swayamdipta
42
3
0
26 Aug 2024
Know Your Limits: A Survey of Abstention in Large Language Models
Know Your Limits: A Survey of Abstention in Large Language Models
Bingbing Wen
Jihan Yao
Shangbin Feng
Chenjun Xu
Yulia Tsvetkov
Bill Howe
Lucy Lu Wang
59
11
0
25 Jul 2024
Perceptions of Linguistic Uncertainty by Language Models and Humans
Perceptions of Linguistic Uncertainty by Language Models and Humans
Catarina G Belém
Markelle Kelly
M. Steyvers
Sameer Singh
Padhraic Smyth
43
3
0
22 Jul 2024
Collaborative Performance Prediction for Large Language Models
Collaborative Performance Prediction for Large Language Models
Qiyuan Zhang
Fuyuan Lyu
Xue Liu
Chen Ma
32
3
0
01 Jul 2024
Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented
  Generation
Synchronous Faithfulness Monitoring for Trustworthy Retrieval-Augmented Generation
Di Wu
Jia-Chen Gu
Fan Yin
Nanyun Peng
Kai-Wei Chang
HILM
58
1
0
19 Jun 2024
Counterfactual Debating with Preset Stances for Hallucination Elimination of LLMs
Counterfactual Debating with Preset Stances for Hallucination Elimination of LLMs
Yi Fang
Moxin Li
Wenjie Wang
Hui Lin
Fuli Feng
LRM
65
5
0
17 Jun 2024
Cycles of Thought: Measuring LLM Confidence through Stable Explanations
Cycles of Thought: Measuring LLM Confidence through Stable Explanations
Evan Becker
Stefano Soatto
45
6
0
05 Jun 2024
Language Model Cascades: Token-level uncertainty and beyond
Language Model Cascades: Token-level uncertainty and beyond
Neha Gupta
Harikrishna Narasimhan
Wittawat Jitkrittum
A. S. Rawat
A. Menon
Sanjiv Kumar
UQLM
53
42
0
15 Apr 2024
Multi-Perspective Consistency Enhances Confidence Estimation in Large
  Language Models
Multi-Perspective Consistency Enhances Confidence Estimation in Large Language Models
Pei Wang
Yejie Wang
Muxi Diao
Keqing He
Guanting Dong
Weiran Xu
24
0
0
17 Feb 2024
Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via
  Self-Evaluation
Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via Self-Evaluation
Xiaoying Zhang
Baolin Peng
Ye Tian
Jingyan Zhou
Lifeng Jin
Linfeng Song
Haitao Mi
Helen Meng
HILM
42
43
0
14 Feb 2024
Incoherent Probability Judgments in Large Language Models
Incoherent Probability Judgments in Large Language Models
Jian-Qiao Zhu
Thomas L. Griffiths
83
6
0
30 Jan 2024
A Survey of Confidence Estimation and Calibration in Large Language
  Models
A Survey of Confidence Estimation and Calibration in Large Language Models
Jiahui Geng
Fengyu Cai
Yuxia Wang
Heinz Koeppl
Preslav Nakov
Iryna Gurevych
UQCV
41
54
0
14 Nov 2023
1