Papers
Communities
Events
Blog
Pricing
Search
Open menu
Home
Papers
2402.12563
Cited By
Confidence Matters: Revisiting Intrinsic Self-Correction Capabilities of Large Language Models
19 February 2024
Loka Li
Zhenhao Chen
Guan-Hong Chen
Yixuan Zhang
Yusheng Su
Eric P. Xing
Anton van den Hengel
LRM
Re-assign community
ArXiv
PDF
HTML
Papers citing
"Confidence Matters: Revisiting Intrinsic Self-Correction Capabilities of Large Language Models"
17 / 17 papers shown
Title
Meta-Thinking in LLMs via Multi-Agent Reinforcement Learning: A Survey
Ahsan Bilal
Muhammad Ahmed Mohsin
Muhammad Umer
Muhammad Awais Khan Bangash
Muhammad Ali Jamshed
LLMAG
LRM
AI4CE
53
0
0
20 Apr 2025
A Data-Centric Approach for Safe and Secure Large Language Models against Threatening and Toxic Content
Chaima Njeh
Haïfa Nakouri
Fehmi Jaafar
22
0
0
19 Apr 2025
Entropy-Based Adaptive Weighting for Self-Training
Xiaoxuan Wang
Yihe Deng
Mingyu Derek Ma
Wei Wang
LRM
52
0
0
31 Mar 2025
Causality for Large Language Models
Anpeng Wu
Kun Kuang
Minqin Zhu
Yingrong Wang
Yujia Zheng
Kairong Han
Yangqiu Song
Guangyi Chen
Fei Wu
Anton van den Hengel
LRM
46
7
0
20 Oct 2024
Latent Space Chain-of-Embedding Enables Output-free LLM Self-Evaluation
Yiming Wang
Pei Zhang
Baosong Yang
Derek F. Wong
Rui-cang Wang
LRM
50
4
0
17 Oct 2024
Re-Thinking Process Mining in the AI-Based Agents Era
Alessandro Berti
Mayssa Maatallah
U. Jessen
Michal Sroka
S. Ghannouchi
LLMAG
25
3
0
14 Aug 2024
Validation Requirements for AI-based Intervention-Evaluation in Aging and Longevity Research and Practice
G. Fuellen
Anton Y Kulaga
Sebastian Lobentanzer
Maximilian Unfried
Roberto Avelar
Daniel Palmer
Brian K. Kennedy
31
1
0
11 Aug 2024
FIRE: A Dataset for Feedback Integration and Refinement Evaluation of Multimodal Models
Pengxiang Li
Zhi Gao
Bofei Zhang
Tao Yuan
Yuwei Wu
Mehrtash Harandi
Yunde Jia
Song-Chun Zhu
Qing Li
VLM
MLLM
45
3
0
16 Jul 2024
On the Intrinsic Self-Correction Capability of LLMs: Uncertainty and Latent Concept
Guangliang Liu
Haitao Mao
Bochuan Cao
Zhiyu Xue
K. Johnson
Jiliang Tang
Rongrong Wang
LRM
34
9
0
04 Jun 2024
When Can LLMs Actually Correct Their Own Mistakes? A Critical Survey of Self-Correction of LLMs
Ryo Kamoi
Yusen Zhang
Nan Zhang
Jiawei Han
Rui Zhang
LRM
47
57
0
03 Jun 2024
A Theoretical Understanding of Self-Correction through In-context Alignment
Yifei Wang
Yuyang Wu
Zeming Wei
Stefanie Jegelka
Yisen Wang
LRM
41
13
0
28 May 2024
Reinforcement Learning from Multi-role Debates as Feedback for Bias Mitigation in LLMs
Ruoxi Cheng
Haoxuan Ma
Shuirong Cao
Jiaqi Li
Aihua Pei
Zhiqiang Wang
Pengliang Ji
Haoyu Wang
Jiaqi Huo
AI4CE
29
6
0
15 Apr 2024
GPT-4 Doesn't Know It's Wrong: An Analysis of Iterative Prompting for Reasoning Problems
Kaya Stechly
Matthew Marquez
Subbarao Kambhampati
LRM
165
84
0
19 Oct 2023
Cumulative Reasoning with Large Language Models
Yifan Zhang
Jingqin Yang
Yang Yuan
Andrew Chi-Chih Yao
ReLM
ELM
LRM
AI4CE
36
69
0
08 Aug 2023
ReAct: Synergizing Reasoning and Acting in Language Models
Shunyu Yao
Jeffrey Zhao
Dian Yu
Nan Du
Izhak Shafran
Karthik Narasimhan
Yuan Cao
LLMAG
ReLM
LRM
246
2,494
0
06 Oct 2022
Improving alignment of dialogue agents via targeted human judgements
Amelia Glaese
Nat McAleese
Maja Trkebacz
John Aslanides
Vlad Firoiu
...
John F. J. Mellor
Demis Hassabis
Koray Kavukcuoglu
Lisa Anne Hendricks
G. Irving
ALM
AAML
227
502
0
28 Sep 2022
Training language models to follow instructions with human feedback
Long Ouyang
Jeff Wu
Xu Jiang
Diogo Almeida
Carroll L. Wainwright
...
Amanda Askell
Peter Welinder
Paul Christiano
Jan Leike
Ryan J. Lowe
OSLM
ALM
313
11,953
0
04 Mar 2022
1