ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.11046
14
0

The Effects of Data Augmentation on Confidence Estimation for LLMs

21 May 2025
Rui Wang
Renyu Zhu
Minmin Lin
R. Wu
Tangjie Lv
Changjie Fan
Haobo Wang
ArXiv (abs)PDFHTML
Main:4 Pages
1 Figures
Bibliography:3 Pages
6 Tables
Appendix:1 Pages
Abstract

Confidence estimation is crucial for reflecting the reliability of large language models (LLMs), particularly in the widely used closed-source models. Utilizing data augmentation for confidence estimation is viable, but discussions focus on specific augmentation techniques, limiting its potential. We study the impact of different data augmentation methods on confidence estimation. Our findings indicate that data augmentation strategies can achieve better performance and mitigate the impact of overconfidence. We investigate the influential factors related to this and discover that, while preserving semantic information, greater data diversity enhances the effectiveness of augmentation. Furthermore, the impact of different augmentation strategies varies across different range of application. Considering parameter transferability and usability, the random combination of augmentations is a promising choice.

View on arXiv
@article{wang2025_2506.11046,
  title={ The Effects of Data Augmentation on Confidence Estimation for LLMs },
  author={ Rui Wang and Renyu Zhu and Minmin Lin and Runze Wu and Tangjie Lv and Changjie Fan and Haobo Wang },
  journal={arXiv preprint arXiv:2506.11046},
  year={ 2025 }
}
Comments on this paper