41
0

Spontaneous Speech Variables for Evaluating LLMs Cognitive Plausibility

Main:8 Pages
11 Figures
Bibliography:4 Pages
3 Tables
Appendix:4 Pages
Abstract

The achievements of Large Language Models in Natural Language Processing, especially for high-resource languages, call for a better understanding of their characteristics from a cognitive perspective. Researchers have attempted to evaluate artificial models by testing their ability to predict behavioral (e.g., eye-tracking fixations) and physiological (e.g., brain responses) variables during language processing (e.g., reading/listening). In this paper, we propose using spontaneous speech corpora to derive production variables (speech reductions, prosodic prominences) and applying them in a similar fashion. More precisely, we extract. We then test models trained with a standard procedure on different pretraining datasets (written, spoken, and mixed genres) for their ability to predict these two variables. Our results show that, after some fine-tuning, the models can predict these production variables well above baselines. We also observe that spoken genre training data provides more accurate predictions than written genres. These results contribute to the broader effort of using high-quality speech corpora as benchmarks for LLMs.

View on arXiv
@article{wang2025_2505.16277,
  title={ Spontaneous Speech Variables for Evaluating LLMs Cognitive Plausibility },
  author={ Sheng-Fu Wang and Laurent Prevot and Jou-an Chi and Ri-Sheng Huang and Shu-Kai Hsieh },
  journal={arXiv preprint arXiv:2505.16277},
  year={ 2025 }
}
Comments on this paper

We use cookies and other tracking technologies to improve your browsing experience on our website, to show you personalized content and targeted ads, to analyze our website traffic, and to understand where our visitors are coming from. See our policy.