What Prompts Don't Say: Understanding and Managing Underspecification in LLM Prompts

Building LLM-powered software requires developers to communicate their requirements through natural language, but developer prompts are frequently underspecified, failing to fully capture many user-important requirements. In this paper, we present an in-depth analysis of prompt underspecification, showing that while LLMs can often (41.1%) guess unspecified requirements by default, such behavior is less robust: Underspecified prompts are 2x more likely to regress over model or prompt changes, sometimes with accuracy drops by more than 20%. We then demonstrate that simply adding more requirements to a prompt does not reliably improve performance, due to LLMs' limited instruction-following capabilities and competing constraints, and standard prompt optimizers do not offer much help. To address this, we introduce novel requirements-aware prompt optimization mechanisms that can improve performance by 4.8% on average over baselines that naively specify everything in the prompt. Beyond prompt optimization, we envision that effectively managing prompt underspecification requires a broader process, including proactive requirements discovery, evaluation, and monitoring.
View on arXiv@article{yang2025_2505.13360, title={ What Prompts Don't Say: Understanding and Managing Underspecification in LLM Prompts }, author={ Chenyang Yang and Yike Shi and Qianou Ma and Michael Xieyang Liu and Christian Kästner and Tongshuang Wu }, journal={arXiv preprint arXiv:2505.13360}, year={ 2025 } }