SAKURA: On the Multi-hop Reasoning of Large Audio-Language Models Based on Speech and Audio Information

Large audio-language models (LALMs) extend the large language models with multimodal understanding in speech, audio, etc. While their performances on speech and audio-processing tasks are extensively studied, their reasoning abilities remain underexplored. Particularly, their multi-hop reasoning, the ability to recall and integrate multiple facts, lacks systematic evaluation. Existing benchmarks focus on general speech and audio-processing tasks, conversational abilities, and fairness but overlook this aspect. To bridge this gap, we introduce SAKURA, a benchmark assessing LALMs' multi-hop reasoning based on speech and audio information. Results show that LALMs struggle to integrate speech/audio representations for multi-hop reasoning, even when they extract the relevant information correctly, highlighting a fundamental challenge in multimodal reasoning. Our findings expose a critical limitation in LALMs, offering insights and resources for future research.
View on arXiv@article{yang2025_2505.13237, title={ SAKURA: On the Multi-hop Reasoning of Large Audio-Language Models Based on Speech and Audio Information }, author={ Chih-Kai Yang and Neo Ho and Yen-Ting Piao and Hung-yi Lee }, journal={arXiv preprint arXiv:2505.13237}, year={ 2025 } }