18
0
v1v2 (latest)

ARGUS: Hallucination and Omission Evaluation in Video-LLMs

Main:8 Pages
20 Figures
Bibliography:3 Pages
3 Tables
Appendix:15 Pages
Abstract

Video large language models have not yet been widely deployed, largely due to their tendency to hallucinate. Typical benchmarks for Video-LLMs rely simply on multiple-choice questions. Unfortunately, VideoLLMs hallucinate far more aggressively on freeform text generation tasks like video captioning than they do on multiple choice verification tasks. To address this weakness, we propose ARGUS, a VideoLLM benchmark that measures freeform video captioning performance. By comparing VideoLLM outputs to human ground truth captions, ARGUS quantifies dual metrics. First, we measure the rate of hallucinations in the form of incorrect statements about video content or temporal relationships. Second, we measure the rate at which the model omits important descriptive details. Together, these dual metrics form a comprehensive view of video captioning performance.

View on arXiv
@article{rawal2025_2506.07371,
  title={ ARGUS: Hallucination and Omission Evaluation in Video-LLMs },
  author={ Ruchit Rawal and Reza Shirkavand and Heng Huang and Gowthami Somepalli and Tom Goldstein },
  journal={arXiv preprint arXiv:2506.07371},
  year={ 2025 }
}
Comments on this paper