ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.14908
51
0

SegSub: Evaluating Robustness to Knowledge Conflicts and Hallucinations in Vision-Language Models

19 February 2025
Peter Carragher
Nikitha Rao
Abhinand Jha
R Raghav
Kathleen M. Carley
    VLM
ArXivPDFHTML
Abstract

Vision language models (VLM) demonstrate sophisticated multimodal reasoning yet are prone to hallucination when confronted with knowledge conflicts, impeding their deployment in information-sensitive contexts. While existing research addresses robustness in unimodal models, the multimodal domain lacks systematic investigation of cross-modal knowledge conflicts. This research introduces \segsub, a framework for applying targeted image perturbations to investigate VLM resilience against knowledge conflicts. Our analysis reveals distinct vulnerability patterns: while VLMs are robust to parametric conflicts (20% adherence rates), they exhibit significant weaknesses in identifying counterfactual conditions (<30% accuracy) and resolving source conflicts (<1% accuracy). Correlations between contextual richness and hallucination rate (r = -0.368, p = 0.003) reveal the kinds of images that are likely to cause hallucinations. Through targeted fine-tuning on our benchmark dataset, we demonstrate improvements in VLM knowledge conflict detection, establishing a foundation for developing hallucination-resilient multimodal systems in information-sensitive environments.

View on arXiv
@article{carragher2025_2502.14908,
  title={ SegSub: Evaluating Robustness to Knowledge Conflicts and Hallucinations in Vision-Language Models },
  author={ Peter Carragher and Nikitha Rao and Abhinand Jha and R Raghav and Kathleen M. Carley },
  journal={arXiv preprint arXiv:2502.14908},
  year={ 2025 }
}
Comments on this paper