Interpreting figurative language such as sarcasm across multi-modal inputs presents unique challenges, often requiring task-specific fine-tuning and extensive reasoning steps. However, current Chain-of-Thought approaches do not efficiently leverage the same cognitive processes that enable humans to identify sarcasm. We present IRONIC, an in-context learning framework that leverages Multi-modal Coherence Relations to analyze referential, analogical and pragmatic image-text linkages. Our experiments show that IRONIC achieves state-of-the-art performance on zero-shot Multi-modal Sarcasm Detection across different baselines. This demonstrates the need for incorporating linguistic and cognitive insights into the design of multi-modal reasoning strategies. Our code is available at:this https URL
View on arXiv@article{ramakrishnan2025_2505.16258, title={ IRONIC: Coherence-Aware Reasoning Chains for Multi-Modal Sarcasm Detection }, author={ Aashish Anantha Ramakrishnan and Aadarsh Anantha Ramakrishnan and Dongwon Lee }, journal={arXiv preprint arXiv:2505.16258}, year={ 2025 } }