ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2502.17669
40
0

Towards Human Cognition: Visual Context Guides Syntactic Priming in Fusion-Encoded Models

24 February 2025
Bushi Xiao
Michael Bennie
Jayetri Bardhan
Daisy Zhe Wang
ArXivPDFHTML
Abstract

We introduced PRISMATIC, the first multimodal structural priming dataset, and proposed a reference-free evaluation metric that assesses priming effects without predefined target sentences. Using this metric, we constructed and tested models with different multimodal encoding architectures (dual encoder and fusion encoder) to investigate their structural preservation capabilities. Our findings show that models with both encoding methods demonstrate comparable syntactic priming effects. However, only fusion-encoded models exhibit robust positive correlations between priming effects and visual similarity, suggesting a cognitive process more aligned with human psycholinguistic patterns. This work provides new insights into evaluating and understanding how syntactic information is processed in multimodal language models.

View on arXiv
@article{xiao2025_2502.17669,
  title={ Towards Human Cognition: Visual Context Guides Syntactic Priming in Fusion-Encoded Models },
  author={ Bushi Xiao and Michael Bennie and Jayetri Bardhan and Daisy Zhe Wang },
  journal={arXiv preprint arXiv:2502.17669},
  year={ 2025 }
}
Comments on this paper