ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2410.00332
23
1

Vision Language Models Know Law of Conservation without Understanding More-or-Less

1 October 2024
Dezhi Luo
Haiyun Lyu
Qingying Gao
Haoran Sun
Yijiang Li
Hokin Deng
ArXivPDFHTML
Abstract

Understanding law of conservation is a critical milestone in human cognitive development considered to be supported by the apprehension of quantitative concepts and the reversibility of operations. To assess whether this critical component of human intelligence has emerged in Vision Language Models, we have curated the ConserveBench, a battery of 365 cognitive experiments across four dimensions of physical quantities: volume, solid quantity, length, and number. The former two involve transformational tasks which require reversibility understanding. The latter two involve non-transformational tasks which assess quantity understanding. Surprisingly, we find that while Vision Language Models are generally good at transformational tasks, they tend to fail at non-transformational tasks. There is a dissociation between understanding the reversibility of operations and understanding the concept of quantity, which both are believed to be the cornerstones of understanding law of conservation in humans. \href\href{this https URL}{Website}\href

View on arXiv
@article{luo2025_2410.00332,
  title={ Vision Language Models Know Law of Conservation without Understanding More-or-Less },
  author={ Dezhi Luo and Haiyun Lyu and Qingying Gao and Haoran Sun and Yijiang Li and Hokin Deng },
  journal={arXiv preprint arXiv:2410.00332},
  year={ 2025 }
}
Comments on this paper