ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2504.02765
32
0

Robot-Led Vision Language Model Wellbeing Assessment of Children

3 April 2025
N. I. Abbasi
Fethiye Irmak Dogan
Guy Laban
Joanna Anderson
Tamsin Ford
Peter B. Jones
Hatice Gunes
ArXivPDFHTML
Abstract

This study presents a novel robot-led approach to assessing children's mental wellbeing using a Vision Language Model (VLM). Inspired by the Child Apperception Test (CAT), the social robot NAO presented children with pictorial stimuli to elicit their verbal narratives of the images, which were then evaluated by a VLM in accordance with CAT assessment guidelines. The VLM's assessments were systematically compared to those provided by a trained psychologist. The results reveal that while the VLM demonstrates moderate reliability in identifying cases with no wellbeing concerns, its ability to accurately classify assessments with clinical concern remains limited. Moreover, although the model's performance was generally consistent when prompted with varying demographic factors such as age and gender, a significantly higher false positive rate was observed for girls, indicating potential sensitivity to gender attribute. These findings highlight both the promise and the challenges of integrating VLMs into robot-led assessments of children's wellbeing.

View on arXiv
@article{abbasi2025_2504.02765,
  title={ Robot-Led Vision Language Model Wellbeing Assessment of Children },
  author={ Nida Itrat Abbasi and Fethiye Irmak Dogan and Guy Laban and Joanna Anderson and Tamsin Ford and Peter B. Jones and Hatice Gunes },
  journal={arXiv preprint arXiv:2504.02765},
  year={ 2025 }
}
Comments on this paper