28
0

Speech as a Multimodal Digital Phenotype for Multi-Task LLM-based Mental Health Prediction

Main:5 Pages
1 Figures
Bibliography:1 Pages
Abstract

Speech is a noninvasive digital phenotype that can offer valuable insights into mental health conditions, but it is often treated as a single modality. In contrast, we propose the treatment of patient speech data as a trimodal multimedia data source for depression detection. This study explores the potential of large language model-based architectures for speech-based depression prediction in a multimodal regime that integrates speech-derived text, acoustic landmarks, and vocal biomarkers. Adolescent depression presents a significant challenge and is often comorbid with multiple disorders, such as suicidal ideation and sleep disturbances. This presents an additional opportunity to integrate multi-task learning (MTL) into our study by simultaneously predicting depression, suicidal ideation, and sleep disturbances using the multimodal formulation. We also propose a longitudinal analysis strategy that models temporal changes across multiple clinical interactions, allowing for a comprehensive understanding of the conditions' progression. Our proposed approach, featuring trimodal, longitudinal MTL is evaluated on the Depression Early Warning dataset. It achieves a balanced accuracy of 70.8%, which is higher than each of the unimodal, single-task, and non-longitudinal methods.

View on arXiv
@article{ali2025_2505.23822,
  title={ Speech as a Multimodal Digital Phenotype for Multi-Task LLM-based Mental Health Prediction },
  author={ Mai Ali and Christopher Lucasius and Tanmay P. Patel and Madison Aitken and Jacob Vorstman and Peter Szatmari and Marco Battaglia and Deepa Kundur },
  journal={arXiv preprint arXiv:2505.23822},
  year={ 2025 }
}
Comments on this paper