ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2505.17233
33
0
v1v2 (latest)

Semantic-Aware Interpretable Multimodal Music Auto-Tagging

22 May 2025
Andreas Patakis
Vassilis Lyberatos
Spyridon Kantarelis
Edmund Dervakos
Giorgos Stamou
    VLM
ArXiv (abs)PDFHTML
Main:4 Pages
2 Figures
Bibliography:1 Pages
2 Tables
Abstract

Music auto-tagging is essential for organizing and discovering music in extensive digital libraries. While foundation models achieve exceptional performance in this domain, their outputs often lack interpretability, limiting trust and usability for researchers and end-users alike. In this work, we present an interpretable framework for music auto-tagging that leverages groups of musically meaningful multimodal features, derived from signal processing, deep learning, ontology engineering, and natural language processing. To enhance interpretability, we cluster features semantically and employ an expectation maximization algorithm, assigning distinct weights to each group based on its contribution to the tagging process. Our method achieves competitive tagging performance while offering a deeper understanding of the decision-making process, paving the way for more transparent and user-centric music tagging systems.

View on arXiv
@article{patakis2025_2505.17233,
  title={ Semantic-Aware Interpretable Multimodal Music Auto-Tagging },
  author={ Andreas Patakis and Vassilis Lyberatos and Spyridon Kantarelis and Edmund Dervakos and Giorgos Stamou },
  journal={arXiv preprint arXiv:2505.17233},
  year={ 2025 }
}
Comments on this paper