ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2503.21775
68
0

StyleMotif: Multi-Modal Motion Stylization using Style-Content Cross Fusion

27 March 2025
Ziyu Guo
Young Yoon Lee
Joseph Liu
Yizhak Ben-Shabat
Victor Zordan
Mubbasir Kapadia
    DiffM
    VGen
ArXivPDFHTML
Abstract

We present StyleMotif, a novel Stylized Motion Latent Diffusion model, generating motion conditioned on both content and style from multiple modalities. Unlike existing approaches that either focus on generating diverse motion content or transferring style from sequences, StyleMotif seamlessly synthesizes motion across a wide range of content while incorporating stylistic cues from multi-modal inputs, including motion, text, image, video, and audio. To achieve this, we introduce a style-content cross fusion mechanism and align a style encoder with a pre-trained multi-modal model, ensuring that the generated motion accurately captures the reference style while preserving realism. Extensive experiments demonstrate that our framework surpasses existing methods in stylized motion generation and exhibits emergent capabilities for multi-modal motion stylization, enabling more nuanced motion synthesis. Source code and pre-trained models will be released upon acceptance. Project Page:this https URL

View on arXiv
@article{guo2025_2503.21775,
  title={ StyleMotif: Multi-Modal Motion Stylization using Style-Content Cross Fusion },
  author={ Ziyu Guo and Young Yoon Lee and Joseph Liu and Yizhak Ben-Shabat and Victor Zordan and Mubbasir Kapadia },
  journal={arXiv preprint arXiv:2503.21775},
  year={ 2025 }
}
Comments on this paper