ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.10568
108
0

DreamActor-H1: High-Fidelity Human-Product Demonstration Video Generation via Motion-designed Diffusion Transformers

12 June 2025
Lizhen Wang
Zhurong Xia
T. Hu
P. Wang
Pengfei Wang
Zerong Zheng
Ming Zhou
    DiffMVGen
ArXiv (abs)PDFHTML
Main:10 Pages
7 Figures
Bibliography:5 Pages
2 Tables
Abstract

In e-commerce and digital marketing, generating high-fidelity human-product demonstration videos is important for effective product presentation. However, most existing frameworks either fail to preserve the identities of both humans and products or lack an understanding of human-product spatial relationships, leading to unrealistic representations and unnatural interactions. To address these challenges, we propose a Diffusion Transformer (DiT)-based framework. Our method simultaneously preserves human identities and product-specific details, such as logos and textures, by injecting paired human-product reference information and utilizing an additional masked cross-attention mechanism. We employ a 3D body mesh template and product bounding boxes to provide precise motion guidance, enabling intuitive alignment of hand gestures with product placements. Additionally, structured text encoding is used to incorporate category-level semantics, enhancing 3D consistency during small rotational changes across frames. Trained on a hybrid dataset with extensive data augmentation strategies, our approach outperforms state-of-the-art techniques in maintaining the identity integrity of both humans and products and generating realistic demonstration motions. Project page:this https URL.

View on arXiv
@article{wang2025_2506.10568,
  title={ DreamActor-H1: High-Fidelity Human-Product Demonstration Video Generation via Motion-designed Diffusion Transformers },
  author={ Lizhen Wang and Zhurong Xia and Tianshu Hu and Pengrui Wang and Pengfei Wang and Zerong Zheng and Ming Zhou },
  journal={arXiv preprint arXiv:2506.10568},
  year={ 2025 }
}
Comments on this paper