ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.03118
44
0

HumanRAM: Feed-forward Human Reconstruction and Animation Model using Transformers

3 June 2025
Zhiyuan Yu
Zhe Li
Hujun Bao
Can Yang
Xiaowei Zhou
    3DH
ArXiv (abs)PDFHTML
Main:9 Pages
8 Figures
Bibliography:4 Pages
5 Tables
Abstract

3D human reconstruction and animation are long-standing topics in computer graphics and vision. However, existing methods typically rely on sophisticated dense-view capture and/or time-consuming per-subject optimization procedures. To address these limitations, we propose HumanRAM, a novel feed-forward approach for generalizable human reconstruction and animation from monocular or sparse human images. Our approach integrates human reconstruction and animation into a unified framework by introducing explicit pose conditions, parameterized by a shared SMPL-X neural texture, into transformer-based large reconstruction models (LRM). Given monocular or sparse input images with associated camera parameters and SMPL-X poses, our model employs scalable transformers and a DPT-based decoder to synthesize realistic human renderings under novel viewpoints and novel poses. By leveraging the explicit pose conditions, our model simultaneously enables high-quality human reconstruction and high-fidelity pose-controlled animation. Experiments show that HumanRAM significantly surpasses previous methods in terms of reconstruction accuracy, animation fidelity, and generalization performance on real-world datasets. Video results are available atthis https URL.

View on arXiv
@article{yu2025_2506.03118,
  title={ HumanRAM: Feed-forward Human Reconstruction and Animation Model using Transformers },
  author={ Zhiyuan Yu and Zhe Li and Hujun Bao and Can Yang and Xiaowei Zhou },
  journal={arXiv preprint arXiv:2506.03118},
  year={ 2025 }
}
Comments on this paper