ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2506.09411
79
0

Synthetic Human Action Video Data Generation with Pose Transfer

11 June 2025
Vaclav Knapp
Matyas Bohacek
ArXiv (abs)PDFHTML
Main:7 Pages
13 Figures
Bibliography:3 Pages
1 Tables
Appendix:7 Pages
Abstract

In video understanding tasks, particularly those involving human motion, synthetic data generation often suffers from uncanny features, diminishing its effectiveness for training. Tasks such as sign language translation, gesture recognition, and human motion understanding in autonomous driving have thus been unable to exploit the full potential of synthetic data. This paper proposes a method for generating synthetic human action video data using pose transfer (specifically, controllable 3D Gaussian avatar models). We evaluate this method on the Toyota Smarthome and NTU RGB+D datasets and show that it improves performance in action recognition tasks. Moreover, we demonstrate that the method can effectively scale few-shot datasets, making up for groups underrepresented in the real training data and adding diverse backgrounds. We open-source the method along with RANDOM People, a dataset with videos and avatars of novel human identities for pose transfer crowd-sourced from the internet.

View on arXiv
@article{knapp2025_2506.09411,
  title={ Synthetic Human Action Video Data Generation with Pose Transfer },
  author={ Vaclav Knapp and Matyas Bohacek },
  journal={arXiv preprint arXiv:2506.09411},
  year={ 2025 }
}
Comments on this paper