TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning Diverse Stochastic Human-Action Generators by Lea...

Learning Diverse Stochastic Human-Action Generators by Learning Smooth Latent Transitions

Zhenyi Wang, Ping Yu, Yang Zhao, Ruiyi Zhang, Yufan Zhou, Junsong Yuan, Changyou Chen

2019-12-21AAAI 2019 12Action GenerationMotion GenerationHuman action generation
PaperPDFCode(official)

Abstract

Human-motion generation is a long-standing challenging task due to the requirement of accurately modeling complex and diverse dynamic patterns. Most existing methods adopt sequence models such as RNN to directly model transitions in the original action space. Due to high dimensionality and potential noise, such modeling of action transitions is particularly challenging. In this paper, we focus on skeleton-based action generation and propose to model smooth and diverse transitions on a latent space of action sequences with much lower dimensionality. Conditioned on a latent sequence, actions are generated by a frame-wise decoder shared by all latent action-poses. Specifically, an implicit RNN is defined to model smooth latent sequences, whose randomness (diversity) is controlled by noise from the input. Different from standard action-prediction methods, our model can generate action sequences from pure noise without any conditional action poses. Remarkably, it can also generate unseen actions from mixed classes during training. Our model is learned with a bi-directional generative-adversarial-net framework, which not only can generate diverse action sequences of a particular class or mix classes, but also learns to classify action sequences within the same model. Experimental results show the superiority of our method in both diverse action-sequence generation and classification, relative to existing methods.

Results

TaskDatasetMetricValueModel
Activity RecognitionNTU RGB+D 2DMMDa (CS)0.338c-SkeletonGAN
Activity RecognitionNTU RGB+D 2DMMDa (CV)0.371c-SkeletonGAN
Activity RecognitionNTU RGB+D 2DMMDs (CS)0.402c-SkeletonGAN
Activity RecognitionNTU RGB+D 2DMMDs (CV)0.398c-SkeletonGAN
Activity RecognitionHuman3.6MMMDa0.195Learning Diverse Stochastic Human-Action Generators by Learning Smooth Latent Transitions
Activity RecognitionHuman3.6MMMDs0.218Learning Diverse Stochastic Human-Action Generators by Learning Smooth Latent Transitions
Human action generationNTU RGB+D 2DMMDa (CS)0.338c-SkeletonGAN
Human action generationNTU RGB+D 2DMMDa (CV)0.371c-SkeletonGAN
Human action generationNTU RGB+D 2DMMDs (CS)0.402c-SkeletonGAN
Human action generationNTU RGB+D 2DMMDs (CV)0.398c-SkeletonGAN
Human action generationHuman3.6MMMDa0.195Learning Diverse Stochastic Human-Action Generators by Learning Smooth Latent Transitions
Human action generationHuman3.6MMMDs0.218Learning Diverse Stochastic Human-Action Generators by Learning Smooth Latent Transitions

Related Papers

VITA: Vision-to-Action Flow Matching Policy2025-07-17SnapMoGen: Human Motion Generation from Expressive Texts2025-07-12Go to Zero: Towards Zero-shot Motion Generation with Million-scale Data2025-07-09Motion Generation: A Survey of Generative Approaches and Benchmarks2025-07-07A Unified Transformer-Based Framework with Pretraining For Whole Body Grasping Motion Generation2025-07-01WorldVLA: Towards Autoregressive Action World Model2025-06-26Flow-Based Single-Step Completion for Efficient and Expressive Policy Learning2025-06-26Parallels Between VLA Model Post-Training and Human Motor Learning: Progress, Challenges, and Trends2025-06-26