TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MotionLCM: Real-time Controllable Motion Generation via La...

MotionLCM: Real-time Controllable Motion Generation via Latent Consistency Model

Wenxun Dai, Ling-Hao Chen, Jingbo Wang, Jinpeng Liu, Bo Dai, Yansong Tang

2024-04-30Motion GenerationMotion Synthesis
PaperPDFCode(official)

Abstract

This work introduces MotionLCM, extending controllable motion generation to a real-time level. Existing methods for spatial-temporal control in text-conditioned motion generation suffer from significant runtime inefficiency. To address this issue, we first propose the motion latent consistency model (MotionLCM) for motion generation, building on the motion latent diffusion model. By adopting one-step (or few-step) inference, we further improve the runtime efficiency of the motion latent diffusion model for motion generation. To ensure effective controllability, we incorporate a motion ControlNet within the latent space of MotionLCM and enable explicit control signals (i.e., initial motions) in the vanilla motion space to further provide supervision for the training process. By employing these techniques, our approach can generate human motions with text and control signals in real-time. Experimental results demonstrate the remarkable generation and controlling capabilities of MotionLCM while maintaining real-time runtime efficiency.

Results

TaskDatasetMetricValueModel
Pose TrackingHumanML3DDiversity9.607MotionLCM (4-step)
Pose TrackingHumanML3DFID0.304MotionLCM (4-step)
Pose TrackingHumanML3DMultimodality2.259MotionLCM (4-step)
Pose TrackingHumanML3DR Precision Top30.798MotionLCM (4-step)
Motion SynthesisHumanML3DDiversity9.607MotionLCM (4-step)
Motion SynthesisHumanML3DFID0.304MotionLCM (4-step)
Motion SynthesisHumanML3DMultimodality2.259MotionLCM (4-step)
Motion SynthesisHumanML3DR Precision Top30.798MotionLCM (4-step)
10-shot image generationHumanML3DDiversity9.607MotionLCM (4-step)
10-shot image generationHumanML3DFID0.304MotionLCM (4-step)
10-shot image generationHumanML3DMultimodality2.259MotionLCM (4-step)
10-shot image generationHumanML3DR Precision Top30.798MotionLCM (4-step)
3D Human Pose TrackingHumanML3DDiversity9.607MotionLCM (4-step)
3D Human Pose TrackingHumanML3DFID0.304MotionLCM (4-step)
3D Human Pose TrackingHumanML3DMultimodality2.259MotionLCM (4-step)
3D Human Pose TrackingHumanML3DR Precision Top30.798MotionLCM (4-step)

Related Papers

SnapMoGen: Human Motion Generation from Expressive Texts2025-07-12Go to Zero: Towards Zero-shot Motion Generation with Million-scale Data2025-07-09Motion Generation: A Survey of Generative Approaches and Benchmarks2025-07-07DeepGesture: A conversational gesture synthesis system based on emotions and semantics2025-07-03A Unified Transformer-Based Framework with Pretraining For Whole Body Grasping Motion Generation2025-07-01VolumetricSMPL: A Neural Volumetric Body Model for Efficient Interactions, Contacts, and Collisions2025-06-29DuetGen: Music Driven Two-Person Dance Generation via Hierarchical Masked Modeling2025-06-23PlanMoGPT: Flow-Enhanced Progressive Planning for Text to Motion Synthesis2025-06-22