TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Digital Life Project: Autonomous 3D Characters with Social...

Digital Life Project: Autonomous 3D Characters with Social Intelligence

Zhongang Cai, Jianping Jiang, Zhongfei Qing, Xinying Guo, Mingyuan Zhang, Zhengyu Lin, Haiyi Mei, Chen Wei, Ruisi Wang, Wanqi Yin, Xiangyu Fan, Han Du, Liang Pan, Peng Gao, Zhitao Yang, Yang Gao, Jiaqi Li, Tianxiang Ren, Yukun Wei, Xiaogang Wang, Chen Change Loy, Lei Yang, Ziwei Liu

2023-12-07CVPR 2024 1Motion CaptioningMotion GenerationMotion Synthesis
PaperPDF

Abstract

In this work, we present Digital Life Project, a framework utilizing language as the universal medium to build autonomous 3D characters, who are capable of engaging in social interactions and expressing with articulated body motions, thereby simulating life in a digital environment. Our framework comprises two primary components: 1) SocioMind: a meticulously crafted digital brain that models personalities with systematic few-shot exemplars, incorporates a reflection process based on psychology principles, and emulates autonomy by initiating dialogue topics; 2) MoMat-MoGen: a text-driven motion synthesis paradigm for controlling the character's digital body. It integrates motion matching, a proven industry technique to ensure motion quality, with cutting-edge advancements in motion generation for diversity. Extensive experiments demonstrate that each module achieves state-of-the-art performance in its respective domain. Collectively, they enable virtual characters to initiate and sustain dialogues autonomously, while evolving their socio-psychological states. Concurrently, these characters can perform contextually relevant bodily movements. Additionally, a motion captioning module further allows the virtual character to recognize and appropriately respond to human players' actions. Homepage: https://digital-life-project.com/

Results

TaskDatasetMetricValueModel
Pose TrackingInterHumanFID5.674MoMat-MoGen
Pose TrackingInterHumanMMDist3.79MoMat-MoGen
Pose TrackingInterHumanMModality1.295MoMat-MoGen
Pose TrackingInterHumanR-Precision Top30.666MoMat-MoGen
Motion SynthesisInterHumanFID5.674MoMat-MoGen
Motion SynthesisInterHumanMMDist3.79MoMat-MoGen
Motion SynthesisInterHumanMModality1.295MoMat-MoGen
Motion SynthesisInterHumanR-Precision Top30.666MoMat-MoGen
10-shot image generationInterHumanFID5.674MoMat-MoGen
10-shot image generationInterHumanMMDist3.79MoMat-MoGen
10-shot image generationInterHumanMModality1.295MoMat-MoGen
10-shot image generationInterHumanR-Precision Top30.666MoMat-MoGen
3D Human Pose TrackingInterHumanFID5.674MoMat-MoGen
3D Human Pose TrackingInterHumanMMDist3.79MoMat-MoGen
3D Human Pose TrackingInterHumanMModality1.295MoMat-MoGen
3D Human Pose TrackingInterHumanR-Precision Top30.666MoMat-MoGen

Related Papers

SnapMoGen: Human Motion Generation from Expressive Texts2025-07-12Go to Zero: Towards Zero-shot Motion Generation with Million-scale Data2025-07-09Motion Generation: A Survey of Generative Approaches and Benchmarks2025-07-07DeepGesture: A conversational gesture synthesis system based on emotions and semantics2025-07-03A Unified Transformer-Based Framework with Pretraining For Whole Body Grasping Motion Generation2025-07-01VolumetricSMPL: A Neural Volumetric Body Model for Efficient Interactions, Contacts, and Collisions2025-06-29DuetGen: Music Driven Two-Person Dance Generation via Hierarchical Masked Modeling2025-06-23PlanMoGPT: Flow-Enhanced Progressive Planning for Text to Motion Synthesis2025-06-22