TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SCANimate: Weakly Supervised Learning of Skinned Clothed A...

SCANimate: Weakly Supervised Learning of Skinned Clothed Avatar Networks

Shunsuke Saito, Jinlong Yang, Qianli Ma, Michael J. Black

2021-04-07CVPR 2021 13D Human Reconstruction
PaperPDFCodeCode(official)

Abstract

We present SCANimate, an end-to-end trainable framework that takes raw 3D scans of a clothed human and turns them into an animatable avatar. These avatars are driven by pose parameters and have realistic clothing that moves and deforms naturally. SCANimate does not rely on a customized mesh template or surface mesh registration. We observe that fitting a parametric 3D body model, like SMPL, to a clothed human scan is tractable while surface registration of the body topology to the scan is often not, because clothing can deviate significantly from the body shape. We also observe that articulated transformations are invertible, resulting in geometric cycle consistency in the posed and unposed shapes. These observations lead us to a weakly supervised learning method that aligns scans into a canonical pose by disentangling articulated deformations without template-based surface registration. Furthermore, to complete missing regions in the aligned scans while modeling pose-dependent deformations, we introduce a locally pose-aware implicit function that learns to complete and model geometry with learned pose correctives. In contrast to commonly used global pose embeddings, our local pose conditioning significantly reduces long-range spurious correlations and improves generalization to unseen poses, especially when training data is limited. Our method can be applied to pose-aware appearance modeling to generate a fully textured avatar. We demonstrate our approach on various clothing types with different amounts of training data, outperforming existing solutions and other variants in terms of fidelity and generality in every setting. The code is available at https://scanimate.is.tue.mpg.de.

Results

TaskDatasetMetricValueModel
Reconstruction4D-DRESSChamfer (cm)0.965SCANimate_Inner
Reconstruction4D-DRESSIoU0.918SCANimate_Inner
Reconstruction4D-DRESSNormal Consistency0.854SCANimate_Inner
Reconstruction4D-DRESSChamfer (cm)1.237SCANimate_Outer
Reconstruction4D-DRESSIoU0.912SCANimate_Outer
Reconstruction4D-DRESSNormal Consistency0.828SCANimate_Outer

Related Papers

PF-LHM: 3D Animatable Avatar Reconstruction from Pose-free Articulated Human Images2025-06-16SMPL Normal Map Is All You Need for Single-view Textured Human Reconstruction2025-06-15HumanRAM: Feed-forward Human Reconstruction and Animation Model using Transformers2025-06-03Link to the Past: Temporal Propagation for Fast 3D Human Reconstruction from Monocular Video2025-05-12DeClotH: Decomposable 3D Cloth and Human Body Reconstruction from a Single Image2025-03-25CHROME: Clothed Human Reconstruction with Occlusion-Resilience and Multiview-Consistency from a Single Image2025-03-19LHM: Large Animatable Human Reconstruction Model from a Single Image in Seconds2025-03-13MVD-HuGaS: Human Gaussians from a Single Image via 3D Human Multi-view Diffusion Prior2025-03-11