TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Vid2Avatar: 3D Avatar Reconstruction from Videos in the Wi...

Vid2Avatar: 3D Avatar Reconstruction from Videos in the Wild via Self-supervised Scene Decomposition

Chen Guo, Tianjian Jiang, Xu Chen, Jie Song, Otmar Hilliges

2023-02-22CVPR 2023 1Surface Reconstruction3D Human Reconstruction
PaperPDFCode(official)

Abstract

We present Vid2Avatar, a method to learn human avatars from monocular in-the-wild videos. Reconstructing humans that move naturally from monocular in-the-wild videos is difficult. Solving it requires accurately separating humans from arbitrary backgrounds. Moreover, it requires reconstructing detailed 3D surface from short video sequences, making it even more challenging. Despite these challenges, our method does not require any groundtruth supervision or priors extracted from large datasets of clothed human scans, nor do we rely on any external segmentation modules. Instead, it solves the tasks of scene decomposition and surface reconstruction directly in 3D by modeling both the human and the background in the scene jointly, parameterized via two separate neural fields. Specifically, we define a temporally consistent human representation in canonical space and formulate a global optimization over the background model, the canonical human shape and texture, and per-frame human pose parameters. A coarse-to-fine sampling strategy for volume rendering and novel objectives are introduced for a clean separation of dynamic human and static background, yielding detailed and robust 3D human geometry reconstructions. We evaluate our methods on publicly available datasets and show improvements over prior art.

Results

TaskDatasetMetricValueModel
Reconstruction4D-DRESSChamfer (cm)2.87Vid2Avatar_Inner
Reconstruction4D-DRESSIoU0.772Vid2Avatar_Inner
Reconstruction4D-DRESSNormal Consistency0.75Vid2Avatar_Inner
Reconstruction4D-DRESSChamfer (cm)4.027Vid2Avatar_Outer
Reconstruction4D-DRESSIoU0.745Vid2Avatar_Outer
Reconstruction4D-DRESSNormal Consistency0.683Vid2Avatar_Outer

Related Papers

A Mixed-Primitive-based Gaussian Splatting Method for Surface Reconstruction2025-07-15High-Fidelity and Generalizable Neural Surface Reconstruction with Sparse Feature Volumes2025-07-08LoomNet: Enhancing Multi-View Image Generation via Latent Space Weaving2025-07-07HiNeuS: High-fidelity Neural Surface Mitigating Low-texture and Reflective Ambiguity2025-06-30SOF: Sorted Opacity Fields for Fast Unbounded Surface Reconstruction2025-06-233D Gaussian Splatting for Fine-Detailed Surface Reconstruction in Large-Scale Scene2025-06-21Wavelet-based Global Orientation and Surface Reconstruction for Point Clouds2025-06-19HRGS: Hierarchical Gaussian Splatting for Memory-Efficient High-Resolution 3D Reconstruction2025-06-17