TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MODNet: Real-Time Trimap-Free Portrait Matting via Objecti...

MODNet: Real-Time Trimap-Free Portrait Matting via Objective Decomposition

Zhanghan Ke, Jiayu Sun, Kaican Li, Qiong Yan, Rynson W. H. Lau

2020-11-24Image MattingVideo Matting
PaperPDFCodeCode(official)CodeCode(official)CodeCodeCodeCodeCode

Abstract

Existing portrait matting methods either require auxiliary inputs that are costly to obtain or involve multiple stages that are computationally expensive, making them less suitable for real-time applications. In this work, we present a light-weight matting objective decomposition network (MODNet) for portrait matting in real-time with a single input image. The key idea behind our efficient design is by optimizing a series of sub-objectives simultaneously via explicit constraints. In addition, MODNet includes two novel techniques for improving model efficiency and robustness. First, an Efficient Atrous Spatial Pyramid Pooling (e-ASPP) module is introduced to fuse multi-scale features for semantic estimation. Second, a self-supervised sub-objectives consistency (SOC) strategy is proposed to adapt MODNet to real-world data to address the domain shift problem common to trimap-free methods. MODNet is easy to be trained in an end-to-end manner. It is much faster than contemporaneous methods and runs at 67 frames per second on a 1080Ti GPU. Experiments show that MODNet outperforms prior trimap-free methods by a large margin on both Adobe Matting Dataset and a carefully designed photographic portrait matting (PPM-100) benchmark proposed by us. Further, MODNet achieves remarkable results on daily photos and videos. Our code and models are available at https://github.com/ZHKKKe/MODNet, and the PPM-100 benchmark is released at https://github.com/ZHKKKe/PPM.

Results

TaskDatasetMetricValueModel
Image MattingAMDMAD0.81MODNet+
Image MattingAMDMSE0.0024MODNet+
Image MattingPPM-100MAD0.97MODNet+ (Our)
Image MattingPPM-100MSE0.0046MODNet+ (Our)

Related Papers

Post-Training Quantization for Video Matting2025-06-12MP-Mat: A 3D-and-Instance-Aware Human Matting and Editing Framework with Multiplane Representation2025-04-20MaSS13K: A Matting-level Semantic Segmentation Benchmark2025-03-24VRMDiff: Text-Guided Video Referring Matting Generation of Diffusion2025-03-11Path-Adaptive Matting for Efficient Inference Under Various Computational Cost Constraints2025-03-05Object-Aware Video Matting with Cross-Frame Guidance2025-03-03Enhancing Image Matting in Real-World Scenes with Mask-Guided Iterative Refinement2025-02-24Efficient Portrait Matte Creation With Layer Diffusion and Connectivity Priors2025-01-27