TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Trimap-guided Feature Mining and Fusion Network for Natura...

Trimap-guided Feature Mining and Fusion Network for Natural Image Matting

Weihao Jiang, Dongdong Yu, Zhaozhi Xie, Yaoyi Li, Zehuan Yuan, Hongtao Lu

2021-12-01Image Matting
PaperPDFCode(official)

Abstract

Utilizing trimap guidance and fusing multi-level features are two important issues for trimap-based matting with pixel-level prediction. To utilize trimap guidance, most existing approaches simply concatenate trimaps and images together to feed a deep network or apply an extra network to extract more trimap guidance, which meets the conflict between efficiency and effectiveness. For emerging content-based feature fusion, most existing matting methods only focus on local features which lack the guidance of a global feature with strong semantic information related to the interesting object. In this paper, we propose a trimap-guided feature mining and fusion network consisting of our trimap-guided non-background multi-scale pooling (TMP) module and global-local context-aware fusion (GLF) modules. Considering that trimap provides strong semantic guidance, our TMP module focuses effective feature mining on interesting objects under the guidance of trimap without extra parameters. Furthermore, our GLF modules use global semantic information of interesting objects mined by our TMP module to guide an effective global-local context-aware multi-level feature fusion. In addition, we build a common interesting object matting (CIOM) dataset to advance high-quality image matting. Particularly, results on the Composition-1k and our CIOM show that our TMFNet achieves 13% and 25% relative improvement on SAD, respectively, against a strong baseline with fewer parameters and 14% fewer FLOPs. Experimental results on the Composition-1k test set, Alphamatting benchmark, and our CIOM test set demonstrate that our method outperforms state-of-the-art approaches. Our code and models are available at https://github.com/Serge-weihao/TMF-Matting.

Results

TaskDatasetMetricValueModel
Image MattingComposition-1KConn17.6TMFNet
Image MattingComposition-1KGrad6.7TMFNet
Image MattingComposition-1KMSE3.6TMFNet
Image MattingComposition-1KSAD22.1TMFNet

Related Papers

Post-Training Quantization for Video Matting2025-06-12MP-Mat: A 3D-and-Instance-Aware Human Matting and Editing Framework with Multiplane Representation2025-04-20MaSS13K: A Matting-level Semantic Segmentation Benchmark2025-03-24VRMDiff: Text-Guided Video Referring Matting Generation of Diffusion2025-03-11Path-Adaptive Matting for Efficient Inference Under Various Computational Cost Constraints2025-03-05Object-Aware Video Matting with Cross-Frame Guidance2025-03-03Enhancing Image Matting in Real-World Scenes with Mask-Guided Iterative Refinement2025-02-24Efficient Portrait Matte Creation With Layer Diffusion and Connectivity Priors2025-01-27