TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Explicit Visual Prompting for Universal Foreground Segment...

Explicit Visual Prompting for Universal Foreground Segmentations

Weihuang Liu, Xi Shen, Chi-Man Pun, Xiaodong Cun

2023-05-29Foreground SegmentationDefocus Blur DetectionShadow DetectionCamouflaged Object Segmentationparameter-efficient fine-tuningSalient Object DetectionImage Manipulation Detection
PaperPDFCode(official)Code

Abstract

Foreground segmentation is a fundamental problem in computer vision, which includes salient object detection, forgery detection, defocus blur detection, shadow detection, and camouflage object detection. Previous works have typically relied on domain-specific solutions to address accuracy and robustness issues in those applications. In this paper, we present a unified framework for a number of foreground segmentation tasks without any task-specific designs. We take inspiration from the widely-used pre-training and then prompt tuning protocols in NLP and propose a new visual prompting model, named Explicit Visual Prompting (EVP). Different from the previous visual prompting which is typically a dataset-level implicit embedding, our key insight is to enforce the tunable parameters focusing on the explicit visual content from each individual image, i.e., the features from frozen patch embeddings and high-frequency components. Our method freezes a pre-trained model and then learns task-specific knowledge using a few extra parameters. Despite introducing only a small number of tunable parameters, EVP achieves superior performance than full fine-tuning and other parameter-efficient fine-tuning methods. Experiments in fourteen datasets across five tasks show the proposed method outperforms other task-specific methods while being considerably simple. The proposed method demonstrates the scalability in different architectures, pre-trained weights, and tasks. The code is available at: https://github.com/NiFangBaAGe/Explicit-Visual-Prompt.

Results

TaskDatasetMetricValueModel
Object DetectionCODMAE0.029EVPv2
Object DetectionCODS-Measure0.843EVPv2
Object DetectionCODWeighted F-Measure0.746EVPv2
Object DetectionCAMOMAE0.058EVPv2
Object DetectionCAMOS-Measure0.848EVPv2
Object DetectionCAMOWeighted F-Measure0.786EVPv2
3DCODMAE0.029EVPv2
3DCODS-Measure0.843EVPv2
3DCODWeighted F-Measure0.746EVPv2
3DCAMOMAE0.058EVPv2
3DCAMOS-Measure0.848EVPv2
3DCAMOWeighted F-Measure0.786EVPv2
Camouflaged Object SegmentationCODMAE0.029EVPv2
Camouflaged Object SegmentationCODS-Measure0.843EVPv2
Camouflaged Object SegmentationCODWeighted F-Measure0.746EVPv2
Camouflaged Object SegmentationCAMOMAE0.058EVPv2
Camouflaged Object SegmentationCAMOS-Measure0.848EVPv2
Camouflaged Object SegmentationCAMOWeighted F-Measure0.786EVPv2
Object SegmentationCODMAE0.029EVPv2
Object SegmentationCODS-Measure0.843EVPv2
Object SegmentationCODWeighted F-Measure0.746EVPv2
Object SegmentationCAMOMAE0.058EVPv2
Object SegmentationCAMOS-Measure0.848EVPv2
Object SegmentationCAMOWeighted F-Measure0.786EVPv2
Salient Object DetectionECSSDE-measure0.957EVPv2
Salient Object DetectionECSSDMAE0.028EVPv2
Salient Object DetectionECSSDS-measure0.935EVPv2
Salient Object DetectionECSSDmax_F10.958EVPv2
Salient Object DetectionDUT-OMRONE-measure0.895EVPv2
Salient Object DetectionDUT-OMRONMAE0.047EVPv2
Salient Object DetectionDUT-OMRONS-measure0.862EVPv2
Salient Object DetectionDUT-OMRONmax_F10.857EVPv2
Salient Object DetectionDUTS-TEE-measure0.948EVPv2
Salient Object DetectionDUTS-TEMAE0.027EVPv2
Salient Object DetectionDUTS-TESmeasure0.915EVPv2
Salient Object DetectionDUTS-TEmax_F10.923EVPv2
Salient Object DetectionPASCAL-SE-measure0.917EVPv2
Salient Object DetectionPASCAL-SMAE0.053EVPv2
Salient Object DetectionPASCAL-SS-measure0.879EVPv2
Salient Object DetectionPASCAL-Smax_F10.869EVPv2
2D ClassificationCODMAE0.029EVPv2
2D ClassificationCODS-Measure0.843EVPv2
2D ClassificationCODWeighted F-Measure0.746EVPv2
2D ClassificationCAMOMAE0.058EVPv2
2D ClassificationCAMOS-Measure0.848EVPv2
2D ClassificationCAMOWeighted F-Measure0.786EVPv2
2D Object DetectionCODMAE0.029EVPv2
2D Object DetectionCODS-Measure0.843EVPv2
2D Object DetectionCODWeighted F-Measure0.746EVPv2
2D Object DetectionCAMOMAE0.058EVPv2
2D Object DetectionCAMOS-Measure0.848EVPv2
2D Object DetectionCAMOWeighted F-Measure0.786EVPv2
16kCODMAE0.029EVPv2
16kCODS-Measure0.843EVPv2
16kCODWeighted F-Measure0.746EVPv2
16kCAMOMAE0.058EVPv2
16kCAMOS-Measure0.848EVPv2
16kCAMOWeighted F-Measure0.786EVPv2

Related Papers

Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17LoSiA: Efficient High-Rank Fine-Tuning via Subnet Localization and Optimization2025-07-06DC-TTA: Divide-and-Conquer Framework for Test-Time Adaptation of Interactive Segmentation2025-06-29Seg-R1: Segmentation Can Be Surprisingly Simple with Reinforcement Learning2025-06-27Exploring Adapter Design Tradeoffs for Low Resource Music Generation2025-06-26WordCon: Word-level Typography Control in Scene Text Rendering2025-06-26Optimising Language Models for Downstream Tasks: A Post-Training Perspective2025-06-26Progtuning: Progressive Fine-tuning Framework for Transformer-based Language Models2025-06-26