TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/ColorMAE: Exploring data-independent masking strategies in...

ColorMAE: Exploring data-independent masking strategies in Masked AutoEncoders

Carlos Hinojosa, Shuming Liu, Bernard Ghanem

2024-07-17Image ClassificationSemantic SegmentationInstance SegmentationObject Detection
PaperPDFCode(official)

Abstract

Masked AutoEncoders (MAE) have emerged as a robust self-supervised framework, offering remarkable performance across a wide range of downstream tasks. To increase the difficulty of the pretext task and learn richer visual representations, existing works have focused on replacing standard random masking with more sophisticated strategies, such as adversarial-guided and teacher-guided masking. However, these strategies depend on the input data thus commonly increasing the model complexity and requiring additional calculations to generate the mask patterns. This raises the question: Can we enhance MAE performance beyond random masking without relying on input data or incurring additional computational costs? In this work, we introduce a simple yet effective data-independent method, termed ColorMAE, which generates different binary mask patterns by filtering random noise. Drawing inspiration from color noise in image processing, we explore four types of filters to yield mask patterns with different spatial and semantic priors. ColorMAE requires no additional learnable parameters or computational overhead in the network, yet it significantly enhances the learned representations. We provide a comprehensive empirical evaluation, demonstrating our strategy's superiority in downstream tasks compared to random masking. Notably, we report an improvement of 2.72 in mIoU in semantic segmentation tasks relative to baseline MAE implementations.

Results

TaskDatasetMetricValueModel
Semantic SegmentationADE20KValidation mIoU49.3ColorMAE-Green-ViTB-1600
Object DetectionCOCOboxAP50.1ColorMAE-Green-ViTB-1600
Object DetectionCOCOboxAP5070.7ColorMAE-Green-ViTB-1600
Object DetectionCOCOboxAP7554.7ColorMAE-Green-ViTB-1600
3DCOCOboxAP50.1ColorMAE-Green-ViTB-1600
3DCOCOboxAP5070.7ColorMAE-Green-ViTB-1600
3DCOCOboxAP7554.7ColorMAE-Green-ViTB-1600
Instance SegmentationCOCOmaskAP44.4ColorMAE-Green-ViTB-1600
Instance SegmentationCOCOmaskAP5067.8ColorMAE-Green-ViTB-1600
Instance SegmentationCOCOmaskAP7548ColorMAE-Green-ViTB-1600
2D ClassificationCOCOboxAP50.1ColorMAE-Green-ViTB-1600
2D ClassificationCOCOboxAP5070.7ColorMAE-Green-ViTB-1600
2D ClassificationCOCOboxAP7554.7ColorMAE-Green-ViTB-1600
2D Object DetectionCOCOboxAP50.1ColorMAE-Green-ViTB-1600
2D Object DetectionCOCOboxAP5070.7ColorMAE-Green-ViTB-1600
2D Object DetectionCOCOboxAP7554.7ColorMAE-Green-ViTB-1600
10-shot image generationADE20KValidation mIoU49.3ColorMAE-Green-ViTB-1600
16kCOCOboxAP50.1ColorMAE-Green-ViTB-1600
16kCOCOboxAP5070.7ColorMAE-Green-ViTB-1600
16kCOCOboxAP7554.7ColorMAE-Green-ViTB-1600

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21Automatic Classification and Segmentation of Tunnel Cracks Based on Deep Learning and Visual Explanations2025-07-18Adversarial attacks to image classification systems using evolutionary algorithms2025-07-17Efficient Adaptation of Pre-trained Vision Transformer underpinned by Approximately Orthogonal Fine-Tuning Strategy2025-07-17Federated Learning for Commercial Image Sources2025-07-17MUPAX: Multidimensional Problem Agnostic eXplainable AI2025-07-17DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17