TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Target-aware Dual Adversarial Learning and a Multi-scenari...

Target-aware Dual Adversarial Learning and a Multi-scenario Multi-Modality Benchmark to Fuse Infrared and Visible for Object Detection

JinYuan Liu, Xin Fan, Zhanbo Huang, Guanyao Wu, Risheng Liu, Wei Zhong, Zhongxuan Luo

2022-03-30CVPR 2022 1Infrared And Visible Image FusionSemantic Segmentation2D Object Detectionobject-detectionObject Detection
PaperPDFCode(official)Code

Abstract

This study addresses the issue of fusing infrared and visible images that appear differently for object detection. Aiming at generating an image of high visual quality, previous approaches discover commons underlying the two modalities and fuse upon the common space either by iterative optimization or deep networks. These approaches neglect that modality differences implying the complementary information are extremely important for both fusion and subsequent detection task. This paper proposes a bilevel optimization formulation for the joint problem of fusion and detection, and then unrolls to a target-aware Dual Adversarial Learning (TarDAL) network for fusion and a commonly used detection network. The fusion network with one generator and dual discriminators seeks commons while learning from differences, which preserves structural information of targets from the infrared and textural details from the visible. Furthermore, we build a synchronized imaging system with calibrated infrared and optical sensors, and collect currently the most comprehensive benchmark covering a wide range of scenarios. Extensive experiments on several public datasets and our benchmark demonstrate that our method outputs not only visually appealing fusion but also higher detection mAP than the state-of-the-art approaches.

Results

TaskDatasetMetricValueModel
Semantic SegmentationFMB DatasetmIoU48.1TarDAL (RGB-Infrared)
Object DetectionMultispectral DatasetmAP@0.578TarDAL
Object DetectionIndia Driving DatasetmAP@0.581.1
3DMultispectral DatasetmAP@0.578TarDAL
3DIndia Driving DatasetmAP@0.581.1
2D ClassificationMultispectral DatasetmAP@0.578TarDAL
2D ClassificationIndia Driving DatasetmAP@0.581.1
2D Object DetectionDroneVehicletest/mAP43.3TarDAL
2D Object DetectionDroneVehicletest/mAP5072.6TarDAL
2D Object DetectionMultispectral DatasetmAP@0.578TarDAL
2D Object DetectionIndia Driving DatasetmAP@0.581.1
10-shot image generationFMB DatasetmIoU48.1TarDAL (RGB-Infrared)
16kMultispectral DatasetmAP@0.578TarDAL
16kIndia Driving DatasetmAP@0.581.1

Related Papers

SeC: Advancing Complex Video Object Segmentation via Progressive Concept Construction2025-07-21DiffOSeg: Omni Medical Image Segmentation via Multi-Expert Collaboration Diffusion Model2025-07-17SCORE: Scene Context Matters in Open-Vocabulary Remote Sensing Instance Segmentation2025-07-17Unified Medical Image Segmentation with State Space Modeling Snake2025-07-17A Privacy-Preserving Semantic-Segmentation Method Using Domain-Adaptation Technique2025-07-17A Real-Time System for Egocentric Hand-Object Interaction Detection in Industrial Domains2025-07-17RS-TinyNet: Stage-wise Feature Fusion Network for Detecting Tiny Objects in Remote Sensing Images2025-07-17Decoupled PROB: Decoupled Query Initialization Tasks and Objectness-Class Learning for Open World Object Detection2025-07-17