TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Revisiting RGBT Tracking Benchmarks from the Perspective o...

Revisiting RGBT Tracking Benchmarks from the Perspective of Modality Validity: A New Benchmark, Problem, and Method

Zhangyong Tang, Tianyang Xu, ZhenHua Feng, XueFeng Zhu, He Wang, Pengcheng Shao, Chunyang Cheng, Xiao-Jun Wu, Muhammad Awais, Sara Atito, Josef Kittler

2024-04-30Rgb-T Tracking
PaperPDFCode(official)

Abstract

RGBT tracking draws increasing attention due to its robustness in multi-modality warranting (MMW) scenarios, such as nighttime and bad weather, where relying on a single sensing modality fails to ensure stable tracking results. However, the existing benchmarks predominantly consist of videos collected in common scenarios where both RGB and thermal infrared (TIR) information are of sufficient quality. This makes the data unrepresentative of severe imaging conditions, leading to tracking failures in MMW scenarios. To bridge this gap, we present a new benchmark, MV-RGBT, captured specifically in MMW scenarios. In contrast with the existing datasets, MV-RGBT comprises more object categories and scenes, providing a diverse and challenging benchmark. Furthermore, for severe imaging conditions of MMW scenarios, a new problem is posed, namely \textit{when to fuse}, to stimulate the development of fusion strategies for such data. We propose a new method based on a mixture of experts, namely MoETrack, as a baseline fusion strategy. In MoETrack, each expert generates independent tracking results along with the corresponding confidence score, which is used to control the fusion process. Extensive experimental results demonstrate the significant potential of MV-RGBT in advancing RGBT tracking and elicit the conclusion that fusion is not always beneficial, especially in MMW scenarios. Significantly, the proposed MoETrack method achieves new state-of-the-art results not only on MV-RGBT, but also on standard benchmarks, such as RGBT234, LasHeR, and the short-term split of VTUAV (VTUAV-ST). More information of MV-RGBT and the source code of MoETrack will be released at https://github.com/Zhangyong-Tang/MoETrack.

Results

TaskDatasetMetricValueModel
Visual TrackingLasHeRPrecision72.1MoETrack
Visual TrackingLasHeRSuccess57.8MoETrack
Visual TrackingGTOTPrecision93.6MoETrack
Visual TrackingGTOTSuccess78.4MoETrack
Visual TrackingRGBT234Precision88.1MoETrack
Visual TrackingRGBT234Success65.1MoETrack

Related Papers

Lightweight RGB-T Tracking with Mobile Vision Transformers2025-06-23Modality-Guided Dynamic Graph Fusion and Temporal Diffusion for Self-Supervised RGB-T Tracking2025-05-06Breaking Shallow Limits: Task-Driven Pixel Fusion for Gap-free RGBT Tracking2025-03-14Adaptive Perception for Unified Visual Multi-modal Object Tracking2025-02-10BTMTrack: Robust RGB-T Tracking via Dual-template Bridging and Temporal-Modal Candidate Elimination2025-01-07PURA: Parameter Update-Recovery Test-Time Adaption for RGB-T Tracking2025-01-01SUTrack: Towards Simple and Unified Single Object Tracking2024-12-26Exploiting Multimodal Spatial-temporal Patterns for Video Object Tracking2024-12-20