TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Learning to Fuse Asymmetric Feature Maps in Siamese Trackers

Learning to Fuse Asymmetric Feature Maps in Siamese Trackers

Wencheng Han, Xingping Dong, Fahad Shahbaz Khan, Ling Shao, Jianbing Shen

2020-12-04CVPR 2021 1Visual Object TrackingVisual TrackingVideo Object Tracking
PaperPDFCode(official)

Abstract

Recently, Siamese-based trackers have achieved promising performance in visual tracking. Most recent Siamese-based trackers typically employ a depth-wise cross-correlation (DW-XCorr) to obtain multi-channel correlation information from the two feature maps (target and search region). However, DW-XCorr has several limitations within Siamese-based tracking: it can easily be fooled by distractors, has fewer activated channels, and provides weak discrimination of object boundaries. Further, DW-XCorr is a handcrafted parameter-free module and cannot fully benefit from offline learning on large-scale data. We propose a learnable module, called the asymmetric convolution (ACM), which learns to better capture the semantic correlation information in offline training on large-scale data. Different from DW-XCorr and its predecessor(XCorr), which regard a single feature map as the convolution kernel, our ACM decomposes the convolution operation on a concatenated feature map into two mathematically equivalent operations, thereby avoiding the need for the feature maps to be of the same size (width and height)during concatenation. Our ACM can incorporate useful prior information, such as bounding-box size, with standard visual features. Furthermore, ACM can easily be integrated into existing Siamese trackers based on DW-XCorror XCorr. To demonstrate its generalization ability, we integrate ACM into three representative trackers: SiamFC, SiamRPN++, and SiamBAN. Our experiments reveal the benefits of the proposed ACM, which outperforms existing methods on six tracking benchmarks. On the LaSOT test set, our ACM-based tracker obtains a significant improvement of 5.8% in terms of success (AUC), over the baseline.

Results

TaskDatasetMetricValueModel
VideoNT-VOT211AUC35.8SiamBAN-ACM
VideoNT-VOT211Precision48.31SiamBAN-ACM
Object TrackingLaSOTAUC57.2SiamBAN-ACM
Object TrackingLaSOTNormalized Precision65.3SiamBAN-ACM
Object TrackingLaSOTPrecision58.7SiamBAN-ACM
Object TrackingTrackingNetAccuracy75.3SiamBAN-ACM
Object TrackingTrackingNetNormalized Precision81SiamBAN-ACM
Object TrackingTrackingNetPrecision71.2SiamBAN-ACM
Object TrackingNT-VOT211AUC35.8SiamBAN-ACM
Object TrackingNT-VOT211Precision48.31SiamBAN-ACM
Visual Object TrackingLaSOTAUC57.2SiamBAN-ACM
Visual Object TrackingLaSOTNormalized Precision65.3SiamBAN-ACM
Visual Object TrackingLaSOTPrecision58.7SiamBAN-ACM
Visual Object TrackingTrackingNetAccuracy75.3SiamBAN-ACM
Visual Object TrackingTrackingNetNormalized Precision81SiamBAN-ACM
Visual Object TrackingTrackingNetPrecision71.2SiamBAN-ACM

Related Papers

HiM2SAM: Enhancing SAM2 with Hierarchical Motion Estimation and Memory Optimization towards Long-term Tracking2025-07-10What You Have is What You Track: Adaptive and Robust Multimodal Tracking2025-07-08UMDATrack: Unified Multi-Domain Adaptive Tracking Under Adverse Weather Conditions2025-07-01Mamba-FETrack V2: Revisiting State Space Model for Frame-Event based Visual Object Tracking2025-06-30R1-Track: Direct Application of MLLMs to Visual Object Tracking via Reinforcement Learning2025-06-27Exploiting Lightweight Hierarchical ViT and Dynamic Framework for Efficient Visual Tracking2025-06-25Comparison of Two Methods for Stationary Incident Detection Based on Background Image2025-06-17Towards Effective and Efficient Adversarial Defense with Diffusion Models for Robust Visual Tracking2025-05-31