TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Towards Seamless Adaptation of Pre-trained Models for Visu...

Towards Seamless Adaptation of Pre-trained Models for Visual Place Recognition

Feng Lu, Lijun Zhang, Xiangyuan Lan, Shuting Dong, YaoWei Wang, Chun Yuan

2024-02-22Visual Place RecognitionRe-Ranking
PaperPDFCode(official)

Abstract

Recent studies show that vision models pre-trained in generic visual learning tasks with large-scale data can provide useful feature representations for a wide range of visual perception problems. However, few attempts have been made to exploit pre-trained foundation models in visual place recognition (VPR). Due to the inherent difference in training objectives and data between the tasks of model pre-training and VPR, how to bridge the gap and fully unleash the capability of pre-trained models for VPR is still a key issue to address. To this end, we propose a novel method to realize seamless adaptation of pre-trained models for VPR. Specifically, to obtain both global and local features that focus on salient landmarks for discriminating places, we design a hybrid adaptation method to achieve both global and local adaptation efficiently, in which only lightweight adapters are tuned without adjusting the pre-trained model. Besides, to guide effective adaptation, we propose a mutual nearest neighbor local feature loss, which ensures proper dense local features are produced for local matching and avoids time-consuming spatial verification in re-ranking. Experimental results show that our method outperforms the state-of-the-art methods with less training data and training time, and uses about only 3% retrieval runtime of the two-stage VPR methods with RANSAC-based spatial verification. It ranks 1st on the MSLS challenge leaderboard (at the time of submission). The code is released at https://github.com/Lu-Feng/SelaVPR.

Results

TaskDatasetMetricValueModel
Visual Place RecognitionNordlandRecall@186.6SelaVPR
Visual Place RecognitionNordlandRecall@594SelaVPR
Visual Place RecognitionSt LuciaRecall@199.8SelaVPR
Visual Place RecognitionPittsburgh-250k-testRecall@195.7SelaVPR
Visual Place RecognitionPittsburgh-250k-testRecall@1098.8SelaVPR
Visual Place RecognitionPittsburgh-250k-testRecall@599.2SelaVPR
Visual Place RecognitionPittsburgh-30k-testRecall@192.8SelaVPR
Visual Place RecognitionPittsburgh-30k-testRecall@597.7SelaVPR
Visual Place RecognitionTokyo247Recall@194SelaVPR
Visual Place RecognitionTokyo247Recall@1096.8SelaVPR
Visual Place RecognitionTokyo247Recall@597.5SelaVPR
Visual Place RecognitionMapillary valRecall@190.8SelaVPR
Visual Place RecognitionMapillary valRecall@1097.2SelaVPR
Visual Place RecognitionMapillary valRecall@596.4SelaVPR
Visual Place RecognitionMapillary testRecall@173.5SelaVPR
Visual Place RecognitionMapillary testRecall@1090.6SelaVPR
Visual Place RecognitionMapillary testRecall@587.5SelaVPR

Related Papers

Visual Place Recognition for Large-Scale UAV Applications2025-07-20Overview of the TalentCLEF 2025: Skill and Job Title Intelligence for Human Capital Management2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17CATVis: Context-Aware Thought Visualization2025-07-15Query-Based Adaptive Aggregation for Multi-Dataset Joint Training Toward Universal Visual Place Recognition2025-07-04SAMURAI: Shape-Aware Multimodal Retrieval for 3D Object Identification2025-06-26RAG-VisualRec: An Open Resource for Vision- and Text-Enhanced Retrieval-Augmented Generation in Recommendation2025-06-25IRanker: Towards Ranking Foundation Model2025-06-25