TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/MixVPR: Feature Mixing for Visual Place Recognition

MixVPR: Feature Mixing for Visual Place Recognition

Amar Ali-bey, Brahim Chaib-Draa, Philippe Giguère

2023-03-03Metric LearningVisual Place RecognitionAutonomous DrivingImage Retrieval
PaperPDFCode(official)

Abstract

Visual Place Recognition (VPR) is a crucial part of mobile robotics and autonomous driving as well as other computer vision tasks. It refers to the process of identifying a place depicted in a query image using only computer vision. At large scale, repetitive structures, weather and illumination changes pose a real challenge, as appearances can drastically change over time. Along with tackling these challenges, an efficient VPR technique must also be practical in real-world scenarios where latency matters. To address this, we introduce MixVPR, a new holistic feature aggregation technique that takes feature maps from pre-trained backbones as a set of global features. Then, it incorporates a global relationship between elements in each feature map in a cascade of feature mixing, eliminating the need for local or pyramidal aggregation as done in NetVLAD or TransVPR. We demonstrate the effectiveness of our technique through extensive experiments on multiple large-scale benchmarks. Our method outperforms all existing techniques by a large margin while having less than half the number of parameters compared to CosPlace and NetVLAD. We achieve a new all-time high recall@1 score of 94.6% on Pitts250k-test, 88.0% on MapillarySLS, and more importantly, 58.4% on Nordland. Finally, our method outperforms two-stage retrieval techniques such as Patch-NetVLAD, TransVPR and SuperGLUE all while being orders of magnitude faster. Our code and trained models are available at https://github.com/amaralibey/MixVPR.

Results

TaskDatasetMetricValueModel
Visual Place RecognitionNardo-Air RRecall@176.06MixVPR
Visual Place RecognitionOxford RobotCar DatasetRecall@190.05MixVPR
Visual Place RecognitionNardo-AirRecall@132.39MixVPR
Visual Place RecognitionNordlandRecall@176MixVPR
Visual Place RecognitionNordlandRecall@589.2MixVPR
Visual Place RecognitionMid-Atlantic RidgeRecall@125.74MixVPR
Visual Place RecognitionSt LuciaRecall@199.66MixVPR
Visual Place RecognitionPittsburgh-250k-testRecall@194.6MixVPR
Visual Place RecognitionPittsburgh-250k-testRecall@1099MixVPR
Visual Place RecognitionPittsburgh-250k-testRecall@598.3MixVPR
Visual Place RecognitionHawkinsRecall@125.42MixVPR
Visual Place RecognitionLaurel CavernsRecall@129.46MixVPR
Visual Place RecognitionGardens PointRecall@191.5MixVPR
Visual Place RecognitionSPEDRecall@185.2MixVPR
Visual Place RecognitionSPEDRecall@1094.6MixVPR
Visual Place RecognitionSPEDRecall@592.1MixVPR
Visual Place RecognitionPittsburgh-30k-testRecall@191.52MixVPR
Visual Place RecognitionPittsburgh-30k-testRecall@595.9MixVPR
Visual Place RecognitionVP-AirRecall@110.31MixVPR
Visual Place RecognitionMapillary valRecall@188.2MixVPR
Visual Place RecognitionMapillary valRecall@1094.3MixVPR
Visual Place RecognitionMapillary valRecall@593.1MixVPR
Visual Place RecognitionMapillary testRecall@164MixVPR
Visual Place RecognitionMapillary testRecall@1080.6MixVPR
Visual Place RecognitionMapillary testRecall@575.9MixVPR
Visual Place Recognition17 PlacesRecall@163.79MixVPR
Visual Place RecognitionBaidu MallRecall@164.44MixVPR

Related Papers

Visual Place Recognition for Large-Scale UAV Applications2025-07-20GEMINUS: Dual-aware Global and Scene-Adaptive Mixture-of-Experts for End-to-End Autonomous Driving2025-07-19AGENTS-LLM: Augmentative GENeration of Challenging Traffic Scenarios with an Agentic LLM Framework2025-07-18Unsupervised Ground Metric Learning2025-07-17World Model-Based End-to-End Scene Generation for Accident Anticipation in Autonomous Driving2025-07-17Orbis: Overcoming Challenges of Long-Horizon Prediction in Driving World Models2025-07-17Channel-wise Motion Features for Efficient Motion Segmentation2025-07-17LaViPlan : Language-Guided Visual Path Planning with RLVR2025-07-17