TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Google Landmarks Dataset v2 -- A Large-Scale Benchmark for...

Google Landmarks Dataset v2 -- A Large-Scale Benchmark for Instance-Level Recognition and Retrieval

Tobias Weyand, Andre Araujo, Bingyi Cao, Jack Sim

2020-04-03Transfer LearningLandmark RecognitionRetrievalImage Retrieval
PaperPDFCodeCode(official)CodeCodeCode

Abstract

While image retrieval and instance recognition techniques are progressing rapidly, there is a need for challenging datasets to accurately measure their performance -- while posing novel challenges that are relevant for practical applications. We introduce the Google Landmarks Dataset v2 (GLDv2), a new benchmark for large-scale, fine-grained instance recognition and image retrieval in the domain of human-made and natural landmarks. GLDv2 is the largest such dataset to date by a large margin, including over 5M images and 200k distinct instance labels. Its test set consists of 118k images with ground truth annotations for both the retrieval and recognition tasks. The ground truth construction involved over 800 hours of human annotator work. Our new dataset has several challenging properties inspired by real world applications that previous datasets did not consider: An extremely long-tailed class distribution, a large fraction of out-of-domain test photos and large intra-class variability. The dataset is sourced from Wikimedia Commons, the world's largest crowdsourced collection of landmark photos. We provide baseline results for both recognition and retrieval tasks based on state-of-the-art methods as well as competitive results from a public challenge. We further demonstrate the suitability of the dataset for transfer learning by showing that image embeddings trained on it achieve competitive retrieval performance on independent datasets. The dataset images, ground-truth and metric scoring code are available at https://github.com/cvdfoundation/google-landmark.

Results

TaskDatasetMetricValueModel
Image RetrievalROxford (Medium)mAP74.2ResNet101+ArcFace GLDv2-train-clean
Image RetrievalRParis (Medium)mAP84.9ResNet101+ArcFace GLDv2-train-clean
Image RetrievalRParis (Hard)mAP70.3ResNet101+ArcFace GLDv2-train-clean
Image RetrievalGoogle Landmarks Dataset v2 (retrieval, testing)mAP@10024.15ResNet101+ArcFace GLDv2-train-clean
Image RetrievalROxford (Hard)mAP51.6ResNet101+ArcFace GLDv2-train-clean
Image RetrievalGoogle Landmarks Dataset v2 (retrieval, validation)mAP@10022.2ResNet101+ArcFace GLDv2-train-clean
Landmark RecognitionGoogle Landmarks Dataset v2 (recognition, testing)microAP56.35DELG global+SP
Landmark RecognitionGoogle Landmarks Dataset v2 (recognition, validation)microAP55.01DELG global+SP

Related Papers

RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17From Roots to Rewards: Dynamic Tree Reasoning with RL2025-07-17HapticCap: A Multimodal Dataset and Task for Understanding User Experience of Vibration Haptic Signals2025-07-17A Survey of Context Engineering for Large Language Models2025-07-17MCoT-RE: Multi-Faceted Chain-of-Thought and Re-Ranking for Training-Free Zero-Shot Composed Image Retrieval2025-07-17FAR-Net: Multi-Stage Fusion Network with Enhanced Semantic Alignment and Adaptive Reconciliation for Composed Image Retrieval2025-07-17Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows2025-07-16