TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/University-1652: A Multi-view Multi-source Benchmark for D...

University-1652: A Multi-view Multi-source Benchmark for Drone-based Geo-localization

Zhedong Zheng, Yunchao Wei, Yi Yang

2020-02-27Visual Localizationgeo-localizationDrone navigationDrone-view target localizationImage-Based Localization
PaperPDFCode(official)CodeCode

Abstract

We consider the problem of cross-view geo-localization. The primary challenge of this task is to learn the robust feature against large viewpoint changes. Existing benchmarks can help, but are limited in the number of viewpoints. Image pairs, containing two viewpoints, e.g., satellite and ground, are usually provided, which may compromise the feature learning. Besides phone cameras and satellites, in this paper, we argue that drones could serve as the third platform to deal with the geo-localization problem. In contrast to the traditional ground-view images, drone-view images meet fewer obstacles, e.g., trees, and could provide a comprehensive view when flying around the target place. To verify the effectiveness of the drone platform, we introduce a new multi-view multi-source benchmark for drone-based geo-localization, named University-1652. University-1652 contains data from three platforms, i.e., synthetic drones, satellites and ground cameras of 1,652 university buildings around the world. To our knowledge, University-1652 is the first drone-based geo-localization dataset and enables two new tasks, i.e., drone-view target localization and drone navigation. As the name implies, drone-view target localization intends to predict the location of the target place via drone-view images. On the other hand, given a satellite-view query image, drone navigation is to drive the drone to the area of interest in the query. We use this dataset to analyze a variety of off-the-shelf CNN features and propose a strong CNN baseline on this challenging dataset. The experiments show that University-1652 helps the model to learn the viewpoint-invariant features and also has good generalization ability in the real-world scenario.

Results

TaskDatasetMetricValueModel
Object LocalizationcvusaRecall@143.91Instance Loss
Object LocalizationcvusaRecall@1074.58Instance Loss
Object LocalizationcvusaRecall@566.38Instance Loss
Object LocalizationcvusaRecall@top1%91.78Instance Loss
Image RetrievalUniversity-1652AP58.74Instance Loss
Image RetrievalUniversity-1652Recall@171.18Instance Loss
Image RetrievalUniversity-1652AP63.13Instance Loss
Image RetrievalUniversity-1652Recall@158.49Instance Loss
Content-Based Image RetrievalUniversity-1652AP58.74Instance Loss
Content-Based Image RetrievalUniversity-1652Recall@171.18Instance Loss
Content-Based Image RetrievalUniversity-1652AP63.13Instance Loss
Content-Based Image RetrievalUniversity-1652Recall@158.49Instance Loss

Related Papers

Kaleidoscopic Background Attack: Disrupting Pose Estimation with Multi-Fold Radial Symmetry Textures2025-07-14Evaluating Attribute Confusion in Fashion Text-to-Image Generation2025-07-09Grid-Reg: Grid-Based SAR and Optical Image Registration Across Platforms2025-07-06MatChA: Cross-Algorithm Matching with Feature Augmentation2025-06-27OracleFusion: Assisting the Decipherment of Oracle Bone Script with Structurally Constrained Semantic Typography2025-06-26Semantic and Feature Guided Uncertainty Quantification of Visual Localization for Autonomous Vehicles2025-06-18Recognition through Reasoning: Reinforcing Image Geo-localization with Large Vision-Language Models2025-06-17A Two-stage Optimization Method for Wide-range Single-electron Quantum Magnetic Sensing2025-06-16