TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Beyond Geo-localization: Fine-grained Orientation of Stree...

Beyond Geo-localization: Fine-grained Orientation of Street-view Images by Cross-view Matching with Satellite Imagery with Supplementary Materials

Wenmiao Hu, Yichen Zhang, Yuxuan Liang, Yifang Yin, Andrei Georgescu, An Tran, Hannes Kruppa, See-Kiong Ng, Roger Zimmermann

2023-07-07geo-localizationImage-Based Localization
PaperPDF

Abstract

Street-view imagery provides us with novel experiences to explore different places remotely. Carefully calibrated street-view images (e.g. Google Street View) can be used for different downstream tasks, e.g. navigation, map features extraction. As personal high-quality cameras have become much more affordable and portable, an enormous amount of crowdsourced street-view images are uploaded to the internet, but commonly with missing or noisy sensor information. To prepare this hidden treasure for "ready-to-use" status, determining missing location information and camera orientation angles are two equally important tasks. Recent methods have achieved high performance on geo-localization of street-view images by cross-view matching with a pool of geo-referenced satellite imagery. However, most of the existing works focus more on geo-localization than estimating the image orientation. In this work, we re-state the importance of finding fine-grained orientation for street-view images, formally define the problem and provide a set of evaluation metrics to assess the quality of the orientation estimation. We propose two methods to improve the granularity of the orientation estimation, achieving 82.4% and 72.3% accuracy for images with estimated angle errors below 2 degrees for CVUSA and CVACT datasets, corresponding to 34.9% and 28.2% absolute improvement compared to previous works. Integrating fine-grained orientation estimation in training also improves the performance on geo-localization, giving top 1 recall 95.5%/85.5% and 86.8%/80.4% for orientation known/unknown tests on the two datasets.

Results

TaskDatasetMetricValueModel
Object LocalizationcvusaRecall@195.43GeoDTR
Object LocalizationcvusaRecall@1099.34GeoDTR
Object LocalizationcvusaRecall@598.86GeoDTR
Object LocalizationcvusaRecall@top1%99.86GeoDTR
Object LocalizationcvactRecall@186.21GeoDTR
Object LocalizationcvactRecall@1 (%)98.77GeoDTR
Object LocalizationcvactRecall@1096.72GeoDTR
Object LocalizationcvactRecall@595.44GeoDTR

Related Papers

Grid-Reg: Grid-Based SAR and Optical Image Registration Across Platforms2025-07-06Recognition through Reasoning: Reinforcing Image Geo-localization with Large Vision-Language Models2025-06-17A Two-stage Optimization Method for Wide-range Single-electron Quantum Magnetic Sensing2025-06-16GRE Suite: Geo-localization Inference via Fine-Tuned Vision-Language Models and Enhanced Reasoning Chains2025-05-24Object-level Cross-view Geo-localization with Location Enhancement and Multi-Head Cross Attention2025-05-23Robust Cross-View Geo-Localization via Content-Viewpoint Disentanglement2025-05-17A Unified Hierarchical Framework for Fine-grained Cross-view Geo-localization over Large-scale Scenarios2025-05-12Geolocating Earth Imagery from ISS: Integrating Machine Learning with Astronaut Photography for Enhanced Geographic Mapping2025-04-29