TAT-VPR: Ternary Adaptive Transformer for Dynamic and Efficient Visual Place Recognition
Oliver Grainge, Michael Milford, Indu Bodala, Sarvapali D. Ramchurn, Shoaib Ehsan
2025-05-22Visual Place Recognition
Abstract
TAT-VPR is a ternary-quantized transformer that brings dynamic accuracy-efficiency trade-offs to visual SLAM loop-closure. By fusing ternary weights with a learned activation-sparsity gate, the model can control computation by up to 40% at run-time without degrading performance (Recall@1). The proposed two-stage distillation pipeline preserves descriptor quality, letting it run on micro-UAV and embedded SLAM stacks while matching state-of-the-art localization accuracy.
Related Papers
Visual Place Recognition for Large-Scale UAV Applications2025-07-20Query-Based Adaptive Aggregation for Multi-Dataset Joint Training Toward Universal Visual Place Recognition2025-07-04Adversarial Attacks and Detection in Visual Place Recognition for Safer Robot Navigation2025-06-19Astra: Toward General-Purpose Mobile Robots via Hierarchical Multimodal Learning2025-06-06HypeVPR: Exploring Hyperbolic Space for Perspective to Equirectangular Visual Place Recognition2025-06-05Place Recognition: A Comprehensive Review, Current Challenges and Future Directions2025-05-20MMS-VPR: Multimodal Street-Level Visual Place Recognition Dataset and Benchmark2025-05-18Geolocating Earth Imagery from ISS: Integrating Machine Learning with Astronaut Photography for Enhanced Geographic Mapping2025-04-29