TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Reverse Knowledge Distillation: Training a Large Model usi...

Reverse Knowledge Distillation: Training a Large Model using a Small One for Retinal Image Matching on Limited Data

Sahar Almahfouz Nasser, Nihar Gupte, Amit Sethi

2023-07-20Image RegistrationKeypoint detection and image matchingKeypoint DetectionKnowledge Distillation
PaperPDFCode(official)

Abstract

Retinal image matching plays a crucial role in monitoring disease progression and treatment response. However, datasets with matched keypoints between temporally separated pairs of images are not available in abundance to train transformer-based model. We propose a novel approach based on reverse knowledge distillation to train large models with limited data while preventing overfitting. Firstly, we propose architectural modifications to a CNN-based semi-supervised method called SuperRetina that help us improve its results on a publicly available dataset. Then, we train a computationally heavier model based on a vision transformer encoder using the lighter CNN-based model, which is counter-intuitive in the field knowledge-distillation research where training lighter models based on heavier ones is the norm. Surprisingly, such reverse knowledge distillation improves generalization even further. Our experiments suggest that high-dimensional fitting in representation space may prevent overfitting unlike training directly to match the final output. We also provide a public dataset with annotations for retinal image keypoint detection and matching to help the research community develop algorithms for retinal image applications.

Results

TaskDatasetMetricValueModel
Image RegistrationFIREmAUC0.761LKRetina

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21fastWDM3D: Fast and Accurate 3D Healthy Tissue Inpainting2025-07-17cIDIR: Conditioned Implicit Neural Representation for Regularized Deformable Image Registration2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Are Vision Foundation Models Ready for Out-of-the-Box Medical Image Registration?2025-07-15KptLLM++: Towards Generic Keypoint Comprehension with Large Language Model2025-07-15GKNet: Graph-based Keypoints Network for Monocular Pose Estimation of Non-cooperative Spacecraft2025-07-15