TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/A Comprehensive Study on Torchvision Pre-trained Models fo...

A Comprehensive Study on Torchvision Pre-trained Models for Fine-grained Inter-species Classification

Feras Albardi, H M Dipu Kabir, Md Mahbub Islam Bhuiyan, Parham M. Kebria, Abbas Khosravi, Saeid Nahavandi

2021-10-14Transfer LearningFine-Grained Image Classification
PaperPDFCode

Abstract

This study aims to explore different pre-trained models offered in the Torchvision package which is available in the PyTorch library. And investigate their effectiveness on fine-grained images classification. Transfer Learning is an effective method of achieving extremely good performance with insufficient training data. In many real-world situations, people cannot collect sufficient data required to train a deep neural network model efficiently. Transfer Learning models are pre-trained on a large data set, and can bring a good performance on smaller datasets with significantly lower training time. Torchvision package offers us many models to apply the Transfer Learning on smaller datasets. Therefore, researchers may need a guideline for the selection of a good model. We investigate Torchvision pre-trained models on four different data sets: 10 Monkey Species, 225 Bird Species, Fruits 360, and Oxford 102 Flowers. These data sets have images of different resolutions, class numbers, and different achievable accuracies. We also apply their usual fully-connected layer and the Spinal fully-connected layer to investigate the effectiveness of SpinalNet. The Spinal fully-connected layer brings better performance in most situations. We apply the same augmentation for different models for the same data set for a fair comparison. This paper may help future Computer Vision researchers in choosing a proper Transfer Learning model.

Results

TaskDatasetMetricValueModel
Image Classification10 Monkey SpeciesAccuracy99.26Inception-v3 (Spinal FC)
Image Classification10 Monkey SpeciesAccuracy99.26WideResNet-101(Spinal FC)
Image Classification10 Monkey SpeciesAccuracy98.9VGG-19_bn
Image ClassificationFruits-360Accuracy (%)99.98ResNeXt-101
Image ClassificationBird-225Accuracy99.56WideResNet-101 (Spinal FC)
Image ClassificationBird-225Accuracy99.38WideResNet-101
Image ClassificationOxford 102 FlowersAccuracy98.36DenseNet-201(Spinal FC)
Image ClassificationOxford 102 FlowersAccuracy98.29DenseNet-201
Fine-Grained Image Classification10 Monkey SpeciesAccuracy99.26Inception-v3 (Spinal FC)
Fine-Grained Image Classification10 Monkey SpeciesAccuracy99.26WideResNet-101(Spinal FC)
Fine-Grained Image Classification10 Monkey SpeciesAccuracy98.9VGG-19_bn
Fine-Grained Image ClassificationFruits-360Accuracy (%)99.98ResNeXt-101
Fine-Grained Image ClassificationBird-225Accuracy99.56WideResNet-101 (Spinal FC)
Fine-Grained Image ClassificationBird-225Accuracy99.38WideResNet-101
Fine-Grained Image ClassificationOxford 102 FlowersAccuracy98.36DenseNet-201(Spinal FC)
Fine-Grained Image ClassificationOxford 102 FlowersAccuracy98.29DenseNet-201

Related Papers

RaMen: Multi-Strategy Multi-Modal Learning for Bundle Construction2025-07-18Disentangling coincident cell events using deep transfer learning and compressive sensing2025-07-17Best Practices for Large-Scale, Pixel-Wise Crop Mapping and Transfer Learning Workflows2025-07-16Robust-Multi-Task Gradient Boosting2025-07-15Calibrated and Robust Foundation Models for Vision-Language and Medical Image Tasks Under Distribution Shift2025-07-12The Bayesian Approach to Continual Learning: An Overview2025-07-11Contrastive and Transfer Learning for Effective Audio Fingerprinting through a Real-World Evaluation Protocol2025-07-08Agent KB: Leveraging Cross-Domain Experience for Agentic Problem Solving2025-07-08