TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/SalNAS: Efficient Saliency-prediction Neural Architecture ...

SalNAS: Efficient Saliency-prediction Neural Architecture Search with self-knowledge distillation

Chakkrit Termritthikun, Ayaz Umer, Suwichaya Suwanwimolkul, Feng Xia, Ivan Lee

2024-07-29Saliency PredictionNeural Architecture SearchPredictionKnowledge Distillation
PaperPDFCode(official)

Abstract

Recent advancements in deep convolutional neural networks have significantly improved the performance of saliency prediction. However, the manual configuration of the neural network architectures requires domain knowledge expertise and can still be time-consuming and error-prone. To solve this, we propose a new Neural Architecture Search (NAS) framework for saliency prediction with two contributions. Firstly, a supernet for saliency prediction is built with a weight-sharing network containing all candidate architectures, by integrating a dynamic convolution into the encoder-decoder in the supernet, termed SalNAS. Secondly, despite the fact that SalNAS is highly efficient (20.98 million parameters), it can suffer from the lack of generalization. To solve this, we propose a self-knowledge distillation approach, termed Self-KD, that trains the student SalNAS with the weighted average information between the ground truth and the prediction from the teacher model. The teacher model, while sharing the same architecture, contains the best-performing weights chosen by cross-validation. Self-KD can generalize well without the need to compute the gradient in the teacher model, enabling an efficient training system. By utilizing Self-KD, SalNAS outperforms other state-of-the-art saliency prediction models in most evaluation rubrics across seven benchmark datasets while being a lightweight model. The code will be available at https://github.com/chakkritte/SalNAS

Results

TaskDatasetMetricValueModel
Saliency DetectionSALICONAUC0.87SalNAS-XL + Self-KD
Saliency DetectionSALICONCC0.909SalNAS-XL + Self-KD
Saliency DetectionSALICONIG0.913SalNAS-XL + Self-KD
Saliency DetectionSALICONKLD0.195SalNAS-XL + Self-KD
Saliency DetectionSALICONNSS2.019SalNAS-XL + Self-KD
Saliency DetectionSALICONSIM0.796SalNAS-XL + Self-KD
Saliency DetectionSALICONsAUC0.749SalNAS-XL + Self-KD
Saliency PredictionSALICONAUC0.87SalNAS-XL + Self-KD
Saliency PredictionSALICONCC0.909SalNAS-XL + Self-KD
Saliency PredictionSALICONIG0.913SalNAS-XL + Self-KD
Saliency PredictionSALICONKLD0.195SalNAS-XL + Self-KD
Saliency PredictionSALICONNSS2.019SalNAS-XL + Self-KD
Saliency PredictionSALICONSIM0.796SalNAS-XL + Self-KD
Saliency PredictionSALICONsAUC0.749SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONAUC0.87SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONCC0.909SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONIG0.913SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONKLD0.195SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONNSS2.019SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONSIM0.796SalNAS-XL + Self-KD
Few-Shot Transfer Learning for Saliency PredictionSALICONsAUC0.749SalNAS-XL + Self-KD

Related Papers

Multi-Strategy Improved Snake Optimizer Accelerated CNN-LSTM-Attention-Adaboost for Trajectory Prediction2025-07-21Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21DASViT: Differentiable Architecture Search for Vision Transformer2025-07-17Uncertainty-Aware Cross-Modal Knowledge Distillation with Prototype Learning for Multimodal Brain-Computer Interfaces2025-07-17DVFL-Net: A Lightweight Distilled Video Focal Modulation Network for Spatio-Temporal Action Recognition2025-07-16Generative Click-through Rate Prediction with Applications to Search Advertising2025-07-15HanjaBridge: Resolving Semantic Ambiguity in Korean LLMs via Hanja-Augmented Pre-Training2025-07-15Feature Distillation is the Better Choice for Model-Heterogeneous Federated Learning2025-07-14