TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Attentions Help CNNs See Better: Attention-based Hybrid Im...

Attentions Help CNNs See Better: Attention-based Hybrid Image Quality Assessment Network

Shanshan Lao, Yuan Gong, Shuwei Shi, Sidi Yang, Tianhe Wu, Jiahao Wang, Weihao Xia, Yujiu Yang

2022-04-22Video Quality AssessmentImage Quality Assessment
PaperPDFCodeCode(official)Code

Abstract

Image quality assessment (IQA) algorithm aims to quantify the human perception of image quality. Unfortunately, there is a performance drop when assessing the distortion images generated by generative adversarial network (GAN) with seemingly realistic texture. In this work, we conjecture that this maladaptation lies in the backbone of IQA models, where patch-level prediction methods use independent image patches as input to calculate their scores separately, but lack spatial relationship modeling among image patches. Therefore, we propose an Attention-based Hybrid Image Quality Assessment Network (AHIQ) to deal with the challenge and get better performance on the GAN-based IQA task. Firstly, we adopt a two-branch architecture, including a vision transformer (ViT) branch and a convolutional neural network (CNN) branch for feature extraction. The hybrid architecture combines interaction information among image patches captured by ViT and local texture details from CNN. To make the features from shallow CNN more focused on the visually salient region, a deformable convolution is applied with the help of semantic information from the ViT branch. Finally, we use a patch-wise score prediction module to obtain the final score. The experiments show that our model outperforms the state-of-the-art methods on four standard IQA datasets and AHIQ ranked first on the Full Reference (FR) track of the NTIRE 2022 Perceptual Image Quality Assessment Challenge.

Results

TaskDatasetMetricValueModel
Video UnderstandingMSU FR VQA DatabaseKLCC0.8015AHIQ
Video UnderstandingMSU FR VQA DatabaseSRCC0.937AHIQ
Video Quality AssessmentMSU FR VQA DatabaseKLCC0.8015AHIQ
Video Quality AssessmentMSU FR VQA DatabaseSRCC0.937AHIQ
Image Quality AssessmentMSU FR VQA DatabaseSRCC0.937AHIQ
VideoMSU FR VQA DatabaseKLCC0.8015AHIQ
VideoMSU FR VQA DatabaseSRCC0.937AHIQ

Related Papers

Visual-Language Model Knowledge Distillation Method for Image Quality Assessment2025-07-21DeQA-Doc: Adapting DeQA-Score to Document Image Quality Assessment2025-07-17Text-Visual Semantic Constrained AI-Generated Image Quality Assessment2025-07-144KAgent: Agentic Any Image to 4K Super-Resolution2025-07-09Bridging Video Quality Scoring and Justification via Large Multimodal Models2025-06-26FundaQ-8: A Clinically-Inspired Scoring Framework for Automated Fundus Image Quality Assessment2025-06-25MS-IQA: A Multi-Scale Feature Fusion Network for PET/CT Image Quality Assessment2025-06-25Enhanced Dermatology Image Quality Assessment via Cross-Domain Training2025-06-19