TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Generative Low-bitwidth Data Free Quantization

Generative Low-bitwidth Data Free Quantization

Shoukai Xu, Haokun Li, Bohan Zhuang, Jing Liu, JieZhang Cao, Chuangrun Liang, Mingkui Tan

2020-03-07ECCV 2020 8Data Free QuantizationQuantization
PaperPDFCodeCode(official)Code

Abstract

Neural network quantization is an effective way to compress deep models and improve their execution latency and energy efficiency, so that they can be deployed on mobile or embedded devices. Existing quantization methods require original data for calibration or fine-tuning to get better performance. However, in many real-world scenarios, the data may not be available due to confidential or private issues, thereby making existing quantization methods not applicable. Moreover, due to the absence of original data, the recently developed generative adversarial networks (GANs) cannot be applied to generate data. Although the full-precision model may contain rich data information, such information alone is hard to exploit for recovering the original data or generating new meaningful data. In this paper, we investigate a simple-yet-effective method called Generative Low-bitwidth Data Free Quantization (GDFQ) to remove the data dependence burden. Specifically, we propose a knowledge matching generator to produce meaningful fake data by exploiting classification boundary knowledge and distribution information in the pre-trained model. With the help of generated data, we can quantize a model by learning knowledge from the pre-trained model. Extensive experiments on three data sets demonstrate the effectiveness of our method. More critically, our method achieves much higher accuracy on 4-bit quantization than the existing data free quantization method. Code is available at https://github.com/xushoukai/GDFQ.

Results

TaskDatasetMetricValueModel
QuantizationCIFAR-100CIFAR-100 W4A4 Top-1 Accuracy43.12ResNet-20 CIFAR-100
QuantizationCIFAR-100CIFAR-100 W5A5 Top-1 Accuracy64.03ResNet-20 CIFAR-100
QuantizationCIFAR-100CIFAR-100 W6A6 Top-1 Accuracy68.63ResNet-20 CIFAR-100
QuantizationCIFAR-100CIFAR-100 W8A8 Top-1 Accuracy70.29ResNet-20 CIFAR-100
QuantizationCIFAR10CIFAR-10 W4A4 Top-1 Accuracy85.2ResNet-20 CIFAR-10
QuantizationCIFAR10CIFAR-10 W5A5 Top-1 Accuracy92.39ResNet-20 CIFAR-10
QuantizationCIFAR10CIFAR-10 W6A6 Top-1 Accuracy93.38ResNet-20 CIFAR-10
QuantizationCIFAR10CIFAR-10 W8A8 Top-1 Accuracy93.92ResNet-20 CIFAR-10

Related Papers

Efficient Deployment of Spiking Neural Networks on SpiNNaker2 for DVS Gesture Recognition Using Neuromorphic Intermediate Representation2025-09-04An End-to-End DNN Inference Framework for the SpiNNaker2 Neuromorphic MPSoC2025-07-18Task-Specific Audio Coding for Machines: Machine-Learned Latent Features Are Codes for That Machine2025-07-17Angle Estimation of a Single Source with Massive Uniform Circular Arrays2025-07-17Quantized Rank Reduction: A Communications-Efficient Federated Learning Scheme for Network-Critical Applications2025-07-15MGVQ: Could VQ-VAE Beat VAE? A Generalizable Tokenizer with Multi-group Quantization2025-07-14Lightweight Federated Learning over Wireless Edge Networks2025-07-13Vision Foundation Models as Effective Visual Tokenizers for Autoregressive Image Generation2025-07-11