TasksSotADatasetsPapersMethodsSubmitAbout
Papers With Code 2

A community resource for machine learning research: papers, code, benchmarks, and state-of-the-art results.

Explore

Notable BenchmarksAll SotADatasetsPapersMethods

Community

Submit ResultsAbout

Data sourced from the PWC Archive (CC-BY-SA 4.0). Built by the community, for the community.

Papers/Qimera: Data-free Quantization with Synthetic Boundary Sup...

Qimera: Data-free Quantization with Synthetic Boundary Supporting Samples

Kanghyun Choi, Deokki Hong, Noseong Park, Youngsok Kim, Jinho Lee

2021-11-04NeurIPS 2021 12Data Free QuantizationQuantizationDisentanglement
PaperPDFCodeCode(official)

Abstract

Model quantization is known as a promising method to compress deep neural networks, especially for inferences on lightweight mobile or edge devices. However, model quantization usually requires access to the original training data to maintain the accuracy of the full-precision models, which is often infeasible in real-world scenarios for security and privacy issues. A popular approach to perform quantization without access to the original data is to use synthetically generated samples, based on batch-normalization statistics or adversarial learning. However, the drawback of such approaches is that they primarily rely on random noise input to the generator to attain diversity of the synthetic samples. We find that this is often insufficient to capture the distribution of the original data, especially around the decision boundaries. To this end, we propose Qimera, a method that uses superposed latent embeddings to generate synthetic boundary supporting samples. For the superposed embeddings to better reflect the original distribution, we also propose using an additional disentanglement mapping layer and extracting information from the full-precision model. The experimental results show that Qimera achieves state-of-the-art performances for various settings on data-free quantization. Code is available at https://github.com/iamkanghyunchoi/qimera.

Results

TaskDatasetMetricValueModel
QuantizationCIFAR-100CIFAR-100 W4A4 Top-1 Accuracy65.1ResNet-20 CIFAR-100
QuantizationCIFAR-100CIFAR-100 W5A5 Top-1 Accuracy69.02ResNet-20 CIFAR-100
QuantizationCIFAR10CIFAR-10 W4A4 Top-1 Accuracy91.26ResNet-20 CIFAR-10
QuantizationCIFAR10CIFAR-10 W5A5 Top-1 Accuracy93.46ResNet-20 CIFAR-10

Related Papers

Efficient Deployment of Spiking Neural Networks on SpiNNaker2 for DVS Gesture Recognition Using Neuromorphic Intermediate Representation2025-09-04An End-to-End DNN Inference Framework for the SpiNNaker2 Neuromorphic MPSoC2025-07-18CSD-VAR: Content-Style Decomposition in Visual Autoregressive Models2025-07-18Task-Specific Audio Coding for Machines: Machine-Learned Latent Features Are Codes for That Machine2025-07-17Angle Estimation of a Single Source with Massive Uniform Circular Arrays2025-07-17Quantized Rank Reduction: A Communications-Efficient Federated Learning Scheme for Network-Critical Applications2025-07-15MGVQ: Could VQ-VAE Beat VAE? A Generalizable Tokenizer with Multi-group Quantization2025-07-14Lightweight Federated Learning over Wireless Edge Networks2025-07-13