Knowledge Distillation

18 benchmarks4240 papers

Knowledge distillation is the process of transferring knowledge from a large model to a smaller one. While large models (such as very deep neural networks or ensembles of many models) have higher knowledge capacity than small models, this capacity might not be fully utilized.

Benchmarks

Knowledge Distillation on CIFAR-100

Knowledge Distillation on QNLI

Knowledge Distillation on SQuAD

Knowledge Distillation on COCO 2017 val

Knowledge Distillation on PASCAL VOC

Knowledge Distillation on Cityscapes

Knowledge Distillation on CloudEval-YAML

Knowledge Distillation on KITTI

Knowledge Distillation on Wiki-40B