Metric: Mismatched (higher is better)
| # | Model↕ | Mismatched▼ | Extra Data | Paper | Date↕ | Code |
|---|---|---|---|---|---|---|
| 1 | TinyBERT-6 67M | 84.5 | No | TinyBERT: Distilling BERT for Natural Language U... | 2019-09-23 | Code |
| 2 | BERT-Large-uncased-PruneOFA (90% unstruct sparse) | 84.2 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 3 | BERT-Large-uncased-PruneOFA (90% unstruct sparse, QAT Int8) | 84.08 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 4 | BERT-Base-uncased-PruneOFA (85% unstruct sparse) | 83.67 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 5 | BERT-Base-uncased-PruneOFA (85% unstruct sparse, QAT Int8) | 82.51 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 6 | BERT-Base-uncased-PruneOFA (90% unstruct sparse) | 82.43 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 7 | DistilBERT-uncased-PruneOFA (85% unstruct sparse) | 82.03 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 8 | DistilBERT-uncased-PruneOFA (90% unstruct sparse) | 81.47 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 9 | DistilBERT-uncased-PruneOFA (85% unstruct sparse, QAT Int8) | 81.14 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |
| 10 | DistilBERT-uncased-PruneOFA (90% unstruct sparse, QAT Int8) | 80.4 | No | Prune Once for All: Sparse Pre-Trained Language ... | 2021-11-10 | Code |