Metric: % info (higher is better)
| # | Model↕ | % info▼ | Extra Data | Paper | Date↕ | Code |
|---|---|---|---|---|---|---|
| 1 | Alpaca 7B + Inference Time Intervention (ITI) | 97.7 | No | - | - | - |
| 2 | GPT-3 175B | 97.55 | No | TruthfulQA: Measuring How Models Mimic Human Fal... | 2021-09-08 | Code |
| 3 | LLaMA 7B + Inference Time Intervention (ITI) | 93.8 | No | - | - | - |
| 4 | GPT-J 6B | 89.96 | No | TruthfulQA: Measuring How Models Mimic Human Fal... | 2021-09-08 | Code |
| 5 | GPT-2 1.5B | 89.84 | No | TruthfulQA: Measuring How Models Mimic Human Fal... | 2021-09-08 | Code |
| 6 | Vicuna 7B + Inference Time Intervention (ITI) | 83.5 | No | - | - | - |
| 7 | UnifiedQA 3B | 64.5 | No | TruthfulQA: Measuring How Models Mimic Human Fal... | 2021-09-08 | Code |
| 8 | LLaMA 65B | 53 | No | LLaMA: Open and Efficient Foundation Language Mo... | 2023-02-27 | Code |
| 9 | LLaMA 33B | 48 | No | LLaMA: Open and Efficient Foundation Language Mo... | 2023-02-27 | Code |
| 10 | LLaMA 13B | 41 | No | LLaMA: Open and Efficient Foundation Language Mo... | 2023-02-27 | Code |
| 11 | LLaMA 7B | 29 | No | LLaMA: Open and Efficient Foundation Language Mo... | 2023-02-27 | Code |