General Knowledge
This task aims to evaluate the ability of a model to answer general-knowledge questions.
Source: BIG-bench
Papers
Showing 26–50 of 399 papers
Benchmark Results
| # | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| 1 | Chinchilla-70B (few-shot, k=5) | Accuracy | 94.3 | — | Unverified |
| 2 | Gopher-280B (few-shot, k=5) | Accuracy | 93.9 | — | Unverified |
| 3 | Chinchilla-70B (few-shot, k=5) | Accuracy | 85.7 | — | Unverified |
| 4 | Gopher-280B (few-shot, k=5) | Accuracy | 84.8 | — | Unverified |
| 5 | Gopher-280B (few-shot, k=5) | Accuracy | 84.2 | — | Unverified |
| 6 | Gopher-280B (few-shot, k=5) | Accuracy | 84.1 | — | Unverified |
| 7 | Gopher-280B (few-shot, k=5) | Accuracy | 83.9 | — | Unverified |
| 8 | Gopher-280B (few-shot, k=5) | Accuracy | 83.3 | — | Unverified |
| 9 | Gopher-280B (few-shot, k=5) | Accuracy | 81.8 | — | Unverified |
| 10 | Gopher-280B (few-shot, k=5) | Accuracy | 81 | — | Unverified |