SOTAVerified

General Knowledge

This task aims to evaluate the ability of a model to answer general-knowledge questions.

Source: BIG-bench

Papers

Showing 151175 of 399 papers

TitleStatusHype
Benchmarking Generative Models on Computational Thinking Tests in Elementary Visual Programming0
Learning from Natural Language Explanations for Generalizable Entity Matching0
RAD: A Comprehensive Dataset for Benchmarking the Robustness of Image Anomaly DetectionCode1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
F-LMM: Grounding Frozen Large Multimodal ModelsCode2
Generative Explore-Exploit: Training-free Optimization of Generative Recommender Systems using LLM Optimizers0
HYDRA: Model Factorization Framework for Black-Box LLM PersonalizationCode1
ContextFlow++: Generalist-Specialist Flow-based Generative Models with Mixed-Variable Context EncodingCode0
Slow and Steady Wins the Race: Maintaining Plasticity with Hare and Tortoise NetworksCode1
CRoFT: Robust Fine-Tuning with Concurrent Optimization for OOD Generalization and Open-Set OOD DetectionCode1
SOK-Bench: A Situated Video Reasoning Benchmark with Aligned Open-World Knowledge0
Health Index Estimation Through Integration of General Knowledge with Unsupervised LearningCode1
MoST: Multi-modality Scene Tokenization for Motion Prediction0
Towards Generalizable Agents in Text-Based Educational Environments: A Study of Integrating RL with LLMs0
Enhancing Action Recognition from Low-Quality Skeleton Data via Part-Level Knowledge Distillation0
Evaluating Consistency and Reasoning Capabilities of Large Language Models0
Learning Electromagnetic Metamaterial Physics With ChatGPT0
When Life gives you LLMs, make LLM-ADE: Large Language Models with Adaptive Data Engineering0
Pretraining and Updates of Domain-Specific LLM: A Case Study in the Japanese Business Domain0
Knowledge graphs for empirical concept retrievalCode0
Eraser: Jailbreaking Defense in Large Language Models via Unlearning Harmful KnowledgeCode0
BEAR: A Unified Framework for Evaluating Relational Knowledge in Causal and Masked Language ModelsCode1
Benchmarking Large Language Models for Persian: A Preliminary Study Focusing on ChatGPTCode1
Prompt Learning via Meta-RegularizationCode1
Juru: Legal Brazilian Large Language Model from Reputable Sources0
Show:102550
← PrevPage 7 of 16Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Chinchilla-70B (few-shot, k=5)Accuracy94.3Unverified
2Gopher-280B (few-shot, k=5)Accuracy93.9Unverified
3Chinchilla-70B (few-shot, k=5)Accuracy 85.7Unverified
4Gopher-280B (few-shot, k=5)Accuracy 84.8Unverified
5Gopher-280B (few-shot, k=5)Accuracy84.2Unverified
6Gopher-280B (few-shot, k=5)Accuracy 84.1Unverified
7Gopher-280B (few-shot, k=5)Accuracy 83.9Unverified
8Gopher-280B (few-shot, k=5)Accuracy83.3Unverified
9Gopher-280B (few-shot, k=5)Accuracy 81.8Unverified
10Gopher-280B (few-shot, k=5)Accuracy 81Unverified