SOTAVerified

Common Sense Reasoning

Common sense reasoning tasks are intended to require the model to go beyond pattern recognition. Instead, the model should use "common sense" or world knowledge to make inferences.

Papers

Showing 151200 of 939 papers

TitleStatusHype
Guiding Pretraining in Reinforcement Learning with Large Language ModelsCode1
HAZARD Challenge: Embodied Decision Making in Dynamically Changing EnvironmentsCode1
Evaluation Toolkit For Robustness Testing Of Automatic Essay Scoring SystemsCode1
ByteSized32: A Corpus and Challenge Task for Generating Task-Specific World Models Expressed as Text GamesCode1
Human Parity on CommonsenseQA: Augmenting Self-Attention with External AttentionCode1
Comprehensive Visual Question Answering on Point Clouds through Compositional Scene ManipulationCode1
A Semantic Space is Worth 256 Language Descriptions: Make Stronger Segmentation Models with Descriptive PropertiesCode1
Expressive Scene Graph Generation Using Commonsense Knowledge Infusion for Visual Understanding and ReasoningCode1
Exploring the Benefits of Training Expert Language Models over Instruction TuningCode1
Conversational Word Embedding for Retrieval-Based Dialog SystemCode1
Generating similes effortlessly like a Pro: A Style Transfer Approach for Simile GenerationCode1
KETM:A Knowledge-Enhanced Text Matching methodCode1
Event2Mind for Russian: Understanding Emotions and Intents in Texts. Corpus and Model for EvaluationCode1
LLM-Coordination: Evaluating and Analyzing Multi-agent Coordination Abilities in Large Language ModelsCode1
Language Models are Unsupervised Multitask LearnersCode1
Common Sense Beyond English: Evaluating and Improving Multilingual Language Models for Commonsense ReasoningCode1
EventPlus: A Temporal Event Understanding PipelineCode1
Boosting Theory-of-Mind Performance in Large Language Models via PromptingCode1
Common Sense Enhanced Knowledge-based Recommendation with Large Language ModelCode1
Large Language Models as Evaluators for Recommendation ExplanationsCode1
Blow the Dog Whistle: A Chinese Dataset for Cant Understanding with Common Sense and World KnowledgeCode1
Evaluating and Analyzing Relationship Hallucinations in Large Vision-Language ModelsCode1
Evidence-Aware Inferential Text Generation with Vector Quantised Variational AutoEncoderCode1
Do Multilingual Language Models Think Better in English?Code1
DomainRAG: A Chinese Benchmark for Evaluating Domain-specific Retrieval-Augmented GenerationCode1
DialogSum: A Real-Life Scenario Dialogue Summarization DatasetCode1
Mitigating the Alignment Tax of RLHFCode1
Exploring AIGC Video Quality: A Focus on Visual Harmony, Video-Text Consistency and Domain Distribution GapCode1
KagNet: Knowledge-Aware Graph Networks for Commonsense ReasoningCode1
Playing Codenames with Language Graphs and Word EmbeddingsCode1
An Improved Neural Baseline for Temporal Relation Extraction0
Benchmarks for Automated Commonsense Reasoning: A Survey0
Belief Scene Graphs: Expanding Partial Scenes with Objects through Computation of Expectation0
An Hymn of an even Deeper Sentiment Analysis0
LLM-Advisor: An LLM Benchmark for Cost-efficient Path Planning across Multiple Terrains0
Behind the Scenes of an Evolving Event Cloze Test0
DEEPYANG at SemEval-2020 Task 4: Using the Hidden Layer State of BERT Model for Differentiating Common Sense0
A Hierarchical Bayesian Model for Unsupervised Induction of Script Knowledge0
Deep Unsupervised Hashing with Latent Semantic Components0
DELTA: Decomposed Efficient Long-Term Robot Task Planning using Large Language Models0
Design and Implementation of Linked Planning Domain Definition Language0
Back to Square One: Artifact Detection, Training and Commonsense Disentanglement in the Winograd Schema0
Baby Intuitions Benchmark (BIB): Discerning the goals, preferences, and actions of others0
An End-to-End Multi-task Learning Model for Fact Checking0
A vision-grounded dataset for predicting typical locations for verbs0
An Application of Pseudo-Log-Likelihoods to Natural Language Scoring0
Abductive Reasoning as Self-Supervision for Common Sense Question Answering0
A Vision for Semantically Enriched Data Science0
A very preliminary analysis of DALL-E 20
An Aposteriorical Clusterability Criterion for k-Means++ and Simplicity of Clustering0
Show:102550
← PrevPage 4 of 19Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ST-MoE-32B 269B (fine-tuned)Accuracy96.1Unverified
2Unicorn 11B (fine-tuned)Accuracy91.3Unverified
3CompassMTL 567M with TailorAccuracy90.5Unverified
4CompassMTL 567MAccuracy89.6Unverified
5UnifiedQA 11B (fine-tuned)Accuracy89.4Unverified
6Claude 3 Opus (5-shot)Accuracy88.5Unverified
7GPT-4 (5-shot)Accuracy87.5Unverified
8ExDeBERTa 567MAccuracy87Unverified
9LLaMA-2 13B + MixLoRAAccuracy86.3Unverified
10LLaMA3 8B+MoSLoRAAccuracy85.8Unverified
#ModelMetricClaimedVerifiedStatus
1GPT-4 (few-shot, k=25)Accuracy96.4Unverified
2PaLM 2 (few-shot, CoT, SC)Accuracy95.1Unverified
3Shivaay (4B, few-shot, k=8)Accuracy91.04Unverified
4StupidLLMAccuracy91.03Unverified
5Claude 2 (few-shot, k=5)Accuracy91Unverified
6Claude 1.3 (few-shot, k=5)Accuracy90Unverified
7PaLM 540B (Self Improvement, Self Consistency)Accuracy89.8Unverified
8PaLM 540B (Self Consistency)Accuracy88.7Unverified
9PaLM 540B (Self Improvement, CoT Prompting)Accuracy88.3Unverified
10PaLM 540B (Self Improvement, Standard-Prompting)Accuracy87.2Unverified
#ModelMetricClaimedVerifiedStatus
1ST-MoE-32B 269B (fine-tuned)Accuracy95.2Unverified
2LLaMA 3 8B+MoSLoRA (fine-tuned)Accuracy90.5Unverified
3PaLM 2-L (1-shot)Accuracy89.7Unverified
4PaLM 2-M (1-shot)Accuracy88Unverified
5LLaMA-3 8B + MixLoRAAccuracy86.5Unverified
6Camelidae-8×34BAccuracy86.2Unverified
7PaLM 2-S (1-shot)Accuracy85.6Unverified
8LLaMA 65B + CFG (0-shot)Accuracy84.2Unverified
9GAL 120B (0-shot)Accuracy83.8Unverified
10LLaMA-2 13B + MixLoRAAccuracy83.5Unverified
#ModelMetricClaimedVerifiedStatus
1Turing NLR v5 XXL 5.4B (fine-tuned)EM95.9Unverified
2ST-MoE-32B 269B (fine-tuned)EM95.1Unverified
3T5-11BF194.1Unverified
4DeBERTa-1.5BEM94.1Unverified
5PaLM 540B (finetuned)EM94Unverified
6Vega v2 6B (fine-tuned)EM93.9Unverified
7PaLM 2-L (one-shot)F193.8Unverified
8T5-XXL 11B (fine-tuned)EM93.4Unverified
9PaLM 2-M (one-shot)F192.4Unverified
10PaLM 2-S (one-shot)F192.1Unverified