SOTAVerified

Explanation Generation

Papers

Showing 2650 of 235 papers

TitleStatusHype
TE2Rules: Explaining Tree Ensembles using RulesCode1
End-to-End Multimodal Fact-Checking and Explanation Generation: A Challenging Dataset and ModelsCode1
Structured, flexible, and robust: benchmarking and improving large language models towards more human-like behavior in out-of-distribution reasoning tasksCode1
CLEVR-X: A Visual Reasoning Dataset for Natural Language ExplanationsCode1
REX: Reasoning-aware and Grounded ExplanationCode1
AR-BERT: Aspect-relation enhanced Aspect-level Sentiment Classification with Multi-modal ExplanationsCode1
Learn-Explain-Reinforce: Counterfactual Reasoning and Its Guidance to Reinforce an Alzheimer's Disease Diagnosis ModelCode1
Zorro: Valid, Sparse, and Stable Explanations in Graph Neural NetworksCode1
Faithfully Explainable Recommendation via Neural Logic ReasoningCode1
Explain and Predict, and then Predict AgainCode1
Towards Interpretable Natural Language Understanding with Explanations as Latent VariablesCode1
Explainable Automated Fact-Checking for Public Health ClaimsCode1
Leakage-Adjusted Simulatability: Can Models Generate Non-Trivial Explanations of Their Behavior in Natural Language?Code1
QED: A Framework and Dataset for Explanations in Question AnsweringCode1
Hierarchical Interaction Summarization and Contrastive Prompting for Explainable Recommendations0
The Future is Agentic: Definitions, Perspectives, and Open Challenges of Multi-Agent Recommender Systems0
RealFactBench: A Benchmark for Evaluating Large Language Models in Real-World Fact-CheckingCode0
LiTEx: A Linguistic Taxonomy of Explanations for Understanding Within-Label Variation in Natural Language InferenceCode0
Does Rationale Quality Matter? Enhancing Mental Disorder Detection via Selective Reasoning DistillationCode0
Multimodal RAG-driven Anomaly Detection and Classification in Laser Powder Bed Fusion using Large Language Models0
SNAPE-PM: Building and Utilizing Dynamic Partner Models for Adaptive Explanation GenerationCode0
Towards Budget-Friendly Model-Agnostic Explanation Generation for Large Language Models0
Generating Skyline Explanations for Graph Neural Networks0
Harnessing LLMs Explanations to Boost Surrogate Models in Tabular Data Classification0
ChartQA-X: Generating Explanations for Charts0
Show:102550
← PrevPage 2 of 10Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1VLIS (Lynx)Accuracy80Unverified
2VLIS (LLaVA)Accuracy73Unverified
3Ground-truth Caption -> GPT3 (Oracle)Human (%)68Unverified
4Predicted Caption -> GPT3Human (%)33Unverified
5BLIP2 FlanT5-XXL (Fine-tuned)Human (%)27Unverified
6BLIP2 FlanT5-XL (Fine-tuned)Human (%)15Unverified
7BLIP2 FlanT5-XXL (Zero-shot)Human (%)0Unverified
#ModelMetricClaimedVerifiedStatus
1PJ-XB487.4Unverified
2FMB478.8Unverified
#ModelMetricClaimedVerifiedStatus
1OFA-XHuman Explanation Rating85.7Unverified
2OFA-X-MTHuman Explanation Rating80.4Unverified
#ModelMetricClaimedVerifiedStatus
1OFA-X-MTHuman Explanation Rating77.3Unverified
2OFA-XHuman Explanation Rating68.9Unverified
#ModelMetricClaimedVerifiedStatus
1OFA-XHuman Explanation Rating89.5Unverified
2OFA-X-MTHuman Explanation Rating87.8Unverified