SOTAVerified

Explanation Generation

Papers

Showing 2130 of 235 papers

TitleStatusHype
Do Vision & Language Decoders use Images and Text equally? How Self-consistent are their Explanations?Code1
End-to-End Multimodal Fact-Checking and Explanation Generation: A Challenging Dataset and ModelsCode1
Explainable Automated Fact-Checking for Public Health ClaimsCode1
Explainable Legal Case Matching via Inverse Optimal Transport-based Rationale ExtractionCode1
LLMRec: Benchmarking Large Language Models on Recommendation TaskCode1
Faithfully Explainable Recommendation via Neural Logic ReasoningCode1
Explain and Predict, and then Predict AgainCode1
QED: A Framework and Dataset for Explanations in Question AnsweringCode1
LLM4Vis: Explainable Visualization Recommendation using ChatGPTCode1
Sim2Word: Explaining Similarity with Representative Attribute Words via Counterfactual ExplanationsCode1
Show:102550
← PrevPage 3 of 24Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1VLIS (Lynx)Accuracy80Unverified
2VLIS (LLaVA)Accuracy73Unverified
3Ground-truth Caption -> GPT3 (Oracle)Human (%)68Unverified
4Predicted Caption -> GPT3Human (%)33Unverified
5BLIP2 FlanT5-XXL (Fine-tuned)Human (%)27Unverified
6BLIP2 FlanT5-XL (Fine-tuned)Human (%)15Unverified
7BLIP2 FlanT5-XXL (Zero-shot)Human (%)0Unverified
#ModelMetricClaimedVerifiedStatus
1PJ-XB487.4Unverified
2FMB478.8Unverified
#ModelMetricClaimedVerifiedStatus
1OFA-XHuman Explanation Rating85.7Unverified
2OFA-X-MTHuman Explanation Rating80.4Unverified
#ModelMetricClaimedVerifiedStatus
1OFA-X-MTHuman Explanation Rating77.3Unverified
2OFA-XHuman Explanation Rating68.9Unverified
#ModelMetricClaimedVerifiedStatus
1OFA-XHuman Explanation Rating89.5Unverified
2OFA-X-MTHuman Explanation Rating87.8Unverified