SOTAVerified

Fact Checking

Papers

Showing 201225 of 669 papers

TitleStatusHype
Automatic Fact-Checking with Frame-Semantics0
Fine-Grained Appropriate Reliance: Human-AI Collaboration with a Multi-Step Transparent Decision Workflow for Complex Task Decomposition0
Zero-shot and Few-shot Learning with Instruction-following LLMs for Claim Matching in Automated Fact-checking0
From Scarcity to Capability: Empowering Fake News Detection in Low-Resource Languages with LLMsCode0
Tracking the Takes and Trajectories of English-Language News Narratives across Trustworthy and Worrisome WebsitesCode0
Improving Factuality with Explicit Working Memory0
Evaluating the Performance of Large Language Models in Scientific Claim Detection and Classification0
Logical Consistency of Large Language Models in Fact-checking0
Face the Facts! Evaluating RAG-based Fact-checking Pipelines in Realistic SettingsCode0
ViFactCheck: A New Benchmark Dataset and Methods for Multi-domain News Fact-Checking in Vietnamese0
Self-Adaptive Paraphrasing and Preference Learning for Improved Claim Verifiability0
Exploring Multidimensional Checkworthiness: Designing AI-assisted Claim Prioritization for Human Fact-checkers0
LLMs as Debate Partners: Utilizing Genetic Algorithms and Adversarial Search for Adaptive ArgumentsCode0
Multimodal Fact-Checking with Vision Language Models: A Probing Classifier based Solution with Embedding StrategiesCode0
Anatomically-Grounded Fact Checking of Automated Chest X-ray Reports0
Do LLMs Understand Ambiguity in Text? A Case Study in Open-world Question Answering0
ZeFaV: Boosting Large Language Models for Zero-shot Fact VerificationCode0
Supporting Automated Fact-checking across Topics: Similarity-driven Gradual Topic Learning for Claim Detection0
Fact or Fiction? Can LLMs be Reliable Annotators for Political Truths?0
Evaluating Large Language Model Capability in Vietnamese Fact-Checking Data Generation0
Ev2R: Evaluating Evidence Retrieval in Automated Fact-Checking0
VERITAS: A Unified Approach to Reliability Evaluation0
Provenance: A Light-weight Fact-checker for Retrieval Augmented LLM Generation Output0
InFact: A Strong Baseline for Automated Fact-Checking0
LEAF: Learning and Evaluation Augmented by Fact-Checking to Improve Factualness in Large Language Models0
Show:102550
← PrevPage 9 of 27Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1monoT5-3BnDCG@100.78Unverified
2SGPT-BE-5.8BnDCG@100.75Unverified
3BM25+CEnDCG@100.69Unverified
4SGPT-CE-6.1BnDCG@100.68Unverified
5ColBERTnDCG@100.67Unverified
#ModelMetricClaimedVerifiedStatus
1SGPT-BE-5.8BnDCG@100.31Unverified
2monoT5-3BnDCG@100.28Unverified
3BM25+CEnDCG@100.25Unverified
4SGPT-CE-6.1BnDCG@100.16Unverified
#ModelMetricClaimedVerifiedStatus
1monoT5-3BnDCG@100.85Unverified
2BM25+CEnDCG@100.82Unverified
3SGPT-BE-5.8BnDCG@100.78Unverified
4SGPT-CE-6.1BnDCG@100.73Unverified
#ModelMetricClaimedVerifiedStatus
1HerOQuestion Only score0.48Unverified
2CTU AICQuestion Only score0.46Unverified
3InFactQuestion Only score0.45Unverified
#ModelMetricClaimedVerifiedStatus
1Abc0..5sec2Unverified
#ModelMetricClaimedVerifiedStatus
1MA-CINPrecision0.26Unverified
#ModelMetricClaimedVerifiedStatus
1FDHNAccuracy (Test)0.7Unverified