SOTAVerified

Fact Checking

Papers

Showing 101150 of 669 papers

TitleStatusHype
Truth or Mirage? Towards End-to-End Factuality Evaluation with LLM-OasisCode1
Do LLMs Understand Ambiguity in Text? A Case Study in Open-world Question Answering0
ZeFaV: Boosting Large Language Models for Zero-shot Fact VerificationCode0
Evaluating Large Language Model Capability in Vietnamese Fact-Checking Data Generation0
Supporting Automated Fact-checking across Topics: Similarity-driven Gradual Topic Learning for Claim Detection0
Ev2R: Evaluating Evidence Retrieval in Automated Fact-Checking0
Fact or Fiction? Can LLMs be Reliable Annotators for Political Truths?0
VERITAS: A Unified Approach to Reliability Evaluation0
InFact: A Strong Baseline for Automated Fact-Checking0
Provenance: A Light-weight Fact-checker for Retrieval Augmented LLM Generation Output0
LEAF: Learning and Evaluation Augmented by Fact-Checking to Improve Factualness in Large Language Models0
SubjECTive-QA: Measuring Subjectivity in Earnings Call Transcripts' QA Through Six-Dimensional Feature AnalysisCode0
Belief in the Machine: Investigating Epistemological Blind Spots of Language ModelsCode1
A Debate-Driven Experiment on LLM Hallucinations and Accuracy0
ChunkRAG: Novel LLM-Chunk Filtering Method for RAG Systems0
Health Misinformation in Social Networks: A Survey of IT Approaches0
ViMGuard: A Novel Multi-Modal System for Video Misinformation Guarding0
Learning to Generate and Evaluate Fact-checking Explanations with Transformers0
Augmenting the Veracity and Explanations of Complex Fact Checking via Iterative Self-Revision with LLMs0
ChronoFact: Timeline-based Temporal Fact Verification0
Real-time Fake News from Adversarial FeedbackCode0
Decomposition Dilemmas: Does Claim Decomposition Boost or Burden Fact-Checking Performance?0
FIRE: Fact-checking with Iterative Retrieval and VerificationCode1
HerO at AVeriTeC: The Herd of Open Large Language Models for Verifying Real-World ClaimsCode1
AIC CTU system at AVeriTeC: Re-framing automated fact-checking as a simple RAG taskCode0
A Comparative Study of Translation Bias and Accuracy in Multilingual Large Language Models for Cross-Language Claim VerificationCode0
ECIS-VQG: Generation of Entity-centric Information-seeking Questions from VideosCode0
Contrastive Learning to Improve Retrieval for Real-world Fact Checking0
LRQ-Fact: LLM-Generated Relevant Questions for Multimodal Fact-Checking0
Take It Easy: Label-Adaptive Self-Rationalization for Fact Verification and Explanation GenerationCode0
Loki: An Open-Source Tool for Fact VerificationCode5
How Entangled is Factuality and Deception in German?0
Multimodal Misinformation Detection by Learning from Synthetic Data with Multimodal LLMs0
HybridFC: A Hybrid Fact-Checking Approach for Knowledge GraphsCode0
LLMs Will Always Hallucinate, and We Need to Live With This0
LoraMap: Harnessing the Power of LoRA Connections0
Grounding Fallacies Misrepresenting Scientific Publications in EvidenceCode0
Evidence-backed Fact Checking using RAG and Few-Shot In-Context Learning with LLMsCode0
"Image, Tell me your story!" Predicting the original meta-context of visual misinformationCode1
CommunityKG-RAG: Leveraging Community Structures in Knowledge Graphs for Advanced Retrieval-Augmented Generation in Fact-CheckingCode0
Web Retrieval Agents for Evidence-Based Misinformation DetectionCode0
Zero-Shot Learning and Key Points Are All You Need for Automated Fact-CheckingCode0
KGV: Integrating Large Language Models with Knowledge Graphs for Cyber Threat Intelligence Credibility Assessment0
LiveFC: A System for Live Fact-Checking of Audio Streams0
Crowd Intelligence for Early Misinformation Prediction on Social MediaCode0
Medical Graph RAG: Towards Safe Medical Large Language Model via Graph Retrieval-Augmented GenerationCode4
OpenFactCheck: A Unified Framework for Factuality Evaluation of LLMsCode1
The Implications of Open Generative Models in Human-Centered Data Science Work: A Case Study with Fact-Checking Organizations0
QuestGen: Effectiveness of Question Generation Methods for Fact-Checking ApplicationsCode0
Cost-Effective Hallucination Detection for LLMs0
Show:102550
← PrevPage 3 of 14Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1monoT5-3BnDCG@100.78Unverified
2SGPT-BE-5.8BnDCG@100.75Unverified
3BM25+CEnDCG@100.69Unverified
4SGPT-CE-6.1BnDCG@100.68Unverified
5ColBERTnDCG@100.67Unverified
#ModelMetricClaimedVerifiedStatus
1SGPT-BE-5.8BnDCG@100.31Unverified
2monoT5-3BnDCG@100.28Unverified
3BM25+CEnDCG@100.25Unverified
4SGPT-CE-6.1BnDCG@100.16Unverified
#ModelMetricClaimedVerifiedStatus
1monoT5-3BnDCG@100.85Unverified
2BM25+CEnDCG@100.82Unverified
3SGPT-BE-5.8BnDCG@100.78Unverified
4SGPT-CE-6.1BnDCG@100.73Unverified
#ModelMetricClaimedVerifiedStatus
1HerOQuestion Only score0.48Unverified
2CTU AICQuestion Only score0.46Unverified
3InFactQuestion Only score0.45Unverified
#ModelMetricClaimedVerifiedStatus
1Abc0..5sec2Unverified
#ModelMetricClaimedVerifiedStatus
1MA-CINPrecision0.26Unverified
#ModelMetricClaimedVerifiedStatus
1FDHNAccuracy (Test)0.7Unverified