SOTAVerified

TruthfulQA

Papers

Showing 5160 of 80 papers

TitleStatusHype
When Hindsight is Not 20/20: Testing Limits on Reflective Thinking in Large Language ModelsCode0
PoLLMgraph: Unraveling Hallucinations in Large Language Models via State Transition DynamicsCode0
PRobELM: Plausibility Ranking Evaluation for Language Models0
Non-Linear Inference Time Intervention: Improving LLM TruthfulnessCode1
In-Context Sharpness as Alerts: An Inner Representation Perspective for Hallucination MitigationCode2
TruthX: Alleviating Hallucinations by Editing Large Language Models in Truthful SpaceCode2
SaGE: Evaluating Moral Consistency in Large Language ModelsCode0
LLMAuditor: A Framework for Auditing Large Language Models Using Human-in-the-Loop0
Self-Alignment for Factuality: Mitigating Hallucinations in LLMs via Self-Evaluation0
GRATH: Gradual Self-Truthifying for Large Language Models0
Show:102550
← PrevPage 6 of 8Next →

No leaderboard results yet.