SOTAVerified

TruthfulQA

Papers

Showing 7180 of 80 papers

TitleStatusHype
LACIE: Listener-Aware Finetuning for Confidence Calibration in Large Language ModelsCode0
SaGE: Evaluating Moral Consistency in Large Language ModelsCode0
Unsupervised Elicitation of Language ModelsCode0
VarBench: Robust Language Model Benchmarking Through Dynamic Variable PerturbationCode0
Enhancing Language Model Factuality via Activation-Based Confidence Calibration and Guided DecodingCode0
DeLTa: A Decoding Strategy based on Logit Trajectory Prediction Improves Factuality and Reasoning AbilityCode0
Truth Knows No Language: Evaluating Truthfulness Beyond EnglishCode0
Truth NeuronsCode0
CHAIR -- Classifier of Hallucination as ImproverCode0
Just Ask for Calibration: Strategies for Eliciting Calibrated Confidence Scores from Language Models Fine-Tuned with Human FeedbackCode0
Show:102550
← PrevPage 8 of 8Next →

No leaderboard results yet.