SOTAVerified

TruthfulQA

Papers

Showing 2650 of 80 papers

TitleStatusHype
A Debate-Driven Experiment on LLM Hallucinations and Accuracy0
Evaluating Consistencies in LLM responses through a Semantic Clustering of Question Answering0
Iter-AHMCL: Alleviate Hallucination for Large Language Model via Iterative Model-level Contrastive Learning0
SkillAggregation: Reference-free LLM-Dependent Aggregation0
Benchmark Inflation: Revealing LLM Performance Gaps Using Retro-Holdouts0
NoVo: Norm Voting off Hallucinations with Attention Heads in Large Language ModelsCode0
Towards Multilingual LLM Evaluation for European Languages0
A test suite of prompt injection attacks for LLM-based machine translationCode0
Efficiently Deploying LLMs with Controlled Risk0
Integrative Decoding: Improve Factuality via Implicit Self-consistencyCode1
Teuken-7B-Base & Teuken-7B-Instruct: Towards European LLMs0
Selective Self-Rehearsal: A Fine-Tuning Approach to Improve Generalization in Large Language Models0
Lower Layer Matters: Alleviating Hallucination via Multi-Layer Fusion Contrastive Decoding with Truthfulness Refocused0
LokiLM: Technical Report0
metabench -- A Sparse Benchmark to Measure General Ability in Large Language ModelsCode0
VarBench: Robust Language Model Benchmarking Through Dynamic Variable PerturbationCode0
Steering Without Side Effects: Improving Post-Deployment Control of Language ModelsCode0
Enhancing Language Model Factuality via Activation-Based Confidence Calibration and Guided DecodingCode0
LACIE: Listener-Aware Finetuning for Confidence Calibration in Large Language ModelsCode0
Multi-Reference Preference Optimization for Large Language Models0
Machine Unlearning in Large Language ModelsCode1
Instruction Tuning With Loss Over InstructionsCode1
RLHF Workflow: From Reward Modeling to Online RLHFCode5
Harmonic LLMs are Trustworthy0
Student Data Paradox and Curious Case of Single Student-Tutor Model: Regressive Side Effects of Training LLMs for Personalized Learning0
Show:102550
← PrevPage 2 of 4Next →

No leaderboard results yet.