SOTAVerified

Instruction Following

Instruction following is the basic task of the model. This task is dedicated to evaluating the ability of the large model to follow human instructions. It is hoped that the model can generate controllable and safe answers.

Papers

Showing 2130 of 1135 papers

TitleStatusHype
HalLoc: Token-level Localization of Hallucinations for Vision Language ModelsCode0
AC/DC: LLM-based Audio Comprehension via Dialogue Continuation0
Magistral0
Conversational Search: From Fundamentals to Frontiers in the LLM Era0
Discovering Hierarchical Latent Capabilities of Language Models via Causal Representation LearningCode0
Alzheimer's Dementia Detection Using Perplexity from Paired Large Language Models0
VerIF: Verification Engineering for Reinforcement Learning in Instruction FollowingCode2
EIFBENCH: Extremely Complex Instruction Following Benchmark for Large Language ModelsCode0
RHealthTwin: Towards Responsible and Multimodal Digital Twins for Personalized Well-being0
LLaVA-c: Continual Improved Visual Instruction Tuning0
Show:102550
← PrevPage 3 of 114Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1AutoIF (Llama3 70B)Inst-level loose-accuracy90.4Unverified
2AutoIF (Qwen2 72B)Inst-level loose-accuracy88Unverified
3GPT-4Inst-level loose-accuracy85.37Unverified
4PaLM 2 SInst-level loose-accuracy59.11Unverified