SOTAVerified

Instruction Following

Instruction following is the basic task of the model. This task is dedicated to evaluating the ability of the large model to follow human instructions. It is hoped that the model can generate controllable and safe answers.

Papers

Showing 901910 of 1135 papers

TitleStatusHype
Finding Fantastic Experts in MoEs: A Unified Study for Expert Dropping Strategies and Observations0
Fine-Tuning Enhances Existing Mechanisms: A Case Study on Entity Tracking0
Fine-Tuning Large Vision-Language Models as Decision-Making Agents via Reinforcement Learning0
FLAME: Factuality-Aware Alignment for Large Language Models0
Don't Half-listen: Capturing Key-part Information in Continual Instruction Tuning0
Domain Adaptation of VLM for Soccer Video Understanding0
FlowKV: Enhancing Multi-Turn Conversational Coherence in LLMs via Isolated Key-Value Cache Management0
GuideBench: Benchmarking Domain-Oriented Guideline Following for LLM Agents0
FollowEval: A Multi-Dimensional Benchmark for Assessing the Instruction-Following Capability of Large Language Models0
StyleAR: Customizing Multimodal Autoregressive Model for Style-Aligned Text-to-Image Generation0
Show:102550
← PrevPage 91 of 114Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1AutoIF (Llama3 70B)Inst-level loose-accuracy90.4Unverified
2AutoIF (Qwen2 72B)Inst-level loose-accuracy88Unverified
3GPT-4Inst-level loose-accuracy85.37Unverified
4PaLM 2 SInst-level loose-accuracy59.11Unverified