SOTAVerified

Instruction Following

Instruction following is the basic task of the model. This task is dedicated to evaluating the ability of the large model to follow human instructions. It is hoped that the model can generate controllable and safe answers.

Papers

Showing 761770 of 1135 papers

TitleStatusHype
Reminding Multimodal Large Language Models of Object-aware Knowledge with Retrieved Tags0
Rethinking Bottlenecks in Safety Fine-Tuning of Vision Language Models0
Rethinking Predictive Modeling for LLM Routing: When Simple kNN Beats Complex Learned Routers0
Retrieval Augmented Chest X-Ray Report Generation using OpenAI GPT models0
URO-Bench: A Comprehensive Benchmark for End-to-End Spoken Dialogue Models0
RevisEval: Improving LLM-as-a-Judge via Response-Adapted References0
Revisiting the Superficial Alignment Hypothesis0
Shuttle Between the Instructions and the Parameters of Large Language Models0
A Systematic Examination of Preference Learning through the Lens of Instruction-Following0
A Survey of Reinforcement Learning Informed by Natural Language0
Show:102550
← PrevPage 77 of 114Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1AutoIF (Llama3 70B)Inst-level loose-accuracy90.4Unverified
2AutoIF (Qwen2 72B)Inst-level loose-accuracy88Unverified
3GPT-4Inst-level loose-accuracy85.37Unverified
4PaLM 2 SInst-level loose-accuracy59.11Unverified