SOTAVerified

Instruction Following

Instruction following is the basic task of the model. This task is dedicated to evaluating the ability of the large model to follow human instructions. It is hoped that the model can generate controllable and safe answers.

Papers

Showing 961970 of 1135 papers

TitleStatusHype
ChatSpot: Bootstrapping Multimodal LLMs via Precise Referring Instruction Tuning0
AlpaGasus: Training A Better Alpaca with Fewer DataCode1
BuboGPT: Enabling Visual Grounding in Multi-Modal LLMsCode2
Do Emergent Abilities Exist in Quantized Large Language Models: An Empirical StudyCode1
Exploring the Integration of Large Language Models into Automatic Speech Recognition Systems: An Empirical Study0
MMBench: Is Your Multi-modal Model an All-around Player?Code5
Instruction Mining: Instruction Data Selection for Tuning Large Language Models0
Opening up ChatGPT: Tracking openness, transparency, and accountability in instruction-tuned text generatorsCode1
Becoming self-instruct: introducing early stopping criteria for minimal instruct tuning0
What Matters in Training a GPT4-Style Language Model with Multimodal Inputs?Code2
Show:102550
← PrevPage 97 of 114Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1AutoIF (Llama3 70B)Inst-level loose-accuracy90.4Unverified
2AutoIF (Qwen2 72B)Inst-level loose-accuracy88Unverified
3GPT-4Inst-level loose-accuracy85.37Unverified
4PaLM 2 SInst-level loose-accuracy59.11Unverified