SOTAVerified

parameter-efficient fine-tuning

Parameter-Efficient Fine-Tuning (PEFT) is a technique used to adapt pre-trained models to new tasks with minimal changes to the model's parameters. This approach is particularly useful in scenarios where computational resources are limited or when it is desirable to maintain the original model's performance on the initial task.

Papers

Showing 691700 of 935 papers

TitleStatusHype
Updating CLIP to Prefer Descriptions Over CaptionsCode0
A Parameter-efficient Language Extension Framework for Multilingual ASR0
An Improved Empirical Fisher Approximation for Natural Gradient Descent0
A Survey of Recent Backdoor Attacks and Defenses in Large Language Models0
Efficient Differentially Private Fine-Tuning of Diffusion Models0
Time Sensitive Knowledge Editing through Efficient Finetuning0
VHDL-Eval: A Framework for Evaluating Large Language Models in VHDL Code Generation0
Hypernetworks for Personalizing ASR to Atypical Speech0
Choice of PEFT Technique in Continual Learning: Prompt Tuning is Not All You Need0
Adapter-X: A Novel General Parameter-Efficient Fine-Tuning Framework for Vision0
Show:102550
← PrevPage 70 of 94Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1LLaMA2-7bAccuracy (% )82.63Unverified
2LLaMA2-7bAccuracy (% )82.63Unverified
3LLaMA2-7bAccuracy (% )81.93Unverified
4LLaMA2-7bAccuracy (% )80.28Unverified
#ModelMetricClaimedVerifiedStatus
1LLaMA2-7bAccuracy (% )76.68Unverified
2LLaMA2-7bAccuracy (% )76.67Unverified
3LLaMA2-7bAccuracy (% )76.27Unverified
#ModelMetricClaimedVerifiedStatus
1LLaMA2-7bAccuracy (% )70.8Unverified
2LLaMA2-7bAccuracy (% )70.09Unverified
3LLaMA2-7bAccuracy (% )69.85Unverified