SOTAVerified

Model extraction

Model extraction attacks, aka model stealing attacks, are used to extract the parameters from the target model. Ideally, the adversary will be able to steal and replicate a model that will have a very similar performance to the target model.

Papers

Showing 121130 of 176 papers

TitleStatusHype
High Accuracy and High Fidelity Extraction of Neural Networks0
HODA: Protecting DNNs Against Model Extraction Attacks via Hardness of Samples0
HoneypotNet: Backdoor Attacks Against Model Extraction0
Increasing the Cost of Model Extraction with Calibrated Proof of Work0
Interpretability via Model Extraction0
Interpreting Blackbox Models via Model Extraction0
Killing One Bird with Two Stones: Model Extraction and Attribute Inference Attacks against BERT-based APIs0
Noisy Data Meets Privacy: Training Local Models with Post-Processed Remote Queries0
Learnable Linguistic Watermarks for Tracing Model Extraction Attacks on Large Language Models0
Leveraging Extracted Model Adversaries for Improved Black Box Attacks0
Show:102550
← PrevPage 13 of 18Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1three-step-originalExact Match0.17Unverified