SOTAVerified

Model extraction

Model extraction attacks, aka model stealing attacks, are used to extract the parameters from the target model. Ideally, the adversary will be able to steal and replicate a model that will have a very similar performance to the target model.

Papers

Showing 3140 of 176 papers

TitleStatusHype
"Yes, My LoRD." Guiding Language Model Extraction with Locality Reinforced DistillationCode1
VidModEx: Interpretable and Efficient Black Box Model Extraction for High-Dimensional SpacesCode0
Enhancing TinyML Security: Study of Adversarial Attack Transferability0
QUEEN: Query Unlearning against Model Extraction0
Privacy Implications of Explainable AI in Data-Driven Systems0
Beyond Slow Signs in High-fidelity Model ExtractionCode0
GENIE: Watermarking Graph Neural Networks for Link Prediction0
Watermarking Counterfactual ExplanationsCode0
Noisy Data Meets Privacy: Training Local Models with Post-Processed Remote Queries0
DeepNcode: Encoding-Based Protection against Bit-Flip Attacks on Neural Networks0
Show:102550
← PrevPage 4 of 18Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1three-step-originalExact Match0.17Unverified