SOTAVerified

Model extraction

Model extraction attacks, aka model stealing attacks, are used to extract the parameters from the target model. Ideally, the adversary will be able to steal and replicate a model that will have a very similar performance to the target model.

Papers

Showing 110 of 176 papers

TitleStatusHype
Entangled Threats: A Unified Kill Chain Model for Quantum Machine Learning Security0
CEGA: A Cost-Effective Approach for Graph-Based Model Extraction and AcquisitionCode0
Navigating the Deep: Signature Extraction on Deep Neural Networks0
Explore the vulnerability of black-box models via diffusion models0
GradEscape: A Gradient-Based Evader Against AI-Generated Text Detectors0
MISLEADER: Defending against Model Extraction with Ensembles of Distilled ModelsCode0
Evaluating Query Efficiency and Accuracy of Transfer Learning-based Model Extraction Attack in Federated Learning0
On the interplay of Explainability, Privacy and Predictive Performance with Explanation-assisted Model Extraction0
Better Decisions through the Right Causal World Model0
CopyQNN: Quantum Neural Network Extraction Attack under Varying Quantum Noise0
Show:102550
← PrevPage 1 of 18Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1three-step-originalExact Match0.17Unverified