SOTAVerified

Model extraction

Model extraction attacks, aka model stealing attacks, are used to extract the parameters from the target model. Ideally, the adversary will be able to steal and replicate a model that will have a very similar performance to the target model.

Papers

Showing 125 of 176 papers

TitleStatusHype
Safety at Scale: A Comprehensive Survey of Large Model SafetyCode3
Are You Copying My Model? Protecting the Copyright of Large Language Models for EaaS via Backdoor WatermarkCode1
Black-Box Attacks on Sequential Recommenders via Data-Free Model ExtractionCode1
Protecting Language Generation Models via Invisible WatermarkingCode1
Neural Honeytrace: A Robust Plug-and-Play Watermarking Framework against Model Extraction AttacksCode1
MEME: Generating RNN Model Explanations via Model ExtractionCode1
Data-Free Model ExtractionCode1
MARLeME: A Multi-Agent Reinforcement Learning Model Extraction LibraryCode1
MEME: Generating RNN Model Explanations via Model ExtractionCode1
Now You See Me (CME): Concept-based Model ExtractionCode1
Watermarking Vision-Language Pre-trained Models for Multi-modal Embedding as a ServiceCode1
ATOM: A Framework of Detecting Query-Based Model Extraction Attacks for Graph Neural NetworksCode1
"Yes, My LoRD." Guiding Language Model Extraction with Locality Reinforced DistillationCode1
Model Extraction and Adversarial Transferability, Your BERT is Vulnerable!Code1
MEA-Defender: A Robust Watermark against Model Extraction AttackCode1
Cryptanalytic Extraction of Neural Network ModelsCode1
Entangled Watermarks as a Defense against Model ExtractionCode1
FedRolex: Model-Heterogeneous Federated Learning with Rolling Sub-Model ExtractionCode1
An anatomy-based V1 model: Extraction of Low-level Features, Reduction of distortion and a V1-inspired SOM0
Adversarial Exploitation of Policy Imitation0
A Knowledge Representation Approach to Automated Mathematical Modelling0
AUTOLYCUS: Exploiting Explainable AI (XAI) for Model Extraction Attacks against Interpretable Models0
A Desynchronization-Based Countermeasure Against Side-Channel Analysis of Neural Networks0
Beyond Labeling Oracles: What does it mean to steal ML models?0
Student Surpasses Teacher: Imitation Attack for Black-Box NLP APIs0
Show:102550
← PrevPage 1 of 8Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1three-step-originalExact Match0.17Unverified