SOTAVerified

Backdoor Attack

Backdoor attacks inject maliciously constructed data into a training set so that, at test time, the trained model misclassifies inputs patched with a backdoor trigger as an adversarially-desired target class.

Papers

Showing 201250 of 523 papers

TitleStatusHype
Model Pairing Using Embedding Translation for Backdoor Attack Detection on Open-Set Classification TasksCode0
Low-Frequency Black-Box Backdoor Attack via Evolutionary Algorithm0
Mitigating Fine-tuning based Jailbreak Attack with Backdoor Enhanced Safety AlignmentCode1
Whispers in Grammars: Injecting Covert Backdoors to Compromise Dense Retrieval SystemsCode0
VL-Trojan: Multimodal Instruction Backdoor Attacks against Autoregressive Visual Language Models0
Defending Against Weight-Poisoning Backdoor Attacks for Parameter-Efficient Fine-Tuning0
Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery DetectionCode1
Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based AgentsCode2
Backdoor Attack against One-Class Sequential Anomaly Detection ModelsCode0
Test-Time Backdoor Attacks on Multimodal Large Language ModelsCode2
OrderBkd: Textual backdoor attack through repositioningCode0
The last Dance : Robust backdoor attack via diffusion models and bayesian approach0
DisDet: Exploring Detectability of Backdoor Attack on Diffusion Models0
Model Supply Chain Poisoning: Backdooring Pre-trained Models via Embedding IndistinguishabilityCode1
BackdoorBench: A Comprehensive Benchmark and Analysis of Backdoor Learning0
BadChain: Backdoor Chain-of-Thought Prompting for Large Language ModelsCode2
Universal Vulnerabilities in Large Language Models: Backdoor Attacks for In-context Learning0
Inferring Properties of Graph Neural Networks0
The Stronger the Diffusion Model, the Easier the Backdoor: Data Poisoning to Induce Copyright Breaches Without Adjusting Finetuning Pipeline0
TEN-GUARD: Tensor Decomposition for Backdoor Attack Detection in Deep Neural Networks0
Object-oriented backdoor attack against image captioning0
Effective backdoor attack on graph neural networks in link prediction tasks0
Spy-Watermark: Robust Invisible Watermarking for Backdoor AttackCode0
The Art of Deception: Robust Backdoor Attack using Dynamic Stacking of Triggers0
Imperio: Language-Guided Backdoor Attacks for Arbitrary Model Control0
Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision TransfomersCode1
Backdoor Attack on Unpaired Medical Image-Text Foundation Models: A Pilot Study on MedCLIPCode0
Does Few-shot Learning Suffer from Backdoor Attacks?0
Is It Possible to Backdoor Face Forgery Detection with Natural Triggers?0
A clean-label graph backdoor attack method in node classification task0
SSL-OTA: Unveiling Backdoor Threats in Self-Supervised Learning for Object Detection0
Punctuation Matters! Stealthy Backdoor Attack for Language Models0
BadRL: Sparse Targeted Backdoor Attack Against Reinforcement LearningCode0
FlowMur: A Stealthy and Practical Audio Backdoor Attack with Limited KnowledgeCode1
Towards Sample-specific Backdoor Attack with Clean Labels via Attribute Trigger0
TARGET: Template-Transferable Backdoor Attack Against Prompt-based NLP Models via GPT40
Rethinking Backdoor Attacks on Dataset Distillation: A Kernel Method Perspective0
Universal Jailbreak Backdoors from Poisoned Human FeedbackCode1
Attacks on fairness in Federated LearningCode0
BadCLIP: Dual-Embedding Guided Backdoor Attack on Multimodal Contrastive LearningCode1
RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with Human Feedback in Large Language Models0
Tabdoor: Backdoor Vulnerabilities in Transformer-based Neural Networks for Tabular Data0
From Trojan Horses to Castle Walls: Unveiling Bilateral Data Poisoning Effects in Diffusion ModelsCode0
Label Poisoning is All You NeedCode1
CBD: A Certified Backdoor Detector Based on Local Dominant ProbabilityCode0
PoisonPrompt: Backdoor Attack on Prompt-based Large Language ModelsCode1
WaveAttack: Asymmetric Frequency Obfuscation-based Backdoor Attacks Against Deep Neural Networks0
Demystifying Poisoning Backdoor Attacks from a Statistical Perspective0
Invisible Threats: Backdoor Attack in OCR Systems0
Composite Backdoor Attacks Against Large Language ModelsCode1
Show:102550
← PrevPage 5 of 11Next →

No leaderboard results yet.