SOTAVerified

Backdoor Attack

Backdoor attacks inject maliciously constructed data into a training set so that, at test time, the trained model misclassifies inputs patched with a backdoor trigger as an adversarially-desired target class.

Papers

Showing 150 of 523 papers

TitleStatusHype
Agent Security Bench (ASB): Formalizing and Benchmarking Attacks and Defenses in LLM-based AgentsCode3
AgentPoison: Red-teaming LLM Agents via Poisoning Memory or Knowledge BasesCode3
BAPLe: Backdoor Attacks on Medical Foundational Models using Prompt LearningCode2
An LLM-Assisted Easy-to-Trigger Backdoor Attack on Code Completion Models: Injecting Disguised Vulnerabilities against Strong DetectionCode2
Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based AgentsCode2
Test-Time Backdoor Attacks on Multimodal Large Language ModelsCode2
BadChain: Backdoor Chain-of-Thought Prompting for Large Language ModelsCode2
Backdoor Learning: A SurveyCode2
To Think or Not to Think: Exploring the Unthinking Vulnerability in Large Reasoning ModelsCode1
Invisible Backdoor Attack against Self-supervised LearningCode1
CL-Attack: Textual Backdoor Attacks via Cross-Lingual TriggersCode1
BadCM: Invisible Backdoor Attack Against Cross-Modal LearningCode1
BadMerging: Backdoor Attacks Against Model MergingCode1
Uncertainty is Fragile: Manipulating Uncertainty in Large Language ModelsCode1
T2IShield: Defending Against Backdoors on Text-to-Image Diffusion ModelsCode1
Invisible Backdoor Attacks on Diffusion ModelsCode1
Fast-FedUL: A Training-Free Federated Unlearning with Provable Skew ResilienceCode1
Towards Imperceptible Backdoor Attack in Self-supervised LearningCode1
Rethinking Graph Backdoor Attacks: A Distribution-Preserving PerspectiveCode1
Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision TransformersCode1
Beyond Traditional Threats: A Persistent Backdoor Attack on Federated LearningCode1
Exploring Backdoor Vulnerabilities of Chat ModelsCode1
LOTUS: Evasive and Resilient Backdoor Attacks through Sub-PartitioningCode1
Generating Potent Poisons and Backdoors from Scratch with Guided DiffusionCode1
Mask-based Invisible Backdoor Attacks on Object DetectionCode1
BadEdit: Backdooring large language models by model editingCode1
Mitigating Fine-tuning based Jailbreak Attack with Backdoor Enhanced Safety AlignmentCode1
Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery DetectionCode1
Model Supply Chain Poisoning: Backdooring Pre-trained Models via Embedding IndistinguishabilityCode1
Not All Prompts Are Secure: A Switchable Backdoor Attack Against Pre-trained Vision TransfomersCode1
FlowMur: A Stealthy and Practical Audio Backdoor Attack with Limited KnowledgeCode1
Universal Jailbreak Backdoors from Poisoned Human FeedbackCode1
BadCLIP: Dual-Embedding Guided Backdoor Attack on Multimodal Contrastive LearningCode1
Label Poisoning is All You NeedCode1
PoisonPrompt: Backdoor Attack on Prompt-based Large Language ModelsCode1
Composite Backdoor Attacks Against Large Language ModelsCode1
VDC: Versatile Data Cleanser based on Visual-Linguistic Inconsistency by Multimodal Large Language ModelsCode1
PatchBackdoor: Backdoor Attack against Deep Neural Networks without Model ModificationCode1
BAGM: A Backdoor Attack for Manipulating Text-to-Image Generative ModelsCode1
Backdooring Instruction-Tuned Large Language Models with Virtual Prompt InjectionCode1
You Can Backdoor Personalized Federated LearningCode1
Risk-optimized Outlier Removal for Robust 3D Point Cloud ClassificationCode1
Towards Stealthy Backdoor Attacks against Speech Recognition via Elements of SoundCode1
FedDefender: Backdoor Attack Defense in Federated LearningCode1
Bkd-FedGNN: A Benchmark for Classification Backdoor Attacks on Federated Graph Neural NetworkCode1
VillanDiffusion: A Unified Backdoor Attack Framework for Diffusion ModelsCode1
Backdoor Attack with Sparse and Invisible TriggerCode1
Text-to-Image Diffusion Models can be Easily Backdoored through Multimodal Data PoisoningCode1
UNICORN: A Unified Backdoor Trigger Inversion FrameworkCode1
Influencer Backdoor Attack on Semantic SegmentationCode1
Show:102550
← PrevPage 1 of 11Next →

No leaderboard results yet.