SOTAVerified

backdoor defense

Papers

Showing 150 of 131 papers

TitleStatusHype
REFINE: Inversion-Free Backdoor Defense via Model ReprogrammingCode4
Watch Out for Your Agents! Investigating Backdoor Threats to LLM-Based AgentsCode2
Gracefully Filtering Backdoor Samples for Generative Large Language Models without RetrainingCode1
CROW: Eliminating Backdoors from Large Language Models via Internal Consistency RegularizationCode1
BackdoorMBTI: A Backdoor Learning Multimodal Benchmark Tool Kit for Backdoor Defense EvaluationCode1
Uncovering, Explaining, and Mitigating the Superficial Safety of Backdoor DefenseCode1
Fisher Information guided Purification against Backdoor AttacksCode1
VFLIP: A Backdoor Defense for Vertical Federated Learning via Identification and PurificationCode1
Lockdown: Backdoor Defense for Federated Learning with Isolated Subspace TrainingCode1
Lockdown: Backdoor Defense for Federated Learning with Isolated Subspace TrainingCode1
Reconstructive Neuron Pruning for Backdoor DefenseCode1
Text-to-Image Diffusion Models can be Easily Backdoored through Multimodal Data PoisoningCode1
Backdoor Defense via Adaptively Splitting Poisoned DatasetCode1
Black-box Backdoor Defense via Zero-shot Image PurificationCode1
Backdoor Defense via Deconfounded Representation LearningCode1
ASSET: Robust Backdoor Data Detection Across a Multiplicity of Deep Learning ParadigmsCode1
Backdoor Attacks for Remote Sensing Data with Wavelet TransformCode1
Effective Backdoor Defense by Exploiting Sensitivity of Poisoned SamplesCode1
FLIP: A Provable Defense Framework for Backdoor Mitigation in Federated LearningCode1
Expose Backdoors on the Way: A Feature-Based Efficient Defense against Textual Backdoor AttacksCode1
Trap and Replace: Defending Backdoor Attacks by Trapping Them into an Easy-to-Replace SubnetworkCode1
MM-BD: Post-Training Detection of Backdoor Attacks with Arbitrary Backdoor Pattern Types Using a Maximum Margin StatisticCode1
Eliminating Backdoor Triggers for Deep Neural Networks Using Attention Relation Graph DistillationCode1
Backdoor Defense via Decoupling the Training ProcessCode1
FIBA: Frequency-Injection based Backdoor Attack in Medical Image AnalysisCode1
LIRA: Learnable, Imperceptible and Robust Backdoor AttacksCode1
ONION: A Simple and Effective Defense Against Textual Backdoor AttacksCode1
Towards Probabilistic Verification of Machine UnlearningCode1
Clean-Label Backdoor Attacks on Video Recognition ModelsCode1
CLIP-Guided Backdoor Defense through Entropy-Based Poisoned Dataset SeparationCode0
CUBA: Controlled Untargeted Backdoor Attack against Deep Neural Networks0
InverTune: Removing Backdoors from Multimodal Contrastive Learning Models via Trigger Inversion and Activation Tuning0
TED-LaST: Towards Robust Backdoor Defense Against Adaptive Attacks0
SRD: Reinforcement-Learned Semantic Perturbation for Backdoor Defense in VLMs0
Robust Anti-Backdoor Instruction Tuning in LVLMs0
Spectral Insights into Data-Oblivious Critical Layers in Large Language Models0
Heterogeneous Graph Backdoor Attack0
FL-PLAS: Federated Learning with Partial Layer Aggregation for Backdoor Defense Against High-Ratio Malicious ClientsCode0
Cert-SSB: Toward Certified Sample-Specific Backdoor DefenseCode0
TrojanDam: Detection-Free Backdoor Defense in Federated Learning through Proactive Model Robustification utilizing OOD DataCode0
Backdoor Defense in Diffusion Models via Spatial Attention Unlearning0
Decoupled Distillation to Erase: A General Unlearning Method for Any Class-centric Tasks0
Prototype Guided Backdoor Defense0
Gungnir: Exploiting Stylistic Features in Images for Backdoor Attacks on Diffusion ModelsCode0
A Dual-Purpose Framework for Backdoor Defense and Backdoor Amplification in Diffusion Models0
Class-Conditional Neural Polarizer: A Lightweight and Effective Backdoor Defense by Purifying Poisoned Features0
Gradient Norm-based Fine-Tuning for Backdoor Defense in Automatic Speech Recognition0
PCAP-Backdoor: Backdoor Poisoning Generator for Network Traffic in CPS/IoT Environments0
Towards Backdoor Stealthiness in Model Parameter SpaceCode0
Backdoor Token Unlearning: Exposing and Defending Backdoors in Pretrained Language ModelsCode0
Show:102550
← PrevPage 1 of 3Next →

No leaderboard results yet.