SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 101150 of 492 papers

TitleStatusHype
Does Low Rank Adaptation Lead to Lower Robustness against Training-Time Attacks?Code0
Towards Robust Spiking Neural Networks:Mitigating Heterogeneous Training Vulnerability via Dominant Eigencomponent Projection0
Sybil-based Virtual Data Poisoning Attacks in Federated Learning0
Stealthy LLM-Driven Data Poisoning Attacks Against Embedding-Based Retrieval-Augmented Recommender Systems0
Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents0
Adversarial Robustness of Deep Learning Models for Inland Water Body Segmentation from SAR ImagesCode0
What's Pulling the Strings? Evaluating Integrity and Attribution in AI Training and Inference through Concept Shift0
A Geometric Approach to Problems in Optimization and Data Science0
Investigating cybersecurity incidents using large language models in latest-generation wireless networks0
ControlNET: A Firewall for RAG-based LLM System0
Diversity-aware Dual-promotion Poisoning Attack on Sequential Recommendation0
Sky of Unlearning (SoUL): Rewiring Federated Machine Unlearning via Selective Pruning0
Clean Image May be Dangerous: Data Poisoning Attacks Against Deep Hashing0
Optimizing ML Training with Metagradient Descent0
Policy Teaching via Data Poisoning in Learning from Human Preferences0
Targeted Data Poisoning for Black-Box Audio Datasets Ownership Verification0
Silent Branding Attack: Trigger-free Data Poisoning Attack on Text-to-Image Diffusion Models0
PoisonedParrot: Subtle Data Poisoning Attacks to Elicit Copyright-Infringing Content from Large Language Models0
Poisoning Attacks to Local Differential Privacy Protocols for Trajectory Data0
Data Poisoning Attacks to Locally Differentially Private Range Query Protocols0
Approaching the Harm of Gradient Attacks While Only Flipping Labels0
Atlas: A Framework for ML Lifecycle Provenance & Transparency0
No, of course I can! Refusal Mechanisms Can Be Exploited Using Harmless Fine-Tuning Data0
FedNIA: Noise-Induced Activation Analysis for Mitigating Data Poisoning in FL0
Swallowing the Poison Pills: Insights from Vulnerability Disparity Among LLMs0
Keeping up with dynamic attackers: Certifying robustness to adaptive online data poisoningCode0
Multi-Faceted Studies on Data Poisoning can Advance LLM DevelopmentCode0
A Robust Attack: Displacement Backdoor Attack0
DROP: Poison Dilution via Knowledge Distillation for Federated LearningCode0
Filter, Obstruct and Dilute: Defending Against Backdoor Attacks on Semi-Supervised Learning0
Detection of Physiological Data Tampering Attacks with Quantum Machine Learning0
TrojanTime: Backdoor Attacks on Time Series Classification0
Provably effective detection of effective data poisoning attacks0
Poison-RAG: Adversarial Data Poisoning Attacks on Retrieval-Augmented Generation in Recommender SystemsCode0
A Novel Pearson Correlation-Based Merging Algorithm for Robust Distributed Machine Learning with Heterogeneous Data0
Fortifying Federated Learning Towards Trustworthiness via Auditable Data Valuation and Verifiable Client Contribution0
Cut the Deadwood Out: Post-Training Model Purification with Selective Module Substitution0
Attacks on the neural network and defense methods0
Trading Devil RL: Backdoor attack via Stock market, Bayesian Optimization and Reinforcement Learning0
From Vulnerabilities to Remediation: A Systematic Literature Review of LLMs in Code Security0
One Pixel is All I Need0
BiCert: A Bilinear Mixed Integer Programming Formulation for Precise Certified Bounds Against Data Poisoning Attacks0
Deep Learning Model Security: Threats and Defenses0
Learning to Forget using Hypernetworks0
Efficient and Private: Memorisation under differentially private parameter-efficient fine-tuning in language models0
Adversarial Data Poisoning Attacks on Quantum Machine Learning in the NISQ Era0
Delta-Influence: Unlearning Poisons via Influence FunctionsCode0
Reliable Poisoned Sample Detection against Backdoor Attacks Enhanced by Sharpness Aware Minimization0
SAFELOC: Overcoming Data Poisoning Attacks in Heterogeneous Federated Machine Learning for Indoor Localization0
Learning from Convolution-based Unlearnable DatasetsCode0
Show:102550
← PrevPage 3 of 10Next →

No leaderboard results yet.