SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 376400 of 492 papers

TitleStatusHype
On the Adversarial Risk of Test Time Adaptation: An Investigation into Realistic Test-Time Data Poisoning0
On the Effectiveness of Poisoning against Unsupervised Domain Adaptation0
RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with Human Feedback in Large Language Models0
On the Relevance of Byzantine Robust Optimization Against Data Poisoning0
On the Robustness of Graph Reduction Against GNN Backdoor0
A Study of Backdoors in Instruction Fine-tuned Language Models0
Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents0
Optimizing ML Training with Metagradient Descent0
Oriole: Thwarting Privacy against Trustworthy Deep Learning Models0
OVLA: Neural Network Ownership Verification using Latent Watermarks0
PACOL: Poisoning Attacks Against Continual Learners0
Partner in Crime: Boosting Targeted Poisoning Attacks against Federated Learning0
Pick your Poison: Undetectability versus Robustness in Data Poisoning Attacks0
PoisHygiene: Detecting and Mitigating Poisoning Attacks in Neural Networks0
PoisonedEncoder: Poisoning the Unlabeled Pre-training Data in Contrastive Learning0
PoisonedParrot: Subtle Data Poisoning Attacks to Elicit Copyright-Infringing Content from Large Language Models0
Poisoning Attacks and Defenses on Artificial Intelligence: A Survey0
Poisoning Attacks to Local Differential Privacy Protocols for Trajectory Data0
Poisoning Deep Reinforcement Learning Agents with In-Distribution Triggers0
Poisoning Programs by Un-Repairing Code: Security Concerns of AI-generated Code0
Policy Teaching via Data Poisoning in Learning from Human Preferences0
Post-Training Overfitting Mitigation in DNN Classifiers0
Practical Data Poisoning Attack against Next-Item Recommendation0
SLSGD: Secure and Efficient Distributed On-device Machine Learning0
Practical Poisoning Attacks on Neural Networks0
Show:102550
← PrevPage 16 of 20Next →

No leaderboard results yet.