SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 201250 of 492 papers

TitleStatusHype
One Pixel is All I Need0
Data Poisoning to Fake a Nash Equilibrium in Markov Games0
Online Data Poisoning Attack0
Online Data Poisoning Attacks0
On Optimal Learning Under Targeted Data Poisoning0
On Practical Aspects of Aggregation Defenses against Data Poisoning Attacks0
On the Adversarial Risk of Test Time Adaptation: An Investigation into Realistic Test-Time Data Poisoning0
On the Effectiveness of Poisoning against Unsupervised Domain Adaptation0
RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with Human Feedback in Large Language Models0
On the Relevance of Byzantine Robust Optimization Against Data Poisoning0
On the Robustness of Graph Reduction Against GNN Backdoor0
A Study of Backdoors in Instruction Fine-tuned Language Models0
Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents0
Optimizing ML Training with Metagradient Descent0
Oriole: Thwarting Privacy against Trustworthy Deep Learning Models0
OVLA: Neural Network Ownership Verification using Latent Watermarks0
PACOL: Poisoning Attacks Against Continual Learners0
Partner in Crime: Boosting Targeted Poisoning Attacks against Federated Learning0
Pick your Poison: Undetectability versus Robustness in Data Poisoning Attacks0
PoisHygiene: Detecting and Mitigating Poisoning Attacks in Neural Networks0
PoisonedEncoder: Poisoning the Unlabeled Pre-training Data in Contrastive Learning0
PoisonedParrot: Subtle Data Poisoning Attacks to Elicit Copyright-Infringing Content from Large Language Models0
Poisoning Attacks and Defenses on Artificial Intelligence: A Survey0
Poisoning Attacks to Local Differential Privacy Protocols for Trajectory Data0
Poisoning Deep Reinforcement Learning Agents with In-Distribution Triggers0
Poisoning Programs by Un-Repairing Code: Security Concerns of AI-generated Code0
Policy Teaching via Data Poisoning in Learning from Human Preferences0
Post-Training Overfitting Mitigation in DNN Classifiers0
Practical Data Poisoning Attack against Next-Item Recommendation0
SLSGD: Secure and Efficient Distributed On-device Machine Learning0
Practical Poisoning Attacks on Neural Networks0
Precision Guided Approach to Mitigate Data Poisoning Attacks in Federated Learning0
Pre-trained Encoders in Self-Supervised Learning Improve Secure and Privacy-preserving Supervised Learning0
Preventing Unauthorized Use of Proprietary Data: Poisoning for Secure Dataset Release0
PrivacyGAN: robust generative image privacy0
Nightshade: Prompt-Specific Poisoning Attacks on Text-to-Image Generative Models0
Property Inference From Poisoning0
Protecting against simultaneous data poisoning attacks0
Protecting Proprietary Data: Poisoning for Secure Dataset Release0
Provably effective detection of effective data poisoning attacks0
Provably Reliable Conformal Prediction Sets in the Presence of Data Poisoning0
Proving Data-Poisoning Robustness in Decision Trees0
Purifying Large Language Models by Ensembling a Small Language Model0
QTrojan: A Circuit Backdoor Against Quantum Neural Networks0
Reaching Data Confidentiality and Model Accountability on the CalTrain0
Recursive Euclidean Distance Based Robust Aggregation Technique For Federated Learning0
Redactor: A Data-centric and Individualized Defense Against Inference Attacks0
FedPrompt: Communication-Efficient and Privacy Preserving Prompt Tuning in Federated Learning0
Regularisation Can Mitigate Poisoning Attacks: A Novel Analysis Based on Multiobjective Bilevel Optimisation0
Regularization Helps with Mitigating Poisoning Attacks: Distributionally-Robust Machine Learning Using the Wasserstein Distance0
Show:102550
← PrevPage 5 of 10Next →

No leaderboard results yet.