SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 351400 of 492 papers

TitleStatusHype
INK: Inheritable Natural Backdoor Attack Against Model Distillation0
Learning and Unlearning of Fabricated Knowledge in Language Models0
Learning to Forget using Hypernetworks0
Local Model Poisoning Attacks to Byzantine-Robust Federated Learning0
Maximal adversarial perturbations for obfuscation: Hiding certain attributes while preserving rest0
Mendata: A Framework to Purify Manipulated Training Data0
Mitigating backdoor attacks in LSTM-based Text Classification Systems by Backdoor Keyword Identification0
Mitigating Data Poisoning in Text Classification with Differential Privacy0
Mitigating the Impact of Adversarial Attacks in Very Deep Networks0
Mixed Strategy Game Model Against Data Poisoning Attacks0
ML Attack Models: Adversarial Attacks and Data Poisoning Attacks0
How to Backdoor HyperNetwork in Personalized Federated Learning?0
Multi-Trigger Poisoning Amplifies Backdoor Vulnerabilities in LLMs0
Privacy and Copyright Protection in Generative AI: A Lifecycle Perspective0
Neural network fragile watermarking with no model performance degradation0
Neuromimetic metaplasticity for adaptive continual learning0
No, of course I can! Refusal Mechanisms Can Be Exploited Using Harmless Fine-Tuning Data0
Reclaiming "Open AI" -- AI Model Serving Can Be Open Access, Yet Monetizable and Loyal0
On Defending Against Label Flipping Attacks on Malware Detection Systems0
One Pixel is All I Need0
Data Poisoning to Fake a Nash Equilibrium in Markov Games0
Online Data Poisoning Attack0
Online Data Poisoning Attacks0
On Optimal Learning Under Targeted Data Poisoning0
On Practical Aspects of Aggregation Defenses against Data Poisoning Attacks0
On the Adversarial Risk of Test Time Adaptation: An Investigation into Realistic Test-Time Data Poisoning0
On the Effectiveness of Poisoning against Unsupervised Domain Adaptation0
RLHFPoison: Reward Poisoning Attack for Reinforcement Learning with Human Feedback in Large Language Models0
On the Relevance of Byzantine Robust Optimization Against Data Poisoning0
On the Robustness of Graph Reduction Against GNN Backdoor0
A Study of Backdoors in Instruction Fine-tuned Language Models0
Open Challenges in Multi-Agent Security: Towards Secure Systems of Interacting AI Agents0
Optimizing ML Training with Metagradient Descent0
Oriole: Thwarting Privacy against Trustworthy Deep Learning Models0
OVLA: Neural Network Ownership Verification using Latent Watermarks0
PACOL: Poisoning Attacks Against Continual Learners0
Partner in Crime: Boosting Targeted Poisoning Attacks against Federated Learning0
Pick your Poison: Undetectability versus Robustness in Data Poisoning Attacks0
PoisHygiene: Detecting and Mitigating Poisoning Attacks in Neural Networks0
PoisonedEncoder: Poisoning the Unlabeled Pre-training Data in Contrastive Learning0
PoisonedParrot: Subtle Data Poisoning Attacks to Elicit Copyright-Infringing Content from Large Language Models0
Poisoning Attacks and Defenses on Artificial Intelligence: A Survey0
Poisoning Attacks to Local Differential Privacy Protocols for Trajectory Data0
Poisoning Deep Reinforcement Learning Agents with In-Distribution Triggers0
Poisoning Programs by Un-Repairing Code: Security Concerns of AI-generated Code0
Policy Teaching via Data Poisoning in Learning from Human Preferences0
Post-Training Overfitting Mitigation in DNN Classifiers0
Practical Data Poisoning Attack against Next-Item Recommendation0
SLSGD: Secure and Efficient Distributed On-device Machine Learning0
Practical Poisoning Attacks on Neural Networks0
Show:102550
← PrevPage 8 of 10Next →

No leaderboard results yet.