SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 351400 of 492 papers

TitleStatusHype
Get a Model! Model Hijacking Attack Against Machine Learning Models0
Mitigating Data Poisoning in Text Classification with Differential Privacy0
CoProtector: Protect Open-Source Code against Unauthorized Training Usage with Data PoisoningCode0
Poison Forensics: Traceback of Data Poisoning Attacks in Neural Networks0
Defending Against Backdoor Attacks Using Ensembles of Weak Learners0
Defending Backdoor Data Poisoning Attacks by Using Noisy Label Defense Algorithm0
DP-InstaHide: Data Augmentations Provably Enhance Guarantees Against Dataset Manipulations0
Protecting Proprietary Data: Poisoning for Secure Dataset Release0
Backdoor Attack and Defense for Deep Regression0
Excess Capacity and Backdoor PoisoningCode0
Certifiers Make Neural Networks Vulnerable to Availability Attacks0
ABC-FL: Anomalous and Benign client Classification in Federated Learning0
Classification Auto-Encoder based Detector against Diverse Data Poisoning AttacksCode0
Adversarial Attacks Against Deep Reinforcement Learning Framework in Internet of Vehicles0
Derivative-free Alternating Projection Algorithms for General Nonconvex-Concave Minimax Problems0
Putting words into the system’s mouth: A targeted attack on neural machine translation using monolingual data poisoningCode0
Fairness-aware Summarization for Justified Decision-Making0
Putting words into the system's mouth: A targeted attack on neural machine translation using monolingual data poisoningCode0
Understanding the Limits of Unsupervised Domain Adaptation via Data PoisoningCode0
Poisoning Attack against Estimating from Pairwise ComparisonsCode0
On the Effectiveness of Poisoning against Unsupervised Domain Adaptation0
Data Poisoning Won’t Save You From Facial Recognition0
Poisoning Deep Reinforcement Learning Agents with In-Distribution Triggers0
Gradient-based Data Subversion Attack Against Binary Classifiers0
A Gradient Method for Multilevel Optimization0
A BIC-based Mixture Model Defense against Data Poisoning Attacks on Classifiers0
Fooling Partial Dependence via Data PoisoningCode0
De-Pois: An Attack-Agnostic Defense against Data Poisoning Attacks0
Incompatibility Clustering as a Defense Against Backdoor Poisoning AttacksCode0
Influence Based Defense Against Data Poisoning Attacks in Online Learning0
FedCom: A Byzantine-Robust Local Model Aggregation Rule Using Data Commitment for Federated Learning0
Defending Against Adversarial Denial-of-Service Data Poisoning Attacks0
The Hammer and the Nut: Is Bilevel Optimization Really Needed to Poison Linear Classifiers?Code0
Data-Driven Control and Data-Poisoning attacks in Buildings: the KTH Live-In Lab case study0
Robust learning under clean-label attack0
Oriole: Thwarting Privacy against Trustworthy Deep Learning Models0
Data Poisoning Attacks and Defenses to Crowdsourcing Systems0
Preventing Unauthorized Use of Proprietary Data: Poisoning for Secure Dataset Release0
Saving Stochastic Bandits from Poisoning Attacks via Limited Data Verification0
Reinforcement Learning For Data Poisoning on Graph Neural Networks0
Adversarial Poisoning Attacks and Defense for General Multi-Class Models Based On Synthetic Reduced Nearest Neighbors0
Generating Fake Cyber Threat Intelligence Using Transformer-Based Models0
Property Inference From Poisoning0
Adversarial Vulnerability of Active Transfer Learning0
Data Poisoning Attacks to Deep Learning Based Recommender Systems0
Just How Toxic is Data Poisoning? A Benchmark for Backdoor and Data Poisoning Attacks0
CLEAR: Clean-Up Sample-Targeted Backdoor in Neural Networks0
Sself: Robust Federated Learning against Stragglers and Adversaries0
Active Learning Under Malicious Mislabeling and Poisoning Attacks0
Federated Unlearning0
Show:102550
← PrevPage 8 of 10Next →

No leaderboard results yet.