SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 301310 of 492 papers

TitleStatusHype
Temporal Robustness against Data Poisoning0
The Price of Tailoring the Index to Your Data: Poisoning Attacks on Learned Index Structures0
The Stronger the Diffusion Model, the Easier the Backdoor: Data Poisoning to Induce Copyright Breaches Without Adjusting Finetuning Pipeline0
Data Poisoning Attack against Knowledge Graph Embedding0
Towards Multi-Objective Statistically Fair Federated Learning0
Towards Poisoning Fair Representations0
Towards Poisoning of Deep Learning Algorithms with Back-gradient Optimization0
Poison Forensics: Traceback of Data Poisoning Attacks in Neural Networks0
Trading Devil Final: Backdoor attack via Stock market and Bayesian Optimization0
Trading Devil RL: Backdoor attack via Stock market, Bayesian Optimization and Reinforcement Learning0
Show:102550
← PrevPage 31 of 50Next →

No leaderboard results yet.