SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 341350 of 492 papers

TitleStatusHype
Collaborative Self Organizing Map with DeepNNs for Fake Task Prevention in Mobile Crowdsensing0
An Equivalence Between Data Poisoning and Byzantine Gradient AttacksCode0
Redactor: A Data-centric and Individualized Defense Against Inference Attacks0
Improved Certified Defenses against Data Poisoning with (Deterministic) Finite AggregationCode0
Towards Multi-Objective Statistically Fair Federated Learning0
How to Backdoor HyperNetwork in Personalized Federated Learning?0
Towards Understanding Quality Challenges of the Federated Learning for Neural Networks: A First Look from the Lens of RobustnessCode0
Compression-Resistant Backdoor Attack against Deep Neural Networks0
Execute Order 66: Targeted Data Poisoning for Reinforcement Learning0
ML Attack Models: Adversarial Attacks and Data Poisoning Attacks0
Show:102550
← PrevPage 35 of 50Next →

No leaderboard results yet.