SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 5160 of 492 papers

TitleStatusHype
Availability Attacks Create ShortcutsCode1
Indiscriminate Poisoning Attacks on Unsupervised Contrastive LearningCode1
Just How Toxic is Data Poisoning? A Unified Benchmark for Backdoor and Data Poisoning AttacksCode1
Auditing Differentially Private Machine Learning: How Private is Private SGD?Code1
Be Careful about Poisoned Word Embeddings: Exploring the Vulnerability of the Embedding Layers in NLP ModelsCode1
Backdoor Attacks on Crowd CountingCode1
Adversarial Robustness of Representation Learning for Knowledge GraphsCode1
Not All Poisons are Created Equal: Robust Training against Data PoisoningCode1
Optimistic Verifiable Training by Controlling Hardware NondeterminismCode1
A Distributed Trust Framework for Privacy-Preserving Machine LearningCode1
Show:102550
← PrevPage 6 of 50Next →

No leaderboard results yet.