SOTAVerified

Data Poisoning

Data Poisoning is an adversarial attack that tries to manipulate the training dataset in order to control the prediction behavior of a trained model such that the model will label malicious examples into a desired classes (e.g., labeling spam e-mails as safe).

Source: Explaining Vulnerabilities to Adversarial Machine Learning through Visual Analytics

Papers

Showing 131140 of 492 papers

TitleStatusHype
Class Machine Unlearning for Complex Data via Concepts Inference and Data Poisoning0
Generative AI in Cybersecurity: A Comprehensive Review of LLM Applications and Vulnerabilities0
Fed-Credit: Robust Federated Learning with Credibility Management0
SEEP: Training Dynamics Grounds Latent Representation Search for Mitigating Backdoor Poisoning Attacks0
Concealing Backdoor Model Updates in Federated Learning by Trigger-Optimized Data Poisoning0
Hard Work Does Not Always Pay Off: Poisoning Attacks on Neural Architecture Search0
On the Relevance of Byzantine Robust Optimization Against Data Poisoning0
Dual Model Replacement:invisible Multi-target Backdoor Attack based on Federal Learning0
Data Poisoning Attacks on Off-Policy Policy Evaluation Methods0
Precision Guided Approach to Mitigate Data Poisoning Attacks in Federated Learning0
Show:102550
← PrevPage 14 of 50Next →

No leaderboard results yet.