SOTAVerified

Adversarial Text

Adversarial Text refers to a specialised text sequence that is designed specifically to influence the prediction of a language model. Generally, Adversarial Text attack are carried out on Large Language Models (LLMs). Research on understanding different adversarial approaches can help us build effective defense mechanisms to detect malicious text input and build robust language models.

Papers

Showing 5160 of 114 papers

TitleStatusHype
Adversarial Training: A simple and efficient technique to Improving NLP Robustness0
A Grey-box Text Attack Framework using Explainable AI0
A survey on text generation using generative adversarial networks0
Autonomous LLM-Enhanced Adversarial Attack for Text-to-Motion0
PBI-Attack: Prior-Guided Bimodal Interactive Black-Box Jailbreak Attack for Toxicity Maximization0
CAT-Gen: Improving Robustness in NLP Models via Controlled Adversarial Text Generation0
Commonsense-T2I Challenge: Can Text-to-Image Generation Models Understand Commonsense?0
Continuous Adversarial Text Representation Learning for Affective Recognition0
Data-Driven Mitigation of Adversarial Text Perturbation0
Detecting Adversarial Text Attacks via SHapley Additive exPlanations0
Show:102550
← PrevPage 6 of 12Next →

No leaderboard results yet.