SOTAVerified

Adversarial Text

Adversarial Text refers to a specialised text sequence that is designed specifically to influence the prediction of a language model. Generally, Adversarial Text attack are carried out on Large Language Models (LLMs). Research on understanding different adversarial approaches can help us build effective defense mechanisms to detect malicious text input and build robust language models.

Papers

Showing 101110 of 114 papers

TitleStatusHype
DISCO : efficient unsupervised decoding for discrete natural language problems via convex relaxation0
Enhancing Adversarial Text Attacks on BERT Models with Projected Gradient Descent0
Don't Search for a Search Method -- Simple Heuristics Suffice for Adversarial Text Attacks0
Don’t Search for a Search Method — Simple Heuristics Suffice for Adversarial Text Attacks0
TextDefense: Adversarial Text Detection based on Word Importance Entropy0
Data-Driven Mitigation of Adversarial Text Perturbation0
FastWordBug: A Fast Method To Generate Adversarial Text Against NLP Applications0
Continuous Adversarial Text Representation Learning for Affective Recognition0
Finding a Wolf in Sheep's Clothing: Combating Adversarial Text-To-Image Prompts with Text Summarization0
Fooling OCR Systems with Adversarial Text Images0
Show:102550
← PrevPage 11 of 12Next →

No leaderboard results yet.