SOTAVerified

LLM Jailbreak

Papers

Showing 110 of 24 papers

TitleStatusHype
PandaGuard: Systematic Evaluation of LLM Safety against Jailbreaking AttacksCode2
Derail Yourself: Multi-turn LLM Jailbreak Attack through Self-discovered CluesCode2
JailbreakZoo: Survey, Landscapes, and Horizons in Jailbreaking Large Language and Vision-Language ModelsCode2
JailBreakV: A Benchmark for Assessing the Robustness of MultiModal Large Language Models against Jailbreak AttacksCode2
CySecBench: Generative AI-based CyberSecurity-focused Prompt Dataset for Benchmarking Large Language ModelsCode1
Cognitive Overload Attack:Prompt Injection for Long ContextCode1
Automatic Prompt Optimization with "Gradient Descent" and Beam SearchCode1
CAVGAN: Unifying Jailbreak and Defense of LLMs via Generative Adversarial Attacks on their Internal RepresentationsCode0
LLM Jailbreak Oracle0
SecurityLingua: Efficient Defense of LLM Jailbreak Attacks via Security-Aware Prompt Compression0
Show:102550
← PrevPage 1 of 3Next →

No leaderboard results yet.