llm-jailbreaks topic
agentic_security
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
circle-guard-bench
First-of-its-kind AI benchmark for evaluating the protection capabilities of large language model (LLM) guard systems (guardrails and safeguards)
LMAP
LMAP (large language model mapper) is like NMAP for LLM, is an LLM Vulnerability Scanner and Zero-day Vulnerability Fuzzer.
JailbreakEval
[NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.
GPT-5_Jailbreak_PoC
A working POC of a GPT-5 jailbreak via PROMISQROUTE (Prompt-based Router Open-Mode Manipulation) with a barebones C2 server & agent generation demo.
SemanticSmooth
Implementation of paper 'Defending Large Language Models against Jailbreak Attacks via Semantic Smoothing'