Found 523 repositories(showing 30)
CyberAlbSecOP
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
Tencent
A full-stack AI Red Teaming platform securing AI ecosystems via OpenClaw Security Scan, Agent Scan, Skills Scan, MCP scan, AI Infra scan and LLM jailbreak evaluation.
msoedov
Agentic LLM Vulnerability Scanner / AI red teaming kit 🧪
cyberark
A powerful tool for automated LLM fuzzing. It is designed to help developers and security researchers identify and mitigate potential jailbreaks in their LLM APIs.
yueliu1999
Awesome-Jailbreak-on-LLMs is a collection of state-of-the-art, novel, exciting jailbreak methods on LLMs. It contains papers, codes, datasets, evaluations, and analyses.
Goochbeater
A repo for jailbreaking various LLMs, mainly Claude
langgptai
LLM Jailbreaks, ChatGPT, Claude, Llama, DAN Prompts, Prompt Leaking
deadbits
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
langgptai
Gemini Prompts, Gemini 3 Prompts, jailbreak, LLM Prompts, LangGPT —— by 云中江树
jconorgrogan
Jailbreak for ChatGPT: Predict the future, opine on politics and controversial topics, and assess what is true. May help us understand more about LLM Bias
tml-epfl
Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [ICLR 2025]
SaFo-Lab
[ICLR 2025 Spotlight] The official implementation of our ICLR2025 paper "AutoDAN-Turbo: A Lifelong Agent for Strategy Self-Exploration to Jailbreak LLMs".
CHATS-lab
Persuasive Jailbreaker: we can persuade LLMs to jailbreak them!
SlowLow999
sharing NEW strong AI jailbreaks of multiple vendors (LLMs)
alexisvalentino
DAN - The ‘JAILBREAK’ Version of ChatGPT and How to Use it. (update: this was 3 years ago, might not work at the current state of LLMs check OBLITERATUS)
RICommunity
TAP: An automated jailbreaking method for black-box LLMs
CryptoAILab
[NDSS'25 Best Technical Poster] A collection of automated evaluators for assessing jailbreak attempts.
Yu-Fangxu
[ICML 2024] COLD-Attack: Jailbreaking LLMs with Stealthiness and Controllability
praetorian-inc
LLM security testing framework for detecting prompt injection, jailbreaks, and adversarial attacks — 190+ probes, 28 providers, single Go binary
yueliu1999
[ICML 2025] An official source code for paper "FlipAttack: Jailbreak LLMs via Flipping".
ReversecLabs
Simple Prompt Injection Kit for Evaluation and Exploitation
usail-hkust
Bag of Tricks: Benchmarking of Jailbreak Attacks on LLMs. Empirical tricks for LLM Jailbreaking. (NeurIPS 2024)
WhileBug
Awesome LLM Jailbreak academic papers
arekusandr
Ultra-fast, low latency LLM prompt injection/jailbreak detection ⛓️
sail-sg
[ICML 2024] Agent Smith: A Single Image Can Jailbreak One Million Multimodal LLM Agents Exponentially Fast
allenai
Open One-Stop Moderation Tools for Safety Risks, Jailbreaks, and Refusals of LLMs
uw-nsl
[ACL24] Official Repo of Paper `ArtPrompt: ASCII Art-based Jailbreak Attacks against Aligned LLMs`
SaFo-Lab
[COLM 2024] JailBreakV-28K: A comprehensive benchmark designed to evaluate the transferability of LLM jailbreak attacks to MLLMs, and further assess the robustness and safety of MLLMs against a variety of jailbreak attacks.
Junjie-Chu
This is the public code repository of paper 'Comprehensive Assessment of Jailbreak Attacks Against LLMs'
gally16
LLM Jailbreaking Guide主流大语言模型越狱指南