Found 190 repositories(showing 30)
SaFo-Lab
[ECCV 2024] The official code for "AdaShield: Safeguarding Multimodal Large Language Models from Structure-based Attack via Adaptive Shield Prompting."
a23one
Shields your confidential data in LLM prompts from third party AI providers, allowing you to send with confidence without compromising security and privacy.
mthamil107
Self-learning prompt injection detection engine — 25 input detectors (10 languages), 5 output scanners, PII redaction, red team self-testing, F1: 96.0% with 0% false positives. Docker, GitHub Action, pre-commit, FastAPI/Flask/Django/LangChain/CrewAI/Dify/n8n.
joseph0926
prompt quality management for Claude Code
Bit-Pulse-AI
OpenClaw Prompt Shield: Security framework that protects OpenClaw AI agents from prompt injections, data leaks, and dangerous commands using Azure AI Content Safety and Microsoft Purview DLP.
nooesc
Fast prompt injection detection engine. Rust library + CLI + WASM.
soumyasagiri
Real-time prompt injection, jailbreak and exfiltration detection middleware for LLM APIs. Zero ML dependencies. Sub-2ms latency. 75 tests.
wagner-group
No description available
alexandrughinea
Security middleware that shields AI applications from prompt injection, jailbreaking, and obfuscated attacks through composable defense layers.
ankitlade12
The full-stack safety layer for AI agents. Budget limits, prompt injection shields, PII filtering, output firewalls, and hooks — in 2 lines of code.
KNOWDYN
Open-source semantic security shield that inspects, sanitizes, and blocks malicious prompts to prevent jailbreaks, data exfiltration, and unauthorized tool use in conversational agents.
gmuskan95
Browser extension that detects PII in AI chat inputs and lets you redact before sending.
Pradeep-2901
A 4-class classification system for securing technical LLMs
25thblame
🛡️ Protect your LLM apps from prompt injection attacks. 3 lines to integrate, Detects jailbreaks, prompt extraction, instruction overrides & more.
rango-ramesh
Lightweight prompt injection detection for Python. Fast, easy integration with LangChain and AutoGen.
topazyo
Production-ready security playbook for AI agent deployments. Addresses backup file persistence, authentication bypass, and prompt injection vulnerabilities. Integrates openclaw-detect, openclaw-telemetry, openclaw-shield, and clawguard.
10486-JosephMutua
No description available
L0uisHu
No description available
mergenhan
This repository contains a simple but functional Chrome Extension. It monitors input fields, textareas, and editable elements on websites. When forbidden words are detected, the extension warns the user (or blocks submission). Users can manage the blocked words list from the popup or options page.
Abdulbasith0512
No description available
prabujayant
No description available
Aadhithya-T
> PromptShield is a prompt security middleware for LLMs. It uses a fine-tuned DistilBERT classifier to label incoming prompts as safe, unsafe, suspicious, or jailbreak before forwarding to Google Gemini 2.5 Flash. Built with Python, Transformers, and a vanilla HTML/CSS/JS frontend.
pravin9033
Runtime security firewall for LLM applications.
Dhwanit2501
A Context-Aware Prompt Injection Defense System for LLM Chatbots that detects and neutralizes prompt injection attacks before they reach your LLM.
aniketkarne
No description available
marmer7
No description available
LuciferForge
Lightweight prompt injection detector. 22 attack patterns. Blocks jailbreaks before they reach your model.
frenchieTST
Security audit skill for Claude Code — 18 automated checks for prompt injection, data exfiltration, supply chain attacks, and persistence mechanisms in skills, MCP servers, and npm packages.
VKD-cyber
making a project to filter prompts to prevent from prompt injection
jinhwak11
No description available