prompt-injection
There are 77 repositories under prompt-injection topic.
CyberAlbSecOP/Awesome_GPT_Super_Prompting
ChatGPT Jailbreaks, GPT Assistants Prompt Leaks, GPTs Prompt Injection, LLM Prompt Security, Super Prompts, Prompt Hack, Prompt Security, Ai Prompt Engineering, Adversarial Machine Learning.
protectai/llm-guard
The Security Toolkit for LLM Interactions
protectai/rebuff
LLM Prompt Injection Detector
whylabs/langkit
🔍 LangKit: An open-source toolkit for monitoring Large Language Models (LLMs). 📚 Extracts signals from prompts & responses, ensuring safety & security. 🛡️ Features include text quality, relevance metrics, & sentiment analysis. 📊 A comprehensive tool for LLM observability. 👀
abilzerian/LLM-Prompt-Library
Advanced Code and Text Manipulation Prompts for Various LLMs. Suitable for Siri, GPT-4o, Claude, Llama3, Gemini, and other high-performance open-source LLMs.
utkusen/promptmap
automatically tests prompt injection attacks on ChatGPT instances
mintyfrankie/brilliant-CV
💼 another CV template for your job application, yet powered by Typst and more
tldrsec/prompt-injection-defenses
Every practical and proposed defense against prompt injection.
zacfrulloni/Prompt-Engineering-Holy-Grail
# Prompt Engineering Hub ⭐️ If you find this helpful, give it a star to show your support! This repository is a one-stop resource for prompt engineering. Also available on: https://promptengineeringhub.dev/
deadbits/vigil-llm
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
automorphic-ai/aegis
Self-hardening firewall for large language models
langgptai/Awesome-Multimodal-Prompts
Prompts of GPT-4V & DALL-E3 to full utilize the multi-modal ability. GPT4V Prompts, DALL-E3 Prompts.
dropbox/llm-security
Dropbox LLM Security research code and results
liu00222/Open-Prompt-Injection
This repository provides implementation to formalize and benchmark Prompt Injection attacks and defenses
yunwei37/prompt-hacker-collections
prompt attack-defense, prompt Injection, reverse engineering notes and examples | 提示词对抗、破解例子与笔记
lakeraai/pint-benchmark
A benchmark for prompt injection detection systems.
pasquini-dario/project_mantis
Project Mantis: Hacking Back the AI-Hacker; Prompt Injection as a Defense Against LLM-driven Cyberattacks
HumanCompatibleAI/tensor-trust
A prompt injection game to collect data for robust ML research
gdalmau/lakera-gandalf-solutions
My inputs for the LLM Gandalf made by Lakera
GPTSafe/PromptGuard
Build production ready apps for GPT using Node.js & TypeScript
sinanw/llm-security-prompt-injection
This project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using classical ML algorithms, a trained LLM model, and a fine-tuned LLM model.
LostOxygen/llm-confidentiality
Whispers in the Machine: Confidentiality in LLM-integrated Systems
microsoft/gandalf_vs_gandalf
Turning Gandalf against itself. Use LLMs to automate playing Lakera Gandalf challenge without needing to set up an account with a platform provider.
MaxMLang/pytector
A Python package designed to detect prompt injection in text inputs utilizing state-of-the-art machine learning models from Hugging Face. The main focus is on ease of use, enabling developers to integrate security features into their applications with minimal effort.
TrustAI-laboratory/Learn-Prompt-Hacking
This is The most comprehensive prompt hacking course available, which record our progress on a prompt engineering and prompt hacking course.
probonodev/jailbreak
jailbreakme.xyz is an open-source decentralized app (dApp) where users are challenged to try and jailbreak pre-existing LLMs in order to find weaknesses and be rewarded. 🏆
grepstrength/WideOpenAI
Short list of indirect prompt injection attacks for OpenAI-based models.
SemanticBrainCorp/SemanticShield
The Security Toolkit for managing Generative AI(especially LLMs) and Supervised Learning processes(Learning and Inference).
wearetyomsmnv/Awesome-LLMSecOps
LLM | Security | Operations in one github repo with good links and pictures.
ZapDos7/lakera-gandalf
My solutions for Lakera's Gandalf
lakeraai/chainguard
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
shaialon/ai-security-demos
🤯 AI Security EXPOSED! Live Demos Showing Hidden Risks of 🤖 Agentic AI Flows: 💉Prompt Injection, ☣️ Data Poisoning. Watch the recorded session:
qroa/QROA
QROA: A Black-Box Query-Response Optimization Attack on LLMs
levitation-opensource/Manipulative-Expression-Recognition
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.
peluche/deck-of-many-prompts
Manual Prompt Injection / Red Teaming Tool
Safetorun/PromptDefender
A prompt defence is a multi-layer defence that can be used to protect your applications against prompt injection attacks.