llm-security
There are 39 repositories under llm-security topic.
Giskard-AI/giskard
🐢 Open-Source Evaluation & Testing for LLMs and ML models
pathwaycom/llm-app
LLM App templates for RAG, knowledge mining, and stream analytics. Ready to run with Docker,⚡in sync with your data sources.
protectai/llm-guard
The Security Toolkit for LLM Interactions
mariocandela/beelzebub
A secure low code honeypot framework, leveraging AI for System Virtualization.
EasyJailbreak/EasyJailbreak
An easy-to-use Python framework to generate adversarial jailbreak prompts.
msoedov/agentic_security
Agentic LLM Vulnerability Scanner
chawins/llm-sp
Papers and resources related to the security and privacy of LLMs 🤖
deadbits/vigil-llm
⚡ Vigil ⚡ Detect prompt injections, jailbreaks, and other potentially risky Large Language Model (LLM) inputs
liu00222/Open-Prompt-Injection
Formalizing and Benchmarking Prompt Injection Attacks and Defenses
arekusandr/last_layer
Ultra-fast, low latency LLM prompt injection/jailbreak detection ⛓️
yevh/TaaC-AI
AI-driven Threat modeling-as-a-Code (TaaC-AI)
ZenGuard-AI/fast-llm-security-guardrails
The fastest && easiest LLM security and privacy guardrails for GenAI apps.
raga-ai-hub/raga-llm-hub
Framework for LLM evaluation, guardrails and security
lakeraai/pint-benchmark
A benchmark for prompt injection detection systems.
pdparchitect/llm-hacking-database
This repository contains various attack against Large Language Models.
NaniDAO/ie
intents engine
microsoft/BIPIA
A benchmark for evaluating the robustness of LLMs and defenses to indirect prompt injection attacks.
briland/LLM-security-and-privacy
LLM security and privacy
llm-platform-security/SecGPT
SecGPT: An execution isolation architecture for LLM-based systems
llm-platform-security/chatgpt-plugin-eval
LLM Platform Security: Applying a Systematic Evaluation Framework to OpenAI's ChatGPT Plugins
leondz/lm_risk_cards
Risks and targets for assessing LLMs & LLM vulnerabilities
LostOxygen/llm-confidentiality
Whispers in the Machine: Confidentiality in LLM-integrated Systems
lakeraai/chainguard
Guard your LangChain applications against prompt injection with Lakera ChainGuard.
SEC-CAFE/handbook
安全手册,企业安全实践、攻防与安全研究知识库
azminewasi/Awesome-LLMs-ICLR-24
It is a comprehensive resource hub compiling all LLM papers accepted at the International Conference on Learning Representations (ICLR) in 2024.
sinanw/llm-security-prompt-injection
This project investigates the security of large language models by performing binary classification of a set of input prompts to discover malicious prompts. Several approaches have been analyzed using classical ML algorithms, a trained LLM model, and a fine-tuned LLM model.
levitation-opensource/Manipulative-Expression-Recognition
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.
M507/HackMeGPT
Vulnerable LLM Application
balavenkatesh3322/guardrails-demo
LLM Security Project with Llama Guard
AiShieldsOrg/AiShieldsWeb
AiShields is an open-source Artificial Intelligence Data Input and Output Sanitizer
CyberAlbSecOP/MINOTAUR_Impossible_GPT_Security_Challenge
MINOTAUR: The STRONGEST Secure Prompt EVER! Prompt Security Challenge, Impossible GPT Security, Prompts Cybersecurity, Prompting Vulnerabilities, FlowGPT, Secure Prompting, Secure LLMs, Prompt Hacker, Cutting-edge Ai Security, Unbreakable GPT Agent, Anti GPT Leak, System Prompt Security.
rohilrg/CatchPromptInjection
This repo focus on how to deal with prompt injection problem faced by LLMs
lastlayer/last-layer-vercel
Example of running last_layer with FastAPI on vercel
mickymultani/TestingGemma2B
Evaluation of Google's Instruction Tuned Gemma-2B, an open-source Large Language Model (LLM). Aimed at understanding the breadth of the model's knowledge, its reasoning capabilities, and adherence to ethical guardrails, this project presents a systematic assessment across a diverse array of domains.
minuva/fast-prompt-attack-detect
User prompt attack detection system
nodite/llm-guard-ts
The Security Toolkit for LLM Interactions (TS version)