Guardrail ML is an alignment toolkit to use LLMs safely and securely. Our firewall scans prompts and LLM behaviors for risks to bring your AI app from prototype to production with confidence.
- 🚀mitigate LLM security and safety risks
- 📝customize and ensure LLM behaviors are safe and secure
- 💸monitor incidents, costs, and responsible AI metrics
- 🛠️ firewall that safeguards against CVEs and improves with each attack
- 🤖 reduce and measure ungrounded additions (hallucinations) with tools
- 🛡️ multi-layered defense with heuristic detectors, LLM-based, vector DB
-
Guardrail API Key and set
env
variable asGUARDRAIL_API_KEY
-
To install guardrail, use the Python Package Index (PyPI) as follows:
pip install guardrail-ml
Firewall
- Prompt Injections
- Factual Consistency
- Factuality Tool
- Toxicity Detector
- Regex Detector
- Stop Patterns Detector
- Malware URL Detector
- PII Anonymize
- Secrets
- DoS Tokens
- Harmful Detector
- Relevance
- Contradictions
- Text Quality
- Language
- Bias
- Adversarial Prompt Generation
- Attack Signature
Integrations
- OpenAI Completion
- LangChain
- LlamaIndex
- Cohere
- HuggingFace
Old Quickstart v0.0.1 (08/03/23)
4-bit QLoRA of llama-v2-7b
with dolly-15k
(07/21/23):