llm-test
There are 10 repositories under llm-test topic.
uptrain-ai/uptrain
UpTrain is an open-source unified platform to evaluate and improve Generative AI applications. We provide grades for 20+ preconfigured checks (covering language, code, embedding use-cases), perform root cause analysis on failure cases and give insights on how to resolve them.
georgian-io/LLM-Finetuning-Toolkit
Toolkit for fine-tuning, ablating and unit-testing open-source LLMs.
JohnSnowLabs/langtest
Deliver safe & effective language models
athina-ai/athina-sdk
LLM Testing SDK that helps you write and run tests to monitor your LLM app in production
levitation-opensource/Manipulative-Expression-Recognition
MER is a software that identifies and highlights manipulative communication in text from human conversations and AI-generated responses. MER benchmarks language models for manipulative expressions, fostering development of transparency and safety in AI. It also supports manipulation victims by detecting manipulative patterns in human communication.
Coldwave96/LLM-Sec-Evaluation
Scripts for evaluating LLM security abilities.
prompt-foundry/typescript-sdk
The prompt engineering, prompt management, and prompt evaluation tool for TypeScript, JavaScript, and NodeJS.
awesome-software/nlptest
Deliver safe & effective language models
prompt-foundry/go-sdk
The prompt engineering, prompt management, and prompt evaluation tool for Go.
awesome-software/promptfoo
Test your prompts. Evaluate and compare LLM outputs, catch regressions, and improve prompt quality.