braintrustdata/autoevals
AutoEvals is a tool for quickly and easily evaluating AI model outputs using best practices.
PythonMIT
Issues
- 5
Basic ValidJSON Question
#105 opened - 1
Supported scores
#101 opened - 1
Azure Openai is not well configured
#100 opened - 3
Is there a way to access LLM token usage
#96 opened - 4
Support of Azure Open AI models and API
#89 opened - 2
- 3
(`autoevals` JS) Better support and documentation for using context-based evaluators in `Eval` run
#82 opened - 2
- 0
- 1
Question about deps
#61 opened - 3
General Question about the Evaluator LLM
#51 opened - 3
- 2
- 7
Factuality Evaluator failing
#28 opened - 0
Add a `make test` github action
#4 opened