llm-evaluation-metrics
There are 3 repositories under llm-evaluation-metrics topic.
confident-ai/deepeval
The LLM Evaluation Framework
zhuohaoyu/KIEval
[ACL'24] A Knowledge-grounded Interactive Evaluation Framework for Large Language Models
ritwickbhargav80/quick-llm-model-evaluations
This repo is for an streamlit application that provides a user-friendly interface for evaluating large language models (LLMs) using the beyondllm package.