/lm-evaluation-harness-bg

A framework for few-shot evaluation of language models, extended with benchmarks by https://insait.ai/

Primary LanguagePythonMIT LicenseMIT

Stargazers