/human-eval

Harness for experiments on human-eval dataset. Based On Code for the paper "Evaluating Large Language Models Trained on Code"

Primary LanguagePythonMIT LicenseMIT

Watchers

No one’s watching this repository yet.