greenlight2000/llm-code-eval
LLMCodeEval: An Execution-Based Multilingual Multitask Multidimensional Benchmark for Evaluating Large Language Models on Code Understanding and Generation
Python
Stargazers
No one’s star this repository yet.
LLMCodeEval: An Execution-Based Multilingual Multitask Multidimensional Benchmark for Evaluating Large Language Models on Code Understanding and Generation
Python
No one’s star this repository yet.