Pinned Repositories
bigcode-evaluation-harness
A framework for the evaluation of autoregressive code generation language models.
MultiPL-E
A multi-programming language benchmark for LLMs
NCISurvey
Neural Code Intelligence Survey 2024; Reading lists and resources
DebugBench
The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".
JIOC
LADCF_MCM-
NCISurvey
Neural Code Intelligence Survey 2024; Reading lists and resources
OOP
Related-papers
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
shuaiwang2022's Repositories
shuaiwang2022/DebugBench
The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".
shuaiwang2022/JIOC
shuaiwang2022/LADCF_MCM-
shuaiwang2022/NCISurvey
Neural Code Intelligence Survey 2024; Reading lists and resources
shuaiwang2022/OOP
shuaiwang2022/Related-papers