Pinned Repositories
OPVGCN
AEPWO
SciAssess
SciAssess is a comprehensive benchmark for evaluating Large Language Models' proficiency in scientific literature analysis across various fields, focusing on memorization, comprehension, and analysis.
evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Uni-Mol
Official Repository for the Uni-Mol Series Methods
SciAssess
SciAssess is a comprehensive benchmark for evaluating Large Language Models' proficiency in scientific literature analysis across various fields, focusing on memorization, comprehension, and analysis.
evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
HongshuaiWang1's Repositories
HongshuaiWang1/SciAssess
SciAssess is a comprehensive benchmark for evaluating Large Language Models' proficiency in scientific literature analysis across various fields, focusing on memorization, comprehension, and analysis.
HongshuaiWang1/OPVGCN
HongshuaiWang1/AEPWO