/clinical-llm-evaluation

This repo is created to evaluate the LLMs on 4 tasks including Question-Answer (QA), Summarization, Name Entity Recognition (NER), and Relation Extraction (RE). The goal is to create a general framework to quickly evaluate any Causal Language Models against publicly available medical datasets.

Primary LanguageJupyter NotebookMIT LicenseMIT

Stargazers