Sunkyoung's Stars
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
huggingface/datasets
🤗 The largest hub of ready-to-use datasets for ML models with fast, easy-to-use and efficient data manipulation tools
deepset-ai/haystack
:mag: AI orchestration framework to build customizable, production-ready LLM applications. Connect components (models, vector DBs, file converters) to pipelines or agents that can interact with your data. With advanced retrieval methods, it's best suited for building RAG, question answering, semantic search or conversational agent chatbots.
openai/evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
ivy-llc/ivy
Convert Machine Learning Code Between Frameworks
dsdanielpark/Bard-API
The unofficial python package that returns response of Google Bard through cookie value.
Instruction-Tuning-with-GPT-4/GPT-4-LLM
Instruction Tuning with GPT-4
thunlp/PromptPapers
Must-read papers on prompt-based tuning for pre-trained language models.
mosaicml/llm-foundry
LLM training code for Databricks foundation models
Beomi/KoAlpaca
KoAlpaca: 한국어 명령어를 이해하는 오픈소스 언어모델
EleutherAI/the-pile
AGI-Edgerunners/LLM-Adapters
Code for our EMNLP 2023 Paper: "LLM-Adapters: An Adapter Family for Parameter-Efficient Fine-Tuning of Large Language Models"
hollobit/GenAI_LLM_timeline
ChatGPT, GenerativeAI and LLMs Timeline
SinclairCoder/Instruction-Tuning-Papers
Reading list of Instruction-tuning. A trend starts from Natrural-Instruction (ACL 2022), FLAN (ICLR 2022) and T0 (ICLR 2022).
allenai/dont-stop-pretraining
Code associated with the Don't Stop Pretraining ACL 2020 paper
bigscience-workshop/xmtf
Crosslingual Generalization through Multitask Finetuning
EleutherAI/polyglot
Polyglot: Large Language Models of Well-balanced Competence in Multi-languages
AlignmentResearch/tuned-lens
Tools for understanding how transformer predictions are built layer-by-layer
JohnGiorgi/DeCLUTR
The corresponding code from our paper "DeCLUTR: Deep Contrastive Learning for Unsupervised Textual Representations". Do not hesitate to open an issue if you run into any trouble!
microsoft/Table-Pretraining
ICLR 2022 Paper, SOTA Table Pre-training Model, TAPEX: Table Pre-training via Learning a Neural SQL Executor
krishnap25/mauve
Package to compute Mauve, a similarity score between neural text and human text. Install with `pip install mauve-text`.
bigscience-workshop/lm-evaluation-harness
A framework for few-shot evaluation of autoregressive language models.
joeljang/ELM
[ICML 2023] Exploring the Benefits of Training Expert Language Models over Instruction Tuning
microsoft/TUTA_table_understanding
TUTA and ForTaP for Structure-Aware and Numerical-Reasoning-Aware Table Pre-Training
joeljang/continual-knowledge-learning
[ICLR 2022] Towards Continual Knowledge Learning of Language Models
seonghyeonye/TAPP
[AAAI 2024] Investigating the Effectiveness of Task-Agnostic Prefix Prompt for Instruction Following
boychaboy/KOLD
KOLD: Korean Offensive Language Dataset
seonghyeonye/RoSPr
[EMNLP 2023 Findings] Efficiently Enhancing Zero-Shot Performance of Instruction Following Model via Retrieval of Soft Prompt
swstarlab-infolab/format_converter
Space-efficient graph data converter