justinxzhao's Stars
pytorch/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
openai/openai-cookbook
Examples and guides for using the OpenAI API
lm-sys/FastChat
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.
openai/openai-python
The official Python library for the OpenAI API
Hannibal046/Awesome-LLM
Awesome-LLM: a curated list of Large Language Model
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
horovod/horovod
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet.
dottxt-ai/outlines
Structured Text Generation
Chainlit/chainlit
Build Conversational AI in minutes ⚡️
weaviate/Verba
Retrieval Augmented Generation (RAG) chatbot powered by Weaviate
nat/openplayground
An LLM playground you can run on your laptop
enricoros/big-AGI
AI suite powered by state-of-the-art models and providing advanced AI/AGI functions. It features AI personas, AGI functions, multi-model chats, text-to-image, voice, response streaming, code highlighting and execution, PDF import, presets for developers, much more. Deploy on-prem or in the cloud.
predibase/lorax
Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs
openai/simple-evals
cuthbertLab/music21
music21 is a Toolkit for Computational Musicology
stanford-crfm/helm
Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110). This framework is also used to evaluate text-to-image models in HEIM (https://arxiv.org/abs/2311.04287) and vision-language models in VHELM (https://arxiv.org/abs/2410.07112).
ise-uiuc/magicoder
[ICML'24] Magicoder: Empowering Code Generation with OSS-Instruct
Audiveris/audiveris
Latest generation of Audiveris OMR engine
tatsu-lab/alpaca_eval
An automatic evaluator for instruction-following language models. Human-validated, high-quality, cheap, and fast.
AgentOps-AI/tokencost
Easy token price estimates for 400+ LLMs. TokenOps.
evalplus/evalplus
Rigourous evaluation of LLM-synthesized code - NeurIPS 2023 & COLM 2024
lmarena/arena-hard-auto
Arena-Hard-Auto: An automatic LLM benchmark.
predibase/llm_distillation_playbook
Best practices for distilling large language models.
lukasberglund/reversal_curse
EQ-bench/EQ-Bench
A benchmark for emotional intelligence in large language models
lucasmaystre/choix
Inference algorithms for models based on Luce's choice axiom
naszilla/tabzilla
vinid/NegotiationArena
Sahandfer/EmoBench
This is the official repository for the paper "EmoBench: Evaluating the Emotional Intelligence of Large Language Models"
llm-council/llm-council
LLMs sitting on a council together to decide, by consensus, who among them is the best.