victorsungo's Stars
mlabonne/llm-course
Course to get into Large Language Models (LLMs) with roadmaps and Colab notebooks.
microsoft/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
mckaywrigley/chatbot-ui
AI chat for any model.
facebookresearch/fastText
Library for fast text representation and classification.
Vision-CAIR/MiniGPT-4
Open-sourced codes for MiniGPT-4 and MiniGPT-v2 (https://minigpt-4.github.io, https://minigpt-v2.github.io/)
Hannibal046/Awesome-LLM
Awesome-LLM: a curated list of Large Language Model
mlc-ai/mlc-llm
Universal LLM Deployment Engine with ML Compilation
huggingface/peft
🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.
meta-llama/codellama
Inference code for CodeLlama models
meta-llama/llama-recipes
Scripts for fine-tuning Meta Llama with composable FSDP & PEFT methods to cover single/multi-node GPUs. Supports default & custom datasets for applications such as summarization and Q&A. Supporting a number of candid inference solutions such as HF TGI, VLLM for local or cloud deployment. Demo apps to showcase Meta Llama for WhatsApp & Messenger.
THUDM/ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
QwenLM/Qwen
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
huggingface/trl
Train transformer language models with reinforcement learning.
artidoro/qlora
QLoRA: Efficient Finetuning of Quantized LLMs
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
bigcode-project/starcoder
Home of StarCoder: fine-tuning & inference!
EleutherAI/gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries
xorbitsai/inference
Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop.
InternLM/MindSearch
🔍 An LLM-based Multi-agent Framework of Web Search Engine (like Perplexity.ai Pro and SearchGPT)
huggingface/alignment-handbook
Robust recipes to align language models with human and AI preferences
ricklamers/gpt-code-ui
An open source implementation of OpenAI's ChatGPT Code interpreter
google-research/t5x
ise-uiuc/magicoder
[ICML'24] Magicoder: Empowering Code Generation with OSS-Instruct
WisdomShell/codeshell
A series of code large language models developed by PKU-KCL
EleutherAI/math-lm
pjlab-sys4nlp/llama-moe
⛷️ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)
abacaj/code-eval
Run evaluation on LLMs using human-eval benchmark
mzbac/wizardCoder-vsc
Visual Studio Code extension for WizardCoder