jamindy's Stars
ollama/ollama
Get up and running with Llama 3.2, Mistral, Gemma 2, and other large language models.
ggerganov/llama.cpp
LLM inference in C/C++
binary-husky/gpt_academic
为GPT/GLM等LLM大语言模型提供实用化交互接口,特别优化论文阅读/润色/写作体验,模块化设计,支持自定义快捷按钮&函数插件,支持Python和C++等项目剖析&自译解功能,PDF/LaTex论文翻译&总结功能,支持并行问询多种LLM模型,支持chatglm3等本地模型。接入通义千问, deepseekcoder, 讯飞星火, 文心一言, llama2, rwkv, claude2, moss等。
langgenius/dify
Dify is an open-source LLM app development platform. Dify's intuitive interface combines AI workflow, RAG pipeline, agent capabilities, model management, observability features and more, letting you quickly go from prototype to production.
hiyouga/LLaMA-Factory
Unified Efficient Fine-Tuning of 100+ LLMs (ACL 2024)
mli/paper-reading
深度学习经典、新论文逐段精读
songquanpeng/one-api
OpenAI 接口管理 & 分发系统,支持 Azure、Anthropic Claude、Google PaLM 2 & Gemini、智谱 ChatGLM、百度文心一言、讯飞星火认知、阿里通义千问、360 智脑以及腾讯混元,可用于二次分发管理 key,仅单可执行文件,已打包好 Docker 镜像,一键部署,开箱即用. OpenAI key management & redistribution system, using a single API for all LLMs, and features an English UI.
naklecha/llama3-from-scratch
llama3 implementation one matrix multiplication at a time
ggerganov/ggml
Tensor library for machine learning
chenzomi12/AISystem
AISystem 主要是指AI系统,包括AI芯片、AI编译器、AI推理和训练框架等AI全栈底层技术
NVIDIA/Megatron-LM
Ongoing research training transformer models at scale
liguodongiot/llm-action
本项目旨在分享大模型相关技术原理以及实战经验(大模型工程化、大模型应用落地)
ymcui/Chinese-LLaMA-Alpaca-2
中文LLaMA-2 & Alpaca-2大模型二期项目 + 64K超长上下文模型 (Chinese LLaMA-2 & Alpaca-2 LLMs with 64K long context models)
Fafa-DL/Lhy_Machine_Learning
李宏毅2021/2022/2023春季机器学习课程课件及作业
facebookresearch/fairscale
PyTorch extensions for high performance and large scale training.
Jittor/jittor
Jittor is a high-performance deep learning framework based on JIT compiling and meta-operators.
OpenRLHF/OpenRLHF
An Easy-to-use, Scalable and High-performance RLHF Framework (70B+ PPO Full Tuning & Iterative DPO & LoRA & RingAttention)
microsoft/Megatron-DeepSpeed
Ongoing research training transformer language models at scale, including: BERT & GPT-2
kyegomez/BitNet
Implementation of "BitNet: Scaling 1-bit Transformers for Large Language Models" in pytorch
XueFuzhao/OpenMoE
A family of open-sourced Mixture-of-Experts (MoE) Large Language Models
intelligent-machine-learning/dlrover
DLRover: An Automatic Distributed Deep Learning System
alibaba/Pai-Megatron-Patch
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
HIT-SCIR/Chinese-Mixtral-8x7B
中文Mixtral-8x7B(Chinese-Mixtral-8x7B)
feifeibear/long-context-attention
USP: Unified (a.k.a. Hybrid, 2D) Sequence Parallel Attention for Long Context Transformers Model Training and Inference
Strivin0311/long-llms-learning
A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarks
FlagOpen/FlagScale
FlagScale is a large model toolkit based on open-sourced projects.
sxontheway/Keep-Learning
The record of what I‘ve been through.
hengjiUSTC/learn-llm
MARD1NO/CUDA-PPT
Strivin0311/llms-learning
A repository sharing the literatures about large language models