wanderingai
A wandering adventurer, crafting about large language models, on a mage's journey to source the gems of AI.
Together AISeattle, WA
wanderingai's Stars
huggingface/transformers
🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.
pytorch/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
hiyouga/LLaMA-Factory
Efficiently Fine-Tune 100+ LLMs in WebUI (ACL 2024)
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
phidatahq/phidata
Build AI Assistants with memory, knowledge and tools.
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
nebuly-ai/optimate
A collection of libraries to optimise AI model performances
kellyjonbrazil/jc
CLI tool and python library that converts the output of popular command-line tools, file-types, and common strings to JSON, YAML, or Dictionaries. This allows piping of output to tools like jq and simplifying automation scripts.
jzhang38/TinyLlama
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
jina-ai/reader
Convert any URL to an LLM-friendly input with a simple prefix https://r.jina.ai/
Lightning-AI/lit-llama
Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adapter fine-tuning, pre-training. Apache 2.0-licensed.
pytorch-labs/gpt-fast
Simple and efficient pytorch-native transformer text generation in <1000 LOC of python.
NVIDIA/cutlass
CUDA Templates for Linear Algebra Subroutines
pytorch/torchtune
A Native-PyTorch Library for LLM Fine-tuning
InternLM/xtuner
An efficient, flexible and full-featured toolkit for fine-tuning LLM (InternLM2, Llama3, Phi3, Qwen, Mistral, ...)
agiresearch/AIOS
AIOS: LLM Agent Operating System
Tinche/aiofiles
File support for asyncio
google-research/t5x
peak/s5cmd
Parallel S3 and local filesystem execution tool.
stochasticai/xTuring
Build, customize and control you own LLMs. From data pre-processing to fine-tuning, xTuring provides an easy way to personalize open-source LLMs. Join our discord community: https://discord.gg/TgHXuSJEk6
huggingface/datatrove
Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.
lcompilers/lpython
Python compiler
huggingface/nanotron
Minimalistic large language model 3D-parallelism training
Liuhong99/Sophia
The official implementation of “Sophia: A Scalable Stochastic Second-order Optimizer for Language Model Pre-training”
datadreamer-dev/DataDreamer
DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. 🤖💤
Ananto30/zero
Zero: A simple and fast Python RPC framework
brentyi/tyro
CLI interfaces & config objects, from types
zeux/calm
CUDA/Metal accelerated language model inference
HazyResearch/flash-fft-conv
FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor Cores
NVIDIA/cuda-checkpoint
CUDA checkpoint and restore utility