BoxiangW's Stars
ohmyzsh/ohmyzsh
🙃 A delightful community-driven (with 2,400+ contributors) framework for managing your zsh configuration. Includes 300+ optional plugins (rails, git, macOS, hub, docker, homebrew, node, php, python, etc), 140+ themes to spice up your morning, and an auto-update tool that makes it easy to keep up with the latest updates from the community.
kubernetes/kubernetes
Production-Grade Container Scheduling and Management
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
gradio-app/gradio
Build and share delightful machine learning apps, all in Python. 🌟 Star to support our work!
vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
mbadolato/iTerm2-Color-Schemes
Over 325 terminal color schemes/themes for iTerm/iTerm2. Includes ports to Terminal, Konsole, PuTTY, Xresources, XRDB, Remmina, Termite, XFCE, Tilda, FreeBSD VT, Terminator, Kitty, MobaXterm, LXTerminal, Microsoft's Windows Terminal, Visual Studio, Alacritty, and many more
openai/gpt-2
Code for the paper "Language Models are Unsupervised Multitask Learners"
haotian-liu/LLaVA
[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.
THUDM/ChatGLM2-6B
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
triton-lang/triton
Development repository for the Triton language and compiler
state-spaces/mamba
Mamba SSM architecture
NVIDIA/NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
NVIDIA/TensorRT
NVIDIA® TensorRT™ is an SDK for high-performance deep learning inference on NVIDIA GPUs. This repository contains the open source components of TensorRT.
facebookresearch/xformers
Hackable and optimized Transformers building blocks, supporting a composable construction.
ohmybash/oh-my-bash
A delightful community-driven framework for managing your bash configuration, and an auto-update tool so that makes it easy to keep up with the latest updates from the community.
NVIDIA/FasterTransformer
Transformer related optimization, including BERT, GPT
nelhage/reptyr
Reparent a running program to a new terminal
salesforce/BLIP
PyTorch code for BLIP: Bootstrapping Language-Image Pre-training for Unified Vision-Language Understanding and Generation
togethercomputer/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
attardi/wikiextractor
A tool for extracting plain text from Wikipedia dumps
NVIDIA/nccl
Optimized primitives for collective multi-GPU communication
NVIDIA/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
binance/binance-public-data
Details on how to get Binance public data
r2d4/react-llm
Easy-to-use headless React Hooks to run LLMs in the browser with WebGPU. Just useLLM().
BlackSamorez/tensor_parallel
Automatically split your PyTorch models on multiple GPUs for training & inference
forhaoliu/ringattention
Transformers with Arbitrarily Large Context
NVIDIA/NeMo-Framework-Launcher
Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.
NVIDIA/NVTX
The NVIDIA® Tools Extension SDK (NVTX) is a C-based Application Programming Interface (API) for annotating events, code ranges, and resources in your applications.
sail-sg/zero-bubble-pipeline-parallelism
Zero Bubble Pipeline Parallelism