Pinned Repositories
6.5930-final-project-2023
6s965-fall2022
boxiangw.github.io
My homepage
CS262
Harvard CS262 Introduction to Distributed Computing
flash-attention
Fast and memory-efficient exact attention
ColossalAI
Making large AI models cheaper, faster and more accessible
Megatron-LM
Ongoing research training transformer models at scale
NeMo
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
NeMo-Framework-Launcher
Provides end-to-end model development pipelines for LLMs and Multimodal models that can be launched on-prem or cloud-native.
TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
BoxiangW's Repositories
BoxiangW/CS262
Harvard CS262 Introduction to Distributed Computing
BoxiangW/6.5930-final-project-2023
BoxiangW/6s965-fall2022
BoxiangW/boxiangw.github.io
My homepage
BoxiangW/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
BoxiangW/ColossalAI
Colossal-AI: A Unified Deep Learning System for Large-Scale Parallel Training
BoxiangW/ColossalAI-Benchmark
Performance benchmarking with ColossalAI
BoxiangW/ColossalAI-Examples
Examples of training models with hybrid parallelism using ColossalAI
BoxiangW/ControlNet
Let us control diffusion models!
BoxiangW/dalai
The simplest way to run LLaMA on your local machine
BoxiangW/DeepSpeed
DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.
BoxiangW/latent-diffusion
High-Resolution Image Synthesis with Latent Diffusion Models
BoxiangW/MockingBird
🚀AI拟声: 5秒内克隆您的声音并生成任意语音内容 Clone a voice in 5 seconds to generate arbitrary speech in real-time
BoxiangW/Poker
Fully functional Pokerbot that works on PartyPoker, PokerStars and GGPoker, scraping tables with Open-CV (adaptable via gui) or neural network and making decisions based on a genetic algorithm and montecarlo simulation for poker equity calculation. Binaries can be downloaded with this link:
BoxiangW/SkyComputing
Sky Computing: Accelerating Geo-distributed Computing in Federated Learning
BoxiangW/flash-attention
Fast and memory-efficient exact attention
BoxiangW/llama
Inference code for LLaMA models
BoxiangW/llama.cpp
Port of Facebook's LLaMA model in C/C++
BoxiangW/Megatron-LM
Ongoing research training transformer models at scale
BoxiangW/NeMo
NeMo: a framework for generative AI
BoxiangW/NeMo-Megatron-Launcher
NeMo Megatron launcher and tools
BoxiangW/ohmyzsh
🙃 A delightful community-driven (with 2,100+ contributors) framework for managing your zsh configuration. Includes 300+ optional plugins (rails, git, macOS, hub, docker, homebrew, node, php, python, etc), 140+ themes to spice up your morning, and an auto-update tool so that makes it easy to keep up with the latest updates from the community.
BoxiangW/Pai-Megatron-Patch
The official repo of Pai-Megatron-Patch for LLM & VLM large scale training developed by Alibaba Cloud.
BoxiangW/PCA_linear_autoencoder
BoxiangW/stable-diffusion
A latent text-to-image diffusion model
BoxiangW/TexasSolver
🚀 A very efficient Texas Holdem GTO solver :spades::hearts::clubs::diamonds:
BoxiangW/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
BoxiangW/triton
Development repository for the Triton language and compiler