szha's Stars
f/awesome-chatgpt-prompts
This repo includes ChatGPT prompt curation to use ChatGPT and other LLM tools better.
langchain-ai/langchain
🦜🔗 Build context-aware reasoning applications
3b1b/manim
Animation engine for explanatory math videos
hpcaitech/ColossalAI
Making large AI models cheaper, faster and more accessible
microsoft/autogen
A programming framework for agentic AI 🤖 PyPi: autogen-agentchat Discord: https://aka.ms/autogen-discord Office Hour: https://aka.ms/autogen-officehour
google-research/tuning_playbook
A playbook for systematically maximizing the performance of deep learning models.
microsoft/unilm
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
openai/evals
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
Dao-AILab/flash-attention
Fast and memory-efficient exact attention
google-deepmind/deepmind-research
This repository contains implementations and illustrative code to accompany DeepMind publications
Qiskit/qiskit
Qiskit is an open-source SDK for working with quantum computers at the level of extended quantum circuits, operators, and primitives.
togethercomputer/RedPajama-Data
The RedPajama-Data repository contains code for preparing large datasets for training large language models.
CodedotAl/gpt-code-clippy
Full description can be found here: https://discuss.huggingface.co/t/pretrain-gpt-neo-for-open-source-github-copilot-model/7678?u=ncoop57
bytedance/lightseq
LightSeq: A High Performance Library for Sequence Processing and Generation
alpa-projects/alpa
Training and serving large-scale neural networks with auto parallelization.
pyro-ppl/numpyro
Probabilistic programming with NumPy powered by JAX for autograd and JIT compilation to GPU/TPU/CPU.
basicmi/AI-Chip
A list of ICs and IPs for AI, Machine Learning and Deep Learning.
AetherCortex/Llama-X
Open Academic Research on Improving LLaMA to SOTA LLM
google-deepmind/jraph
A Graph Neural Network Library in Jax
Lightning-AI/lightning-thunder
Make PyTorch models up to 40% faster! Thunder is a source to source compiler for PyTorch. It enables using different hardware executors at once; across one or thousands of GPUs.
facebookresearch/bitsandbytes
Library for 8-bit optimizers and quantization routines.
inverse-scaling/prize
A prize for finding tasks that cause large language models to show inverse scaling
google-research/byt5
epfml/powersgd
Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727
google/xarray-beam
Distributed Xarray with Apache Beam
parasj/checkmate
Training neural networks in TensorFlow 2.0 with 5x less memory
data-apis/dataframe-api
RFC document, tooling and other content related to the dataframe API standard
oneapi-src/oneAPI-tab
oneAPI Technical Advisory Board (TAB) Meeting Notes
awslabs/lorien
aws-samples/aws-efa-eks
Deploying EFA in EKS utilizing GPUDirectRDMA where supported