andrewg42's Stars
bitcoin/bitcoin
Bitcoin Core integration/staging tree
ethereum/go-ethereum
Go implementation of the Ethereum protocol
nlohmann/json
JSON for Modern C++
THUDM/ChatGLM-6B
ChatGLM-6B: An Open Bilingual Dialogue Language Model | 开源双语对话语言模型
Lightning-AI/pytorch-lightning
Pretrain, finetune and deploy AI models on multiple GPUs, TPUs with zero code changes.
meta-llama/llama3
The official Meta Llama 3 GitHub site
openai/gpt-2
Code for the paper "Language Models are Unsupervised Multitask Learners"
tpope/vim-fugitive
fugitive.vim: A Git wrapper so awesome, it should be illegal
Mozilla-Ocho/llamafile
Distribute and run LLMs with a single file.
LlamaFamily/Llama-Chinese
Llama中文社区,Llama3在线体验和微调模型已开放,实时汇总最新Llama3学习资料,已将所有代码更新适配Llama3,构建最好的中文Llama大模型,完全开源可商用
morhetz/gruvbox
Retro groove color scheme for Vim
tpope/vim-surround
surround.vim: Delete/change/add parentheses/quotes/XML-tags/much more with ease
THUDM/ChatGLM3
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型
benfred/py-spy
Sampling profiler for Python programs
itchyny/lightline.vim
A light and configurable statusline/tabline plugin for Vim
halide/Halide
a language for fast, portable data-parallel computation
preservim/nerdcommenter
Vim plugin for intensely nerdy commenting powers
jarro2783/cxxopts
Lightweight C++ command line option parser
zeux/pugixml
Light-weight, simple and fast XML parser for C++ with XPath support
tomasr/molokai
Molokai color scheme for Vim
nihui/waifu2x-ncnn-vulkan
waifu2x converter ncnn version, runs fast on intel / amd / nvidia / apple-silicon GPU with vulkan
DefTruth/Awesome-LLM-Inference
📖A curated list of Awesome LLM Inference Paper with codes, TensorRT-LLM, vLLM, streaming-llm, AWQ, SmoothQuant, WINT8/4, Continuous Batching, FlashAttention, PagedAttention etc.
llvm/torch-mlir
The Torch-MLIR project aims to provide first class support from the PyTorch ecosystem to the MLIR ecosystem.
Evian-Zhang/llvm-ir-tutorial
LLVM IR入门指南
mit-han-lab/smoothquant
[ICML 2023] SmoothQuant: Accurate and Efficient Post-Training Quantization for Large Language Models
triton-inference-server/tensorrtllm_backend
The Triton TensorRT-LLM Backend
NVIDIA/cudnn-frontend
cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it
davidsmorais/kuro
An elegant Microsoft ToDo desktop client for Linux (a fork of Ao)
NVIDIA/k8s-dra-driver
Dynamic Resource Allocation (DRA) for NVIDIA GPUs in Kubernetes
intel/zlib