Pinned Repositories
CMU-10-714
CMU 10-714 Deep-Learning-Systems
Compass_Optimizer
Compass Optimizer (OPT for short), is part of the Zhouyi Compass Neural Network Compiler. The OPT is designed for converting the float Intermediate Representation (IR) generated by the Compass Unified Parser to an optimized quantized or mixed IR which is suited for Zhouyi NPU hardware platforms.
Compass_Unified_Parser
armchina NPU parser
Competitive_Programming
WPLF template
MIT-6.031-Software-Construction
The record of learning 6.031
MIT_6.5940
MIT open course, efficient ML
my-CS-road
py_tutorial
UCB-CS161-sp24
UCB-CS61c-2020summer
wplf's Repositories
wplf/MIT_6.5940
MIT open course, efficient ML
wplf/my-CS-road
wplf/py_tutorial
wplf/UCB-CS161-sp24
wplf/CMU-10-714
CMU 10-714 Deep-Learning-Systems
wplf/Compass_Optimizer
Compass Optimizer (OPT for short), is part of the Zhouyi Compass Neural Network Compiler. The OPT is designed for converting the float Intermediate Representation (IR) generated by the Compass Unified Parser to an optimized quantized or mixed IR which is suited for Zhouyi NPU hardware platforms.
wplf/Compass_Unified_Parser
armchina NPU parser
wplf/Competitive_Programming
WPLF template
wplf/cs-self-learning
计算机自学指南
wplf/how-to-optimize-gemm
row-major matmul optimization
wplf/How_to_optimize_in_GPU
This is a series of GPU optimization topics. Here we will introduce how to optimize the CUDA kernel in detail. I will introduce several basic kernel optimizations, including: elementwise, reduce, sgemv, sgemm, etc. The performance of these kernels is basically at or near the theoretical limit.
wplf/lzzplus2x
lzzkmc_wplf_changed
wplf/MIT-6.031-Software-Construction
The record of learning 6.031
wplf/OI-wiki
:star2: Wiki of OI / ICPC for everyone. (某大型游戏线上攻略,内含炫酷算术魔法)
wplf/onnx
Open standard for machine learning interoperability
wplf/UCB-CS61c-2020summer
wplf/Megatron-LM
Ongoing research training transformer models at scale
wplf/mit-65840
wplf/tinyflow
Tutorial code on how to build your own Deep Learning System in 2k Lines
wplf/TransformerEngine
A library for accelerating Transformer models on NVIDIA GPUs, including using 8-bit floating point (FP8) precision on Hopper and Ada GPUs, to provide better performance with lower memory utilization in both training and inference.
wplf/wplf
This is a special repository about my Github profile.
wplf/wplf.github.io