Pinned Repositories
asfermi
assembler for NVIDIA FERMI. Imported from Google Code
AutoKernel
AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。
bid
C-primer
C-Primer-4th
caffe
Caffe for Structurally Sparse Deep Neural Networks
Constructor
cutlass
CUDA Templates for Linear Algebra Subroutines
deeplearningbook-chinese
Deep Learning Book Chinese Translation
gpgpu-sim_distribution
GPGPU-Sim provides a detailed simulation model of a contemporary GPU (such as NVIDIA's Fermi and GT200 architectures) running CUDA and/or OpenCL workloads and now includes an integrated (and validated) energy model, GPUWattch.
ltj2013's Repositories
ltj2013/gpgpu-sim_distribution
GPGPU-Sim provides a detailed simulation model of a contemporary GPU (such as NVIDIA's Fermi and GT200 architectures) running CUDA and/or OpenCL workloads and now includes an integrated (and validated) energy model, GPUWattch.
ltj2013/asfermi
assembler for NVIDIA FERMI. Imported from Google Code
ltj2013/AutoKernel
AutoKernel 是一个简单易用,低门槛的自动算子优化工具,提高深度学习算法部署效率。
ltj2013/bid
ltj2013/C-primer
ltj2013/C-Primer-4th
ltj2013/caffe
Caffe for Structurally Sparse Deep Neural Networks
ltj2013/Constructor
ltj2013/cutlass
CUDA Templates for Linear Algebra Subroutines
ltj2013/deeplearningbook-chinese
Deep Learning Book Chinese Translation
ltj2013/Fractional-GPUs
Splits single Nvidia GPU into multiple partitions with complete compute and memory isolation (wrt to performace) between the partitions
ltj2013/jitify
A single-header C++ library for simplifying the use of CUDA Runtime Compilation (NVRTC).
ltj2013/ltj2013.github.io
ltj2013/MatrixTranspose
ltj2013/maxas
Assembler for NVIDIA Maxwell architecture
ltj2013/myconfig
ltj2013/onnx
Open standard for machine learning interoperability
ltj2013/ppl.nn
A primitive library for neural network
ltj2013/PPLNN
ltj2013/PPoPP2017_artifact
Third party assembler and GEMM library for NVIDIA Kepler GPU
ltj2013/SJTUThesis
上海交通大学 XeLaTeX 学位论文模板 A XeLaTeX template for Shanghai Jiao Tong University (SJTU) thesis.
ltj2013/TNN
TNN: developed by Tencent Youtu Lab and Guangying Lab, a lightweight and high-performance deep learning framework for mobile inference. TNN is distinguished by several outstanding features, including its cross-platform capability, high performance, model compression and code pruning. Based on ncnn and Rapidnet, TNN further strengthens the support and performance optimization for mobile devices, and also draws on the advantages of good extensibility and high performance from existed open source efforts. TNN has been deployed in multiple Apps from Tencent, such as Mobile QQ, Weishi, Pitu, etc. Contributions are welcome to work in collaborative with us and make TNN a better framework. TNN:由腾讯优图实验室和光影实验室协同打造,移动端高性能、轻量级推理框架,同时拥有跨平台、高性能、模型压缩、代码裁剪等众多突出优势。TNN框架在原有Rapidnet、ncnn框架的基础上进一步加强了移动端设备的支持以及性能优化,同时也借鉴了业界主流开源框架高性能和良好拓展性的优点。目前TNN已经在手Q、微视、P图等应用中落地,欢迎大家参与协同共建,促进TNN推理框架进一步完善。
ltj2013/turingas
Assembler for NVIDIA Volta and Turing GPUs
ltj2013/tvm
bring deep learning workloads to bare metal
ltj2013/vim-deprecated
vim配置文件和插件(不再维护)
ltj2013/zh-google-styleguide
Google 开源项目风格指南 (中文版)