Pinned Repositories
intel-extension-for-transformers
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
lc0
SYCL work
llama.cpp
Port of Facebook's LLaMA model in C/C++
llvm
Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.
TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
triton
Development repository for the Triton language and compiler
vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
xetla
lc0
The rewritten engine, originally for tensorflow. Now all other backends have been ported here.
Velocity-Bench
KateBlueSky's Repositories
KateBlueSky/intel-extension-for-transformers
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
KateBlueSky/lc0
SYCL work
KateBlueSky/llama.cpp
Port of Facebook's LLaMA model in C/C++
KateBlueSky/llvm
Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.
KateBlueSky/TensorRT-LLM
TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that contain state-of-the-art optimizations to perform inference efficiently on NVIDIA GPUs. TensorRT-LLM also contains components to create Python and C++ runtimes that execute those TensorRT engines.
KateBlueSky/triton
Development repository for the Triton language and compiler
KateBlueSky/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
KateBlueSky/xetla