Pinned Repositories
llama.cpp
LLM inference in C/C++
intel-extension-for-transformers
⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Platforms⚡
llvm
Intel staging area for llvm.org contribution. Home for Intel LLVM-based projects.
neural-speed
An innovative library for efficient LLM inference via low-bit quantization
llama.cpp
LLM inference in C/C++
onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator
luoyu-intel's Repositories
luoyu-intel/llama.cpp
LLM inference in C/C++
luoyu-intel/onnxruntime
ONNX Runtime: cross-platform, high performance ML inferencing and training accelerator