Pinned Repositories
ipex-llm
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.
deep_learning_utilities_cpu
Utilities for Intel optimized Deep learning frameworks
docs
TensorFlow documentation
models
Model Zoo for Intel® Architecture: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
relion
Relion-1.4-opt
optimization based on relion-1.4
TensorRT
TensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.
tfjs-website
WebGL-accelerated ML // linear algebra // automatic differentiation for JavaScript.
docs
TensorFlow documentation
jessie-zhao's Repositories
jessie-zhao/deep_learning_utilities_cpu
Utilities for Intel optimized Deep learning frameworks
jessie-zhao/docs
TensorFlow documentation
jessie-zhao/models
Model Zoo for Intel® Architecture: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors
jessie-zhao/pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
jessie-zhao/relion
jessie-zhao/Relion-1.4-opt
optimization based on relion-1.4
jessie-zhao/TensorRT
TensorRT is a C++ library for high performance inference on NVIDIA GPUs and deep learning accelerators.
jessie-zhao/tfjs-website
WebGL-accelerated ML // linear algebra // automatic differentiation for JavaScript.