Pinned Repositories
ipex-llm
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, Phi, MiniCPM, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, GraphRAG, DeepSpeed, vLLM, FastChat, Axolotl, etc.
ai-reference-models
Intel® AI Reference Models: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors and Intel® Data Center GPUs
oneAPI-samples
Samples for Intel® oneAPI Toolkits
pytorch
Tensors and Dynamic neural networks in Python with strong GPU acceleration
ai-documents
GenAIExamples
Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open Platform for Enterprise AI (OPEA) project.
inference_results_v2.1
Model-References
Reference models for Intel(R) Gaudi(R) AI Accelerator
models
Model Zoo for Intel® Architecture: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors
oneAPI-samples
Samples for Intel oneAPI toolkits
wangkl2's Repositories
wangkl2/ai-documents
wangkl2/GenAIExamples
Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open Platform for Enterprise AI (OPEA) project.
wangkl2/inference_results_v2.1
wangkl2/Model-References
Reference models for Intel(R) Gaudi(R) AI Accelerator
wangkl2/models
Model Zoo for Intel® Architecture: contains Intel optimizations for running deep learning workloads on Intel® Xeon® Scalable processors
wangkl2/oneAPI-samples
Samples for Intel oneAPI toolkits