intel-analytics/ipex-llm
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Mixtral, Gemma, Phi, MiniCPM, Qwen-VL, MiniCPM-V, etc.) on Intel XPU (e.g., local PC with iGPU and NPU, discrete GPU such as Arc, Flex and Max); seamlessly integrate with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, vLLM, GraphRAG, DeepSpeed, Axolotl, etc
PythonApache-2.0
Watchers
- 422053362
- acrozesParis / France / Europe
- ambertkNike
- anderson@Kazap
- ArckoDalian Neicun Tech .Ltd
- BENMFeng
- brandontanUtopia
- chenxofhitHunan University of Science And Technology
- codekkembryonic
- dsoul
- greysunBeijing,China
- gxhrid
- ibrahimishagEnsol Biosciences Inc
- jamesyoungAbridged
- jhcloos
- kozo2Tokyo, Japan
- krsreenathaDrona Labs
- kudzai
- laurencecaoShanghai
- leelakrishna
- mahmoodmd
- mailmaheeSan Francisco Bay area
- mloenow
- MultipleCrashesOpenlabs.ai
- orcunyucelNorth Cyprus
- phvuHo Chi Minh city, Viet Nam
- quxiaofengTsinghua University
- saurabhdhupar
- send2vinnie
- stonegaoInnovation Works
- tomycPoland
- verdyr
- waitinfuture
- xuchengbeijing
- yernsunBaidu
- yubingxingicestar