intel-analytics/ipex-llm
Accelerate local LLM inference and finetuning (LLaMA, Mistral, ChatGLM, Qwen, Baichuan, Mixtral, Gemma, etc.) on Intel CPU and GPU (e.g., local PC with iGPU, discrete GPU such as Arc, Flex and Max). A PyTorch LLM library that seamlessly integrates with llama.cpp, Ollama, HuggingFace, LangChain, LlamaIndex, DeepSpeed, vLLM, FastChat, etc.
PythonApache-2.0
Watchers
- 422053362
- acrozesParis / France / Europe
- ambertkNike
- anderson@Kazap
- ArckoDalian Neicun Tech .Ltd
- BENMFeng
- brandontanUtopia
- codekkembryonic
- dsoul
- greysunBeijing,China
- gxhrid
- ibrahimishagEnsol Biosciences Inc
- jamesyoungAbridged
- jhcloos
- kozo2Tokyo, Japan
- krsreenathaDrona Labs
- kudzai
- laurencecaoShanghai
- leelakrishna
- mahmoodmd
- mailmaheeSan Francisco Bay area
- mloenow
- MultipleCrashesOpenlabs.ai
- orcunyucelNorth Cyprus
- phvuHo Chi Minh city, Viet Nam
- quxiaofengTsinghua University
- robbinhanYouzu
- saurabhdhupar
- send2vinnie
- stonegaoInnovation Works
- tomycPoland
- verdyr
- waitinfuture
- xuchengbeijing
- yernsunBaidu
- yubingxingicestar