llm模型导出onnx模型请使用llm-export
当前支持以模型:
model | onnx-fp32 | mnn-int4 |
---|---|---|
chatglm-6b | ||
chatglm2-6b | ||
codegeex2-6b | ||
Qwen-7B-Chat | ||
Baichuan2-7B-Chat | ||
Llama-2-7b-chat |
# <model> like `chatglm-6b`
# linux/macos
./script/download_model.sh <model>
# windows
./script/download_model.ps1 <model>
当前构建状态:
System | build-test |
---|---|
Linux | |
Macos | |
Windows | |
Android |
# linux
./script/linux_build.sh
# macos
./script/macos_build.sh
# windows msvc
./script/windows_build.ps1
# android
./script/android_build.sh
TODO
# linux/macos
./cli_demo # cli demo
./web_demo # web ui demo
# windows
.\Debug\cli_demo.exe
.\Debug\web_demo.exe