- These are my records of testing LLMs locally Using Rust with Candle.
I just modified the sample code to fit my needs, and currently Mistral-7B-Instruct-v0.2-GGUF was tested.
-
Download the model file.
: mistral-7b-instruct-v0.2.Q4_K_M.gguf (TheBloke/Mistral-7B-Instruct-v0.2-GGUF) -
Download the tokenizer file.
: tokenizer.json (mistralai/Mistral-7B-Instruct-v0.2) -
Place them somewhere in root project.
: /model/Mistral-7B-Instruct-v0.2/ -
Run.
cargo run --bin mistral --release -- -p "chat"
- interactive: multiple prompts in an interactive way
- chat: multiple prompts in an interactive way where history is preserved
- else: a single prompt