Allows to convert gguf-ed LMMs to llamafile and upload them to public HF repo You can do that from your local machine and from GitHub Actions
- Convert gguf model to llamafile and upload to huggingface using docker
- GitHub actions for convert gguf model to llamafile and upload to huggingface
- GitHub actions for convert raw model to llamafile and upload to huggingface
- Create ollama file and upload to ollama repo using docker
- GitHub actions to convert gguf model to ollama file and upload to ollama repo
- Create HF repo
- Create Access Key with write permission and save it somewhere
- Create
.env
file and set there corresponding vars:
cp .env.example .env
- Up container
docker compose up -d olmo
- Copy gguf-ed model
docker cp OLMo-1.7-7B-hf.Q8_0.gguf llfiler-olmo-1:/app/
- Connect to container shell
docker exec -it llfiler-olmo-1 /bin/bash
- Convert ggufed model to llamafile:
llamafile-0.8.6/bin/llamafile-convert OLMo-1.7-7B-hf.Q8_0.gguf
- Upload to HF:
huggingface-cli upload "$HF_REPO" "$HF_REPO_FILE"
- Copy
.github/workflows/main.yml
workflow to your repo - Add secret
HF_TOKEN
to your repo secrets - Input
HF_REPO
,HF_REPO_FILE
,REMOTE_GGUF_MODEL
,LLAMAFILE_RELEASE
on workflow start
- Create modelfile:
cp Olmo.Modelfile.example Olmo.Modelfile
- Create
ollama
directory in root of the project (there'll be your secrets and models saved):
mkdir ollama
- Up container
docker compose up -d ollama
- Copy gguf model or download it via curl
docker cp OLMo-1.7-7B-hf.Q8_0.gguf llfiler-ollama-1:/
-
In your
Olmo.Modelfile
replace<path_to_model>
with filename of gguf model -
Create repo for model and add public key from
ollama/id_ed25519.pub
to your account -
Create model with Ollama (by default tag will be
latest
):
ollama create <your_name>/<model_name>:optional_tag -f Modelfile
- Push model to the repo:
ollama push <your_name>/<model_name>:optional_tag