This guide will walk you through the steps to set up the Ollama server, pull the necessary models, and run the application.
- Ensure you have Git installed on your system.
- Python and pip should be installed on your machine.
- A terminal or command prompt to run the commands.
First, download or clone the repository containing the application code.
git clone <repository_url>
cd <repository_folder>
Visit the Ollama website to download the Ollama server. Follow the instructions on the site for installation.
Choose and pull an embedding model. Here, mxbai-embed-large is recommended, but you can choose another if you prefer.
ollama pull mxbai-embed-large
Next, pull the response generation model. You can use gemma2, llama3, or another model of your preference.
ollama pull gemma2
Verify that the models specified in your code match the ones you have pulled onto your device. Open your application code and check the model names.
Start the Ollama server in one terminal window.
ollama serve
Open another terminal window and navigate to the directory containing app.py. Run the application on localhost port 5500.
python app.py
Please note that the response time may vary from 1 to 15 minutes, depending on your device's specifications.
Feel free to contribute to this project by opening issues or submitting pull requests.