python3 -m venv he_qa
source he_qa/bin/activate
pip install -r requirements.txt
usinng GPT4all:
- downlad https://gpt4all.io/index.html
- insntall
- downlad from the UI, mistral-7B-Instruct-v0.1
- change the modle path in code ->
#fileQa.py
model_path="/Users/admin/Library/Application Support/nomic.ai/GPT4All/mistral-7b-instruct-v0.1.Q4_0.gguf"
mistralai/Mistral-7B-Instruct-v0.1
wget https://files.mistral-7b-v0-1.mistral.ai/mistral-7B-Instruct-v0.1.tar
tar -xf mistral-7B-v0.1.tar
python3 app.py load # for load data and preform sematic search
python3 app.py just sematic search on persist data in /data
streamlit run main.py --logger.level debug
ספר לי על האירועים בבסיס בברמינגטון תאר לי את החשוד מי שלח אותו לבסיס
ssh ec2-user@10.3.19.11 -L 2222:10.3.19.11:22
scp -P 2222 ~/Code/QA-your-data ec2-user@localhost:
turnnnlingn
aws --profile General-Admin-PS-127977499263 ssm start-session --target i-081a949ee752c7450 --document-name AWS-StartPortForwardingSessionToRemoteHost --parameters host="localhost",portNumber="80",localPortNumber="80"
curl http://localhost:8080/models/apply -H "Content-Type: application/json" -d '{ "id" : "mistral" }'
curl http://localhost:8080/v1/chat/completions -H "Content-Type: application/json" -d '{ "model": "mistral", "messages": [{"role": "user", "content": "How are you?"}], "temperature": 0.9 }'
To convert the langchain-chat-app service definition from a docker-compose.yml file to a docker run command, you'll need to translate each of the Compose file directives into their docker run command equivalent.
Here is what the docker run command would look like for your service:
docker run -d --name redis
-p 6379:6379
--network="host"
redis/redis-stack:latest
redis-server
docker run -d --name langchain-chat-app
-p 8000:8000
-v "$(pwd)/models/:/models/"
--network="host"
-e REDIS_URL=redis://localhost:6379
-e REDIS_COLLECTION=reports
-e MODEL_PATH=/models/mistral-7B-v0.1/mistral-7b-instruct-v0.1.Q4_0.gguf
langchain-chat-app:latest
chainlit run main_chainlid.py
docker run --rm -it -p 8000:8000 \
-e DEFAULT_MODEL_HG_REPO_ID="TheBloke/Mistral-7B-Instruct-v0.1-GGUF"
-e DEFAULT_MODEL_FILE="mistral-7b-instruct-v0.1.Q4_0.gguf"
ghcr.io/chenhunghan/ialacol:latest