Running www.perplexity.ai like app complete FREE, LOCAL, PRIVATE and NO GPU NEED on any computer [!IMPORTANT]
If you are unable to use this project normally, it is most likely due to issues with your internet connection or your IP, you need free internet connection to use this project normally. ๅฆๆๆจๆ ๆณๆญฃๅธธไฝฟ็จๆญค้กน็ฎ๏ผๅพๅฏ่ฝๆฏ็ฑไบๆจ็ IP ๅญๅจ้ฎ้ข๏ผๆ่ ไฝ ไธ่ฝ่ช็ฑ่ฎฟ้ฎไบ่็ฝใ
FreeAskInternet is a completely free, private and locally running search aggregator & answer generate using LLM, Without GPU needed. The user can ask a question and the system will use searxng to make a multi engine search and combine the search result to the ChatGPT3.5 LLM and generate the answer based on search results. All process running locally and No GPU or OpenAI or Google API keys are needed.
- ๐๏ธ Completely FREE (no need for any API keys)
- ๐ป Completely LOCAL (no GPU need, any computer can run )
- ๐ Completely PRIVATE (all thing running locally, using custom llm)
- ๐ป Runs WITHOUT LLM Hardware (NO GPU NEED!)
- ๐คฉ Using Free ChatGPT3.5 / Qwen / Kimi / ZhipuAI(GLM) API (NO API keys need! Thx OpenAI)
- ๐ต Custom LLM(ollama,llama.cpp) support, Yes we love ollama!
- ๐ Fast and easy to deploy with Docker Compose
- ๐ Web and Mobile friendly interface, designed for Web Search enhanced AI Chat, allowing for easy access from any device.
- index:
- Search based AI Chat:
- Multi LLM models and custom LLM like ollama support:
- System get user input question in FreeAskInternet UI interface( running locally), and call searxng (running locally) to make search on multi search engine.
- crawl search result links content and pass to ChatGPT3.5 / Kimi / Qwen / ZhipuAI / ollama (by using custom llm), ask LLM to answer user question based on this contents as references.
- Stream the answer to Chat UI.
- We support custom LLM setting, so theoretically infinite llm support.
This project is still in its very early days. Expect some bugs.
git clone https://github.com/nashsu/FreeAskInternet.git
cd ./FreeAskInternet
docker-compose up -d
๐ You should now be able to open the web interface on http://localhost:3000. Nothing else is exposed by default.( For old web interface, accessing http://localhost:3030)
How to get Token?
We are using https://github.com/LLM-Red-Team projects to provide those service, you can reference to their readme.
Reference : https://github.com/LLM-Red-Team/kimi-free-api
- start ollama serve
export OLLAMA_HOST=0.0.0.0
ollama serve
- set ollama url in setting: You MUST using your computer's ip address, not localhost/127.0.0.1, because in docker you can't access this address. The model name is the model you want to serve by ollama.
ollama model Reference : https://ollama.com/library
cd ./FreeAskInternet
git pull
docker compose down
docker compose rm backend
docker compose rm free_ask_internet_ui
docker image rm nashsu/free_ask_internet
docker image rm nashsu/free_ask_internet_ui
docker-compose up -d
- ChatGPT-Next-Web : https://github.com/ChatGPTNextWeb/ChatGPT-Next-Web
- FreeGPT35: https://github.com/missuo/FreeGPT35
- Kimi\Qwen\ZhipuAI https://github.com/LLM-Red-Team
- searxng: https://github.com/searxng/searxng
Apache-2.0 license