Welcome to big-AGI 👋, the GPT application for professionals that need function, form,
simplicity, and speed. Powered by the latest models from 12 vendors and
open-source model servers, big-AGI
offers best-in-class Voice and Chat with AI Personas,
visualizations, coding, drawing, calling, and quite more -- all in a polished UX.
Pros use big-AGI. 🚀 Developers love big-AGI. 🤖
Or fork & run on Vercel
👉 roadmap 👉 documentation
big-AGI is an open book; see the ready-to-ship and future ideas in our open roadmap
1.14.0.-.Release.-.2024-02-07.233744.mp4
- Side-by-Side Split Windows: multitask with parallel conversations. #208
- Multi-Chat Mode: message everyone, all at once. #388
- Export tables as CSV: big thanks to @aj47. #392
- Adjustable text size: customize density. #399
- Dev2 Persona Technology Preview
- Better looking chats with improved spacing, fonts, and menus
- More: new video player, LM Studio tutorial (thanks @aj47), MongoDB support (thanks @ranfysvalle02), and speedups
What's New in 1.12.0 · Jan 26, 2024 · AGI Hotline
1.12.0.-.Release.2024-01-26.042616.mp4
- Voice Calls: real-time voice call your personas out of the blue or in relation to a chat #354
- Support OpenAI 0125 Models. #364
- Rename or Auto-Rename chats. #222, #360
- More control over Link Sharing #356
- Accessibility to screen readers #358
- Export chats to Markdown #337
- Paste tables from Excel #286
- Ollama model updates and context window detection fixes #309
What's New in 1.11.0 · Jan 16, 2024 · Singularity
big-agi.mp4
- Find chats: search in titles and content, with frequency ranking. #329
- Commands: command auto-completion (type '/'). #327
- Together AI inference platform support (good speed and newer models). #346
- Persona Creator history, deletion, custom creation, fix llm API timeouts
- Enable adding up to five custom OpenAI-compatible endpoints
- Developer enhancements: new 'Actiles' framework
What's New in 1.10.0 · Jan 6, 2024 · The Year of AGI
- New UI: for both desktop and mobile, sets the stage for future scale. #201
- Conversation Folders: enhanced conversation organization. #321
- LM Studio support and improved token management
- Resizable panes in split-screen conversations.
- Large performance optimizations
- Developer enhancements: new UI framework, updated documentation for proxy settings on browserless/docker
For full details and former releases, check out the changelog.
You can easily configure 100s of AI models in big-AGI:
AI models | supported vendors |
---|---|
Opensource Servers | LocalAI (multimodal) · Ollama · Oobabooga |
Local Servers | LM Studio |
Multimodal services | Azure · Google Gemini · OpenAI |
Language services | Anthropic · Groq · Mistral · OpenRouter · Perplexity · Together AI |
Image services | Prodia (SDXL) |
Speech services | ElevenLabs (Voice synthesis / cloning) |
Add extra functionality with these integrations:
More | integrations |
---|---|
Web Browse | Browserless · Puppeteer-based |
Web Search | Google CSE |
Code Editors | CodePen · StackBlitz · JSFiddle |
Sharing | Paste.gg (Paste chats) |
Tracking | Helicone (LLM Observability) |
- 📢️ Chat with us on Discord
- ⭐ Give us a star on GitHub 👆
- 🚀 Do you like code? You'll love this gem of a project! Pick up a task! - easy to pro
- 💡 Got a feature suggestion? Add your roadmap ideas
- ✨ Deploy your fork for your friends and family, or customize it for work
- Check out some of the big-AGI community projects
Project | Features | GitHub |
---|---|---|
CoolAGI | Code Interpreter, Vision, Mind maps, and much more | nextgen-user/CoolAGI |
HL-GPT | Fully remodeled UI | harlanlewis/nextjs-chatgpt-app |
To download and run this Typescript/React/Next.js project locally, the only prerequisite is Node.js with the npm
package manager.
Clone this repo, install the dependencies (all local), and run the development server (which auto-watches the
files for changes):
git clone https://github.com/enricoros/big-agi.git
cd big-agi
npm install
npm run dev
# You will see something like:
#
# ▲ Next.js 14.1.0
# - Local: http://localhost:3000
# ✓ Ready in 2.6s
The development app will be running on http://localhost:3000
. Development builds have the advantage of not requiring
a build step, but can be slower than production builds. Also, development builds won't have timeout on edge functions.
The production build of the application is optimized for performance and is performed by the npm run build
command,
after installing the required dependencies.
# .. repeat the steps above up to `npm install`, then:
npm run build
next start --port 3000
The app will be running on the specified port, e.g. http://localhost:3000
.
Want to deploy with username/password? See the Authentication guide.
For more detailed information on deploying with Docker, please refer to the docker deployment documentation.
Build and run:
docker build -t big-agi .
docker run -d -p 3000:3000 big-agi
Or run the official container:
- manually:
docker run -d -p 3000:3000 ghcr.io/enricoros/big-agi
- or, with docker-compose:
docker-compose up
or see the documentation for a composer file with integrated browsing
Please refer to the Cloudflare deployment documentation.
Create your GitHub fork, create a Vercel project over that fork, and deploy it. Or press the button below for convenience.
2023-2024 · Enrico Ros x big-AGI · License: MIT · Made with 💙