/dataline

Chat with your data - AI data analysis and visualization on CSV, Postgres, MySQL, Snowflake, SQLite...

Primary LanguageTypeScriptGNU General Public License v3.0GPL-3.0

DataLine logo

💬 Chat with your data using natural language 📊

Introducing DataLine, the simplest and fastest way⚡️ to analyze and visualize your data!
Generate and export charts, tables, reports in seconds with DataLine - Your AI-driven data analysis and visualization tool 🤓

🍿 Watch a quick demo

DataLine logo


Index

Who is this for?

Technical or non-technical people who want to explore data, fast. ⚡️⚡️

It also works for backend developers to speed up drafting queries and explore new DBs with ease. 😎

It's especially well-suited for businesses given its security-first 🔒 and open-source 📖 nature.

What is it?

DataLine is an AI-driven data analysis and visualization tool.

It's privacy-focused, storing everything on your device. No ☁️, only ☀️!

It hides your data from the LLMs used by default, but this can be disabled if the data is not deemed sensitive.

It can connect to a variety of data sources (Postgres, Snowflake, MySQL, Azure SQL Server, Microsoft SQL Server, Excel, SQLite, CSV, sas7bdat, and more), execute queries, generate charts, and allow for copying the results to build reports quickly.

Where is it going?

For now, we're trying to help people get insights out of their data, fast.

This is meant to enable non-technical folks to query data and aid data analysts in getting their jobs done 10x as fast.

But you can still influence the direction we go in. We're building this for you, so you have the biggest say.

Feature Support

  • Broad DB support: Postgres, MySQL, Snowflake, Excel, CSV, SQLite, and more
  • Generating and executing SQL from natural language
  • Ability to modify SQL results, save them, and re-run
  • Better support for explorative questions
  • Querying data files like CSV, Excel, SQLite, sas7bdat (more connection types)
  • Charting via natural language
  • Modifying chart queries and re-rendering/refreshing charts

With a lot more coming soon. You can still influence what we build, so if you're a user and you're down for it, we'd love to interview you! Book some time with one of us here:

Getting started

There are multiple ways of setting up DataLine, simplest being using a binary executable. This allows you to download a file and run it to get started.

A more flexible option is using our hosted Docker image. This allows you to setup authentication and other features if you need them.

Windows

Head over to our releases page, and open the most recent one. There you should find a windows-exe.zip file. Download it, unzip it, and run the DataLine.exe file.

You might get a "Windows protected your PC" message, which is normal (more info -> run anyway). Finally, open http://localhost:7377/ in your browser.

Mac

Homebrew

# install dataline
brew tap ramiawar/dataline
brew install dataline

# run dataline
dataline

DataLine should then be running on port 7377 accessible from your browser: http://localhost:7377

Linux

You can use Homebrew, see the Mac section.

You may also wish to use the binary instead, to do so, follow the instructions in the Windows section, and use the dataline-linux.tar.zip file instead.

Docker

You can also use our official docker image and get started in one command. This is more suitable for business use:

docker run -p 7377:7377 -v dataline:/home/.dataline --name dataline ramiawar/dataline:latest

You can manage this as you would any other container. docker start dataline, docker stop dataline

For updating to a new version, just remove the container and rerun the command. This way the volume is persisted across updates.

docker rm dataline
docker run -p 7377:7377 -v dataline:/home/.dataline --name dataline ramiawar/dataline:latest

To connect to the frontend, you can then visit: http://localhost:7377

Running manually

Check the backend and frontend readmes.

Authentication

DataLine also supports basic auth 🔒 in self-hosted mode 🥳 in case you're hosting it and would like to secure it with a username/password.

Auth is NOT supported ❌ when running the DataLine executable.

To enable authentication on the self-hosted version, add the environment variables AUTH_USERNAME and AUTH_PASSWORD while launching the service. ✅

With Docker

Inject the env vars with the docker run command as follows: docker run -p 7377:7377 -v dataline:/home/.dataline --name dataline -e AUTH_USERNAME=admin -e AUTH_PASSWORD=admin ramiawar/dataline:latest

We plan on supporting multiple user auth in the future, but for now it supports a single user by default.

Startup Quest

Go through the following checklist to explore DataLine's features!

  • Create a sample database connection
  • Create a new chat and rename it
  • Start asking questions about your data and getting answers
  • Refresh the page and re-run some SQL queries
  • Click inside an SQL query, modify it, and save your modification for later!
  • Try to modify your sample DB connection and explore the connection editor page
  • Try asking for a chart!
  • To really challenge it, ask a question that is answered with multiple results (charts, tables, etc.) - example
  • Add a profile picture

Supported databases

See instructions file for more details.

Deployment

The one thing you must configure when deploying DataLine to a custom domain is CORS allowed origins. To do this, add the environment variable ALLOWED_ORIGINS (comma separated list of origins) to your domain name(s).

By default, it is set to http://localhost:7377,http://0.0.0.0:7377 to make it work with local Docker and local binaries.

For example, running the docker image on a remote server with IP 123.123.12.34:

docker run -p 7377:7377 -v dataline:/home/.dataline --name dataline -e ALLOWED_ORIGINS="http://123.123.12.34:7377,https://123.123.12.34:7377" ramiawar/dataline:latest

Excel Support

We support excel files, but they will have to conform to some structure for the time being. We also support multiple sheets - each sheet will be ingested as a separate table.

Right now, we will try to automatically detect the 'header row' and the first column based on some manual data processing (so as to keep things secure). This means that we might detect the wrong things if you have extra rows on top / logos / branding elements.

To ensure the best quality, make sure your first row is the column names, and remove any padding rows/columns from all the sheets. If any sheet fails, the import will fail.

Future improvements to this include optionally allowing LLMs to figure out what the header row is to reduce user effort.