I aim to give users a new way to interact with Google Maps through engaging text-based conversational interfaces.
I'm only a baby bot right now, I need you to feed me with logic, data and inspiration.
The primary motivation of the developers of MapBot is to provide a playground to tech enthusiasts, both beginners and advanced to try algorithms, approaches and ideas while contributing to a real-life project.
- I want to help users in the most comprehensive way.
- I want to give 'geeks' a platform to try out all things 'cool'.
Check out all related information here
- MySQL
- Install the community version of mySQL from the official mySQL documentation page.
- Create root user credentials during installation.
- Verify the installation, running the command
mysql -uroot -p -hlocalhost
should open the mySQL monitor. (Enter the root password when prompted) - StanfordCoreNLP
- Clone the repository
- Create the mapbot database in mySQL
mysql -uroot -p -hlocalhost
- Enter root password when prompted
create database mapbot;
- Verify creation of the database
show databases;
- Unzip the StanfordCoreNLP package in the repository and keep the file names
stanford-corenlp-x.x.x.jar
andstanford-corenlp-x.x.x-models.jar
handy. - Add config.py file to .gitignore to avoid pushing changes made to config
- Run
git rm --cached config.py
- Edit the config.py file with the corresponding values
- user = "root"
- password = <your_root_password>
- host = "localhost"
- database = "mapbot"
- key = <your_Google_Cloud_API_key>
- stanford_path_to_jar = <your_path_to_stanford-corenlp-x.x.x.jar>
- stanford_path_to_models_jar = <your_path_to_stanford-corenlp-x.x.x-models.jar>
- javahome = <your_path_to_jdk_bin_java.exe>
- Install dependencies from
requirements.txt
file. Runpip install -r requirements.txt
- You're all set up, run the
init.py
file.python init.py
- It is recommended that you set this project up in a virtual environment to keep the dependencies separated and for easier debugging. Here's how you can do that -
The analysis folder contains data files for the project. The sentences.csv contains the base training dataset which is used to classify the user's input into three classes - Statement, Question, and Chat. Going through some examples would clarify the difference between statement and chat. The featuresDump.csv is the result of text pre-processing done using the code in features.py and featuresDump.py.
Here's a Medium article with the some superficial explanations, there are some video links too!