/ifood-scraper

iFood scrapper, scraped restaurant data, and examples on how to use the API

Primary LanguageJupyter Notebook

Overview

This project uses the API presented here to scrape data about all the restaurants in iFood using an Scrapy Spider.

Download the complete datasets

I uploaded the November 2020 and February 2021 datasets to Kaggle, see here

Installation and use of the scraper

Remember to download coordinates_list.csv and put it in the same folder as the spider

Install it using pip (preferably on an virtual environment):

$ pip install -r requirements.txt

Then, you can run the spider with:

$ scrapy runspider ifood-spider.py -o output.csv

You can use other output formats. More info on this link.

Some info on the list of coordinates

I used the coordinate - city pair present in this repo to obtain the coordinates of all the cities in Brazil. Then, I slightly moved the coordinates of the 100 biggest cities to generate more points (making it possible to find restaurants that would not be available in the center of the city). The bigger the city, the more points I generated. The population data can be found here