An open source and collaborative framework for extracting the data you need from websites. In a fast, simple, yet extensible way.
See scrapy official page and the official documentation for more details.
For a list of scrapy commands, simply run:
$ docker run -v $(pwd):/runtime/app aciobanu/scrapy
Since the container doesn't provide any persistence, we can use the volumes
(-v) directive to share the current folder with the container.
To start a new project
$ docker run -v $(pwd):/runtime/app aciobanu/scrapy startproject tutorial
This will create a new tutorial
folder in your current path.
To work on the scrapy project:
$ cd tutorial
$ docker run -v $(pwd):/runtime/app aciobanu/scrapy
Continue reading the official tutorial for a more in depth usage manual of scrapy. For more details about Docker and usage options, please see the official documentation page.