Web crawler to get some data from specific pages
There are some third dependencies in project, it is possible install through poetry or in your preferred Python environment by requirements.txt
.
poetry install [--dev]
You can run the crawler calling crawler
module with output arguments. The output arguments are:
-
--print
: to show output in stdout -
--save-json FILENAME
: to save in a file with JSON format -
--save-csv FILENAME
: to save in a file with CSV format
[poetry run] python -m crawler [--print | --save-json | --save-csv]