A small but not smart crawler
- Install Chrome and corresponding version of Chrome WebDriver then configure PATH
-
Install selenium pip install selenium
Get profile path (type chrom://version/
in Chrome to get the profile path) and replace line 16 of spider.py
Change the default download dir in the Chrome and range of accounts in spider.py
at your ease
index.txt
records download index
Run dump_cookies.py
-> Login account -> Rename cookies.pkl and throw it into the cookies directory
Every account can download 100 file from the website due to a server limitation
Run spider.py