cloning all comments of the Stock channel on VOZ
location:
- frontend: app/frontend
- backend: app/backend
# run at root project
## up the related services
docker-compose up -d
# generate db (do this at the first time)
./scripts/clone.sh
# install dependencies
yarn install
# build common libs
yarn build:shared
# start frontend service
yarn start:fe
# start backend service
yarn start:be # run backend server
# prepare env for crawl job (just support centos env)
./scripts/setupEnv.sh
# crawling data
./scripts/crawl.sh
- data/comments.csv
- data/comments.xlsx
- location: ./crawler/databases/*
- update the database backup: ./script/dump.sh
By default, the script is crawling data from Stock with 30 Codes.
./spiders/voz_stock.py
stockCodes=[
...
]