job-information-crawler
this is a linkedin, seek job information crawler
This project shows how to use the Puppeteer framework to work with several other commonly used Node.js libraries, such as csv-writer
for data export, dayjs
for date processing, etc., to implement web page automation tasks.
- linkedin, seek job information crawler
- Web page automation
- CSV file generation
- Environment variable management
- Database interaction
First, make sure you have installed Node.js (the latest stable version is recommended). Then, clone this repository and install dependencies:
git clone https://github.com/kenjiding/job-information-crawler.git
cd job-information-crawler
# .env file
LINKEDIN_EMAIL='your_linkedin_email'
LINKEDIN_PASSWORD='your_linkedin_password'
SEEK_EMAIL='your_seek_email'
SEEK_PASSWORD='your_seek_password'
# your email account for sending email
SENDER_EMAIL='sender_email'
# your email account password
SENDER_EMAIL_PASSWORD='sender_email_password'
# receiving email
SENDER_EMAIL_TARGET='sender_email'
npm install
npm run start
/**
* ./src/index.ts
*
* You can use the following order:
*
*/
start.seak(...).linkedin(...);
start.linkedin(...).seek(...);
start.linkedin(...);
start.seek(...);
// example of search options
/**
* @param {string} username - LinkedIn username
* @param {string} password - LinkedIn password
* @param {string} keywords - Keywords to search
* @param {string} location - Location to search
* @param {string} filterAlreadyApply - filter out jobs that have already been applied, default is true
* @param {string} titleIncludes - Title includes to search
* @param {string[]} ignores - Keywords to ignore
* @param {Object} filter - Filter options
* @param {number} pages - Number of pages to search
*/