Welcome to ChromeHarvester - your sleek and powerful tool for mining the depths of open Chrome tabs! Dive into a sea of data, effortlessly extracting and shuttling tab URLs and source codes to your desired destination, be it a REST endpoint or a cozy SQL database.
📄 Data Extraction: Get the URLs and full HTML source from all your open Chrome tabs.
🌐 Endpoint Integration: Effortlessly post data to your specified REST API endpoint.
💾 Database Storage: Seamlessly save data to your MySQL database.
📂 File Storage: Save tab data to a specified file on your local system for easy access and record-keeping.
⚙️ Customizable: Flexible configurations to fit your workflow.
Node.js MySQL database (optional, for database storage mode)
git clone https://github.com/TonyGlezx/ChromeHarvester.git
cd ChromeHarvester
npm install -g .
To use ChromeHarvester, start Google Chrome with the remote debugging option enabled:
google-chrome --remote-debugging-port=9222
scrape-tabs -e [endpoint]
Replace [endpoint] with your target URL.
scrape-tabs --dbconfig [path to db config file]
Point to your database configuration JSON file.
scrape-tabs -o [file path]
Specify the file path where you want to save the tab data.
--version: Uncover the current version. -e, --endpoint: Designate your endpoint for data dispatch. --dbconfig: Specify your database config file path. -o, --output: Provide the file path to save the tab data. --help: Need assistance? Just ask!
For database adventures, format your dbconfig.json like so:
{
"host": "localhost",
"user": "your_username",
"password": "your_password",
"database": "your_database"
}
For file output mode, simply specify the desired file path using the -o or --output command-line option.
Dive into open-source! Suggestions, improvements, and enhancements are warmly welcomed. Feel free to fork, modify, and send those pull requests, or drop an issue if you spot something amiss!