/web-scraper-projects

This repository contains projects that include web scraping

Primary LanguageJupyter Notebook

Web Scraper Projects

This repository contains a collection of web scraping projects aimed at various purposes. Whether you're interested in monitoring news websites, analyzing data, or fetching real-time data like currency exchange rates, this repository offers a range of tools to help you achieve your goals.

Projects Overview

1. Nan (The country's needs)

Description

This project focuses on gathering and analyzing data related to the country's needs. It includes tools for scraping websites, extracting relevant information, and performing data analysis.

Files

  • data_analyze_v1_3.ipynb: Jupyter Notebook containing data analysis scripts.
  • Data_nan.csv: CSV file containing the scraped data.
  • links.txt: Text file containing links used for scraping.

Source Code

  • src/get_page.py: Python script for fetching web pages.
  • src/main.py: Main script for running the scraping process.
  • src/save_link.py: Script for saving scraped links.

2. News

Description

This project focuses on scraping news articles from specific sources. It provides tools to extract news articles from designated websites.

Source Code

  • src/fars_news.py: Python script for scraping news articles from Fars News website.

3. Price - Dollar

Description

This project aims to fetch real-time data related to the dollar exchange rate. It provides tools to scrape currency exchange information from relevant websites.

Source Code

  • src/get_info.py: Python script for fetching currency exchange information.
  • src/main.py: Main script for running the scraping process.

Getting Started

To get started with any of the projects, follow these steps:

  1. Clone the repository to your local machine.
  2. Navigate to the project directory you're interested in.
  3. Install any necessary dependencies listed in the requirements.txt file.
  4. Run the main script to start scraping data.

Contribution

Contributions to this repository are welcome! Whether it's adding new scraping scripts, improving existing ones, or fixing bugs, feel free to submit pull requests.

License

This repository is licensed under the MIT License.