/trump-lies

Tutorial: Web scraping in Python with Beautiful Soup

Primary LanguageJupyter Notebook

Web scraping the President's lies in 16 lines of Python

This repository contains the Jupyter notebook and dataset from Data School's introductory web scraping tutorial. All that is required to follow along is a basic understanding of the Python programming language.

By the end of the tutorial, you will be able to scrape data from a static web page using the requests and Beautiful Soup libraries, and export that data into a structured text file using the pandas library.

You can also watch the tutorial on YouTube.

Watch the tutorial on YouTube

Motivation

On July 21, 2017, the New York Times updated an opinion article called Trump's Lies, detailing every public lie the President has told since taking office. Because this is a newspaper, the information was (of course) published as a block of text:

Screenshot of the article

This is a great format for human consumption, but it can't easily be understood by a computer. In this tutorial, we'll extract the President's lies from the New York Times article and store them in a structured dataset.

Screenshot of the DataFrame

Outline of the tutorial

  • What is web scraping?
  • Examining the New York Times article
    • Examining the HTML
    • Fact 1: HTML consists of tags
    • Fact 2: Tags can have attributes
    • Fact 3: Tags can be nested
  • Reading the web page into Python
  • Parsing the HTML using Beautiful Soup
    • Collecting all of the records
    • Extracting the date
    • Extracting the lie
    • Extracting the explanation
    • Extracting the URL
    • Recap: Beautiful Soup methods and attributes
  • Building the dataset
    • Applying a tabular data structure
    • Exporting the dataset to a CSV file
  • Summary: 16 lines of Python code
    • Appendix A: Web scraping advice
    • Appendix B: Web scraping resources
    • Appendix C: Alternative syntax for Beautiful Soup

16 lines of Python code

Just want to see the code? Here it is:

import requests  
r = requests.get('https://www.nytimes.com/interactive/2017/06/23/opinion/trumps-lies.html')

from bs4 import BeautifulSoup  
soup = BeautifulSoup(r.text, 'html.parser')  
results = soup.find_all('span', attrs={'class':'short-desc'})

records = []  
for result in results:  
    date = result.find('strong').text[0:-1] + ', 2017'
    lie = result.contents[1][1:-2]
    explanation = result.find('a').text[1:-1]
    url = result.find('a')['href']
    records.append((date, lie, explanation, url))

import pandas as pd  
df = pd.DataFrame(records, columns=['date', 'lie', 'explanation', 'url'])  
df['date'] = pd.to_datetime(df['date'])  
df.to_csv('trump_lies.csv', index=False, encoding='utf-8') 

Want to understand the code? Read the tutorial!