Docker is a container management platform for automating deployments of re-usable environemnets.
Containers are a type of virtualization. They virtualize environments at the operating system level by sharing the base libraries. This removes the need for boot drives and hardware interfaces for each environment. This makes a container use much less resources than a virtual machine (VM) which can make environments more efficient because lack of duplication of resources.
Using Docker we can keep a fresh deployment of our Hadoop/Jupyter/Spark environment readily avaliable. Taking around 2 minutes to remove and launch a new environment with the same initial configuration everytime.
Download Docker Community Edition from Docker
docker pull w261/w261-environment
In the container for W261 we use docker-compose to build our container. Below is a potentially out of date example. Use the one in your class repo.
version: '3'
services:
quickstart.cloudera:
image: w261/w261-environment:latest
hostname: docker.w261
privileged: true
command: bash -c "/root/start-notebook.sh;/usr/bin/docker-quickstart"
ports:
- "8887:8888" # Hue server
- "8889:8889" # jupyter
- "10020:10020" # mapreduce job history server
- "8022:22" # ssh
- "7180:7180" # Cloudera Manager
- "11000:11000" # Oozie
- "50070:50070" # HDFS REST Namenode
- "50075:50075" # hdfs REST Datanode
- "8088:8088" # yarn resource manager webapp address
- "19888:19888" # mapreduce job history webapp address
- "8983:8983" # Solr console
- "8032:8032" # yarn resource manager access
- "8042:8042" # yarn node manager
- "60010:60010" # hbase
- "4040:4040" # Spark UI
- "8080:8080" # Hadoop Job Tracker
tty: true
stdin_open: true
volumes:
- .:/media/notebooks
- version: this item says use v2 syntax
- services: list of containers
- quickstart.cloudera: the name of a container, the label being quickstart.cloudera
- image: use this base container
- hostname: DNS name for the container
- privledged: allow access to other machines such as the local machine
- commands: run this commands on start
- ports: map ports so that services running on the container are accessible from the local computer
- remote port:local port
- tty: allow a shell to be initiated
- stdin_open: allow interactivity with the shell
- volumes: location to map from local computer to the docker container so they can share.
- /local/path:/media/notebook
- quickstart.cloudera: the name of a container, the label being quickstart.cloudera
If we review the bash scripts startup.sh
we can see that the jupyter notebook is launched from the /media/notebook
directory. This is very important for our deployment.
- Install Docker (Restart as needed)
- Go to your class repo folder on your computer
- Run
docker-compose up
- Open your browser and go to
localhost:8889
Using python packages against HDFS
Add the following parameter to Map Reduce Streaming and MRJob commands:
-cmdenv PATH=/opt/anaconda/bin:$PATH
Apply the docker.w261
alias for 127.0.0.1
aka localhost
- Linux & Mac
- Open Terminal
- Open hostfile by running
sudo nano /etc/hosts
- Append the following line, then save:
127.0.0.1 docker.w261
- Refresh DNS with
sudo killall -HUP mDNSResponder
- Windows:
- Open notepad as administrator (otherwise you'll not be able to save the file)
- Open
C:\Windows\System32\drivers\etc\hosts
in notepad. Note the file has no extension - Append the following line, then save:
127.0.0.1 docker.w261
- Refresh DNS by running
ipconfig /flushdns
in command prompt or powershell
Docker needs 2 CPUs and 4 GB of RAM to ensure resource managers don't crash during normal operation.
- Linux
- By default Docker shares the same resources as the local computer.
- Windows
- Right click Docker in the notification area
- Click Settings
- Click Advanced
- Slide Memory to 4096 MB
- Mac OS
- Click Docker in the clock(?) area
- Click Settings
- Click Advanced
- Slide Memory to 4096 MB
from pyspark.sql import SparkSession
app_name = "example_notebook"
master = "local[*]"
spark = SparkSession\
.builder\
.appName(app_name)\
.master(master)\
.getOrCreate()
sc = spark.sparkContext
spark
is the general session manager for dataframes and the newer style introduced in Spark 2.0
sc
is a Spark context sets up internal services and establishes a connection to a Spark execution environment.
- Windows 10 Pro/Education is required to run Docker on Windows. A free license of Windows 10 Education is avaliable to all students through UCB Software Central
- Macs require a computer capable of virtualization to test this run
sysctl kern.hv_support
in a terminal.- If 1 then good to go
- If 0 then you need a new computer