big-data-europe/docker-hadoop-spark-workbench
[EXPERIMENTAL] This repo includes deployment instructions for running HDFS/Spark inside docker containers. Also includes spark-notebook and HDFS FileBrowser.
Makefile
Issues
- 13
- 0
TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources
#72 opened by nguacon90 - 0
- 0
- 0
Unable to query the hive table in container from spark-shell running in Windows
#68 opened by ZosBHAI - 0
beeline
#67 opened by omarelnahas23 - 4
- 2
- 0
swarm makefile traefik:1.7 syntax
#66 opened by alxdembo - 1
How to read data into Spark from HDFS?
#64 opened by aleksandarskrbic - 0
Directory /hadoop/dfs/name is in an inconsistent state: storage directory does not exist or is not accessible.
#65 opened by wdorninger - 2
Great repo! But how can we use Pyspark?
#53 opened by haydenliu - 2
- 0
- 1
- 0
error in spark notebook
#57 opened by gsumar - 1
Why HDFS run without YARN ?
#56 opened by dolenam317 - 4
incompatible clusterID Hadoop
#55 opened by Atahualkpa - 1
How to add a file from local file system?
#54 opened by alexeytochin - 1
Faile to connect to namenode:8020
#52 opened by kkalugerov - 2
Global Values not found
#50 opened by smrazaabbas - 2
Spark Worker not connected to Spark Master
#48 opened by Vzzarr - 1
GlusterFS?
#47 opened by antonkulaga - 2
- 2
latest version of hue and hue.ini in volume
#29 opened by antonkulaga - 1
java.net.UnknownHostException: namenode
#45 opened by C0mmander198 - 4
- 1
- 4
Binding 8081 port error when I execute scale up
#24 opened by VijayQin - 1
sparkContext not found in spark notebook
#36 opened by Henrilin28 - 2
docker-compose scale spark-worker=3 & spark-submit
#38 opened by Vzzarr - 3
command: ["./wait-for-it.sh"] for swarm mode
#39 opened by antonkulaga - 6
Switch spark-notebook to apache zeppelin
#40 opened by earthquakesan - 1
Question: how can i run hdfs command line commands such as "hdfs dfs fs -ls"
#41 opened by tomer-ben-david - 5
Q: How can you scale out on multiple hosts?
#42 opened by cmantas - 1
- 5
How do i get my hadoop home directory?
#35 opened by akinmail - 12
- 2
How to enable Pyspark in Jupyter Notebook
#18 opened by agmistry - 2
Steps compatible to Windows?
#19 opened by prashantladha - 2
- 7
Any way to run this in swarm mode?
#31 opened by nirobayo - 4
MapReduce jobs causes Namenode to stop
#27 opened by georgiosgeorgiadis - 4
Copying files to HDFS
#28 opened by ashemery - 2
Failed on Connection Exception
#26 opened by spaghettifunk - 1
- 1
Hue is not working
#21 opened - 3
Broken hadoop package - missing GLIBC_2.14
#20 opened by marcino239 - 3
Connect spark-notebook to spark cluster
#17 opened by Miguel-Alonso - 3
Unable to view worker details or Jobs
#16 opened by brett--anderson