/ScalaSparkMinioS3

The code contained in this repository shows how to setup a MinIO, using docker-compose, and how to create an Apache Spark job capable of read and write on it.

Primary LanguageScala

ScalaSparkMinioS3

The code contained in this repository shows how to setup a MinIO, using docker-compose, and how to create an Apache Spark job capable of read and write on it.

The Apache Spark code is going to execute in embedded mode, setting master to local, using all computing resources available.

This code can be used as a template in order to setup Spark to work along with MinIO in Scala.

This repository is partially based on this tutorial.

Requirements

In order to execute this code, you are going to need:

  • sbt
  • Java 8+
  • docker
  • on Windows:
    • setup winutil.exe and hadoop.dll, like here.

Setting up services before running the code

docker-compose up -d # starting minio

How to run

sbt run