The workload read data from Softlayer Object Storage with Spark Swift integration, then write data to a DataStore using Spark SQL. The run results, throughput and latency, are stored as CSV file in Softlayer Object Storage too.
- Reference the high level overview
- Reference the workload details in Python using Spark SQL
export marathonIp=MARATHON_IP
curl -i -H 'Content-Type: application/json' -d@config/es/marathon-es.json $marathonIp:8080/v2/apps
- Start ElasticSearch in Marathon JSON
- Start workload in Marathon JSON
- Start workload with file index pattern in Marathon JSON
The image is used for both Spark job submission and Spark executor on Mesos, besides it can be used to start Spark Standalone cluster.