Spark on Tez execution context - running in Docker Janos Matyas 02 November 2014

Last week Hortonworks announced improvements for running Apache Spark at scale by introducing a new pluggable execution context and has open sourced it.

At SequenceIQ we are always trying to work and offer the latest technology solutions for our clients and help them to choose their favorite technology/option. We are running a project called Banzai Pipeline – to be open sourced soon – with the goal (among many others) to abstract and allow our customers to use their favorite big data runtime: MR2, Spark or Tez. Along this process we have dockerized most of the Hadoop ecosystem – we are running MR2, Spark, Storm, Hive, HBase, Pig, Oozie, Drill etc in Docker containers – on bare metal and in the cloud as well (all of these containers have made top downloads on the official Docker repository). For details you can check these older posts/resources:

Name Description Documentation GitHub
Apache Hadoop Pseudo distributed container
Apache Ambari Multi node – full Hadoop stack, blueprint based
Cloudbreak Cloud agnostic Hadoop as a Service
Periscope SLA policy based autoscaling for Hadoop clusters

We have always been big fans on Apache Spark – due to the simplicity of development and at the same time we are big fans of Apache Tez, for reasons we have blogged before.

When the SPARK-3561 has been submitted we were eager to get our hands on the WIP and early implementation – and this time we’d like to help you with a quick ramp-up and easy solution to have a Spark Docker container where the execution context has been changed to Apache Tez and everything is preconfigured. The only thing you will need to do is to follow these easy steps.

Pull the image from the Docker Repository

We suggest to always pull the container from the official Docker repository – as this is always maintained and supported by us.

docker pull sequenceiq/spark-native-yarn

Once you have pulled the container you are ready to run the image.

Run the image

docker run -i -t -h sandbox sequenceiq/spark-native-yarn /etc/ -bash

You have now a fully configured Apache Spark, where the execution context is Apache Tez.

Test the container

We have pushed sample data and tests from the code repository into the Docker container, thus you can start experimenting right away without writing one line of code.

Calculate PI

Simplest example to test with is the PI calculation.

cd /usr/local/spark
./bin/spark-submit --class org.apache.spark.examples.SparkPi --master execution-context:org.apache.spark.tez.TezJobExecutionContext --conf update-classpath=true ./lib/spark-examples-

You should expect something like the following as the result:

Pi is roughly 3.14668

Run a KMeans example

Run the KMeans example using the sample dataset.

./bin/spark-submit --class sample.KMeans --master execution-context:org.apache.spark.tez.TezJobExecutionContext --conf update-classpath=true ./lib/spark-native-yarn-samples-1.0.jar /sample-data/kmeans_data.txt

You should expect something like the following as the result:

Finished iteration (delta = 0.0)
Final centers:
DenseVector(0.15000000000000002, 0.15000000000000002, 0.15000000000000002)
DenseVector(9.2, 9.2, 9.2)
DenseVector(0.0, 0.0, 0.0)
DenseVector(9.05, 9.05, 9.05)

Other examples (Join, Partition By, Source count, Word count)


./bin/spark-submit --class sample.Join --master execution-context:org.apache.spark.tez.TezJobExecutionContext --conf update-classpath=true ./lib/spark-native-yarn-samples-1.0.jar /sample-data/join1.txt /sample-data/join2.txt

Partition By

./bin/spark-submit --class sample.PartitionBy --master execution-context:org.apache.spark.tez.TezJobExecutionContext --conf update-classpath=true ./lib/spark-native-yarn-samples-1.0.jar /sample-data/partitioning.txt

Source count

./bin/spark-submit --class sample.SourceCount --master execution-context:org.apache.spark.tez.TezJobExecutionContext --conf update-classpath=true ./lib/spark-native-yarn-samples-1.0.jar /sample-data/wordcount.txt

Word count

./bin/spark-submit --class sample.WordCount --master execution-context:org.apache.spark.tez.TezJobExecutionContext --conf update-classpath=true ./lib/spark-native-yarn-samples-1.0.jar /sample-data/wordcount.txt 1

Note that the last argument (1) is the number of reducers.

Using the Spark Shell

The Spark shell works out of the box with the new Tez executor context, the only thing you will need to do is run:

./bin/spark-shell --master execution-context:org.apache.spark.tez.TezJobExecutionContext


Right after the next day that SPARK-3561 has been made available we have started to test at scale using Cloudbreak and run performance tests by using the same Spark jobs developed in Banzai (over 50 individual jobs) using the same input sets, cluster size and Scala code – but changing the default Spark context to a Tez context. Follow up with us on LinkedIn, Twitter or Facebook as we will release these test results and the lessons we have learned in the coming weeks.


Recent Posts