...
- Install Spark (either download pre-built Spark, or build assembly from source).
- Install/build a compatible version. Hive root pom.xml's <spark.version> defines what version of Spark it was built/tested with.
- Install/build a compatible distribution. Each version of Spark has several distributions, corresponding with different versions of Hadoop.
- Once Spark is installed, find and keep note of the <spark-assembly-*.jar> location.If you download Spark pre-built you will need to replace the Spark 1.2.x assembly with http://ec2-50-18-79-139.us-west-1.compute.amazonaws.com/data/spark-assembly-1.2.0-SNAPSHOT-hadoop2.3.0-cdh5.1.2.jar
- Note that you must have a version of Spark which does not include the hive jars. Meaning one which was not build with the hive profile.
- Start Spark cluster (both standalone and Spark on YARN are supported).
- Keep note of the <Spark Master URL>. This can be found in Spark master WebUI.
...