-
Notifications
You must be signed in to change notification settings - Fork 0
Running Spark on Mesos
To run on a cluster, Spark uses the Apache Mesos resource manager. Follow the steps below to install Mesos and Spark:
- Download and build Spark using the instructions here.
- Download Mesos 0.9.0 from a mirror.
- Configure Mesos using the
configure
script, passing the location of yourJAVA_HOME
using--with-java-home
. Mesos comes with "template" configure scripts for different platforms, such asconfigure.macosx
, so you can just run the one on your platform if it exists. See the README file in the Mesos directory for other configuration options.-
Note: If you want to run Mesos without installing it into the default paths on your system (e.g. if you don't have administrative privileges to install it), you should also pass the
--prefix
option toconfigure
to tell it where to install. For example, pass--prefix=/home/user/mesos
. By default the prefix is/usr/local
.
-
Note: If you want to run Mesos without installing it into the default paths on your system (e.g. if you don't have administrative privileges to install it), you should also pass the
- Build Mesos using
make
and install it usingmake install
. - Create a file called
spark-env.sh
in Spark'sconf
directory, by copyingconf/spark-env.sh.template
, and add the following lines in it:-
export MESOS_NATIVE_LIBRARY=<path to libmesos.so>
. This path is usually<prefix>/lib/libmesos.so
(/usr/local/lib/libmesos.so
by default). Also, on Mac OS X, the library is calledlibmesos.dylib
instead of.so
. -
export SCALA_HOME=<path to Scala installation>
. You can skip this if Scala is on your system path.
-
- Copy Spark and Mesos to the same paths on all the nodes in the cluster (for Mesos the easiest way is to
make install
on every node). - Configure Mesos for deployment:
- On your master node, edit
<prefix>/var/mesos/deploy/masters
to list your master and<prefix>/var/mesos/deploy/slaves
to list the slaves, where<prefix>
is the prefix where you installed Mesos (/usr/local
by default). - On all nodes, edit
<prefix>/var/mesos/deploy/mesos.conf
and add the linemaster=HOST:5050
, where HOST is your master node's hostname. - Run
<prefix>/sbin/mesos-start-cluster.sh
to start Mesos. If all goes well, you should see Mesos's web UI on port 8080 of the master machine. - See Mesos's README file for more information on deploying it.
- On your master node, edit
- To run a Spark job against the cluster, when you create your
SparkContext
, pass the stringHOST:5050
as the first parameter, whereHOST
is the machine running your Mesos master. In addition, pass the location of Spark on your nodes as the third parameter, and a list of JAR files containing your JAR's code as the fourth (these will automatically get copied to the workers). For example:
new SparkContext("HOST:5050", "My Job Name", "/home/user/spark", List("my-job.jar"))
- Download and build Spark using the instructions here.
- Download either revision 1205738 of Mesos if you're using the master branch of Spark, or the pre-protobuf branch of Mesos if you're using Spark 0.3 or earlier (note that for new users, we recommend the master branch instead of 0.3). For revision 1205738 of Mesos, use:
svn checkout -r 1205738 http://svn.apache.org/repos/asf/incubator/mesos/trunk mesos
For the pre-protobuf branch (for Spark 0.3 and earlier), use:
git clone git://github.com/mesos/mesos cd mesos git checkout --track origin/pre-protobuf
- Configure Mesos using the
configure
script, passing the location of yourJAVA_HOME
using--with-java-home
. Mesos comes with "template" configure scripts for different platforms, such asconfigure.template.macosx
, so you can just run the one on your platform if it exists. See the Mesos wiki for other configuration options. - Build Mesos using
make
. - In Spark's
conf/spark-env.sh
file, addexport MESOS_HOME=<path to Mesos directory>
. If you don't have aspark-env.sh
, copyconf/spark-env.sh.template
. You should also setSCALA_HOME
there if it's not on your system's default path. - Copy Spark and Mesos to the same path on all the nodes in the cluster.
- Configure Mesos for deployment:
- On your master node, edit
MESOS_HOME/conf/masters
to list your master andMESOS_HOME/conf/slaves
to list the slaves. Also, editMESOS_HOME/conf/mesos.conf
and add the linefailover_timeout=1
to change a timeout parameter that is too high by default. - Run
MESOS_HOME/deploy/start-mesos
to start it up. If all goes well, you should see Mesos's web UI on port 8080 of the master machine. - See Mesos's deploy instructions for more information on deploying it.
- On your master node, edit
- To run a Spark job against the cluster, when you create your
SparkContext
, pass the stringmaster@HOST:5050
as the first parameter, whereHOST
is the machine running your Mesos master. In addition, pass the location of Spark on your nodes as the third parameter, and a list of JAR files containing your JAR's code as the fourth (these will automatically get copied to the workers). For example:
new SparkContext("master@HOST:5050", "My Job Name", "/home/user/spark", List("my-job.jar"))
If you want to run Spark on Amazon EC2, there's an easy way to launch a cluster with Mesos, Spark, and HDFS pre-configured: the EC2 launch scripts. This will get you a cluster in about five minutes without any configuration on your part.
You can run Spark and Mesos alongside your existing Hadoop cluster by just launching them as a separate service on the machines. To access Hadoop data from Spark, just use a hdfs:// URL (typically hdfs://<namenode>:9000/path
, but you can find the right URL on your Hadoop Namenode's web UI).
In addition, it is possible to also run Hadoop MapReduce on Mesos, to get better resource isolation and sharing between the two. In this case, Mesos will act as a unified scheduler that assigns cores to either Hadoop or Spark, as opposed to having them share resources via the Linux scheduler on each node. Please refer to the Mesos wiki page on Running Hadoop on Mesos.
In either case, HDFS runs separately from Hadoop MapReduce, without going through Mesos.