mesos / mih

Mesos-in-Hadoop: allows launching a Mesos cluster as a Hadoop job

Geek Repo:Geek Repo

Github PK Tool:Github PK Tool

MIH (Mesos-in-Hadoop) lets you launch a Mesos cluster as a Hadoop job, for
example to easily experiment with Mesos frameworks like Spark on an existing
Hadoop cluster.

MIH requires Hadoop 0.20.203 or later. It should also work with the Cloudera
distribution of Hadoop version 3 or later. On older Hadoops (e.g. 0.20.2),
MIH will run but Hadoop does not properly clean up processes spawned by a job
when the job finishes, so you risk having processes left on the nodes.

To build MIH, export HADOOP_HOME=<path to your Hadoop directory> and run make.

To run, export MESOS_HOME=<location where you installed Mesos on the slaves>
and use HADOOP_HOME/bin/hadoop mih.jar <number of slots>.

MIH will acquire the given number of slots from Hadoop (as a set of long
running map tasks) and launch a Mesos cluster inside them, then print out
the master URL for that cluster. To stop the cluster, kill the corresponding
Hadoop job with HADOOP_HOME/bin/hadoop job -kill.

You can also provide a second argument to mih.jar for the memory to use per
slot (in megabytes). By default, it's 1024 MB.

About

Mesos-in-Hadoop: allows launching a Mesos cluster as a Hadoop job

License:Apache License 2.0


Languages

Language:Java 100.0%