-
Notifications
You must be signed in to change notification settings - Fork 2
Mesos-in-Hadoop: allows launching a Mesos cluster as a Hadoop job
License
mesos/mih
Folders and files
Name | Name | Last commit message | Last commit date | |
---|---|---|---|---|
Repository files navigation
MIH (Mesos-in-Hadoop) lets you launch a Mesos cluster as a Hadoop job, for example to easily experiment with Mesos frameworks like Spark on an existing Hadoop cluster. MIH requires Hadoop 0.20.203 or later. It should also work with the Cloudera distribution of Hadoop version 3 or later. On older Hadoops (e.g. 0.20.2), MIH will run but Hadoop does not properly clean up processes spawned by a job when the job finishes, so you risk having processes left on the nodes. To build MIH, export HADOOP_HOME=<path to your Hadoop directory> and run make. To run, export MESOS_HOME=<location where you installed Mesos on the slaves> and use HADOOP_HOME/bin/hadoop mih.jar <number of slots>. MIH will acquire the given number of slots from Hadoop (as a set of long running map tasks) and launch a Mesos cluster inside them, then print out the master URL for that cluster. To stop the cluster, kill the corresponding Hadoop job with HADOOP_HOME/bin/hadoop job -kill. You can also provide a second argument to mih.jar for the memory to use per slot (in megabytes). By default, it's 1024 MB.
About
Mesos-in-Hadoop: allows launching a Mesos cluster as a Hadoop job
Resources
License
Stars
Watchers
Forks
Releases
No releases published
Packages 0
No packages published