Framework for deploying Hadoop clusters on traditional HPC from userland
Switch branches/tags
Clone or download
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Failed to load latest commit information.


This represents the logical continuation of the myhadoop project by Sriram Krishnan (

myhadoop provides a framework for launching Hadoop clusters within traditional high-performance compute clusters and supercomputers. It allows users to provision and deploy Hadoop clusters within the batch scheduling environment of such systems with minimal expertise required.

Quick Install

Assuming you unpacked myHadoop in /usr/local/myhadoop and your Hadoop binary distribution is located in /usr/local/hadoop-1.2.1:

cd /usr/local/hadoop-1.2.1/conf
patch < /usr/local/myhadoop/myhadoop-1.2.1.patch

That's it.

myHadoop also supports Hadoop 2 via the myhadoop-2.2.0.patch file. This patch has been tested to work against Hadoop 2.6, and it may work with newer versions as well.

See for a more detailed installation guide and a brief introduction to using myHadoop.

About myHadoop

This framework comes with three principal runtimes:

  • bin/ - creates a Hadoop configuration set based on the information presented by a system's batch scheduler. At present, myhadoop interfaces with Torque, SLURM, and Sun Grid Engine.
  • bin/ - cleans up after the Hadoop cluster has been torn down.
  • bin/ - When run from either within a job submission script or an interactive job, it provides a one-command configuration and spinup of a Hadoop cluster and instructions for the user on how to connect to his or her personal cluster.

This script takes a series of template configuration XML files, applies the necessary patches based on the job runtime environment provided by the batch scheduler, and (optionally) formats HDFS. The general syntax is -c /your/new/config/dir -s /path/to/node/local/storage


  • /your/new/config/dir is where you would like your Hadoop cluster's config directory to reside. This is a relatively arbitrary choice, but it will then serve as your Hadoop cluster's $HADOOP_CONF_DIR
  • /path/to/node/local/storage is the location of a non-shared filesystem on each node that can be used to store each node's configuration, state, and HDFS data.

The examples/ directory contains torque.qsub which illustrates how this would look in practice.

Before calling, you MUST have JAVA_HOME defined in your environment and HADOOP_HOME defined in either your environment or your myhadoop/etc/myhadoop.conf file. will look in $HADOOP_HOME/conf for the configuration templates it will use for your personal Hadoop cluster.

This is a courtesy script that simply deletes all of the data created by a Hadoop cluster on all of the cluster's nodes. A proper batch system should do this automatically.

To run, you must have HADOOP_HOME and HADOOP_CONF_DIR defined in your environment.

This is another courtesy script that wraps to simplify the process of creating Hadoop clusters for users. When run from within a batch environment (e.g., a batch job or an interactive job), it will gather the necessary information from the batch system to run, run it, and create a file called "setenv.sourceme" that contains all of the instructions a user will need to connect to his or her newly spawned Hadoop cluster and begin using it.

The examples/ directory contains bootstrap.qsub that illustrates how a user may use in all supported batch environments. He or she simply has to submit this script and wait for "setenv.sourceme" to appear in his or her directory. Once it appears, he or she can "cat" the file and follow the instructions contained within to get to an environment where he or she can begin interacting with his or her cluster using the "hadoop" command.