Fetching contributors…
Cannot retrieve contributors at this time
268 lines (208 sloc) 10.1 KB
# $Id$
# Author: Thilee Subramaniam
# Copyright 2012,2016 Quantcast Corporation. All rights reserved.
# This file is part of Kosmos File System (KFS).
# Licensed under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License. You may obtain a copy of the License at
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# implied. See the License for the specific language governing
# permissions and limitations under the License.
MSTRESS : A framework for metaserver/namenode benchmarking
[1] Framework description
[2] Files in this direcotry
[3] Running benchmark
[4] Setting up DFS metaserver/namenode
[1] Framework
The mstress master would invoke in slave mode on the client hosts
through SSH.
Each mstress slave would invoke the necessary number of load-generating clients,
which would stress the meta server.
| +-------------------+ |
| | +-----+----------------------+
| | (--mode master) +-----+------------------+ |
| +-------------------+ | | |
| (master host) | | |
+-----------------------------+ | |
| |
+--------------------------------------+ | |
| | | |
+-----------+ | +--------------+ +--------------+ | | |
| |<---------+-|mstress_client|<--| |<-+--+ |
| | | +--------------+ |(--mode slave)| | |
| DFS meta | | +--------------+ | |
| server | | (client host 1) | |
| | +--------------------------------------+ |
| | |
| | |
| | +--------------------------------------+ |
| | | +--------------+ +--------------+ | |
| |<-----------|mstress_client|<--| |<-+------+
+-----------+ | +--------------+ |(--mode slave)| |
| +--------------+ |
| (client host 2) |
The clients will do file or directory tree creation, stat, or directory walk as
specified by the benchmark plan.
[2] Files
- CMakeLists.txt
Builds the QFS stress client (C++) and HDFS stress client (Java) along with
the main release build.
- build.xml, ivy.xml, ivysettings.xml
Ant build file and Ivy dependency and settings files, used by CMakeLists.txt
to build the HDFS stress client (Java). Ensure that $JAVA_HOME is set.
Produces the mstress_client binary that actually drives the QFS metaserver.
Builds to $GIT_DIR/build/release/bin/benchmarks/mstress_client
See 'Benchmarking Procedure' below for details.
Produces the java MStress_Client for HDFS namenode. Built using ant to
See 'Benchmarking Procedure' below for details.
Helper script used to deploy mstress to a list of hosts. Will invoke cmake
and make under ./build/.
Used to generate a plan file for benchmarking.
Args: client hosts list, number of clients per client host, file tree depth,
nodes per level etc.
The generated plan file is also copied to the /tmp firectory of the
participating client hosts.
Do ./ --help to see all options.
Used to run the metaserver test with the help of the plan file.
Args: dfs server host & port, planfile etc.
This script invokes on the remote host through SSH. For this
reason, the mstress path should be the same on the participating hosts.
Do ./ --help to see all options.
Essentially a wrapper around and
Args: client hosts list and DFS server:port information.
Do --help to see usage.
Used to run sample benchmarks on given QFS and HDFS servers by launching
clients on localhost. Essentially a wrapper around,
make,, and
Used to clean up the plan files and log files created on participating
Do ./ --help to see usage.
[3] Benchmarking Procedure
In reality, benchmark would use separate physical machines each for compiling,
running the DFS server, running mstress master, and load generating clients.
The procedure below assumes different machines, but one can also run all
on the same box, "localhost".
(1) Setup the QFS metaserver and HDFS namenode with the help of
section [4] "Setting up DFS metaserver/namenode" below.
(2) You should have SSH key authentication set up on the hosts involved so
that the scripts can do password/passphrase-less login.
(3) On the build host, compile and install QFS using the steps described in
(4) Determine the master and load generating client hosts that you want to use
to connect to the DFS server. This could just be "localhost" if you want to
run the benchmark locally.
(5) From the build host in this directory, run `./ hosts..`
to deploy mstress files to the participating hosts under ~/mstress.
(6) On the master host change directory to ~/mstress
Create a plan file using
Do ./ --help to see example usage.
./ -c localhost, -n 3 -t file -l 2 -i 10 -n 139
This will create a plan that creates 2 levels of 10 inodes each by 3
processes on 2 hosts. Since each client creates 110 inodes (10 directories
with 10 files each) and since there are 6 clients (3 x 2), this plan is to
create 660 inodes on the DFS server.
The planfile will pick N files to stat per client such that
(N x client-host-count x clients-per-host) is just enough to meet 139.
The plan file gets copied to the /tmp directory where you run it. It will
also get copied to the participating client hosts in the '-c' option.
(7) Checklist: check the presence of,
- the plan file on master host and client hosts (step 6 does this for you)
- the mstress_client binaries (QFS and HDFS clients) on master and all
client hosts (step 5).
(8) Run the benchmark from the master with
Do ./ --help to see options.
./ -f qfs -s <metahost> -p <metaport> -a </tmp/something.plan>
./ -f hdfs -s <namehost> -p <nameport> -a </tmp/something.plan>
(9) The benchmark name, progress, and time taken will be printed out.
[4] DFS Server Setup
[4.1] QFS Metaserver Setup
You can setup the QFS metaserver using the steps described in
If you want to set up a simple metaserver for local testing, please use the
script ~/code/qfs/examples/sampleservers/
[4.2] HDFS Namenode Setup
This will setup the HDFS namenode to listen on port 40000.
The webUI will run on default port 50070.
The installation used here is based on Cloudera's CDH4 release.
(1) Ensure java is installed, and $JAVA_HOME is set.
(2) Add the following to /etc/yum.repos.d/thirdparty.repo (sudo needed)
name=Cloudera's Distribution for Hadoop, Version 4
gpgkey =
gpgcheck = 1
(3) Install hadoop-hdfs-namenode and update the configs.
sudo yum install hadoop-hdfs-namenode
sudo mv /etc/hadoop/conf /etc/hadoop/conf.orig
sudo cp -r /etc/hadoop/conf.empty /etc/hadoop/conf
(4) Update /etc/hadoop/conf/core-site.xml (enter your server name instead
(5) Edit /etc/hadoop/conf/hdfs-site.xml, fix or ensure that there is
a "file://" prefix to avoid warnings.
(6) Format the namenode:
sudo service hadoop-hdfs-namenode init
(7) Start namenode.
sudo service hadoop-hdfs-namenode start
(8) Now namenode should be running. Confirm this by running,
ps aux | grep java
sudo netstat -pan | grep 40000
(9) To administer the files and directories,
/usr/lib/hadoop/bin/hadoop fs -ls /
(10) The user with write access on this namenode is "hdfs". Therefore, give
write permission to "/" folder (for mstress benchmark to use) by logging
in as "hdfs" user.
sudo bash
su hdfs
JAVA_HOME=<java-home> /usr/lib/hadoop/bin/hadoop fs -chmod 777 /
(11) Now the namenode is ready for running benchmarks.