Spring for Apache Hadoop extends Spring Batch by providing support for reading from and writing to HDFS, running various types of Hadoop jobs (Java MapReduce, Streaming, Hive, Pig), HBase and Cascading interactions. An important goal is to provide excellent support for non-Java based developers to be productive using Spring Hadoop and not have to write any Java code to use the core feature set.
Spring for Apache Hadoop also applies the familiar Spring programming model to Java MapReduce jobs by providing support for dependency injection of simple jobs as well as a POJO based MapReduce programming model that decouples your MapReduce classes from Hadoop specific details such as base classes and data types.
For build dependencies to use in your own projects see our Quick Start page.
Spring for Apache Hadoop uses Gradle as its build system. To build the system simply run:
from the project root folder. This will compile the sources, run the tests and create the artifacts.
By default Spring for Apache Hadoop compiles against the Apache Hadoop 2.2.x stable relase (hadoop22) *.
The following distros and versions are also supported:
- Apache Hadoop 1.2.x (hadoop12)
- Pivotal HD 1.1 (phd1)
- Pivotal HD 2.0 (phd20) *
- Cloudera CDH4 MR1 (cdh4, cdh4mr1)
- Cloudera CDH4 YARN (cdh4yarn)
- Cloudera CDH5 YARN (cdh5, cdh5yarn) *
- Cloudera CDH5 MR1 (cdh5mr1)
- Hortonworks HDP 1.3 (hdp13)
- Hortonworks HDP 2.0 (hdp20) *
- Hortonworks HDP 2.1 (hdp21) *
* The distributions noted with and asterisk will include spring-yarn support in the build.
- For anyone using older distros and versions we recommend using Spring for Apache Hadoop 1.0.x which is maintained here: https://github.com/spring-projects/spring-hadoop/tree/1.0.x
To compile against a specific distro version pass the
-Pdistro=<label> project property, like so:
gradlew -Pdistro=hadoop12 build
Note that the chosen distro is displayed on the screen:
Using Apache Hadoop 1.2.x [1.2.1]
In this case, the specified Hadoop distribution (above Apache Hadoop 1.2.x) is used to create the project binaries.
The results for CI builds are available at Spring Data Hadoop: Project Summary - Spring CI
We are currently running tests against the following distributions:
- Apache Hadoop 1.2.1
- Apache Hadoop 2.2.0
- Cloudera CDH4
- Hortonworks HDP 1.3
- Hortonworks HDP 2.0
- Pivotal HD 1.1
For its testing, Spring for Apache Hadoop expects a pseudo-distributed/local Hadoop instalation available on
localhost configured with a port of
8020 for HDFS. The
local Hadoop setup allows the project classpath to be automatically used by the Hadoop job tracker. These settings can be customized in two ways:
- Build properties
From the command-line, use
hd.fs for the file-system (to avoid confusion, specify the protocol such as 'hdfs://', 's3://', etc - if none is specified,
hdfs:// will be used),
hd.jt for the jobtracker,
hd.rm for the YARN resourcemanager and
hd.hive for the Hive host/port information, to override the defaults. For example to run against HDFS at
dumbo:8020 one would use:
gradlew -Phd.fs=hdfs://dumbo:8020 build
- Properties file
test.properties file under
src/test/resources folder (further tweaks can be applied through
hadoop-ctx.xml file under
Enabling Hbase/Hive/Pig/WebHdfs Tests
Note that by default, only the vanilla Hadoop tests are running - you can enable additional tests (such as Hive or Pig) by adding the tasks
enableAllTests in short). Use
test.properties file for customizing the default location for these services as well.
Disabling test execution
You can disable all tests by skipping the
gradlew -x test
Here are some ways for you to get involved in the community:
- Get involved with the Spring community on the Spring Community Forums. Please help out on the forum by responding to questions and joining the debate.
- Create JIRA tickets for bugs and new features and comment and vote on the ones that you are interested in.
- Watch for upcoming articles on Spring by subscribing to the Spring Blog.
Github is for social coding: if you want to write code, we encourage contributions through pull requests from forks of this repository. If you want to contribute code this way, read the Spring Framework contributor guidelines.