Skip to content

mrkamel/map-reduce-ruby

Repository files navigation

MapReduce

The easiest way to write distributed, larger than memory map-reduce jobs

The MapReduce gem provides the easiest way to write custom, distributed, larger than memory map-reduce jobs by using your local disk and some arbitrary storage layer like s3. You can specify how much memory you are willing to offer and MapReduce will use its buffers accordingly. Finally, you can use your already existing background job system like sidekiq or one of its various alternatives.

Installation

Add this line to your application's Gemfile:

gem 'map-reduce-ruby'

And then execute:

$ bundle install

Or install it yourself as:

$ gem install map-reduce-ruby

Usage

Any map-reduce job consists of an implementation of your map function, your reduce function and worker code. So let's start with an implementation for a word count map-reduce task which fetches txt documents from the web.

class WordCounter
  def map(url)
    HTTP.get(url).to_s.split.each do |word|
      yield(word, 1)
    end
  end

  def reduce(word, count1, count2)
    count1 + count2
  end
end

The #map method takes some key, e.g. a url, and yields an arbitrary amount of key-value pairs. The #reduce method takes the key as well as two values and should return a single reduced value.

Next, we need some worker code to run the mapping part:

class WordCountMapper
  def perform(job_id, mapper_id, url)
    mapper = MapReduce::Mapper.new(WordCounter.new, partitioner: MapReduce::HashPartitioner.new(16), memory_limit: 10.megabytes)
    mapper.map(url)

    mapper.shuffle(chunk_limit: 64) do |partitions|
      partitions.each do |partition, path|
        # store content of the tempfile located at path e.g. on s3:
        bucket.object("map_reduce/jobs/#{job_id}/partitions/#{partition}/chunk.#{mapper_id}.json").put(body: File.open(path))
      end
    end
  end
end

Please note that MapReduce::HashPartitioner.new(16) states that we want to split the dataset into 16 partitions (i.e. 0, 1, ... 15). Finally, we need some worker code to run the reduce part:

class WordCountReducer
  def perform(job_id, partition)
    reducer = MapReduce::Reducer.new(WordCounter.new)

    # fetch all chunks of the partitions e.g. from s3:
    bucket.list(prefix: "map_reduce/jobs/#{job_id}/partitions/#{partition}/").each do |object|
      chunk_path = reducer.add_chunk # returns a path to a tempfile

      object.download_file(temp_path)
    end

    reducer.reduce(chunk_limit: 32) do |word, count|
      # each word with its final count
    end
  end
end

Please note that MapReduce::Reducer#add_chunk returns a path to a tempfile, not a Tempfile object. This allows to limit the number of open file descriptors.

To run your mappers, you can do:

job_id = SecureRandom.hex

list_of_urls.each_with_index do |url, index|
  WordCountMapper.perform_async(job_id, index, url)
end

And to run your reducers:

(0..15).each do |partition|
  WordCountReducer.perform_async(job_id, partition)
end

How to automate running the mappers and reducers in sequence, depends on your background job system. The most simple approach is e.g. to track your mapper state in redis and have a job to start your reducers which waits up until your mappers are finished.

That's it.

Limitations for Keys

You have to make sure that your keys are properly sortable in ruby. Please note:

"key" < nil # comparison of String with nil failed (ArgumentError)

false < true # undefined method `<' for false:FalseClass (NoMethodError)

1 > "key" # comparison of Integer with String failed (ArgumentError

{ "key" => "value1" } < { "key" => "value2" } #=> false
{ "key" => "value1" } > { "key" => "value2" } #=> false
{ "key" => "value1" } <=> { "key" => "value2" } #=> nil

For those reasons, it is recommended to only use strings, numbers and arrays or a combination of those.

Internals

To fully understand the performance details, the following outlines the inner workings of MapReduce. Of course, feel free to check the code as well.

MapReduce::Mapper#map calls your map implementation and adds each yielded key-value pair to an internal buffer up until the memory limit is reached. More concretely, it specifies how big the file size of a temporary chunk can grow in memory up until it must be written to disk. However, ruby of course allocates much more memory for a chunk than the raw file size of the chunk. As a rule of thumb, it allocates 10 times more memory. Still, choosing a value for memory_size depends on the memory size of your container/server, how much worker threads your background queue spawns and how much memory your workers need besides map/reduce. Let's say your container/server has 2 gigabytes of memory and your background framework spawns 5 threads. Theoretically, you might be able to give 300-400 megabytes, but now divide this by 10 and specify a memory_limit of around 30.megabytes, better less. The memory_limit affects how much chunks will be written to disk depending on the data size you are processing and how big these chunks are. The smaller the value, the more chunks and the more chunks, the more runs are needed to merge the chunks. When the memory limit is reached, the buffer is sorted by key and fed through your reduce implementation already, as this can greatly reduce the amount of data already. The result is written to a tempfile. This proceeds up until all key-value pairs are yielded. MapReduce::Mapper#shuffle then reads the first key-value pair of all already sorted chunk tempfiles and adds them to a priority queue using a binomial heap, such that with every pop operation on that heap, we get items sorted by key. When the item returned by pop e.g. belongs to the second chunk, then the next key-value pair of the second chunk is subsequently read and added to the priority queue, up until no more pairs are available. This guarantees that we sort all chunks without fully loading them into memory and is called k-way-merge. With every pop operation, your reduce implementation is continously called up until the key changes between two calls to pop. When the key changes, the key is known to be fully reduced, such that the key is hashed modulo the number of partitions and gets written to the correct partition tempfile (when MapReduce::HashPartitioner is used).

The resulting partition tempfiles need to be stored in some global storage system like s3, such that your mapper workers can upload them and the reducer workers can download them.

MapReduce::Reducer#add_chunk adds and registers a new tempfile path such that your reducer can download a mapper file for the particular partition and write its contents to that tempfile path. MapReduce::Reducer#reduce finally again builds up a priority queue and performs k-way-merge, feeds the key-value pairs into your reduce implementation up until a key change between pop operations occurs and yields the fully reduced key-value pair. At the end #reduce removes all the tempfiles. You can pass a chunk_limit to MapReduce::Reducer#reduce, which is most useful when you run on a system with a limited number of open file descriptors allowed. The chunk_limit ensures that only the specified amount of chunks are processed in a single run. A run basically means: it takes up to chunk_limit chunks, reduces them and pushes the result as a new chunk to the list of chunks to process. Thus, if your number of file descriptors is unlimited, you want to set it to a higher number to avoid the overhead of multiple runs.

Partitioners

Partitioners are used to split the dataset into a specified amount of partitions, which allows to parallelize the work to be done by reducers. MapReduce comes with a HashPartitioner, which takes the number of partitions as an argument and derives the partition number from the key as follows:

class HashPartitioner
  def initialize(num_partitions)
    @num_partitions = num_partitions
  end

  def call(key)
    Digest::SHA1.hexdigest(JSON.generate(key))[0..4].to_i(16) % @num_partitions
  end
end

Thus, writing your own custom partitioner is really easy and, as it follows the interface of callables, could even be expressed as a simple one-liner:

MyPartitioner = proc { |key| Digest::SHA1.hexdigest(JSON.generate(key))[0..4].to_i(16) % 8 }

Semantic Versioning

MapReduce is using Semantic Versioning: SemVer

Development

After checking out the repo, run bin/setup to install dependencies. Then, run rake spec to run the tests. You can also run bin/console for an interactive prompt that will allow you to experiment.

To install this gem onto your local machine, run bundle exec rake install. To release a new version, update the version number in version.rb, and then run bundle exec rake release, which will create a git tag for the version, push git commits and the created tag, and push the .gem file to rubygems.org.

Contributing

Bug reports and pull requests are welcome on GitHub at https://github.com/mrkamel/map-reduce-ruby

License

The gem is available as open source under the terms of the MIT License.

About

The easiest way to write distributed, larger than memory map-reduce jobs

Resources

License

Stars

Watchers

Forks

Packages

No packages published