No description, website, or topics provided.
Jupyter Notebook C++ Python CMake Shell Protocol Buffer Other
Switch branches/tags
Nothing to show
Clone or download
sztomi Merge pull request #1 from kentonv/master
Fix bugs and remove redundant copies
Latest commit 023e6ff Mar 12, 2017

README.md

rpclib benchmarks

This is a set of benchmarks that compares the performance of various RPC libraries/frameworks in specific situations. The benchmark suite is created as part of the rpclib project.

Libraries

The following libraries are or will be benchmarked:

Results

The results can be found in this repository (json files and plots, plus a Jupyter Notebook). I've also written a blog post about this: http://szelei.me/rpc-benchmark-part1/

Compiling and running

Currently, this benchmark set can only be compiled on Linux. Even though all libraries are cross-platform, the conan packages I'm using aren't.

Requirements

  • A reasonably modern Linux distribution (tested on Arch Linux and Ubuntu 16.04)
  • g++ 5.4 or later, or clang-3.8 or later
  • cmake 3.x

Build steps

mkdir build && cd build
conan install .. --build missing
cmake ..
make

Running the benchmarks

I suggest using the run_benchmarks.sh script. It took about 40 minutes no my machine to complete all benchmark runs.

FAQ

What kind of benchmarks are performed?

These are the benchmarks that are implemented or will be implemented for each library from the above list.

  • get_answer: Perform a function call that returns a number and deserialize the result (so that it's available for the user code as an int)
  • get_blob: Perform a function call that returns a large blob and deserialize it into a byte buffer. This is performed in increasing sizes.
  • get_struct: Perform a function call that returns a nested struct with lots of fields and lots of data in it and intentionally bad alignment.

Each library is set up to create local client and server that communicate through a TCP socket.

How are the benchmarks ran?

This benchmark suite uses Google Benchmark to manage the benchmarking. Each library has a fixture that is set up and torn down in each benchmark run once (similarly to unit tests). The setup and teardown performance is not measured. Each microbenchmark is performed in a number of iterations (determined by Google Benchmark). All runs are performed five times. The results are the means, medians and standard deviations of the remaining three measured times.

What is rpclib?

rpclib is a modern msgpack-rpc implementation for C++. It provides a very clean interface inspired by Boost.Python. See rpclib.net for more information.

I think your implementation for library XYZ is suboptimal. Will you change it?

Probably yes, please send a PR. I'm not an expert in the other libraries, so it's entirely possible that I used them wrong.

Why don't you measure against library XYZ?

One or more of the following reasons:

  • I did not know about the library
  • There was no conan package for it and I didn't have time to make one
  • I didn't have time to learn the library
  • The library wasn't to my knowledge widespread enough to be interesting for this comparison.

However, by all means, please let me know if there is one that is worth looking at. Bonus points if you make a conan package and/or pull request with benchmark code for it.