Another benchmark for some python frameworks
Clone or download
Pull request Compare This branch is 17 commits ahead of klen:develop.
Fetching latest commit…
Cannot retrieve the latest commit at this time.
Permalink
Failed to load latest commit information.
benchmarks
deploy
pyfb
scripts
.dockerignore
.gitignore
.vimrc
Dockerfile
Makefile
README.md
db.py
pylama.ini
pytest.ini
requirements.txt
setup.py
tests.py

README.md

Python frameworks' benchmarks

This is a fork of https://github.com/klen/py-frameworks-bench, refactored a bit to enable different kinds of experiments.

Setup

The tests themselves are orchestrated using Make. In order to run these demos, you will require the following tools (left as an excersize for the reader):

The experiment

The execution is wrk benchmarking an application, sending 400 concurrent requests over 10 threads. The type of work depends on the type parameter passed to make, but is typically some remote call: see the Makefile for more information.

Running a Benchmark

The benchmarks themselves are stored in makefiles under the benchmarks directory.

First, start by making your lab, if you haven't already. This starts up the docker container to execute remote calls against.

make lab

For example, running async-vs-threads would be:

make -f benchmarks/async-vs-threaded.makefile main

The benchmark files themselves are very barebones, just running the main benchmark file with different configuration.

Reading results

Results are written to results.csv, regardless of the benchmark. You will have to clear this file in between benchmark runs.

The columns are declared in scripts/cvs-report.lua

Contributors