Skip to content
Switch branches/tags
Go to file
Cannot retrieve contributors at this time


This is the working draft of the system design.


profefe consists of the following components


(a.k.a. cmd/profefe)

A service that receives profiling data (pprof-formatted, i.e. pb.gz, or Go runtime traces). Collector persists the data in a plugable storage. It also provides an API for querying and retrieving stored profiles.

Currently implemented storages:

  • storage/badger, keeps data in Badger DB;
  • storage/s3, stores data in s3-compatible object storage service;
  • storage/gcs, stores data on Google Cloud Storage (GCS);
  • storage/clickhouse (experimental), stores parsed profiling data in ClickHouse DB.


An optional library, that can be integrated into an application instead of net/http/prof. The goal of agent is to periodically scrap pprof data from the running application and to send the data to collector.


Below is the schema of profefe's components interaction:

Schema of profefe components interaction

If integrating the agent into the application is impossible or undesirable, but the application already exposes net/http/pprof server, the pprof data can be scrapped using an external mechanism, that will send the data to collector using its HTTP API.

In the following example, the agent was replaced with an external script that is running inside a cronjob. It periodically requests profiling data from application's pprof server and sends it to collector:

Schema of replacing agent with a cronjob