Skip to content
Permalink
Branch: master
Find file Copy path
Find file Copy path
Fetching contributors…
Cannot retrieve contributors at this time
224 lines (165 sloc) 6.36 KB

profefe - Collect profiling data for long-term analysis

Build Status MIT licensed

profefe continuously collects profiling data from a running Go application and provides an API for querying the profiling samples base on metadata associated with the application.


The project is still in its early state. Feedback and contribution are very welcome.


Why Continuous Profiling?

"Continuous Profiling and Go" describes the motivation behind the project.

How does it work?

See Design Docs documentation.

Quickstart

To build and start profefe collector, run:

> make
> ./BUILD/profefe -addr :10100 -log.level debug -badger.dir /tmp/profefe

2019-06-06T00:07:58.499+0200    info    profefe/main.go:86    server is running    {"addr": ":10100"}

You can build a docker image with the collector running the command:

> make docker-image

More documentation about running profefe in docker can be found in contrib/docker/README.md.


The project includes a fork of Google Stackdriver Profiler's example application, modified to use profefe agent, that sends profiling data to profefe collector.

To start the example, in a separate terminal window, run:

> go run ./examples/hotapp/main.go

After a brief period, the application will start sending CPU profiles to the collector:

send profile: http://localhost:10100/api/0/profiles?service=hotapp-service&labels=version=1.0.0&type=cpu
send profile: http://localhost:10100/api/0/profiles?service=hotapp-service&labels=version=1.0.0&type=cpu
send profile: http://localhost:10100/api/0/profiles?service=hotapp-service&labels=version=1.0.0&type=cpu

To query stored profiling data, make an HTTP call to profefe collector API endpoint (see documentation for collector's HTTP API):

> go tool pprof 'http://localhost:10100/api/0/profiles/merge?service=hotapp-service&type=cpu&from=2019-05-30T11:49:00&to=2019-05-30T12:49:00&labels=version=1.0.0'

Fetching profile over HTTP from http://localhost:10100/api/0/profiles...
Saved profile in /Users/varankinv/pprof/pprof.samples.cpu.001.pb.gz
Type: cpu
Entering interactive mode (type "help" for commands, "o" for options)
(pprof) top
Showing nodes accounting for 43080ms, 99.15% of 43450ms total
Dropped 53 nodes (cum <= 217.25ms)
Showing top 10 nodes out of 12
      flat  flat%   sum%        cum   cum%
   42220ms 97.17% 97.17%    42220ms 97.17%  main.load
     860ms  1.98% 99.15%      860ms  1.98%  runtime.nanotime
         0     0% 99.15%    21050ms 48.45%  main.bar
         0     0% 99.15%    21170ms 48.72%  main.baz
         0     0% 99.15%    42250ms 97.24%  main.busyloop
         0     0% 99.15%    21010ms 48.35%  main.foo1
         0     0% 99.15%    21240ms 48.88%  main.foo2
         0     0% 99.15%    42250ms 97.24%  main.main
         0     0% 99.15%    42250ms 97.24%  runtime.main
         0     0% 99.15%     1020ms  2.35%  runtime.mstart

Note, above we requested all profiling data associated with the given meta-information (service and time period), as a single merged profile.

profefe includes an experimental tool, that allows importing existing pprof data into the collector. While the profefe collector is running, you can use the tool as following:

> ./scripts/pprof_import.sh --service service1 --label region=europe-west3 --label host=backend1 --type cpu -- path/to/cpu.prof

uploading service1-cpu-backend1-20190313-0948Z.prof...OK

HTTP API

Save pprof data

POST /api/0/profiles?service=<service>&type=[cpu|heap|...]&labels=<key=value,key=value>
body pprof.pb.gz

< 200 OK
<
{
  "code": 200,
  "body": {
    "id": <id>,
    "type": <type>,
    ···
  }
}
  • service — service name (string)
  • type — profile type (cpu, heap, block, mutex, goroutine, threadcreate or other)
  • labels — a set of key-value pairs, e.g. "region=europe-west3,dc=fra,ip=1.2.3.4,version=1.0" (Optional)

Query saved meta information

GET /api/0/profiles?service=<service>&type=<type>&from=<created_from>&to=<created_to>&labels=<key=value,key=value>

< 200 OK
<
{
  "code": 200,
  "body": [
    {
      "id": <id>,
      "type": <type>
    },
    ···
  ]
}
  • service — service name
  • type — profile type
  • created_from, created_to — a time window between which pprof data was collected, e.g. "from=2006-01-02T15:04:05"
  • labels — a set of key-value pairs

Query saved pprof data returning it as a single merged profile

GET /api/0/profiles/merge?service=<service>&type=<type>&from=<created_from>&to=<created_to>&labels=<key=value,key=value>

< 200 OK
< pprof.pb.gz

Request parameters are the same as for querying meta information.

Return individual pprof data

GET /api/0/profiles/<id>

< 200 OK
< pprof.pb.gz
  • id - id of stored pprof file; returned with the request for meta information query

Get services for which profiling data is stored

GET /api/0/services

< 200 OK
<
{
  "code": 200,
  "body": [
    <service1>,
    ···
  ]
}

Get profefe server version

GET /api/0/version

< 200 OK
<
{
  "code": 200,
  "body": {
    "version": <version>,
    "commit": <git revision>,
    "build_time": <build timestamp>"
  }
}

Further reading

While the topic of continuous profiling in the production is a bit unrepresented in the public internet, some research or commercial projects are already exist

License

MIT

You can’t perform that action at this time.