Skip to content

Official evaluation metrics hook for experimentation #371

@jonathannorris

Description

@jonathannorris

Context

At KubeCon EU 2026, the experimentation discussion (recap) identified interest in having the SDK produce basic evaluation metrics that analytics vendors can consume for experimentation analysis.

Problem

Today, if you want evaluation metrics (e.g. how many evaluations per flag key, broken down by targeting reason, how many unique targeting keys have seen a given variant), you have to build that yourself. Multiple participants noted that this is a common need across experimentation platforms and the data is straightforward to collect.

Proposal

Define an official evaluation metrics hook that collects basic evaluation data:

  • Evaluation count per flag key
  • Evaluations broken down by reason (e.g. TARGETING_MATCH, DEFAULT, STATIC)
  • Unique targeting keys per flag/variant
  • Variant distribution per flag key

This hook could be published as an official OpenFeature contrib package. Analytics vendors (or an OFREP endpoint) could then consume this data without each needing to implement their own collection logic.

The hook approach is preferred over new API surface since this is buildable with the existing hook model.

Questions

  • Should this be a spec recommendation or just an official contrib hook?
  • What is the minimal set of metrics that provides value across vendors?
  • Should this hook emit OTel metrics, or define its own output interface that vendors implement?

Related

Metadata

Metadata

Assignees

No one assigned

    Labels

    enhancementNew feature or request

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions