This repository contains a Reporting Standard for Algorithms. It consists out of a System Card which contains Model Cards and Assessment Cards.
Inspired by Model Cards for Model Reporting and Papers with Code Model Index this standard almost1 2 3 4 extends the Hugging Face model card metadata specification to allow for:
- More fine-grained information on performance metrics, by extending the
metrics_field
from the Hugging Face metadata specification. - Capturing additional measurements on fairness and bias, which can be partitioned into bar plot like
measurements (such as mean absolute SHAP values) and graph plot like measurements (such as partial dependence). This
is achieved by defining a new field
measurements
. - Capturing assessments (such as
IAMA
and ALTAI).
This is achieved by defining a new field
assessments
.
Following Hugging Face, this proposed standard will be written in YAML.
This standard does not contain all fields present in the Hugging Face metadata specification. The fields that are optional in the Hugging Face specification and are specific to the Hugging Face interface are omitted.
Another difference is that we divide our implementation into three separate parts.
system_card
, containing information about a group of ML-models which accomplish a specific task.model_card
, containing information about a specific data science model.assessment_card
, containing information about a regulatory assessment.
!!! note "Include statements"
These `model_card`s and `assessment_card`s can be included verbatim into a `system_card`, or referenced with an
`!include` statement, allowing for minimal cards to be compact in a single file. Extensive cards can be split up for
readability and maintainability. Our standard allows for the `!include` to be used anywhere.
The standard will be written in YAML. Example YAML files are given in the next section. The standard defines three
cards: a system_card
, a model_card
and an assessment_card
. A system_card
contains information about an
algorithmic system. It can have multiple models and each of these models should have a model_card
. Regulatory
assessments can be processed in an assessment_card
. Note that model_card
's and assessment_card
's can be included
directly into the system_card
or can be included as separate YAML files with help of a YAML-include mechanism. For
clarity the latter is preferred and is also used in the examples in the next section.
Footnotes
-
Deviation from the Hugging Face specification is in the License field. Hugging Face only accepts dataset id's from Hugging Face license list while we accept any license from Open Source License List. ↩
-
Deviation from the Hugging Face specification is in the
model_index:results:dataset
field. Hugging Face only accepts one dataset, while we accept a list of datasets. ↩ -
Deviation from the Hugging Face specification is in the Dataset Type field. Hugging Face only accepts dataset id's from Hugging Face datasets while we also allow for any url pointing to the dataset. ↩
-
For this extension to work relevant metrics (such as for example false positive rate) have to be added to the Hugging Face metrics, possibly this can be done in our organizational namespace. ↩