/
grouping.Rmd
172 lines (122 loc) · 6.5 KB
/
grouping.Rmd
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
---
title: "Grouping behavior in yardstick"
author: "Simon Couch"
date: "`r Sys.Date()`"
output: rmarkdown::html_vignette
vignette: >
%\VignetteIndexEntry{Grouping behavior in yardstick}
%\VignetteEngine{knitr::rmarkdown}
%\VignetteEncoding{UTF-8}
---
```{r setup, include = FALSE}
knitr::opts_chunk$set(
collapse = TRUE,
comment = "#>"
)
```
The 1.3.0 release of yardstick introduced an implementation for _groupwise metrics_. The use case motivating the implementation of this functionality is _fairness metrics_, though groupwise metrics have applications beyond that domain. Fairness metrics quantify the degree of disparity in a metric value across groups. To learn more about carrying out fairness-oriented analyses with tidymodels, see the blog post on the tidymodels website. This vignette will instead focus on groupwise metrics generally, clarifying the meaning of "groupwise" and demonstrating functionality with an example dataset.
<!-- TODO: link to forthcoming tidymodels blog post -->
```{r pkgs, message = FALSE}
library(yardstick)
library(dplyr)
data("hpc_cv")
```
# Group-awareness
Even before the implementation of groupwise metrics, _all_ yardstick metrics had been _group-aware_. When grouped data is passed to a group-aware metric, it will return metric values calculated for each group.
To demonstrate, we'll make use of the `hpc_cv` data set, containing class probabilities and class predictions for a linear discriminant analysis fit to the HPC data set of Kuhn and Johnson (2013). The model is evaluated via 10-fold cross-validation, and the predictions for all folds are included.
```{r hpc-cv}
tibble(hpc_cv)
```
For the purposes of this vignette, we'll also add a column `batch` to the data and select off the columns for the class probabilities, which we don't need.
```{r hpc-modify}
set.seed(1)
hpc <-
tibble(hpc_cv) %>%
mutate(batch = sample(c("a", "b"), nrow(.), replace = TRUE)) %>%
select(-c(VF, F, M, L))
hpc
```
If we wanted to compute the accuracy of the first resampled model, we could write:
```{r acc-1}
hpc %>%
filter(Resample == "Fold01") %>%
accuracy(obs, pred)
```
The metric function returns one row, giving the `.metric`, `.estimator`, and `.estimate` for the whole data set it is passed.
If we instead group the data by fold, metric functions like `accuracy` will know to compute values for each group; in the output, each row will correspond to a Resample.
```{r hpc-cv-2}
hpc %>%
group_by(Resample) %>%
accuracy(obs, pred)
```
Note that the first row, corresponding to `Fold01`, gives the same value as manually filtering for the observations corresponding to the first resample and then computing the accuracy.
This behavior is what we mean by group-awareness. When grouped data is passed to group-aware metric functions, they will return values for each group.
# Groupwise metrics
Groupwise metrics are associated with a data-column such that, when passed data with that column, the metric will temporarily group by that column, compute values for each of the groups defined by the column, and then aggregate the values computed for the temporary grouping back to the level of the input data's grouping.
More concretely, let's turn to an example where there is no pre-existing grouping in the data. Consider the portion of the HPC data pertaining to the first resample:
```{r res-1}
hpc %>%
filter(Resample == "Fold01")
```
Suppose that the `batch`es in the data represent two groups for which model performance ought not to differ. To quantify the degree to which model performance differs for these two groups, we could compute accuracy values for either group separately, and then take their difference. First, computing accuracies:
```{r acc-by-group}
acc_by_group <-
hpc %>%
filter(Resample == "Fold01") %>%
group_by(batch) %>%
accuracy(obs, pred)
acc_by_group
```
Now, taking the difference:
```{r diff-acc}
diff(c(acc_by_group$.estimate[2], acc_by_group$.estimate[1]))
```
Groupwise metrics encode the `group_by()` and aggregation step (in this case, subtraction) shown above into a yardstick metric. We can define a new groupwise metric with the `new_groupwise_metric()` function:
```{r}
accuracy_diff <-
new_groupwise_metric(
fn = accuracy,
name = "accuracy_diff",
aggregate = function(acc_by_group) {
diff(c(acc_by_group$.estimate[2], acc_by_group$.estimate[1]))
}
)
```
* The `fn` argument is the yardstick metric that will be computed for each data group.
* The `name` argument gives the name of the new metric we've created; we'll call ours "accuracy difference."
* The `aggregate` argument is a function defining how to go from `fn` output by group to a single numeric value.
The output, `accuracy_diff`, is a function subclass called a `metric_factory`:
```{r acc-diff-class}
class(accuracy_diff)
```
`accuracy_diff` now knows to take accuracy values for each group and then return the difference between the accuracy for the first and second result as output. The last thing we need to associate with the object is the name of the grouping variable to pass to `group_by()`; we can pass that variable name to `accuracy_diff` to do so:
```{r acc-diff-by}
accuracy_diff_by_batch <- accuracy_diff(batch)
```
The output, `accuracy_diff_by_batch`, is a yardstick metric function like any other:
```{r metric-classes}
class(accuracy)
class(accuracy_diff_by_batch)
```
<!-- TODO: once a print method is added, we can print this out and the meaning of "this is just a yardstick metric" will be clearer -->
We can use the `accuracy_diff_by_batch()` metric in the same way that we would use `accuracy()`. On its own:
```{r ex-acc-diff-by-batch}
hpc %>%
filter(Resample == "Fold01") %>%
accuracy_diff_by_batch(obs, pred)
```
We can also add `accuracy_diff_by_batch()` to metric sets:
```{r ex-acc-diff-by-batch-ms}
acc_ms <- metric_set(accuracy, accuracy_diff_by_batch)
hpc %>%
filter(Resample == "Fold01") %>%
acc_ms(truth = obs, estimate = pred)
```
_Groupwise metrics are group-aware._ When passed data with any grouping variables other than the column passed as the first argument to `accuracy_diff()`---in this case, `group`---`accuracy_diff_by_batch()` will behave like any other yardstick metric. For example:
```{r ex-acc-diff-by-batch-2}
hpc %>%
group_by(Resample) %>%
accuracy_diff_by_batch(obs, pred)
```
Groupwise metrics form the backend of fairness metrics in tidymodels. To learn more about groupwise metrics and their applications in fairness problems, see `new_groupwise_metric()`.
<!-- TODO: link to tidyverse blog post and tidymodels articles. -->