Skip to content

Commit

Permalink
Add more help on explain predictions
Browse files Browse the repository at this point in the history
  • Loading branch information
koenderks committed May 22, 2024
1 parent b9c70ec commit 37a6da4
Show file tree
Hide file tree
Showing 16 changed files with 31 additions and 16 deletions.
3 changes: 2 additions & 1 deletion inst/help/mlClassificationBoosting.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ Boosting works by sequentially adding features to an decision tree ensemble, eac
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).

#### Plots
- Data split: Shows how the data is split into training (and validation), and test set.
Expand Down Expand Up @@ -91,6 +91,7 @@ Generates a new column in your dataset with the class labels of your classificat
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlClassificationKnn.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ K-nearest neighbors is a method of classification that looks at the *k* number o
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).

#### Plots
- Data split: Shows how the data is split into training (and validation), and test set.
Expand Down Expand Up @@ -89,6 +89,7 @@ Generates a new column in your dataset with the class labels of your classificat
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlClassificationLda.md
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@ Linear Discriminant Analysis (LDA) is a method of classification that aims to fi
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Coefficients: Shows the coefficients for the linear discriminants.
- Prior and posterior probabilities: Shows the prior and posterior group probabilities. Prior probabilities are the proportions in the training set.
- Class means training data: Shows the means on every variable for every class in the training data.
Expand Down Expand Up @@ -80,6 +80,7 @@ Generates a new column in your dataset with the class labels of your classificat
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlClassificationNeuralNetwork.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ Feedforward neural networks are predictive algorithms inspired by the biological
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Network weights: Shows the connections in the neural network together with their weights.

#### Plots
Expand Down Expand Up @@ -108,6 +108,7 @@ Generates a new column in your dataset with the class labels of your classificat
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlClassificationRandomForest.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ Random Forest is a method of classification that creates a set of decision trees
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).

#### Plots
- Data split: Shows how the data is split into training (and validation), and test set.
Expand Down Expand Up @@ -89,6 +89,7 @@ Generates a new column in your dataset with the class labels of your classificat
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Breiman, Leo. (2001). Random forests. Machine Learning. Springer, 45(1), 5-32
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlRegressionBoosting.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Boosting works by sequentially adding features to an decision tree ensemble, eac
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).

#### Plots
- Data split: Shows how the data is split into training (and validation), and test set.
Expand Down Expand Up @@ -77,6 +77,7 @@ Generates a new column in your dataset with the values of your regression result
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlRegressionKnn.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ K-nearest neighbors is a method of regression that looks at the *k* number of fe
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).

#### Plots
- Data split: Shows how the data is split into training (and validation), and test set.
Expand Down Expand Up @@ -75,6 +75,7 @@ Generates a new column in your dataset with the values of your regression result
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlRegressionLinear.md
Original file line number Diff line number Diff line change
Expand Up @@ -13,7 +13,7 @@ Linear Regression

#### Tables
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Regression coefficients: Shows a table of the regression coefficients.

#### Plots
Expand Down Expand Up @@ -53,6 +53,7 @@ Generates a new column in your dataset with the values of your regression result
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlRegressionNeuralNetwork.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Feedforward neural networks are predictive algorithms inspired by the biological
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Network weights: Shows the connections in the neural network together with their weights.

#### Plots
Expand Down Expand Up @@ -93,6 +93,7 @@ Generates a new column in your dataset with the values of your regression result
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlRegressionRandomForest.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Random Forest is a method of regression that creates a set of decision trees tha
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).

#### Plots
- Data split: Shows how the data is split into training (and validation), and test set.
Expand Down Expand Up @@ -75,6 +75,7 @@ Generates a new column in your dataset with the values of your regression result
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Breiman, Leo. (2001). Random forests. Machine Learning. Springer, 45(1), 5-32
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlRegressionRegularized.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ Regularized linear regression is an adaptation of linear regression in which the
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Regression coefficients: Gives the regression coefficient for each feature.

#### Plots
Expand Down Expand Up @@ -72,6 +72,7 @@ Generates a new column in your dataset with the values of your regression result
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlclassificationdecisiontree.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ Decision Trees is a supervised learning algorithm that uses a decision tree as a
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Splits in tree: Shows the split variables, their split point, and the number of observations (which are not missing and are of positive weight) sent left or right by the split. It also shows the improvement in deviance given by this split.

#### Plots
Expand Down Expand Up @@ -77,6 +77,7 @@ Generates a new column in your dataset with the class labels of your classificat
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlclassificationsvm.md
Original file line number Diff line number Diff line change
Expand Up @@ -19,7 +19,7 @@ Support Vector Machines is a supervised learning algorithm that maps training ex
- Model performance: Shows commonly used classification evaluation metrics like precision, recall, the F1-score, support and AUC (area under the ROC curve).
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Support vectors: Shows a table containing the data (points) indicated as support vectors by the algorithm.

#### Plots
Expand Down Expand Up @@ -76,6 +76,7 @@ Generates a new column in your dataset with the class labels of your classificat
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
3 changes: 2 additions & 1 deletion inst/help/mlregressiondecisiontree.md
Original file line number Diff line number Diff line change
Expand Up @@ -17,7 +17,7 @@ Decision Trees is a supervised learning algorithm that uses a decision tree as a
- Model performance: Shows commonly used classification evaluation metrics like mean squared error (MSE), root mean squared error (RMSE) and R<sup>2</sup>.
- Feature importance: Shows the available feature importance metrics for the fitted model.
- Permutations: Sets the number of permutations on which the mean dropout loss is based.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables.
- Explain predictions: Shows the decomposition of the model’s prediction into contributions that can be attributed to different explanatory variables. This feature uses the breakdown algoritm from the `ibreakdown` R package. For more details about this method, see Gosiewska and Biecek (2019).
- Splits in tree: Shows the split variables, their split point, and the number of observations (which are not missing and are of positive weight) sent left or right by the split. It also shows the improvement in deviance given by this split.

#### Plots
Expand Down Expand Up @@ -64,6 +64,7 @@ Generates a new column in your dataset with the values of your regression result
### References
-------
- James, G., Witten, D., Hastie, T., & Tibshirani, R. (2013). An introduction to statistical learning. Springer New York.
- Gosiewska, A. & Biecek, P. (2019). Do Not Trust Additive Explanations. <i>ArXiv</i>. https://doi.org/10.48550/arXiv.1903.11420

### R-packages
---
Expand Down
Loading

0 comments on commit 37a6da4

Please sign in to comment.