Skip to content
This repository has been archived by the owner on Mar 12, 2024. It is now read-only.

Auto-ranking of most explicative features #228

Open
danthe3rd opened this issue Feb 3, 2022 · 0 comments
Open

Auto-ranking of most explicative features #228

danthe3rd opened this issue Feb 3, 2022 · 0 comments
Labels
enhancement New feature or request

Comments

@danthe3rd
Copy link
Contributor

danthe3rd commented Feb 3, 2022

Scenario:
I have a grid-search on parameters A, B and C.
For each sample, I have an associated loss which I try to minimize.

I want to know which parameter (A, B or C) has the most influence on the loss automatically.

In python: This can be done by learning a simple RandomForestRegressor (or Classifier depending on the target value type), and then calling permutation_importance to get an importance score for each parameter.
For this to be embedded in HiPlot, it would need to be done in JS (for example with this library?)

https://scikit-learn.org/stable/modules/generated/sklearn.ensemble.RandomForestRegressor.html
https://scikit-learn.org/stable/modules/permutation_importance.html

UI: This could be triggered by right-clicking a column. The result could be displayed by ordering the column by relative importance.
Need a way to select which columns to include/exclude from the calculation,
and to display the correlation score

@danthe3rd danthe3rd added the enhancement New feature or request label Feb 3, 2022
Sign up for free to subscribe to this conversation on GitHub. Already have an account? Sign in.
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant