You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
There might be several mathematical ways to define such a transform, in particular if when passing a X_fake_transformed that contain real numbers that do not actually result from a spline expansion. For instance when:
or when all values of a given row are non-zeros at once...
One possible way would be to decode based on X_fake_transformed.argmax(axis=1) and then using the relative strength of neighboring spline activations to resolve ambiguities.
Describe alternatives you've considered, if relevant
The main alternative is to not implement this. The main question is probably why try to implement this in the first place?
Possible use cases:
fit a GMM model on spline transformed data (to get a more axis-aligned inductive prior), generate samples in the GMM latent space and then recode those samples back into the original space.
fit PCA with a small rank on spline encoded data and then reconstruct back the projected data,
fit k-means in spline space and recode the learned centroids back in the original feature space for inspection.
idem for NMF or dictionary learning components.
Additional context
If #28043 gets merged, missing values support should also be included when using the 'indicator' strategy.
The text was updated successfully, but these errors were encountered:
Describe the workflow you want to enable
I think it should be possible to implement a new method
inverse_transform
such that:Describe your proposed solution
There might be several mathematical ways to define such a transform, in particular if when passing a
X_fake_transformed
that contain real numbers that do not actually result from a spline expansion. For instance when:(X_fake_transformed < 0).any()
(X_fake_transformed > 1).any()
X_fake_transformed.sum(axis=1) != np.ones(n_samples)
.or when all values of a given row are non-zeros at once...
One possible way would be to decode based on
X_fake_transformed.argmax(axis=1)
and then using the relative strength of neighboring spline activations to resolve ambiguities.Describe alternatives you've considered, if relevant
The main alternative is to not implement this. The main question is probably why try to implement this in the first place?
Possible use cases:
fit a GMM model on spline transformed data (to get a more axis-aligned inductive prior), generate samples in the GMM latent space and then recode those samples back into the original space.
fit PCA with a small rank on spline encoded data and then reconstruct back the projected data,
fit k-means in spline space and recode the learned centroids back in the original feature space for inspection.
idem for NMF or dictionary learning components.
Additional context
If #28043 gets merged, missing values support should also be included when using the 'indicator' strategy.
The text was updated successfully, but these errors were encountered: