New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How to build transform-only pipelines #642

Closed
aho81 opened this Issue Aug 3, 2018 · 2 comments

Comments

Projects
None yet
2 participants
@aho81

aho81 commented Aug 3, 2018

I found comments, that there should be a way to use transforms without the need of a trainer/learner (i.e., building a "processing / transform - pipeline instead of an LearningPipeline, cp. #259 (comment)). Unfourtunately, I could not find out, how to achieve this.

In my usecase, I want to determine similarity of documents with n-gram vectorization and cosine distance. The functionalty for featurization is given by the TextFeaturizer (https://docs.microsoft.com/de-de/dotnet/api/microsoft.ml.transforms.textfeaturizer). In this usecase I don't want to do a training (yet), but am interessted in the output in the result of the TextFeaturizer itself.

Accessing the results of partial steps could be helpful for debugging LearningPipelines too (cp. discussion here: #259).
@TomFinley

@Zruty0

This comment has been minimized.

Show comment
Hide comment
@Zruty0

Zruty0 Aug 5, 2018

Member

As far as I know, this is currently not possible.

However, once we build the final API (see #583), you will be able to access the output of transforms without having to train a model.

In fact, the 'trained model' will be just one form of 'transformer', and you will be able to have as many of them chained together as you want (including 0), and mix and match them with other transformers, like TextFeaturizer.

Member

Zruty0 commented Aug 5, 2018

As far as I know, this is currently not possible.

However, once we build the final API (see #583), you will be able to access the output of transforms without having to train a model.

In fact, the 'trained model' will be just one form of 'transformer', and you will be able to have as many of them chained together as you want (including 0), and mix and match them with other transformers, like TextFeaturizer.

@aho81 aho81 closed this Aug 6, 2018

@aho81

This comment has been minimized.

Show comment
Hide comment
@aho81

aho81 Aug 6, 2018

Thanks! I'm looking forward to the final API.

aho81 commented Aug 6, 2018

Thanks! I'm looking forward to the final API.

@aho81 aho81 reopened this Aug 6, 2018

@aho81 aho81 closed this Aug 6, 2018

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment