-
Notifications
You must be signed in to change notification settings - Fork 7
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Should the pipeline interface include path to sample yaml outputs? #61
Comments
I'm not sure I understand. Could you give more context here? is the goal here to set the future location of sample.yaml file in a pipeline interface and then use this path to save the file there instead of in a default spot? |
well, this is a few years old... but yes I believe your interpretation is correct. I believe this is sort of accomplished by the output schema concept... |
more by the input schema, I think. |
so what's the key for the sample yaml path in the pipeline interface, if we even want to proceed?
|
Well, there can be an input sample.yaml, which is in some sense an instance of the object specified by the input schema, and an output sample.yaml which is in some sense an instance of the object specified by the output schema. we could produce both yamls. right now we only produce the first.
those are relative to the path specified in just brainstorming here... |
I think it all makes sense. So, in practice: |
makes sense to me... I think it's a superset of the input yaml. the only reason we make the input yaml is because it's used as an input to the pipeline. In fact... why even make the input yaml? if the output yaml is a superset of it, then it could be used as an input to the pipeline as well... so now, with this model, there is only 1 sample.yaml, which is exactly what you say: input yaml + populated sample attrs defined in the output schema. one question: would this yaml include a property given in the table table that is not specified in the input schema? |
that's right, input sample.yaml is probably not necessary in such a case.
I'd say yes, I can imagine writing a small schema (for example with just required attrs), that does not necessarily cover all the sample attributes. Still I'd expect all my columns from sample table to be accessible in the yaml file. |
I agree. |
for future reference: sample yaml file path is constructed from a template in pipeline interface file pipelines:
pipeline:
sample_yaml_path: >
{sample.sample_name}.yaml # relative to looper.output_dir if |
In #32 we build a pipeline interface section called
summary_results
, which records the location of summarizer results.What about something similar to report the location of sample yaml file from the pipeline?
The text was updated successfully, but these errors were encountered: