You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am new to DVC and I have a question about an use case which I could not find an example of.
The project I am working on requires the same experiment (code) to be run for multiple scenarios (for example, 4 datasets). Each scenario requires separate training and evaluation runs, but all of them use the same stages (except for input files) and output the same files and metrics.
I'd like to have snapshot of the metrics for all scenarios when code changes (but sometimes only for some scenarios, I don't want to run all 4 experiments if a change produces bad results for one experiment).
Does DVC currently support this workflow? How should I organize my pipeline(s)/metrics and execute my experiments?
Is it the right way to build separate pipelines, one for each scenario?
Thanks in advance,
The text was updated successfully, but these errors were encountered:
Build separate pipelines using the same code. You could do that quite conveniently by using separate directories for your experiements and just copying dvcfiles over.
In both cases, metrics comparison will work pretty much the same.
Another option for your scenario would be to implement "best experiment selection" in your own code, so that it goes through the list of input datas and selects the best one.
There is also another feature that we are considering implementing build matrix #1018 . If you are familiar with travis ci, it is a pretty similar thing. Please take a look to see if you find it helpful for your scenario and don't hesitate to leave a comment there 🙂
I am new to DVC and I have a question about an use case which I could not find an example of.
The project I am working on requires the same experiment (code) to be run for multiple scenarios (for example, 4 datasets). Each scenario requires separate training and evaluation runs, but all of them use the same stages (except for input files) and output the same files and metrics.
I'd like to have snapshot of the metrics for all scenarios when code changes (but sometimes only for some scenarios, I don't want to run all 4 experiments if a change produces bad results for one experiment).
Does DVC currently support this workflow? How should I organize my pipeline(s)/metrics and execute my experiments?
Is it the right way to build separate pipelines, one for each scenario?
Thanks in advance,
The text was updated successfully, but these errors were encountered: