Skip to content

Latest commit

 

History

History
35 lines (22 loc) · 1.54 KB

a-pipeline-i-love.md

File metadata and controls

35 lines (22 loc) · 1.54 KB
layout title comments lang permalink
page
What I would love the pipeline to look like
false
en
/a-pipeline-i-love/

The ideal pipeline for me is one that minimizes the time I spend in 2 and 3 this cycle

  1. code a feature/build a new model based on a new idea
  2. test the feature/model
  3. record the metrics and compare to current model with current features

Say, if there's a model with a bunch of features, and I have a metric that I want to optimize. There are also other supporting metrics that I'm interested in.

step1

Then I write some code to create some new features to be tested.

step2

After I created the feature with some code, this process of "creating feature" is fed into the bot and the bot will test this feature by comparing scores of the possible models --- not just the model which is my baseline but a set of potential models --- with and without this new feature.

step3

The score is then compared to current baseline.

step4

If the score bump is better than a predefined number, then I know I want to keep it and investigate more, so the bot will keep it in there for me along with the metric logs.

step5

The part where I spend the most time on should be the part where I think about how to make new features or use the features in a different way to spin up a model, not the actual process of testing this feature against current model.