-
Hey there, Thanks in advance! |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 2 replies
-
Hi, The API support an option to have one The Spark agent doesn't do this right now, but the API is prepared for that. |
Beta Was this translation helpful? Give feedback.
-
That's exactly the point - execution plan represents the static model of the pipeline, sort of a math formula. You can see it as a version of the pipeline. It's only when you execute it, some data flows through it, and the lineage of that data is tracked. There is no logical value in recording the same execution plan over and over again, it only creates unnecessary clutter in the database and make the cumulative lineage hard to analyze. It gets especially perceivable when using "append" writes, in which case the full lineage of the content of the given file would be presented as a union of all past append writes following the last overwrite. If every write comes with its own exec plan, the resulted graph would be a mess. Unfortunately, at the moment the agent cannot reuse / de-duplicate exec plans automatically. We are aware of this limitation and will likely be introducing some label based version mechanism to overcome this problem at least for cases when the structure of the data pipelines doesn't change frequently. |
Beta Was this translation helpful? Give feedback.
That's exactly the point - execution plan represents the static model of the pipeline, sort of a math formula. You can see it as a version of the pipeline. It's only when you execute it, some data flows through it, and the lineage of that data is tracked. There is no logical value in recording the same execution plan over and over again, it only creates unnecessary clutter in the database and make the cumulative lineage hard to analyze. It gets especially perceivable when using "append" writes, in which case the full lineage of the content of the given fi…