You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Using PipesDatabricksClient to run a pre-defined task (one with just a task key, e.g. jobs.SubmitTask(task_key="task_to_run")), the following error is occurring:
dagster_databricks/pipes.py", line 144, in run
**submit_task_dict["new_cluster"].get("spark_env_vars", {}),
What did you expect to happen?
The task to run, even without a provided cluster definition for a new cluster.
How to reproduce?
Create an asset or an op provided with a PipesDatabricksClient and a relevant context.
In general, the code looks like it's unable to run a pre-defined task (one that already exists on a Databricks cluster) – I wonder whether this is intentional. If so, not a great way to cooperate with Databricks-native deployment options.
However, even if the above is intentional for some reason, even if the PipesDatabricksClient expects a "fresh" task definition, it should be able to accept an existing cluster ID. The way the code is written in dagster_pipes, the run method of the client always expect the new_cluster key to be present.
Message from the maintainers
Impacted by this issue? Give it a 👍! We factor engagement into prioritization.
The text was updated successfully, but these errors were encountered:
garethbrickman
changed the title
PipesDataBricksClient not accepting a task definition without a provided
PipesDataBricksClient not accepting a task definition without a provided existing cluster ID
May 23, 2024
garethbrickman
changed the title
PipesDataBricksClient not accepting a task definition without a provided existing cluster ID
PipesDataBricksClient not accepting a task definition with an existing cluster ID
May 23, 2024
Dagster version
1.7.6
What's the issue?
Using
PipesDatabricksClient
to run a pre-defined task (one with just a task key, e.g.jobs.SubmitTask(task_key="task_to_run")
), the following error is occurring:What did you expect to happen?
The task to run, even without a provided cluster definition for a new cluster.
How to reproduce?
PipesDatabricksClient
and a relevantcontext
.Deployment type
None
Deployment details
No response
Additional information
PipesDatabricksClient
expects a "fresh" task definition, it should be able to accept an existing cluster ID. The way the code is written indagster_pipes
, therun
method of the client always expect thenew_cluster
key to be present.Message from the maintainers
Impacted by this issue? Give it a 👍! We factor engagement into prioritization.
The text was updated successfully, but these errors were encountered: