Skip to content
This repository has been archived by the owner on Jun 13, 2023. It is now read-only.

Latest commit

 

History

History
37 lines (29 loc) · 5.96 KB

JobsRunNowRequest.md

File metadata and controls

37 lines (29 loc) · 5.96 KB

JobsRunNowRequest

Properties

Name Type Description Notes
job_id int The ID of the job to be executed [optional]
idempotency_token str An optional token to guarantee the idempotency of job run requests. If a run with the provided token already exists, the request does not create a new run but returns the ID of the existing run instead. If a run with the provided token is deleted, an error is returned. If you specify the idempotency token, upon failure you can retry until the request succeeds. Databricks guarantees that exactly one run is launched with that idempotency token. This token must have at most 64 characters. For more information, see How to ensure idempotency for jobs. [optional]
jar_params List[str] A list of parameters for jobs with Spark JAR tasks, for example `"jar_params": ["john doe", "35"]`. The parameters are used to invoke the main function of the main class specified in the Spark JAR task. If not specified upon `run-now`, it defaults to an empty list. jar_params cannot be specified in conjunction with notebook_params. The JSON representation of this field (for example `{"jar_params":["john doe","35"]}`) cannot exceed 10,000 bytes. Use Task parameter variables to set parameters containing information about job runs. [optional]
notebook_params Dict[str, object] A map from keys to values for jobs with notebook task, for example `"notebook_params": {"name": "john doe", "age": "35"}`. The map is passed to the notebook and is accessible through the dbutils.widgets.get function. If not specified upon `run-now`, the triggered run uses the job’s base parameters. notebook_params cannot be specified in conjunction with jar_params. Use Task parameter variables to set parameters containing information about job runs. The JSON representation of this field (for example `{"notebook_params":{"name":"john doe","age":"35"}}`) cannot exceed 10,000 bytes. [optional]
python_params List[str] A list of parameters for jobs with Python tasks, for example `"python_params": ["john doe", "35"]`. The parameters are passed to Python file as command-line parameters. If specified upon `run-now`, it would overwrite the parameters specified in job setting. The JSON representation of this field (for example `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. Use Task parameter variables to set parameters containing information about job runs. Important These parameters accept only Latin characters (ASCII character set). Using non-ASCII characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese kanjis, and emojis. [optional]
spark_submit_params List[str] A list of parameters for jobs with spark submit task, for example `"spark_submit_params": ["--class", "org.apache.spark.examples.SparkPi"]`. The parameters are passed to spark-submit script as command-line parameters. If specified upon `run-now`, it would overwrite the parameters specified in job setting. The JSON representation of this field (for example `{"python_params":["john doe","35"]}`) cannot exceed 10,000 bytes. Use Task parameter variables to set parameters containing information about job runs. Important These parameters accept only Latin characters (ASCII character set). Using non-ASCII characters returns an error. Examples of invalid, non-ASCII characters are Chinese, Japanese kanjis, and emojis. [optional]
python_named_params object A map from keys to values for jobs with Python wheel task, for example `"python_named_params": {"name": "task", "data": "dbfs:/path/to/data.json"}`. [optional]
pipeline_params RunParametersPipelineParams [optional]
sql_params object A map from keys to values for SQL tasks, for example `"sql_params": {"name": "john doe", "age": "35"}`. The SQL alert task does not support custom parameters. [optional]
dbt_commands List[str] An array of commands to execute for jobs with the dbt task, for example `"dbt_commands": ["dbt deps", "dbt seed", "dbt run"]` [optional]

Example

from databricks_jobs.models.jobs_run_now_request import JobsRunNowRequest

# TODO update the JSON string below
json = "{}"
# create an instance of JobsRunNowRequest from a JSON string
jobs_run_now_request_instance = JobsRunNowRequest.from_json(json)
# print the JSON string representation of the object
print JobsRunNowRequest.to_json()

# convert the object into a dict
jobs_run_now_request_dict = jobs_run_now_request_instance.to_dict()
# create an instance of JobsRunNowRequest from a dict
jobs_run_now_request_form_dict = jobs_run_now_request.from_dict(jobs_run_now_request_dict)

[Back to Model list] [Back to API list] [Back to README]