Skip to content
Closed
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
10 changes: 7 additions & 3 deletions databricks_cli/sdk/service.py
Original file line number Diff line number Diff line change
Expand Up @@ -212,12 +212,13 @@ def list_clusters(self, headers=None):

return self.client.perform_query('GET', '/clusters/list', data=_data, headers=headers)

# init_scripts argument was removed by autogeneration and manually added here.
def create_cluster(self, num_workers=None, autoscale=None, cluster_name=None, spark_version=None,
spark_conf=None, aws_attributes=None, node_type_id=None,
driver_node_type_id=None, ssh_public_keys=None, custom_tags=None,
cluster_log_conf=None, spark_env_vars=None, autotermination_minutes=None,
enable_elastic_disk=None, cluster_source=None, instance_pool_id=None,
headers=None):
cluster_log_conf=None, init_scripts=None, spark_env_vars=None,
autotermination_minutes=None, enable_elastic_disk=None, cluster_source=None,
instance_pool_id=None, headers=None):
_data = {}
if num_workers is not None:
_data['num_workers'] = num_workers
Expand Down Expand Up @@ -247,6 +248,9 @@ def create_cluster(self, num_workers=None, autoscale=None, cluster_name=None, sp
_data['cluster_log_conf'] = cluster_log_conf
if not isinstance(cluster_log_conf, dict):
raise TypeError('Expected databricks.ClusterLogConf() or dict for field cluster_log_conf')
# init_scripts handling was removed by autogeneration and manually added here.
if init_scripts is not None:
_data['init_scripts'] = init_scripts
if spark_env_vars is not None:
_data['spark_env_vars'] = spark_env_vars
if autotermination_minutes is not None:
Expand Down