-
Notifications
You must be signed in to change notification settings - Fork 240
/
azure.synapse.spark.models.SparkBatchJobOptions.yml
227 lines (217 loc) · 6.68 KB
/
azure.synapse.spark.models.SparkBatchJobOptions.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
### YamlMime:PythonClass
uid: azure.synapse.spark.models.SparkBatchJobOptions
name: SparkBatchJobOptions
fullName: azure.synapse.spark.models.SparkBatchJobOptions
module: azure.synapse.spark.models
inheritances:
- msrest.serialization.Model
summary: 'SparkBatchJobOptions.
All required parameters must be populated in order to send to Azure.'
constructor:
syntax: 'SparkBatchJobOptions(*, name: str, file: str, tags: Dict[str, str] | None
= None, artifact_id: str | None = None, class_name: str | None = None, arguments:
List[str] | None = None, jars: List[str] | None = None, python_files: List[str]
| None = None, files: List[str] | None = None, archives: List[str] | None = None,
configuration: Dict[str, str] | None = None, driver_memory: str | None = None,
driver_cores: int | None = None, executor_memory: str | None = None, executor_cores:
int | None = None, executor_count: int | None = None, **kwargs)'
parameters:
- name: tags
description: A set of tags. Dictionary of `<string>`.
isRequired: true
types:
- <xref:dict>[<xref:str>, <xref:str>]
- name: artifact_id
isRequired: true
types:
- <xref:str>
- name: name
description: Required.
isRequired: true
types:
- <xref:str>
- name: file
description: Required.
isRequired: true
types:
- <xref:str>
- name: class_name
isRequired: true
types:
- <xref:str>
- name: arguments
isRequired: true
types:
- <xref:list>[<xref:str>]
- name: jars
isRequired: true
types:
- <xref:list>[<xref:str>]
- name: python_files
isRequired: true
types:
- <xref:list>[<xref:str>]
- name: files
isRequired: true
types:
- <xref:list>[<xref:str>]
- name: archives
isRequired: true
types:
- <xref:list>[<xref:str>]
- name: configuration
description: Dictionary of `<string>`.
isRequired: true
types:
- <xref:dict>[<xref:str>, <xref:str>]
- name: driver_memory
isRequired: true
types:
- <xref:str>
- name: driver_cores
isRequired: true
types:
- <xref:int>
- name: executor_memory
isRequired: true
types:
- <xref:str>
- name: executor_cores
isRequired: true
types:
- <xref:int>
- name: executor_count
isRequired: true
types:
- <xref:int>
keywordOnlyParameters:
- name: name
isRequired: true
- name: file
isRequired: true
- name: tags
isRequired: true
- name: artifact_id
isRequired: true
- name: class_name
isRequired: true
- name: arguments
isRequired: true
- name: jars
isRequired: true
- name: python_files
isRequired: true
- name: files
isRequired: true
- name: archives
isRequired: true
- name: configuration
isRequired: true
- name: driver_memory
isRequired: true
- name: driver_cores
isRequired: true
- name: executor_memory
isRequired: true
- name: executor_cores
isRequired: true
- name: executor_count
isRequired: true
methods:
- uid: azure.synapse.spark.models.SparkBatchJobOptions.as_dict
name: as_dict
summary: "Return a dict that can be JSONify using json.dump.\n\nAdvanced usage might\
\ optionally use a callback as parameter:\n\nKey is the attribute name used in\
\ Python. Attr_desc\nis a dict of metadata. Currently contains 'type' with the\n\
msrest type and 'key' with the RestAPI encoded key.\nValue is the current value\
\ in this object.\n\nThe string returned will be used to serialize the key.\n\
If the return type is a list, this is considered hierarchical\nresult dict.\n\n\
See the three examples in this file:\n\n* attribute_transformer \n\n* full_restapi_key_transformer\
\ \n\n* last_restapi_key_transformer \n\nIf you want XML serialization, you can\
\ pass the kwargs is_xml=True."
signature: as_dict(keep_readonly=True, key_transformer=<function attribute_transformer>,
**kwargs)
parameters:
- name: key_transformer
description: A key transformer function.
types:
- <xref:function>
- name: keep_readonly
defaultValue: 'True'
return:
description: A dict JSON compatible object
types:
- <xref:dict>
- uid: azure.synapse.spark.models.SparkBatchJobOptions.deserialize
name: deserialize
summary: Parse a str using the RestAPI syntax and return a model.
signature: deserialize(data, content_type=None)
parameters:
- name: data
description: A str using RestAPI structure. JSON by default.
isRequired: true
types:
- <xref:str>
- name: content_type
description: JSON by default, set application/xml if XML.
defaultValue: None
types:
- <xref:str>
return:
description: An instance of this model
exceptions:
- type: DeserializationError if something went wrong
- uid: azure.synapse.spark.models.SparkBatchJobOptions.enable_additional_properties_sending
name: enable_additional_properties_sending
signature: enable_additional_properties_sending()
- uid: azure.synapse.spark.models.SparkBatchJobOptions.from_dict
name: from_dict
summary: 'Parse a dict using given key extractor return a model.
By default consider key
extractors (rest_key_case_insensitive_extractor, attribute_key_case_insensitive_extractor
and last_rest_key_case_insensitive_extractor)'
signature: from_dict(data, key_extractors=None, content_type=None)
parameters:
- name: data
description: A dict using RestAPI structure
isRequired: true
types:
- <xref:dict>
- name: content_type
description: JSON by default, set application/xml if XML.
defaultValue: None
types:
- <xref:str>
- name: key_extractors
defaultValue: None
return:
description: An instance of this model
exceptions:
- type: DeserializationError if something went wrong
- uid: azure.synapse.spark.models.SparkBatchJobOptions.is_xml_model
name: is_xml_model
signature: is_xml_model()
- uid: azure.synapse.spark.models.SparkBatchJobOptions.serialize
name: serialize
summary: 'Return the JSON that would be sent to azure from this model.
This is an alias to *as_dict(full_restapi_key_transformer, keep_readonly=False)*.
If you want XML serialization, you can pass the kwargs is_xml=True.'
signature: serialize(keep_readonly=False, **kwargs)
parameters:
- name: keep_readonly
description: If you want to serialize the readonly attributes
defaultValue: 'False'
types:
- <xref:bool>
return:
description: A dict JSON compatible object
types:
- <xref:dict>
- uid: azure.synapse.spark.models.SparkBatchJobOptions.validate
name: validate
summary: Validate this model recursively and return a list of ValidationError.
signature: validate()
return:
description: A list of validation error
types:
- <xref:list>