Skip to content

Commit e6ffca9

Browse files
committed
Generated ts 2024-05-18 for dataworks-public.
1 parent 397688d commit e6ffca9

File tree

35 files changed

+96
-934
lines changed

35 files changed

+96
-934
lines changed

dataworks-public-20240518/package.json

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,6 @@
11
{
22
"name": "@alicloud/dataworks-public20240518",
3-
"version": "7.2.1",
3+
"version": "7.2.2",
44
"description": "",
55
"main": "dist/client.js",
66
"scripts": {

dataworks-public-20240518/src/client.ts

Lines changed: 95 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -594,16 +594,72 @@ export default class Client extends OpenApi {
594594
request.transformationRulesShrink = OpenApiUtil.arrayToStringWithSpecifiedStyle(tmpReq.transformationRules, "TransformationRules", "json");
595595
}
596596

597-
let query = OpenApiUtil.query(request.toMap());
597+
let body : {[key: string ]: any} = { };
598+
if (!$dara.isNull(request.description)) {
599+
body["Description"] = request.description;
600+
}
601+
602+
if (!$dara.isNull(request.destinationDataSourceSettingsShrink)) {
603+
body["DestinationDataSourceSettings"] = request.destinationDataSourceSettingsShrink;
604+
}
605+
606+
if (!$dara.isNull(request.destinationDataSourceType)) {
607+
body["DestinationDataSourceType"] = request.destinationDataSourceType;
608+
}
609+
610+
if (!$dara.isNull(request.jobName)) {
611+
body["JobName"] = request.jobName;
612+
}
613+
614+
if (!$dara.isNull(request.jobSettingsShrink)) {
615+
body["JobSettings"] = request.jobSettingsShrink;
616+
}
617+
618+
if (!$dara.isNull(request.jobType)) {
619+
body["JobType"] = request.jobType;
620+
}
621+
622+
if (!$dara.isNull(request.migrationType)) {
623+
body["MigrationType"] = request.migrationType;
624+
}
625+
626+
if (!$dara.isNull(request.name)) {
627+
body["Name"] = request.name;
628+
}
629+
630+
if (!$dara.isNull(request.projectId)) {
631+
body["ProjectId"] = request.projectId;
632+
}
633+
634+
if (!$dara.isNull(request.resourceSettingsShrink)) {
635+
body["ResourceSettings"] = request.resourceSettingsShrink;
636+
}
637+
638+
if (!$dara.isNull(request.sourceDataSourceSettingsShrink)) {
639+
body["SourceDataSourceSettings"] = request.sourceDataSourceSettingsShrink;
640+
}
641+
642+
if (!$dara.isNull(request.sourceDataSourceType)) {
643+
body["SourceDataSourceType"] = request.sourceDataSourceType;
644+
}
645+
646+
if (!$dara.isNull(request.tableMappingsShrink)) {
647+
body["TableMappings"] = request.tableMappingsShrink;
648+
}
649+
650+
if (!$dara.isNull(request.transformationRulesShrink)) {
651+
body["TransformationRules"] = request.transformationRulesShrink;
652+
}
653+
598654
let req = new $OpenApiUtil.OpenApiRequest({
599-
query: OpenApiUtil.query(query),
655+
body: OpenApiUtil.parseToMap(body),
600656
});
601657
let params = new $OpenApiUtil.Params({
602658
action: "CreateDIJob",
603659
version: "2024-05-18",
604660
protocol: "HTTPS",
605661
pathname: "/",
606-
method: "GET",
662+
method: "POST",
607663
authType: "AK",
608664
style: "RPC",
609665
reqBodyType: "formData",
@@ -10888,16 +10944,50 @@ export default class Client extends OpenApi {
1088810944
request.transformationRulesShrink = OpenApiUtil.arrayToStringWithSpecifiedStyle(tmpReq.transformationRules, "TransformationRules", "json");
1088910945
}
1089010946

10891-
let query = OpenApiUtil.query(request.toMap());
10947+
let query = { };
10948+
if (!$dara.isNull(request.id)) {
10949+
query["Id"] = request.id;
10950+
}
10951+
10952+
let body : {[key: string ]: any} = { };
10953+
if (!$dara.isNull(request.DIJobId)) {
10954+
body["DIJobId"] = request.DIJobId;
10955+
}
10956+
10957+
if (!$dara.isNull(request.description)) {
10958+
body["Description"] = request.description;
10959+
}
10960+
10961+
if (!$dara.isNull(request.jobSettingsShrink)) {
10962+
body["JobSettings"] = request.jobSettingsShrink;
10963+
}
10964+
10965+
if (!$dara.isNull(request.projectId)) {
10966+
body["ProjectId"] = request.projectId;
10967+
}
10968+
10969+
if (!$dara.isNull(request.resourceSettingsShrink)) {
10970+
body["ResourceSettings"] = request.resourceSettingsShrink;
10971+
}
10972+
10973+
if (!$dara.isNull(request.tableMappingsShrink)) {
10974+
body["TableMappings"] = request.tableMappingsShrink;
10975+
}
10976+
10977+
if (!$dara.isNull(request.transformationRulesShrink)) {
10978+
body["TransformationRules"] = request.transformationRulesShrink;
10979+
}
10980+
1089210981
let req = new $OpenApiUtil.OpenApiRequest({
1089310982
query: OpenApiUtil.query(query),
10983+
body: OpenApiUtil.parseToMap(body),
1089410984
});
1089510985
let params = new $OpenApiUtil.Params({
1089610986
action: "UpdateDIJob",
1089710987
version: "2024-05-18",
1089810988
protocol: "HTTPS",
1089910989
pathname: "/",
10900-
method: "GET",
10990+
method: "POST",
1090110991
authType: "AK",
1090210992
style: "RPC",
1090310993
reqBodyType: "formData",

dataworks-public-20240518/src/models/CreateDijobRequest.ts

Lines changed: 0 additions & 81 deletions
Original file line numberDiff line numberDiff line change
@@ -9,131 +9,50 @@ import { CreateDIJobRequestTransformationRules } from "./CreateDijobRequestTrans
99

1010

1111
export class CreateDIJobRequest extends $dara.Model {
12-
/**
13-
* @remarks
14-
* The description of the synchronization task.
15-
*
16-
* @example
17-
* The description of the synchronization task.
18-
*/
1912
description?: string;
2013
/**
2114
* @remarks
22-
* The settings of the destination. Only a single destination is supported.
23-
*
2415
* This parameter is required.
2516
*/
2617
destinationDataSourceSettings?: CreateDIJobRequestDestinationDataSourceSettings[];
2718
/**
2819
* @remarks
29-
* The destination type. Valid values: Hologres, OSS-HDFS, OSS, MaxCompute, LogHub, StarRocks, DataHub, AnalyticDB for MySQL, Kafka, and Hive.
30-
*
3120
* This parameter is required.
32-
*
33-
* @example
34-
* Hologres
3521
*/
3622
destinationDataSourceType?: string;
3723
/**
38-
* @remarks
39-
* This parameter is deprecated and is replaced by the Name parameter.
40-
*
41-
* @example
42-
* mysql_to_holo_sync_8772
43-
*
4424
* @deprecated
4525
*/
4626
jobName?: string;
47-
/**
48-
* @remarks
49-
* The settings for the dimension of the synchronization task. The settings include processing policies for DDL messages, policies for data type mappings between source fields and destination fields, and runtime parameters of the synchronization task.
50-
*/
5127
jobSettings?: CreateDIJobRequestJobSettings;
52-
/**
53-
* @remarks
54-
* The type of the task. This parameter is optional. Valid values:
55-
*
56-
* * DatabaseRealtimeMigration: A real-time synchronization task used to synchronize only full data, only incremental data, or full and incremental data in multiple tables of multiple databases at the source.
57-
* * DatabaseOfflineMigration: A batch synchronization task used to synchronize only full data, only incremental data, or full and incremental data in multiple tables of multiple databases at the source.
58-
* * SingleTableRealtimeMigration: A real-time synchronization task used to synchronize only data in single table at the source.
59-
*
60-
* @example
61-
* DatabaseRealtimeMigration
62-
*/
6328
jobType?: string;
6429
/**
6530
* @remarks
66-
* The synchronization type. Valid values:
67-
*
68-
* * FullAndRealtimeIncremental: full synchronization and real-time incremental synchronization of data in an entire database
69-
* * RealtimeIncremental: real-time incremental synchronization of data in a single table
70-
* * Full: full batch synchronization of data in an entire database
71-
* * OfflineIncremental: batch incremental synchronization of data in an entire database
72-
* * FullAndOfflineIncremental: full synchronization and batch incremental synchronization of data in an entire database
73-
*
7431
* This parameter is required.
75-
*
76-
* @example
77-
* FullAndRealtimeIncremental
7832
*/
7933
migrationType?: string;
80-
/**
81-
* @remarks
82-
* The name of the synchronization task.
83-
*
84-
* @example
85-
* mysql_to_holo_sync_8772
86-
*/
8734
name?: string;
88-
/**
89-
* @remarks
90-
* The DataWorks workspace ID. You can log on to the [DataWorks console](https://workbench.data.aliyun.com/console) and go to the Workspace page to query the ID.
91-
*
92-
* You must configure this parameter to specify the DataWorks workspace to which the API operation is applied.
93-
*
94-
* @example
95-
* 10000
96-
*/
9735
projectId?: number;
9836
/**
9937
* @remarks
100-
* The resource settings.
101-
*
10238
* This parameter is required.
10339
*/
10440
resourceSettings?: CreateDIJobRequestResourceSettings;
10541
/**
10642
* @remarks
107-
* The settings of the source. Only a single source is supported.
108-
*
10943
* This parameter is required.
11044
*/
11145
sourceDataSourceSettings?: CreateDIJobRequestSourceDataSourceSettings[];
11246
/**
11347
* @remarks
114-
* The source type. Valid values: PolarDB, MySQL, Kafka, LogHub, Hologres, Oracle, OceanBase, MongoDB, Redshift, Hive, SQL Server, Doris, and ClickHouse.
115-
*
11648
* This parameter is required.
117-
*
118-
* @example
119-
* MySQL
12049
*/
12150
sourceDataSourceType?: string;
12251
/**
12352
* @remarks
124-
* The list of mappings between rules used to select synchronization objects in the source and transformation rules applied to the selected synchronization objects. Each entry in the list displays a mapping between a rule used to select synchronization objects and a transformation rule applied to the selected synchronization objects.
125-
*
126-
* > [ { "SourceObjectSelectionRules":[ { "ObjectType":"Database", "Action":"Include", "ExpressionType":"Exact", "Expression":"biz_db" }, { "ObjectType":"Schema", "Action":"Include", "ExpressionType":"Exact", "Expression":"s1" }, { "ObjectType":"Table", "Action":"Include", "ExpressionType":"Exact", "Expression":"table1" } ], "TransformationRuleNames":[ { "RuleName":"my_database_rename_rule", "RuleActionType":"Rename", "RuleTargetType":"Schema" } ] } ]
127-
*
12853
* This parameter is required.
12954
*/
13055
tableMappings?: CreateDIJobRequestTableMappings[];
131-
/**
132-
* @remarks
133-
* The list of transformation rules for objects involved in the synchronization task.
134-
*
135-
* > [ { "RuleName":"my_database_rename_rule", "RuleActionType":"Rename", "RuleTargetType":"Schema", "RuleExpression":"{"expression":"${srcDatasoureName}_${srcDatabaseName}"}" } ]
136-
*/
13756
transformationRules?: CreateDIJobRequestTransformationRules[];
13857
static names(): { [key: string]: string } {
13958
return {

dataworks-public-20240518/src/models/CreateDijobRequestDestinationDataSourceSettings.ts

Lines changed: 0 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -3,13 +3,6 @@ import * as $dara from '@darabonba/typescript';
33

44

55
export class CreateDIJobRequestDestinationDataSourceSettings extends $dara.Model {
6-
/**
7-
* @remarks
8-
* The name of the data source.
9-
*
10-
* @example
11-
* holo_datasource_1
12-
*/
136
dataSourceName?: string;
147
static names(): { [key: string]: string } {
158
return {

dataworks-public-20240518/src/models/CreateDijobRequestJobSettings.ts

Lines changed: 0 additions & 41 deletions
Original file line numberDiff line numberDiff line change
@@ -7,51 +7,10 @@ import { CreateDIJobRequestJobSettingsRuntimeSettings } from "./CreateDijobReque
77

88

99
export class CreateDIJobRequestJobSettings extends $dara.Model {
10-
/**
11-
* @remarks
12-
* The channel control settings for the synchronization task. You can configure special channel control settings for the following synchronization links: data synchronization between Hologres data sources and data synchronization from Hologres to Kafka.
13-
*
14-
* 1. Holo2Kafka
15-
*
16-
* * Example: {"destinationChannelSettings":{"kafkaClientProperties":[{"key":"linger.ms","value":"100"}],"keyColumns":["col3"],"writeMode":"canal"}}
17-
* * kafkaClientProperties: the parameters related to a Kafka producer, which are used when you write data to a Kafka data source.
18-
* * keyColumns: the names of Kafka columns to which you want to write data.
19-
* * writeMode: the writing format of the Kafka data source. Valid values: json and canal.
20-
*
21-
* 2. Holo2Holo
22-
*
23-
* * Example: {"destinationChannelSettings":{"conflictMode":"replace","dynamicColumnAction":"replay","writeMode":"replay"}}
24-
* * conflictMode: the policy used to handle a conflict that occurs during data writing to Hologres. Valid values: replace and ignore.
25-
* * writeMode: the mode in which you want to write data to Hologres. Valid values: replay and insert.
26-
* * dynamicColumnAction: the mode in which you want to write data to dynamic columns in a Hologres table. Valid values: replay, insert, and ignore.
27-
*
28-
* @example
29-
* {"structInfo":"MANAGED","storageType":"TEXTFILE","writeMode":"APPEND","partitionColumns":[{"columnName":"pt","columnType":"STRING","comment":""}],"fieldDelimiter":""}
30-
*/
3110
channelSettings?: string;
32-
/**
33-
* @remarks
34-
* The data type mappings between source fields and destination fields.
35-
*
36-
* > "ColumnDataTypeSettings":[ { "SourceDataType":"Bigint", "DestinationDataType":"Text" } ]
37-
*/
3811
columnDataTypeSettings?: CreateDIJobRequestJobSettingsColumnDataTypeSettings[];
39-
/**
40-
* @remarks
41-
* The settings for periodic scheduling.
42-
*/
4312
cycleScheduleSettings?: CreateDIJobRequestJobSettingsCycleScheduleSettings;
44-
/**
45-
* @remarks
46-
* The processing settings for DDL messages.
47-
*
48-
* > "DDLHandlingSettings":[ { "Type":"Insert", "Action":"Normal" } ]
49-
*/
5013
ddlHandlingSettings?: CreateDIJobRequestJobSettingsDdlHandlingSettings[];
51-
/**
52-
* @remarks
53-
* The runtime settings.
54-
*/
5514
runtimeSettings?: CreateDIJobRequestJobSettingsRuntimeSettings[];
5615
static names(): { [key: string]: string } {
5716
return {

dataworks-public-20240518/src/models/CreateDijobRequestJobSettingsColumnDataTypeSettings.ts

Lines changed: 0 additions & 14 deletions
Original file line numberDiff line numberDiff line change
@@ -3,21 +3,7 @@ import * as $dara from '@darabonba/typescript';
33

44

55
export class CreateDIJobRequestJobSettingsColumnDataTypeSettings extends $dara.Model {
6-
/**
7-
* @remarks
8-
* The data type of the destination field. Valid values: bigint, boolean, string, text, datetime, timestamp, decimal, and binary. Different types of data sources support different data types.
9-
*
10-
* @example
11-
* text
12-
*/
136
destinationDataType?: string;
14-
/**
15-
* @remarks
16-
* The data type of the source field. Valid values: Valid values: bigint, boolean, string, text, datetime, timestamp, decimal, and binary. Different types of data sources support different data types.
17-
*
18-
* @example
19-
* bigint
20-
*/
217
sourceDataType?: string;
228
static names(): { [key: string]: string } {
239
return {

dataworks-public-20240518/src/models/CreateDijobRequestJobSettingsCycleScheduleSettings.ts

Lines changed: 0 additions & 17 deletions
Original file line numberDiff line numberDiff line change
@@ -3,24 +3,7 @@ import * as $dara from '@darabonba/typescript';
33

44

55
export class CreateDIJobRequestJobSettingsCycleScheduleSettings extends $dara.Model {
6-
/**
7-
* @remarks
8-
* The synchronization type that requires periodic scheduling. Valid values:
9-
*
10-
* * Full: full synchronization
11-
* * OfflineIncremental: batch incremental synchronization
12-
*
13-
* @example
14-
* Full
15-
*/
166
cycleMigrationType?: string;
17-
/**
18-
* @remarks
19-
* The scheduling parameters.
20-
*
21-
* @example
22-
* bizdate=$bizdate
23-
*/
247
scheduleParameters?: string;
258
static names(): { [key: string]: string } {
269
return {

0 commit comments

Comments
 (0)