/
api_op_CreateBatchInferenceJob.go
163 lines (141 loc) · 4.96 KB
/
api_op_CreateBatchInferenceJob.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
// Code generated by smithy-go-codegen DO NOT EDIT.
package personalize
import (
"context"
awsmiddleware "github.com/aws/aws-sdk-go-v2/aws/middleware"
"github.com/aws/aws-sdk-go-v2/aws/signer/v4"
"github.com/aws/aws-sdk-go-v2/service/personalize/types"
"github.com/aws/smithy-go/middleware"
smithyhttp "github.com/aws/smithy-go/transport/http"
)
// Creates a batch inference job. The operation can handle up to 50 million records
// and the input file must be in JSON format. For more information, see Creating a
// batch inference job
// (https://docs.aws.amazon.com/personalize/latest/dg/creating-batch-inference-job.html).
func (c *Client) CreateBatchInferenceJob(ctx context.Context, params *CreateBatchInferenceJobInput, optFns ...func(*Options)) (*CreateBatchInferenceJobOutput, error) {
if params == nil {
params = &CreateBatchInferenceJobInput{}
}
result, metadata, err := c.invokeOperation(ctx, "CreateBatchInferenceJob", params, optFns, c.addOperationCreateBatchInferenceJobMiddlewares)
if err != nil {
return nil, err
}
out := result.(*CreateBatchInferenceJobOutput)
out.ResultMetadata = metadata
return out, nil
}
type CreateBatchInferenceJobInput struct {
// The Amazon S3 path that leads to the input file to base your recommendations on.
// The input material must be in JSON format.
//
// This member is required.
JobInput *types.BatchInferenceJobInput
// The name of the batch inference job to create.
//
// This member is required.
JobName *string
// The path to the Amazon S3 bucket where the job's output will be stored.
//
// This member is required.
JobOutput *types.BatchInferenceJobOutput
// The ARN of the Amazon Identity and Access Management role that has permissions
// to read and write to your input and output Amazon S3 buckets respectively.
//
// This member is required.
RoleArn *string
// The Amazon Resource Name (ARN) of the solution version that will be used to
// generate the batch inference recommendations.
//
// This member is required.
SolutionVersionArn *string
// The configuration details of a batch inference job.
BatchInferenceJobConfig *types.BatchInferenceJobConfig
// The ARN of the filter to apply to the batch inference job. For more information
// on using filters, see Filtering batch recommendations
// (https://docs.aws.amazon.com/personalize/latest/dg/filter-batch.html).
FilterArn *string
// The number of recommendations to retrieve.
NumResults *int32
// A list of tags
// (https://docs.aws.amazon.com/personalize/latest/dev/tagging-resources.html) to
// apply to the batch inference job.
Tags []types.Tag
noSmithyDocumentSerde
}
type CreateBatchInferenceJobOutput struct {
// The ARN of the batch inference job.
BatchInferenceJobArn *string
// Metadata pertaining to the operation's result.
ResultMetadata middleware.Metadata
noSmithyDocumentSerde
}
func (c *Client) addOperationCreateBatchInferenceJobMiddlewares(stack *middleware.Stack, options Options) (err error) {
err = stack.Serialize.Add(&awsAwsjson11_serializeOpCreateBatchInferenceJob{}, middleware.After)
if err != nil {
return err
}
err = stack.Deserialize.Add(&awsAwsjson11_deserializeOpCreateBatchInferenceJob{}, middleware.After)
if err != nil {
return err
}
if err = addSetLoggerMiddleware(stack, options); err != nil {
return err
}
if err = awsmiddleware.AddClientRequestIDMiddleware(stack); err != nil {
return err
}
if err = smithyhttp.AddComputeContentLengthMiddleware(stack); err != nil {
return err
}
if err = addResolveEndpointMiddleware(stack, options); err != nil {
return err
}
if err = v4.AddComputePayloadSHA256Middleware(stack); err != nil {
return err
}
if err = addRetryMiddlewares(stack, options); err != nil {
return err
}
if err = addHTTPSignerV4Middleware(stack, options); err != nil {
return err
}
if err = awsmiddleware.AddRawResponseToMetadata(stack); err != nil {
return err
}
if err = awsmiddleware.AddRecordResponseTiming(stack); err != nil {
return err
}
if err = addClientUserAgent(stack); err != nil {
return err
}
if err = smithyhttp.AddErrorCloseResponseBodyMiddleware(stack); err != nil {
return err
}
if err = smithyhttp.AddCloseResponseBodyMiddleware(stack); err != nil {
return err
}
if err = addOpCreateBatchInferenceJobValidationMiddleware(stack); err != nil {
return err
}
if err = stack.Initialize.Add(newServiceMetadataMiddleware_opCreateBatchInferenceJob(options.Region), middleware.Before); err != nil {
return err
}
if err = addRequestIDRetrieverMiddleware(stack); err != nil {
return err
}
if err = addResponseErrorMiddleware(stack); err != nil {
return err
}
if err = addRequestResponseLogging(stack, options); err != nil {
return err
}
return nil
}
func newServiceMetadataMiddleware_opCreateBatchInferenceJob(region string) *awsmiddleware.RegisterServiceMetadata {
return &awsmiddleware.RegisterServiceMetadata{
Region: region,
ServiceID: ServiceID,
SigningName: "personalize",
OperationName: "CreateBatchInferenceJob",
}
}