/
api_op_StartFaceDetection.go
156 lines (138 loc) · 5.41 KB
/
api_op_StartFaceDetection.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
// Code generated by smithy-go-codegen DO NOT EDIT.
package rekognition
import (
"context"
awsmiddleware "github.com/aws/aws-sdk-go-v2/aws/middleware"
"github.com/aws/aws-sdk-go-v2/aws/signer/v4"
"github.com/aws/aws-sdk-go-v2/service/rekognition/types"
"github.com/aws/smithy-go/middleware"
smithyhttp "github.com/aws/smithy-go/transport/http"
)
// Starts asynchronous detection of faces in a stored video. Amazon Rekognition
// Video can detect faces in a video stored in an Amazon S3 bucket. Use Video to
// specify the bucket name and the filename of the video. StartFaceDetection
// returns a job identifier (JobId) that you use to get the results of the
// operation. When face detection is finished, Amazon Rekognition Video publishes a
// completion status to the Amazon Simple Notification Service topic that you
// specify in NotificationChannel. To get the results of the face detection
// operation, first check that the status value published to the Amazon SNS topic
// is SUCCEEDED. If so, call GetFaceDetection and pass the job identifier (JobId)
// from the initial call to StartFaceDetection. For more information, see Detecting
// Faces in a Stored Video in the Amazon Rekognition Developer Guide.
func (c *Client) StartFaceDetection(ctx context.Context, params *StartFaceDetectionInput, optFns ...func(*Options)) (*StartFaceDetectionOutput, error) {
if params == nil {
params = &StartFaceDetectionInput{}
}
result, metadata, err := c.invokeOperation(ctx, "StartFaceDetection", params, optFns, c.addOperationStartFaceDetectionMiddlewares)
if err != nil {
return nil, err
}
out := result.(*StartFaceDetectionOutput)
out.ResultMetadata = metadata
return out, nil
}
type StartFaceDetectionInput struct {
// The video in which you want to detect faces. The video must be stored in an
// Amazon S3 bucket.
//
// This member is required.
Video *types.Video
// Idempotent token used to identify the start request. If you use the same token
// with multiple StartFaceDetection requests, the same JobId is returned. Use
// ClientRequestToken to prevent the same job from being accidently started more
// than once.
ClientRequestToken *string
// The face attributes you want returned. DEFAULT - The following subset of facial
// attributes are returned: BoundingBox, Confidence, Pose, Quality and Landmarks.
// ALL - All facial attributes are returned.
FaceAttributes types.FaceAttributes
// An identifier you specify that's returned in the completion notification that's
// published to your Amazon Simple Notification Service topic. For example, you can
// use JobTag to group related jobs and identify them in the completion
// notification.
JobTag *string
// The ARN of the Amazon SNS topic to which you want Amazon Rekognition Video to
// publish the completion status of the face detection operation. The Amazon SNS
// topic must have a topic name that begins with AmazonRekognition if you are using
// the AmazonRekognitionServiceRole permissions policy.
NotificationChannel *types.NotificationChannel
noSmithyDocumentSerde
}
type StartFaceDetectionOutput struct {
// The identifier for the face detection job. Use JobId to identify the job in a
// subsequent call to GetFaceDetection.
JobId *string
// Metadata pertaining to the operation's result.
ResultMetadata middleware.Metadata
noSmithyDocumentSerde
}
func (c *Client) addOperationStartFaceDetectionMiddlewares(stack *middleware.Stack, options Options) (err error) {
err = stack.Serialize.Add(&awsAwsjson11_serializeOpStartFaceDetection{}, middleware.After)
if err != nil {
return err
}
err = stack.Deserialize.Add(&awsAwsjson11_deserializeOpStartFaceDetection{}, middleware.After)
if err != nil {
return err
}
if err = addSetLoggerMiddleware(stack, options); err != nil {
return err
}
if err = awsmiddleware.AddClientRequestIDMiddleware(stack); err != nil {
return err
}
if err = smithyhttp.AddComputeContentLengthMiddleware(stack); err != nil {
return err
}
if err = addResolveEndpointMiddleware(stack, options); err != nil {
return err
}
if err = v4.AddComputePayloadSHA256Middleware(stack); err != nil {
return err
}
if err = addRetryMiddlewares(stack, options); err != nil {
return err
}
if err = addHTTPSignerV4Middleware(stack, options); err != nil {
return err
}
if err = awsmiddleware.AddRawResponseToMetadata(stack); err != nil {
return err
}
if err = awsmiddleware.AddRecordResponseTiming(stack); err != nil {
return err
}
if err = addClientUserAgent(stack); err != nil {
return err
}
if err = smithyhttp.AddErrorCloseResponseBodyMiddleware(stack); err != nil {
return err
}
if err = smithyhttp.AddCloseResponseBodyMiddleware(stack); err != nil {
return err
}
if err = addOpStartFaceDetectionValidationMiddleware(stack); err != nil {
return err
}
if err = stack.Initialize.Add(newServiceMetadataMiddleware_opStartFaceDetection(options.Region), middleware.Before); err != nil {
return err
}
if err = addRequestIDRetrieverMiddleware(stack); err != nil {
return err
}
if err = addResponseErrorMiddleware(stack); err != nil {
return err
}
if err = addRequestResponseLogging(stack, options); err != nil {
return err
}
return nil
}
func newServiceMetadataMiddleware_opStartFaceDetection(region string) *awsmiddleware.RegisterServiceMetadata {
return &awsmiddleware.RegisterServiceMetadata{
Region: region,
ServiceID: ServiceID,
SigningName: "rekognition",
OperationName: "StartFaceDetection",
}
}