-
Notifications
You must be signed in to change notification settings - Fork 124
/
s3_put.go
408 lines (340 loc) · 12.7 KB
/
s3_put.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
package command
import (
"context"
"fmt"
"net/http"
"os"
"path/filepath"
"strconv"
"strings"
"time"
"github.com/aws/aws-sdk-go/aws/endpoints"
"github.com/evergreen-ci/evergreen/model"
"github.com/evergreen-ci/evergreen/model/artifact"
"github.com/evergreen-ci/evergreen/rest/client"
"github.com/evergreen-ci/evergreen/util"
"github.com/evergreen-ci/pail"
"github.com/mitchellh/mapstructure"
"github.com/mongodb/grip"
"github.com/pkg/errors"
)
// A plugin command to put a resource to an s3 bucket and download it to
// the local machine.
type s3put struct {
// AwsKey and AwsSecret are the user's credentials for
// authenticating interactions with s3.
AwsKey string `mapstructure:"aws_key" plugin:"expand"`
AwsSecret string `mapstructure:"aws_secret" plugin:"expand"`
// LocalFile is the local filepath to the file the user
// wishes to store in s3
LocalFile string `mapstructure:"local_file" plugin:"expand"`
// LocalFilesIncludeFilter is an array of expressions that specify what files should be
// included in this upload.
LocalFilesIncludeFilter []string `mapstructure:"local_files_include_filter" plugin:"expand"`
// LocalFilesIncludeFilterPrefix is an optional path to start processing the LocalFilesIncludeFilter, relative to the working directory.
LocalFilesIncludeFilterPrefix string `mapstructure:"local_files_include_filter_prefix" plugin:"expand"`
// RemoteFile is the filepath to store the file to,
// within an s3 bucket. Is a prefix when multiple files are uploaded via LocalFilesIncludeFilter.
RemoteFile string `mapstructure:"remote_file" plugin:"expand"`
// Bucket is the s3 bucket to use when storing the desired file
Bucket string `mapstructure:"bucket" plugin:"expand"`
// Permissions is the ACL to apply to the uploaded file. See:
// http://docs.aws.amazon.com/AmazonS3/latest/dev/acl-overview.html#canned-acl
// for some examples.
Permissions string `mapstructure:"permissions"`
// ContentType is the MIME type of the uploaded file.
// E.g. text/html, application/pdf, image/jpeg, ...
ContentType string `mapstructure:"content_type" plugin:"expand"`
// BuildVariants stores a list of MCI build variants to run the command for.
// If the list is empty, it runs for all build variants.
BuildVariants []string `mapstructure:"build_variants"`
// ResourceDisplayName stores the name of the file that is linked. Is a prefix when
// to the matched file name when multiple files are uploaded.
ResourceDisplayName string `mapstructure:"display_name" plugin:"expand"`
// Visibility determines who can see file links in the UI.
// Visibility can be set to either
// "private", which allows logged-in users to see the file;
// "public", which allows anyone to see the file; or
// "none", which hides the file from the UI for everybody.
// If unset, the file will be public.
Visibility string `mapstructure:"visibility" plugin:"expand"`
// Optional, when set to true, causes this command to be skipped over without an error when
// the path specified in local_file does not exist. Defaults to false, which triggers errors
// for missing files.
Optional string `mapstructure:"optional" plugin:"expand"`
// workDir sets the working directory relative to which s3put should look for files to upload.
// workDir will be empty if an absolute path is provided to the file.
workDir string
skipMissing bool
bucket pail.Bucket
taskdata client.TaskData
base
}
func s3PutFactory() Command { return &s3put{} }
func (s3pc *s3put) Name() string { return "s3.put" }
// s3put-specific implementation of ParseParams.
func (s3pc *s3put) ParseParams(params map[string]interface{}) error {
decoder, err := mapstructure.NewDecoder(&mapstructure.DecoderConfig{
WeaklyTypedInput: true,
Result: s3pc,
})
if err != nil {
return errors.WithStack(err)
}
if err := decoder.Decode(params); err != nil {
return errors.Wrapf(err, "error decoding %s params", s3pc.Name())
}
return s3pc.validate()
}
func (s3pc *s3put) validate() error {
catcher := grip.NewSimpleCatcher()
// make sure the command params are valid
if s3pc.AwsKey == "" {
catcher.Add(errors.New("aws_key cannot be blank"))
}
if s3pc.AwsSecret == "" {
catcher.Add(errors.New("aws_secret cannot be blank"))
}
if s3pc.LocalFile == "" && !s3pc.isMulti() {
catcher.Add(errors.New("local_file and local_files_include_filter cannot both be blank"))
}
if s3pc.LocalFile != "" && s3pc.isMulti() {
catcher.Add(errors.New("local_file and local_files_include_filter cannot both be specified"))
}
if s3pc.skipMissing && s3pc.isMulti() {
catcher.Add(errors.New("cannot use optional upload with local_files_include_filter"))
}
if s3pc.RemoteFile == "" {
catcher.Add(errors.New("remote_file cannot be blank"))
}
if s3pc.ContentType == "" {
catcher.Add(errors.New("content_type cannot be blank"))
}
if s3pc.isMulti() && filepath.IsAbs(s3pc.LocalFile) {
catcher.Add(errors.New("cannot use absolute path with local_files_include_filter"))
}
if !util.StringSliceContains(artifact.ValidVisibilities, s3pc.Visibility) {
catcher.Add(errors.Errorf("invalid visibility setting: %v", s3pc.Visibility))
}
// make sure the bucket is valid
if err := validateS3BucketName(s3pc.Bucket); err != nil {
catcher.Add(errors.Wrapf(err, "%v is an invalid bucket name", s3pc.Bucket))
}
// make sure the s3 permissions are valid
if !validS3Permissions(s3pc.Permissions) {
catcher.Add(errors.Errorf("permissions '%v' are not valid", s3pc.Permissions))
}
return catcher.Resolve()
}
// Apply the expansions from the relevant task config to all appropriate
// fields of the s3put.
func (s3pc *s3put) expandParams(conf *model.TaskConfig) error {
if filepath.IsAbs(s3pc.LocalFile) {
s3pc.workDir = ""
} else {
s3pc.workDir = conf.WorkDir
}
var err error
if err = util.ExpandValues(s3pc, conf.Expansions); err != nil {
return errors.WithStack(err)
}
if s3pc.Optional != "" {
s3pc.skipMissing, err = strconv.ParseBool(s3pc.Optional)
if err != nil {
return errors.WithStack(err)
}
} else {
s3pc.skipMissing = false
}
return nil
}
// isMulti returns whether or not this using the multiple file upload
// capability of the Put command.
func (s3pc *s3put) isMulti() bool {
return (len(s3pc.LocalFilesIncludeFilter) != 0)
}
func (s3pc *s3put) shouldRunForVariant(buildVariantName string) bool {
//No buildvariant filter, so run always
if len(s3pc.BuildVariants) == 0 {
return true
}
//Only run if the buildvariant specified appears in our list.
return util.StringSliceContains(s3pc.BuildVariants, buildVariantName)
}
// Implementation of Execute. Expands the parameters, and then puts the
// resource to s3.
func (s3pc *s3put) Execute(ctx context.Context,
comm client.Communicator, logger client.LoggerProducer, conf *model.TaskConfig) error {
// expand necessary params
if err := s3pc.expandParams(conf); err != nil {
return errors.WithStack(err)
}
// re-validate command here, in case an expansion is not defined
if err := s3pc.validate(); err != nil {
return errors.WithStack(err)
}
// create pail bucket
httpClient := util.GetHTTPClient()
httpClient.Timeout = s3HTTPClientTimeout
defer util.PutHTTPClient(httpClient)
if err := s3pc.createPailBucket(httpClient); err != nil {
return errors.Wrap(err, "problem connecting to s3")
}
ctx, cancel := context.WithCancel(context.Background())
defer cancel()
if err := s3pc.bucket.Check(ctx); err != nil {
return errors.Wrap(err, "invalid bucket")
}
s3pc.taskdata = client.TaskData{ID: conf.Task.Id, Secret: conf.Task.Secret}
if !s3pc.shouldRunForVariant(conf.BuildVariant.Name) {
logger.Task().Infof("Skipping S3 put of local file %v for variant %v",
s3pc.LocalFile, conf.BuildVariant.Name)
return nil
}
if s3pc.isMulti() {
logger.Task().Infof("Putting files matching filter %v into path %v in s3 bucket %v",
s3pc.LocalFilesIncludeFilter, s3pc.RemoteFile, s3pc.Bucket)
} else {
logger.Task().Infof("Putting %s into %s/%s/%s",
s3pc.LocalFile, s3baseURL, s3pc.Bucket, s3pc.RemoteFile)
}
errChan := make(chan error)
go func() {
errChan <- errors.WithStack(s3pc.putWithRetry(ctx, comm, logger))
}()
select {
case err := <-errChan:
return err
case <-ctx.Done():
logger.Execution().Info("Received signal to terminate execution of S3 Put Command")
return nil
}
}
// Wrapper around the Put() function to retry it.
func (s3pc *s3put) putWithRetry(ctx context.Context, comm client.Communicator, logger client.LoggerProducer) error {
backoffCounter := getS3OpBackoff()
var (
err error
uploadedFiles []string
filesList []string
)
timer := time.NewTimer(0)
defer timer.Stop()
retryLoop:
for i := 1; i <= maxS3OpAttempts; i++ {
logger.Task().Infof("performing s3 put to %s of %s [%d of %d]",
s3pc.Bucket, s3pc.RemoteFile,
i, maxS3OpAttempts)
select {
case <-ctx.Done():
return errors.New("s3 put operation canceled")
case <-timer.C:
filesList = []string{s3pc.LocalFile}
if s3pc.isMulti() {
filesList, err = util.BuildFileList(filepath.Join(s3pc.workDir, s3pc.LocalFilesIncludeFilterPrefix), s3pc.LocalFilesIncludeFilter...)
if err != nil {
return errors.Wrapf(err, "error processing filter %s",
strings.Join(s3pc.LocalFilesIncludeFilter, " "))
}
if len(filesList) == 0 {
logger.Task().Info("s3 put include filter matched no files")
return nil
}
}
// reset to avoid duplicated uploaded references
uploadedFiles = []string{}
uploadLoop:
for _, fpath := range filesList {
if ctx.Err() != nil {
return errors.New("s3 put operation canceled")
}
remoteName := s3pc.RemoteFile
if s3pc.isMulti() {
fname := filepath.Base(fpath)
remoteName = fmt.Sprintf("%s%s", s3pc.RemoteFile, fname)
}
fpath = filepath.Join(filepath.Join(s3pc.workDir, s3pc.LocalFilesIncludeFilterPrefix), fpath)
err = s3pc.bucket.Upload(ctx, remoteName, fpath)
if err != nil {
// retry errors other than "file doesn't exist", which we handle differently based on what
// kind of upload it is
if os.IsNotExist(errors.Cause(err)) {
if s3pc.isMulti() {
// try the remaining multi uploads in the group, effectively ignoring this
// error.
continue uploadLoop
} else if s3pc.skipMissing {
// single optional file uploads should return early.
logger.Task().Infof("file %s not found but skip missing true", fpath)
return nil
} else {
// single required uploads should return an error asap.
return errors.Wrapf(err, "missing file %s", fpath)
}
}
// in all other cases, log an error and retry after an interval.
logger.Execution().Error(errors.WithMessage(err, "problem putting s3 file"))
timer.Reset(backoffCounter.Duration())
continue retryLoop
}
uploadedFiles = append(uploadedFiles, fpath)
}
break retryLoop
}
}
if len(uploadedFiles) == 0 && s3pc.skipMissing {
logger.Task().Info("s3 put uploaded no files")
return nil
}
err = errors.WithStack(s3pc.attachFiles(ctx, comm, logger, uploadedFiles, s3pc.RemoteFile))
if err != nil {
return err
}
if len(uploadedFiles) != len(filesList) && !s3pc.skipMissing {
logger.Task().Infof("%d requested, %d uploaded", len(filesList), len(uploadedFiles))
return errors.Errorf("uploaded %d files of %d requested", len(uploadedFiles), len(filesList))
}
return nil
}
// attachTaskFiles is responsible for sending the
// specified file to the API Server. Does not support multiple file putting.
func (s3pc *s3put) attachFiles(ctx context.Context, comm client.Communicator, logger client.LoggerProducer, localFiles []string, remoteFile string) error {
files := []*artifact.File{}
for _, fn := range localFiles {
remoteFileName := filepath.ToSlash(remoteFile)
if s3pc.isMulti() {
remoteFileName = fmt.Sprintf("%s%s", remoteFile, filepath.Base(fn))
}
fileLink := s3baseURL + s3pc.Bucket + "/" + remoteFileName
displayName := s3pc.ResourceDisplayName
if s3pc.isMulti() || displayName == "" {
displayName = fmt.Sprintf("%s %s", s3pc.ResourceDisplayName, filepath.Base(fn))
}
files = append(files, &artifact.File{
Name: displayName,
Link: fileLink,
Visibility: s3pc.Visibility,
})
}
err := comm.AttachFiles(ctx, s3pc.taskdata, files)
if err != nil {
return errors.Wrap(err, "Attach files failed")
}
logger.Execution().Info("API attach files call succeeded")
return nil
}
func (s3pc *s3put) createPailBucket(httpClient *http.Client) error {
if s3pc.bucket != nil {
return nil
}
opts := pail.S3Options{
Credentials: pail.CreateAWSCredentials(s3pc.AwsKey, s3pc.AwsSecret, ""),
Region: endpoints.UsEast1RegionID,
Name: s3pc.Bucket,
Permissions: pail.S3Permissions(s3pc.Permissions),
ContentType: s3pc.ContentType,
}
bucket, err := pail.NewS3MultiPartBucketWithHTTPClient(httpClient, opts)
s3pc.bucket = bucket
return err
}