forked from juju/juju
-
Notifications
You must be signed in to change notification settings - Fork 0
/
uniter.go
1047 lines (957 loc) · 35.9 KB
/
uniter.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
// Copyright 2012-2015 Canonical Ltd.
// Licensed under the AGPLv3, see LICENCE file for details.
package uniter
import (
"fmt"
"os"
"sync"
corecharm "github.com/juju/charm/v9"
"github.com/juju/clock"
"github.com/juju/errors"
"github.com/juju/names/v4"
"github.com/juju/utils/v3"
"github.com/juju/utils/v3/exec"
"github.com/juju/worker/v3"
"github.com/juju/worker/v3/catacomb"
"github.com/juju/juju/agent/tools"
"github.com/juju/juju/api/agent/uniter"
"github.com/juju/juju/core/leadership"
"github.com/juju/juju/core/life"
corelogger "github.com/juju/juju/core/logger"
"github.com/juju/juju/core/lxdprofile"
"github.com/juju/juju/core/machinelock"
"github.com/juju/juju/core/model"
"github.com/juju/juju/core/status"
"github.com/juju/juju/core/watcher"
"github.com/juju/juju/rpc/params"
jworker "github.com/juju/juju/worker"
"github.com/juju/juju/worker/fortress"
"github.com/juju/juju/worker/uniter/actions"
"github.com/juju/juju/worker/uniter/charm"
"github.com/juju/juju/worker/uniter/container"
"github.com/juju/juju/worker/uniter/hook"
uniterleadership "github.com/juju/juju/worker/uniter/leadership"
"github.com/juju/juju/worker/uniter/operation"
"github.com/juju/juju/worker/uniter/reboot"
"github.com/juju/juju/worker/uniter/relation"
"github.com/juju/juju/worker/uniter/remotestate"
"github.com/juju/juju/worker/uniter/resolver"
"github.com/juju/juju/worker/uniter/runcommands"
"github.com/juju/juju/worker/uniter/runner"
"github.com/juju/juju/worker/uniter/runner/context"
"github.com/juju/juju/worker/uniter/runner/jujuc"
"github.com/juju/juju/worker/uniter/secrets"
"github.com/juju/juju/worker/uniter/storage"
"github.com/juju/juju/worker/uniter/upgradeseries"
"github.com/juju/juju/worker/uniter/verifycharmprofile"
)
var (
// ErrCAASUnitDead is the error returned from terminate or init
// if the unit is Dead.
ErrCAASUnitDead = errors.New("unit dead")
)
// A UniterExecutionObserver gets the appropriate methods called when a hook
// is executed and either succeeds or fails. Missing hooks don't get reported
// in this way.
type UniterExecutionObserver interface {
HookCompleted(hookName string)
HookFailed(hookName string)
}
// RebootQuerier is implemented by types that can deliver one-off machine
// reboot notifications to entities.
type RebootQuerier interface {
Query(tag names.Tag) (bool, error)
}
// SecretsClient provides methods used by the remote state watcher, hook context,
// and op callbacks.
type SecretsClient interface {
remotestate.SecretsClient
context.SecretsAccessor
}
// RemoteInitFunc is used to init remote state
type RemoteInitFunc func(remotestate.ContainerRunningStatus, <-chan struct{}) error
// Uniter implements the capabilities of the unit agent, for example running hooks.
type Uniter struct {
catacomb catacomb.Catacomb
st *uniter.State
secretsClient SecretsClient
secretsStoreGetter context.SecretsStoreGetter
paths Paths
unit *uniter.Unit
resources *uniter.ResourcesFacadeClient
payloads *uniter.PayloadFacadeClient
modelType model.ModelType
sidecar bool
enforcedCharmModifiedVersion int
storage *storage.Attachments
clock clock.Clock
relationStateTracker relation.RelationStateTracker
secretsTracker secrets.SecretStateTracker
// Cache the last reported status information
// so we don't make unnecessary api calls.
setStatusMutex sync.Mutex
lastReportedStatus status.Status
lastReportedMessage string
operationFactory operation.Factory
operationExecutor operation.Executor
newOperationExecutor NewOperationExecutorFunc
newProcessRunner runner.NewRunnerFunc
newDeployer charm.NewDeployerFunc
newRemoteRunnerExecutor NewRunnerExecutorFunc
translateResolverErr func(error) error
leadershipTracker leadership.TrackerWorker
charmDirGuard fortress.Guard
hookLock machinelock.Lock
// secretRotateWatcherFunc returns a watcher that triggers when secrets
// owned by this unit ot its application should be rotated.
secretRotateWatcherFunc remotestate.SecretTriggerWatcherFunc
// secretExpiryWatcherFunc returns a watcher that triggers when
// secret revisions owned by this unit or its application should be expired.
secretExpiryWatcherFunc remotestate.SecretTriggerWatcherFunc
Probe Probe
// TODO(axw) move the runListener and run-command code outside of the
// uniter, and introduce a separate worker. Each worker would feed
// operations to a single, synchronized runner to execute.
runListener *RunListener
localRunListener *RunListener
commands runcommands.Commands
commandChannel chan string
// The execution observer is only used in tests at this stage. Should this
// need to be extended, perhaps a list of observers would be needed.
observer UniterExecutionObserver
// updateStatusAt defines a function that will be used to generate signals for
// the update-status hook
updateStatusAt remotestate.UpdateStatusTimerFunc
// containerRunningStatusChannel, if set, is used to signal a change in the
// unit's status. It is passed to the remote state watcher.
containerRunningStatusChannel watcher.NotifyChannel
// containerRunningStatusFunc is used to determine the unit's running status.
containerRunningStatusFunc remotestate.ContainerRunningStatusFunc
// remoteInitFunc is used to init remote charm state.
remoteInitFunc RemoteInitFunc
// isRemoteUnit is true when the unit is remotely deployed.
isRemoteUnit bool
// containerNames will have a list of the workload containers created alongside this
// unit agent.
containerNames []string
workloadEvents container.WorkloadEvents
workloadEventChannel chan string
newPebbleClient NewPebbleClientFunc
// hookRetryStrategy represents configuration for hook retries
hookRetryStrategy params.RetryStrategy
// downloader is the downloader that should be used to get the charm
// archive.
downloader charm.Downloader
// rebootQuerier allows the uniter to detect when the machine has
// rebooted so we can notify the charms accordingly.
rebootQuerier RebootQuerier
logger Logger
// shutdownChannel is passed to the remote state watcher. When true is
// sent on the channel, it causes the uniter to start the shutdown process.
shutdownChannel chan bool
}
// UniterParams hold all the necessary parameters for a new Uniter.
type UniterParams struct {
UniterFacade *uniter.State
ResourcesFacade *uniter.ResourcesFacadeClient
PayloadFacade *uniter.PayloadFacadeClient
SecretsClient SecretsClient
SecretsStoreGetter context.SecretsStoreGetter
UnitTag names.UnitTag
ModelType model.ModelType
LeadershipTrackerFunc func(names.UnitTag) leadership.TrackerWorker
SecretRotateWatcherFunc remotestate.SecretTriggerWatcherFunc
SecretExpiryWatcherFunc remotestate.SecretTriggerWatcherFunc
DataDir string
Downloader charm.Downloader
MachineLock machinelock.Lock
CharmDirGuard fortress.Guard
UpdateStatusSignal remotestate.UpdateStatusTimerFunc
HookRetryStrategy params.RetryStrategy
NewOperationExecutor NewOperationExecutorFunc
NewProcessRunner runner.NewRunnerFunc
NewDeployer charm.NewDeployerFunc
NewRemoteRunnerExecutor NewRunnerExecutorFunc
RemoteInitFunc RemoteInitFunc
RunListener *RunListener
TranslateResolverErr func(error) error
Clock clock.Clock
ContainerRunningStatusChannel watcher.NotifyChannel
ContainerRunningStatusFunc remotestate.ContainerRunningStatusFunc
IsRemoteUnit bool
SocketConfig *SocketConfig
// TODO (mattyw, wallyworld, fwereade) Having the observer here make this approach a bit more legitimate, but it isn't.
// the observer is only a stop gap to be used in tests. A better approach would be to have the uniter tests start hooks
// that write to files, and have the tests watch the output to know that hooks have finished.
Observer UniterExecutionObserver
RebootQuerier RebootQuerier
Logger Logger
Sidecar bool
EnforcedCharmModifiedVersion int
ContainerNames []string
NewPebbleClient NewPebbleClientFunc
}
// NewOperationExecutorFunc is a func which returns an operations.Executor.
type NewOperationExecutorFunc func(string, operation.ExecutorConfig) (operation.Executor, error)
// ProviderIDGetter defines the API to get provider ID.
type ProviderIDGetter interface {
ProviderID() string
Refresh() error
Name() string
}
// NewRunnerExecutorFunc defines the type of the NewRunnerExecutor.
type NewRunnerExecutorFunc func(ProviderIDGetter, Paths) runner.ExecFunc
// NewUniter creates a new Uniter which will install, run, and upgrade
// a charm on behalf of the unit with the given unitTag, by executing
// hooks and operations provoked by changes in st.
func NewUniter(uniterParams *UniterParams) (*Uniter, error) {
startFunc := newUniter(uniterParams)
w, err := startFunc()
return w.(*Uniter), err
}
// StartUniter creates a new Uniter and starts it using the specified runner.
func StartUniter(runner *worker.Runner, params *UniterParams) error {
startFunc := newUniter(params)
params.Logger.Debugf("starting uniter for %q", params.UnitTag.Id())
err := runner.StartWorker(params.UnitTag.Id(), startFunc)
return errors.Annotate(err, "error starting uniter worker")
}
func newUniter(uniterParams *UniterParams) func() (worker.Worker, error) {
translateResolverErr := uniterParams.TranslateResolverErr
if translateResolverErr == nil {
translateResolverErr = func(err error) error { return err }
}
startFunc := func() (worker.Worker, error) {
u := &Uniter{
st: uniterParams.UniterFacade,
resources: uniterParams.ResourcesFacade,
payloads: uniterParams.PayloadFacade,
secretsClient: uniterParams.SecretsClient,
secretsStoreGetter: uniterParams.SecretsStoreGetter,
paths: NewPaths(uniterParams.DataDir, uniterParams.UnitTag, uniterParams.SocketConfig),
modelType: uniterParams.ModelType,
hookLock: uniterParams.MachineLock,
leadershipTracker: uniterParams.LeadershipTrackerFunc(uniterParams.UnitTag),
secretRotateWatcherFunc: uniterParams.SecretRotateWatcherFunc,
secretExpiryWatcherFunc: uniterParams.SecretExpiryWatcherFunc,
charmDirGuard: uniterParams.CharmDirGuard,
updateStatusAt: uniterParams.UpdateStatusSignal,
hookRetryStrategy: uniterParams.HookRetryStrategy,
newOperationExecutor: uniterParams.NewOperationExecutor,
newProcessRunner: uniterParams.NewProcessRunner,
newDeployer: uniterParams.NewDeployer,
newRemoteRunnerExecutor: uniterParams.NewRemoteRunnerExecutor,
remoteInitFunc: uniterParams.RemoteInitFunc,
translateResolverErr: translateResolverErr,
observer: uniterParams.Observer,
clock: uniterParams.Clock,
downloader: uniterParams.Downloader,
containerRunningStatusChannel: uniterParams.ContainerRunningStatusChannel,
containerRunningStatusFunc: uniterParams.ContainerRunningStatusFunc,
isRemoteUnit: uniterParams.IsRemoteUnit,
runListener: uniterParams.RunListener,
rebootQuerier: uniterParams.RebootQuerier,
logger: uniterParams.Logger,
sidecar: uniterParams.Sidecar,
enforcedCharmModifiedVersion: uniterParams.EnforcedCharmModifiedVersion,
containerNames: uniterParams.ContainerNames,
newPebbleClient: uniterParams.NewPebbleClient,
shutdownChannel: make(chan bool, 1),
}
plan := catacomb.Plan{
Site: &u.catacomb,
Work: func() error {
return u.loop(uniterParams.UnitTag)
},
}
if u.modelType == model.CAAS {
// For CAAS models, make sure the leadership tracker is killed when the Uniter
// dies.
plan.Init = append(plan.Init, u.leadershipTracker)
}
if err := catacomb.Invoke(plan); err != nil {
return nil, errors.Trace(err)
}
return u, nil
}
return startFunc
}
func (u *Uniter) loop(unitTag names.UnitTag) (err error) {
defer func() {
// If this is a CAAS unit, then dead errors are fairly normal ways to exit
// the uniter main loop, but the parent operator agent needs to keep running.
errorString := "<unknown>"
if err != nil {
errorString = err.Error()
}
if errors.Cause(err) == ErrCAASUnitDead {
errorString = err.Error()
err = nil
}
if u.runListener != nil {
u.runListener.UnregisterRunner(unitTag.Id())
}
if u.localRunListener != nil {
u.localRunListener.UnregisterRunner(unitTag.Id())
}
u.logger.Infof("unit %q shutting down: %s", unitTag.Id(), errorString)
}()
if err := u.init(unitTag); err != nil {
switch cause := errors.Cause(err); cause {
case resolver.ErrLoopAborted:
return u.catacomb.ErrDying()
case ErrCAASUnitDead:
// Normal exit from the loop as we don't want it restarted.
return nil
case jworker.ErrTerminateAgent:
return err
default:
return errors.Annotatef(err, "failed to initialize uniter for %q", unitTag)
}
}
u.logger.Infof("unit %q started", u.unit)
canApplyCharmProfile, charmURL, charmModifiedVersion, err := u.charmState()
if err != nil {
return errors.Trace(err)
}
// Check we are running the correct charm version.
if u.sidecar && u.enforcedCharmModifiedVersion != -1 {
if charmModifiedVersion != u.enforcedCharmModifiedVersion {
u.logger.Infof("remote charm modified version (%d) does not match agent's (%d)",
charmModifiedVersion, u.enforcedCharmModifiedVersion)
return u.stopUnitError()
}
}
var watcher *remotestate.RemoteStateWatcher
u.logger.Infof("hooks are retried %v", u.hookRetryStrategy.ShouldRetry)
retryHookChan := make(chan struct{}, 1)
// TODO(katco): 2016-08-09: This type is deprecated: lp:1611427
retryHookTimer := utils.NewBackoffTimer(utils.BackoffTimerConfig{
Min: u.hookRetryStrategy.MinRetryTime,
Max: u.hookRetryStrategy.MaxRetryTime,
Jitter: u.hookRetryStrategy.JitterRetryTime,
Factor: u.hookRetryStrategy.RetryTimeFactor,
Func: func() {
// Don't try to send on the channel if it's already full
// This can happen if the timer fires off before the event is consumed
// by the resolver loop
select {
case retryHookChan <- struct{}{}:
default:
}
},
Clock: u.clock,
})
defer func() {
// Whenever we exit the uniter we want to stop a potentially
// running timer so it doesn't trigger for nothing.
retryHookTimer.Reset()
}()
restartWatcher := func() error {
if watcher != nil {
// watcher added to catacomb, will kill uniter if there's an error.
_ = worker.Stop(watcher)
}
var err error
watcher, err = remotestate.NewWatcher(
remotestate.WatcherConfig{
State: remotestate.NewAPIState(u.st),
LeadershipTracker: u.leadershipTracker,
SecretsClient: u.secretsClient,
SecretRotateWatcherFunc: u.secretRotateWatcherFunc,
SecretExpiryWatcherFunc: u.secretExpiryWatcherFunc,
UnitTag: unitTag,
UpdateStatusChannel: u.updateStatusAt,
CommandChannel: u.commandChannel,
RetryHookChannel: retryHookChan,
ContainerRunningStatusChannel: u.containerRunningStatusChannel,
ContainerRunningStatusFunc: u.containerRunningStatusFunc,
ModelType: u.modelType,
Logger: u.logger.Child("remotestate"),
CanApplyCharmProfile: canApplyCharmProfile,
Sidecar: u.sidecar,
EnforcedCharmModifiedVersion: u.enforcedCharmModifiedVersion,
WorkloadEventChannel: u.workloadEventChannel,
InitialWorkloadEventIDs: u.workloadEvents.EventIDs(),
ShutdownChannel: u.shutdownChannel,
})
if err != nil {
return errors.Trace(err)
}
if err := u.catacomb.Add(watcher); err != nil {
return errors.Trace(err)
}
return nil
}
onIdle := func() error {
opState := u.operationExecutor.State()
if opState.Kind != operation.Continue {
// We should only set idle status if we're in
// the "Continue" state, which indicates that
// there is nothing to do and we're not in an
// error state.
return nil
}
return setAgentStatus(u, status.Idle, "", nil)
}
clearResolved := func() error {
if err := u.unit.ClearResolved(); err != nil {
return errors.Trace(err)
}
watcher.ClearResolvedMode()
return nil
}
if u.modelType == model.CAAS && u.isRemoteUnit {
if u.containerRunningStatusChannel == nil {
return errors.NotValidf("ContainerRunningStatusChannel missing for CAAS remote unit")
}
if u.containerRunningStatusFunc == nil {
return errors.NotValidf("ContainerRunningStatusFunc missing for CAAS remote unit")
}
}
var rebootDetected bool
if u.modelType == model.IAAS {
if rebootDetected, err = u.rebootQuerier.Query(unitTag); err != nil {
return errors.Annotatef(err, "could not check reboot status for %q", unitTag)
}
} else if u.modelType == model.CAAS && u.sidecar {
rebootDetected = true
}
rebootResolver := reboot.NewResolver(u.logger, rebootDetected)
for {
if err = restartWatcher(); err != nil {
err = errors.Annotate(err, "(re)starting watcher")
break
}
cfg := ResolverConfig{
ModelType: u.modelType,
ClearResolved: clearResolved,
ReportHookError: u.reportHookError,
ShouldRetryHooks: u.hookRetryStrategy.ShouldRetry,
StartRetryHookTimer: retryHookTimer.Start,
StopRetryHookTimer: retryHookTimer.Reset,
Actions: actions.NewResolver(
u.logger.Child("actions"),
),
VerifyCharmProfile: verifycharmprofile.NewResolver(
u.logger.Child("verifycharmprofile"),
u.modelType,
),
UpgradeSeries: upgradeseries.NewResolver(
u.logger.Child("upgradeseries"),
),
Reboot: rebootResolver,
Leadership: uniterleadership.NewResolver(
u.logger.Child("leadership"),
),
CreatedRelations: relation.NewCreatedRelationResolver(
u.relationStateTracker, u.logger.ChildWithLabels("relation", corelogger.CMR)),
Relations: relation.NewRelationResolver(
u.relationStateTracker, u.unit, u.logger.ChildWithLabels("relation", corelogger.CMR)),
Storage: storage.NewResolver(
u.logger.Child("storage"), u.storage, u.modelType),
Commands: runcommands.NewCommandsResolver(
u.commands, watcher.CommandCompleted,
),
Secrets: secrets.NewSecretsResolver(
u.logger.ChildWithLabels("secrets", corelogger.SECRETS),
u.secretsTracker,
watcher.RotateSecretCompleted,
watcher.ExpireRevisionCompleted,
),
Logger: u.logger,
}
if u.modelType == model.CAAS && u.isRemoteUnit {
cfg.OptionalResolvers = append(cfg.OptionalResolvers, container.NewRemoteContainerInitResolver())
}
if len(u.containerNames) > 0 {
cfg.OptionalResolvers = append(cfg.OptionalResolvers, container.NewWorkloadHookResolver(
u.logger.Child("workload"),
u.workloadEvents,
watcher.WorkloadEventCompleted),
)
}
uniterResolver := NewUniterResolver(cfg)
// We should not do anything until there has been a change
// to the remote state. The watcher will trigger at least
// once initially.
select {
case <-u.catacomb.Dying():
return u.catacomb.ErrDying()
case <-watcher.RemoteStateChanged():
}
localState := resolver.LocalState{
CharmURL: charmURL,
CharmModifiedVersion: charmModifiedVersion,
UpgradeSeriesStatus: model.UpgradeSeriesNotStarted,
// CAAS remote units should trigger remote update of the charm every start.
OutdatedRemoteCharm: u.isRemoteUnit,
}
for err == nil {
err = resolver.Loop(resolver.LoopConfig{
Resolver: uniterResolver,
Watcher: watcher,
Executor: u.operationExecutor,
Factory: u.operationFactory,
Abort: u.catacomb.Dying(),
OnIdle: onIdle,
CharmDirGuard: u.charmDirGuard,
CharmDir: u.paths.State.CharmDir,
Logger: u.logger.Child("resolver"),
}, &localState)
err = u.translateResolverErr(err)
switch cause := errors.Cause(err); cause {
case nil:
// Loop back around.
case resolver.ErrLoopAborted:
err = u.catacomb.ErrDying()
case operation.ErrNeedsReboot:
err = jworker.ErrRebootMachine
case operation.ErrHookFailed:
// Loop back around. The resolver can tell that it is in
// an error state by inspecting the operation state.
err = nil
case resolver.ErrTerminate:
err = u.terminate()
case resolver.ErrRestart:
// make sure we update the two values used above in
// creating LocalState.
charmURL = localState.CharmURL
charmModifiedVersion = localState.CharmModifiedVersion
// leave err assigned, causing loop to break
case jworker.ErrTerminateAgent:
// terminate agent
default:
// We need to set conflicted from here, because error
// handling is outside of the resolver's control.
if operation.IsDeployConflictError(cause) {
localState.Conflicted = true
err = setAgentStatus(u, status.Error, "upgrade failed", nil)
} else {
reportAgentError(u, "resolver loop error", err)
}
}
}
if errors.Cause(err) != resolver.ErrRestart {
break
}
}
return err
}
func (u *Uniter) verifyCharmProfile(url string) error {
curl, err := corecharm.ParseURL(url)
if err != nil {
return errors.Trace(err)
}
// NOTE: this is very similar code to verifyCharmProfile.NextOp,
// if you make changes here, check to see if they are needed there.
ch, err := u.st.Charm(curl)
if err != nil {
return errors.Trace(err)
}
required, err := ch.LXDProfileRequired()
if err != nil {
return errors.Trace(err)
}
if !required {
// If no lxd profile is required for this charm, move on.
u.logger.Debugf("no lxd profile required for %s", curl)
return nil
}
profile, err := u.unit.LXDProfileName()
if err != nil {
return errors.Trace(err)
}
if profile == "" {
if err := u.unit.SetUnitStatus(status.Waiting, "required charm profile not yet applied to machine", nil); err != nil {
return errors.Trace(err)
}
u.logger.Debugf("required lxd profile not found on machine")
return errors.NotFoundf("required charm profile on machine")
}
// double check profile revision matches charm revision.
rev, err := lxdprofile.ProfileRevision(profile)
if err != nil {
return errors.Trace(err)
}
if rev != curl.Revision {
if err := u.unit.SetUnitStatus(status.Waiting, fmt.Sprintf("required charm profile %q not yet applied to machine", profile), nil); err != nil {
return errors.Trace(err)
}
u.logger.Debugf("charm is revision %d, charm profile has revision %d", curl.Revision, rev)
return errors.NotFoundf("required charm profile, %q, on machine", profile)
}
u.logger.Debugf("required lxd profile %q FOUND on machine", profile)
if err := u.unit.SetUnitStatus(status.Waiting, status.MessageInitializingAgent, nil); err != nil {
return errors.Trace(err)
}
return nil
}
// charmState returns data for the local state setup.
// While gathering the data, look for interrupted Install or pending
// charm upgrade, execute if found.
func (u *Uniter) charmState() (bool, string, int, error) {
// Install is a special case, as it must run before there
// is any remote state, and before the remote state watcher
// is started.
var charmURL string
var charmModifiedVersion int
canApplyCharmProfile, err := u.unit.CanApplyLXDProfile()
if err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, err
}
opState := u.operationExecutor.State()
if opState.Kind == operation.Install {
u.logger.Infof("resuming charm install")
if canApplyCharmProfile {
// Note: canApplyCharmProfile will be false for a CAAS model.
// Verify the charm profile before proceeding.
if err := u.verifyCharmProfile(opState.CharmURL); err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, err
}
}
op, err := u.operationFactory.NewInstall(opState.CharmURL)
if err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, errors.Trace(err)
}
if err := u.operationExecutor.Run(op, nil); err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, errors.Trace(err)
}
charmURL = opState.CharmURL
return canApplyCharmProfile, charmURL, charmModifiedVersion, nil
}
// No install needed, find the curl and start.
curl, err := u.unit.CharmURL()
if err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, errors.Trace(err)
}
charmURL = curl
app, err := u.unit.Application()
if err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, errors.Trace(err)
}
// TODO (hml) 25-09-2020 - investigate
// This assumes that the uniter is not restarting after an application
// changed notification, with changes to CharmModifiedVersion, but before
// it could be acted on.
charmModifiedVersion, err = app.CharmModifiedVersion()
if err != nil {
return canApplyCharmProfile, charmURL, charmModifiedVersion, errors.Trace(err)
}
return canApplyCharmProfile, charmURL, charmModifiedVersion, nil
}
func (u *Uniter) terminate() error {
unitWatcher, err := u.unit.Watch()
if err != nil {
return errors.Trace(err)
}
if err := u.catacomb.Add(unitWatcher); err != nil {
return errors.Trace(err)
}
for {
select {
case <-u.catacomb.Dying():
return u.catacomb.ErrDying()
case _, ok := <-unitWatcher.Changes():
if !ok {
return errors.New("unit watcher closed")
}
if err := u.unit.Refresh(); err != nil {
return errors.Trace(err)
}
if hasSubs, err := u.unit.HasSubordinates(); err != nil {
return errors.Trace(err)
} else if hasSubs {
continue
}
// The unit is known to be Dying; so if it didn't have subordinates
// just above, it can't acquire new ones before this call.
if err := u.unit.EnsureDead(); err != nil {
return errors.Trace(err)
}
return u.stopUnitError()
}
}
}
// stopUnitError returns the error to use when exiting from stopping the unit.
// For IAAS models, we want to terminate the agent, as each unit is run by
// an individual agent for that unit.
func (u *Uniter) stopUnitError() error {
u.logger.Debugf("u.modelType: %s", u.modelType)
if u.modelType == model.CAAS && !u.sidecar {
return ErrCAASUnitDead
}
return jworker.ErrTerminateAgent
}
func (u *Uniter) init(unitTag names.UnitTag) (err error) {
switch u.modelType {
case model.IAAS, model.CAAS:
// known types, all good
default:
return errors.Errorf("unknown model type %q", u.modelType)
}
// If we started up already dead, we should not progress further.
// If we become Dead immediately after starting up, we may well
// complete any operations in progress before detecting it,
// but that race is fundamental and inescapable,
// whereas this one is not.
u.unit, err = u.st.Unit(unitTag)
if err != nil {
if errors.IsNotFound(err) {
return u.stopUnitError()
}
return errors.Trace(err)
}
if u.unit.Life() == life.Dead {
return u.stopUnitError()
}
// If initialising for the first time after deploying, update the status.
currentStatus, err := u.unit.UnitStatus()
if err != nil {
return errors.Trace(err)
}
// TODO(fwereade/wallyworld): we should have an explicit place in the model
// to tell us when we've hit this point, instead of piggybacking on top of
// status and/or status history.
// If the previous status was waiting for machine, we transition to the next step.
if currentStatus.Status == string(status.Waiting) &&
(currentStatus.Info == status.MessageWaitForMachine || currentStatus.Info == status.MessageInstallingAgent) {
if err := u.unit.SetUnitStatus(status.Waiting, status.MessageInitializingAgent, nil); err != nil {
return errors.Trace(err)
}
}
if err := tools.EnsureSymlinks(u.paths.ToolsDir, u.paths.ToolsDir, jujuc.CommandNames()); err != nil {
return err
}
relStateTracker, err := relation.NewRelationStateTracker(
relation.RelationStateTrackerConfig{
State: u.st,
Unit: u.unit,
Tracker: u.leadershipTracker,
NewLeadershipContext: context.NewLeadershipContext,
CharmDir: u.paths.State.CharmDir,
Abort: u.catacomb.Dying(),
Logger: u.logger.Child("relation"),
})
if err != nil {
return errors.Annotatef(err, "cannot create relation state tracker")
}
u.relationStateTracker = relStateTracker
u.commands = runcommands.NewCommands()
u.commandChannel = make(chan string)
storageAttachments, err := storage.NewAttachments(
u.st, unitTag, u.unit, u.catacomb.Dying(),
)
if err != nil {
return errors.Annotatef(err, "cannot create storage hook source")
}
u.storage = storageAttachments
secretsTracker, err := secrets.NewSecrets(
u.secretsClient, unitTag, u.unit, u.logger.ChildWithLabels("secrets", corelogger.SECRETS),
)
if err != nil {
return errors.Annotatef(err, "cannot create secrets tracker")
}
u.secretsTracker = secretsTracker
if err := charm.ClearDownloads(u.paths.State.BundlesDir); err != nil {
u.logger.Warningf(err.Error())
}
charmLogger := u.logger.Child("charm")
deployer, err := u.newDeployer(
u.paths.State.CharmDir,
u.paths.State.DeployerDir,
charm.NewBundlesDir(
u.paths.State.BundlesDir,
u.downloader,
charmLogger),
charmLogger,
)
if err != nil {
return errors.Annotatef(err, "cannot create deployer")
}
contextFactory, err := context.NewContextFactory(context.FactoryConfig{
State: u.st,
SecretsClient: u.secretsClient,
SecretsStoreGetter: u.secretsStoreGetter,
Unit: u.unit,
Resources: u.resources,
Payloads: u.payloads,
Tracker: u.leadershipTracker,
GetRelationInfos: u.relationStateTracker.GetInfo,
Storage: u.storage,
Paths: u.paths,
Clock: u.clock,
Logger: u.logger.Child("context"),
})
if err != nil {
return err
}
var remoteExecutor runner.ExecFunc
if u.newRemoteRunnerExecutor != nil {
remoteExecutor = u.newRemoteRunnerExecutor(u.unit, u.paths)
}
runnerFactory, err := runner.NewFactory(
u.paths, contextFactory, u.newProcessRunner, remoteExecutor,
)
if err != nil {
return errors.Trace(err)
}
u.operationFactory = operation.NewFactory(operation.FactoryParams{
Deployer: deployer,
RunnerFactory: runnerFactory,
Callbacks: &operationCallbacks{u},
State: u.st,
Abort: u.catacomb.Dying(),
MetricSpoolDir: u.paths.GetMetricsSpoolDir(),
Logger: u.logger.Child("operation"),
})
charmURL, err := u.getApplicationCharmURL()
if err != nil {
return errors.Trace(err)
}
initialState := operation.State{
Kind: operation.Install,
Step: operation.Queued,
CharmURL: charmURL,
}
operationExecutor, err := u.newOperationExecutor(u.unit.Name(), operation.ExecutorConfig{
StateReadWriter: u.unit,
InitialState: initialState,
AcquireLock: u.acquireExecutionLock,
Logger: u.logger.Child("operation"),
})
if err != nil {
return errors.Trace(err)
}
u.operationExecutor = operationExecutor
// Ensure we have an agent directory to to write the socket.
if err := os.MkdirAll(u.paths.State.BaseDir, 0755); err != nil {
return errors.Trace(err)
}
socket := u.paths.Runtime.LocalJujuExecSocket.Server
u.logger.Debugf("starting local juju-exec listener on %v", socket)
u.localRunListener, err = NewRunListener(socket, u.logger)
if err != nil {
return errors.Annotate(err, "creating juju run listener")
}
rlw := NewRunListenerWrapper(u.localRunListener, u.logger)
if err := u.catacomb.Add(rlw); err != nil {
return errors.Trace(err)
}
commandRunner, err := NewChannelCommandRunner(ChannelCommandRunnerConfig{
Abort: u.catacomb.Dying(),
Commands: u.commands,
CommandChannel: u.commandChannel,
})
if err != nil {
return errors.Annotate(err, "creating command runner")
}
u.localRunListener.RegisterRunner(u.unit.Name(), commandRunner)
if u.runListener != nil {
u.runListener.RegisterRunner(u.unit.Name(), commandRunner)
}
u.workloadEvents = container.NewWorkloadEvents()
u.workloadEventChannel = make(chan string)
if len(u.containerNames) > 0 {
pebblePoller := NewPebblePoller(u.logger, u.clock, u.containerNames, u.workloadEventChannel, u.workloadEvents, u.newPebbleClient)
if err := u.catacomb.Add(pebblePoller); err != nil {
return errors.Trace(err)
}
}
return nil
}
func (u *Uniter) Kill() {
u.catacomb.Kill(nil)
}
func (u *Uniter) Wait() error {
return u.catacomb.Wait()
}
func (u *Uniter) getApplicationCharmURL() (string, error) {
// TODO(fwereade): pretty sure there's no reason to make 2 API calls here.
app, err := u.st.Application(u.unit.ApplicationTag())
if err != nil {
return "", err
}
charmURL, _, err := app.CharmURL()
return charmURL, err
}
// RunCommands executes the supplied commands in a hook context.
func (u *Uniter) RunCommands(args RunCommandsArgs) (results *exec.ExecResponse, err error) {
// TODO(axw) drop this when we move the run-listener to an independent
// worker. This exists purely for the tests.
return u.localRunListener.RunCommands(args)
}
// acquireExecutionLock acquires the machine-level execution lock, and
// returns a func that must be called to unlock it. It's used by operation.Executor
// when running operations that execute external code.
func (u *Uniter) acquireExecutionLock(action, executionGroup string) (func(), error) {
// We want to make sure we don't block forever when locking, but take the
// Uniter's catacomb into account.
spec := machinelock.Spec{
Cancel: u.catacomb.Dying(),
Worker: fmt.Sprintf("%s uniter", u.unit.Name()),
Comment: action,
Group: executionGroup,
}
releaser, err := u.hookLock.Acquire(spec)
if err != nil {
return nil, errors.Trace(err)
}
return releaser, nil
}
func (u *Uniter) reportHookError(hookInfo hook.Info) error {
// Set the agent status to "error". We must do this here in case the
// hook is interrupted (e.g. unit agent crashes), rather than immediately
// after attempting a runHookOp.
hookName := string(hookInfo.Kind)
hookMessage := string(hookInfo.Kind)
statusData := map[string]interface{}{}
if hookInfo.Kind.IsRelation() {
statusData["relation-id"] = hookInfo.RelationId
if hookInfo.RemoteUnit != "" {
statusData["remote-unit"] = hookInfo.RemoteUnit
}
relationName, err := u.relationStateTracker.Name(hookInfo.RelationId)