forked from argoproj/argo-workflows
/
estimator_factory.go
89 lines (81 loc) · 3.26 KB
/
estimator_factory.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
package estimation
import (
"fmt"
"time"
"k8s.io/apimachinery/pkg/apis/meta/v1/unstructured"
"k8s.io/apimachinery/pkg/labels"
"k8s.io/client-go/tools/cache"
"github.com/argoproj/argo-workflows/v3/persist/sqldb"
wfv1 "github.com/argoproj/argo-workflows/v3/pkg/apis/workflow/v1alpha1"
"github.com/argoproj/argo-workflows/v3/workflow/common"
"github.com/argoproj/argo-workflows/v3/workflow/controller/indexes"
"github.com/argoproj/argo-workflows/v3/workflow/hydrator"
"github.com/argoproj/argo-workflows/v3/workflow/util"
)
type EstimatorFactory interface {
// ALWAYS return as estimator, even if it also returns an error.
NewEstimator(wf *wfv1.Workflow) (Estimator, error)
}
type estimatorFactory struct {
wfInformer cache.SharedIndexInformer
hydrator hydrator.Interface
wfArchive sqldb.WorkflowArchive
}
var _ EstimatorFactory = &estimatorFactory{}
func NewEstimatorFactory(wfInformer cache.SharedIndexInformer, hydrator hydrator.Interface, wfArchive sqldb.WorkflowArchive) EstimatorFactory {
return &estimatorFactory{wfInformer, hydrator, wfArchive}
}
func (f *estimatorFactory) NewEstimator(wf *wfv1.Workflow) (Estimator, error) {
defaultEstimator := &estimator{wf: wf}
for labelName, indexName := range map[string]string{
common.LabelKeyWorkflowTemplate: indexes.WorkflowTemplateIndex,
common.LabelKeyClusterWorkflowTemplate: indexes.ClusterWorkflowTemplateIndex,
common.LabelKeyCronWorkflow: indexes.CronWorkflowIndex,
} {
labelValue, exists := wf.Labels[labelName]
if exists {
objs, err := f.wfInformer.GetIndexer().ByIndex(indexName, indexes.MetaNamespaceLabelIndex(wf.Namespace, labelValue))
if err != nil {
return defaultEstimator, fmt.Errorf("failed to list workflows by index: %v", err)
}
var newestUn *unstructured.Unstructured
for _, obj := range objs {
un, ok := obj.(*unstructured.Unstructured)
if !ok {
return defaultEstimator, fmt.Errorf("failed convert object to unstructured")
}
if un.GetLabels()[common.LabelKeyPhase] != string(wfv1.NodeSucceeded) {
continue
}
// we use `creationTimestamp` because it's fast
if newestUn == nil || un.GetCreationTimestamp().After(newestUn.GetCreationTimestamp().Time) {
newestUn = un
}
}
if newestUn != nil {
newestWf, err := util.FromUnstructured(newestUn)
if err != nil {
return defaultEstimator, fmt.Errorf("failed convert unstructured to workflow: %w", err)
}
err = f.hydrator.Hydrate(newestWf)
if err != nil {
return defaultEstimator, fmt.Errorf("failed hydrate last workflow: %w", err)
}
return &estimator{wf, newestWf}, nil
}
// we failed to find a base-line in the live set, so we now look in the archive
requirements, err := labels.ParseToRequirements(common.LabelKeyPhase + "=" + string(wfv1.NodeSucceeded) + "," + labelName + "=" + labelValue)
if err != nil {
return defaultEstimator, fmt.Errorf("failed to parse selector to requirements: %v", err)
}
workflows, err := f.wfArchive.ListWorkflows(wf.Namespace, time.Time{}, time.Time{}, requirements, 1, 0)
if err != nil {
return defaultEstimator, fmt.Errorf("failed to list archived workflows: %v", err)
}
if len(workflows) > 0 {
return &estimator{wf, &workflows[0]}, nil
}
}
}
return defaultEstimator, nil
}