forked from coreos/fleet
-
Notifications
You must be signed in to change notification settings - Fork 1
/
engine.go
131 lines (101 loc) · 3.06 KB
/
engine.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
package engine
import (
"errors"
log "github.com/coreos/fleet/third_party/github.com/golang/glog"
"github.com/coreos/fleet/event"
"github.com/coreos/fleet/job"
"github.com/coreos/fleet/machine"
"github.com/coreos/fleet/registry"
)
type Engine struct {
registry *registry.Registry
eStream *registry.EventStream
eBus *event.EventBus
machine *machine.Machine
// keeps a picture of the load in the cluster for more intelligent scheduling
clust *cluster
stop chan bool
}
func New(reg *registry.Registry, eStream *registry.EventStream, mach *machine.Machine) *Engine {
eBus := event.NewEventBus()
e := &Engine{reg, eStream, eBus, mach, newCluster(), nil}
hdlr := NewEventHandler(e)
eBus.AddListener("engine", hdlr)
return e
}
func (e *Engine) Run() {
e.stop = make(chan bool)
go e.eBus.Listen(e.stop)
go e.eStream.Stream(0, e.eBus.Channel, e.stop)
}
func (e *Engine) Stop() {
log.Info("Stopping Engine")
close(e.stop)
}
func (e *Engine) GetJobsScheduledToMachine(machID string) []job.Job {
var jobs []job.Job
for _, j := range e.registry.GetAllJobs() {
tgt := e.registry.GetJobTarget(j.Name)
if tgt == "" || tgt != machID {
continue
}
jobs = append(jobs, j)
}
return jobs
}
func (e *Engine) OfferJob(j job.Job) error {
log.V(1).Infof("Attempting to lock Job(%s)", j.Name)
mutex := e.lockJob(j.Name)
if mutex == nil {
log.V(1).Infof("Could not lock Job(%s)", j.Name)
return errors.New("Could not lock Job")
}
defer mutex.Unlock()
log.V(1).Infof("Claimed Job(%s)", j.Name)
machineIDs, err := e.partitionCluster(&j)
if err != nil {
log.Errorf("Failed partitioning cluster for Job(%s): %v", j.Name, err)
return err
}
offer := job.NewOfferFromJob(j, machineIDs)
err = e.registry.CreateJobOffer(offer)
if err == nil {
log.Infof("Published JobOffer(%s)", offer.Job.Name)
}
return err
}
func (e *Engine) ResolveJobOffer(jobName string, machID string) error {
log.V(1).Infof("Attempting to lock JobOffer(%s)", jobName)
mutex := e.lockJobOffer(jobName)
if mutex == nil {
log.V(1).Infof("Could not lock JobOffer(%s)", jobName)
return errors.New("Could not lock JobOffer")
}
defer mutex.Unlock()
log.V(1).Infof("Claimed JobOffer(%s)", jobName)
err := e.registry.ResolveJobOffer(jobName)
if err != nil {
log.Errorf("Failed resolving JobOffer(%s): %v", jobName, err)
return err
}
err = e.registry.ScheduleJob(jobName, machID)
if err != nil {
log.Errorf("Failed scheduling Job(%s): %v", jobName, err)
return err
}
log.Infof("Scheduled Job(%s) to Machine(%s)", jobName, machID)
return nil
}
func (e *Engine) RemoveUnitState(jobName string) {
e.registry.RemoveUnitState(jobName)
}
func (e *Engine) lockJobOffer(jobName string) *registry.TimedResourceMutex {
return e.registry.LockJobOffer(jobName, e.machine.State().ID)
}
func (e *Engine) lockJob(jobName string) *registry.TimedResourceMutex {
return e.registry.LockJob(jobName, e.machine.State().ID)
}
// Pass-through to Registry.LockMachine
func (e *Engine) LockMachine(machID string) *registry.TimedResourceMutex {
return e.registry.LockMachine(machID, e.machine.State().ID)
}