/
juggler.rb
182 lines (141 loc) · 3.29 KB
/
juggler.rb
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
$:.unshift File.join(File.dirname(__FILE__),'..','lib')
require 'rubygems'
require 'bundler'
require 'digest/sha1'
require 'json'
Bundler.require
require 'juggler/cli'
module Juggler
def self.redis
@redis ||= Redis.new(:host => @config['redis_host'], :port => @config['redis_port'])
end
def self.config
@config
end
DEFAULT_OPTIONS =
{
'redis_host' => 'localhost',
'redis_port' => 6379
}.freeze
def self.configure(options)
@config = DEFAULT_OPTIONS.merge(options)
AWS::S3::Base.establish_connection!(
:access_key_id => @config['access_key_id'],
:secret_access_key => @config['secret_access_key']
)
end
def self.queue_bucket
AWS::S3::Bucket.find(self.config['queue_bucket_name'])
end
def self.processed_bucket
AWS::S3::Bucket.find(self.config['processed_bucket_name'])
end
def self.queue
queue_bucket.objects.reject {|o| o.path.include?("---LOCKED---")}
end
def self.processor=(processor)
@processor = processor
end
def self.processor
@processor
end
class Worker
def self.run
new.run
end
def initialize
end
def run
Juggler.queue.each do |object|
Juggler::Job.run(object)
end
end
end
class Job
def self.run(object)
job = new(object)
job.run
job
end
def initialize(object)
@object = lock(object)
self.status = 'queued'
end
def run
if (result = perform(StringIO.new(@object.value)))
cleanup @object
to_write = result.is_a?(Array) ? result : [ result ]
to_write.each do |io|
AWS::S3::S3Object.store(sha1(io), io, Juggler.config['processed_bucket_name'])
end
self.status = 'completed'
else
self.status = 'failed'
end
end
def id
Digest::SHA1.hexdigest(@object.value).to_s
end
def status=(value)
Juggler.redis.hset 'juggler.status', self.id, value
end
def status
Juggler.redis.hget 'juggler.status', self.id
end
def progress=(value)
Juggler.redis.hset 'juggler.progress', self.id, value
end
def progress
(Juggler.redis.hget 'juggler.progress', self.id).to_f
end
def set_data(key, value)
Juggler.redis.hset 'juggler.data', self.id, JSON.dump(data.merge(key => value))
end
def data
str = Juggler.redis.hget 'juggler.data', self.id
str ? JSON.parse(str) : {}
end
def sha1(io)
sha1 = Digest::SHA1.new
counter = 0
while (!io.eof)
buffer = io.readpartial(4096)
sha1.update(buffer)
end
return sha1.hexdigest
end
def lock(object)
parts = File.split(object.path)
name = "---LOCKED---#{parts.last}"
object.rename(name)
Juggler.queue_bucket[name]
end
def perform(io)
Juggler.processor.new(self).run(io)
end
def cleanup(object)
object.delete
end
end
class Processor
def initialize(job)
@job = job
end
attr_reader :job
def progress(value)
job.progress = value
end
def set(key, value)
job.set_data key, value
end
def get(key)
job.data[key]
end
end
class PassthroughProcessor < Processor
def run(io)
io
end
end
end
Juggler.processor = Juggler::PassthroughProcessor