goworker is a Resque-compatible, Go-based background worker. It allows you to push jobs into a queue using an expressive language like Ruby while harnessing the efficiency and concurrency of Go to minimize job latency and cost.
goworker workers can run alongside Ruby Resque clients so that you can keep all but your most resource-intensive jobs in Ruby.
Forked from https://github.com/benmanns/goworker and includes some commits from other forks. The following modifications have been added to this fork:
- Configuration parameters for initialization
- Removed usage of command line arguements / flags
- Enqueue method which adds jobs to the queue featuring a dedupe option
To install goworker, use
go get github.com/kylebush/goworker
to install the package, and then from your worker
import "github.com/kylebush/goworker"
To create a worker, write a function matching the signature
func(string, ...interface{}) error
and register it using
goworker.Register("MyClass", myFunc)
Here is a simple worker that prints its arguments:
package main
import (
"fmt"
"github.com/kylebush/goworker"
)
func myFunc(queue string, args ...interface{}) error {
fmt.Printf("From %s, %v", queue, args)
return
}
func init() {
goworker.Register("MyClass", myFunc)
}
func main() {
if err := goworker.Work(); err != nil {
fmt.Println("Error:", err)
}
}
To create workers that share a database pool or other resources, use a closure to share variables.
package main
import (
"github.com/kylebush/goworker"
)
func newMyFunc(uri string) {
foo := NewFoo(uri)
return func(queue string, args ...interface{}) error {
foo.Bar(args)
return nil
}
}
func init() {
goworker.Register("MyClass", newMyFunc())
}
func main() {
if err := goworker.Work(); err != nil {
fmt.Println("Error:", err)
}
}
goworker worker functions receive the queue they are serving and a slice of interfaces. To use them as parameters to other functions, use Go type assertions to convert them into usable types.
// Expecting (int, string, float64)
func myFunc(queue, args ...interface{}) error {
id, ok := args[0].(int)
if !ok {
return errorInvalidParam
}
name, ok := args[1].(string)
if !ok {
return errorInvalidParam
}
weight, ok := args[2].(float64)
if !ok {
return errorInvalidParam
}
doSomething(id, name, weight)
return nil
}
You can enqueue jobs and optionally specify that those jobs are deduped - only allowing unique jobs to be added to the queue:
args := make([]interface{}, 2)
args[0] = "hi"
args[1] = "there"
goworker.Enqueue("myqueue", "MyClass", args, true)
will enqueue a job with arguements hi
and there
for the MyClass
worker onto the myqueue
queue. Job will be enqueued only if the job for MyClass
worker with arguements hi
and there
do not already exist in the queue.
For testing, it is helpful to use the redis-cli
program to insert jobs onto the Redis queue:
redis-cli -r 100 RPUSH resque:queue:myqueue '{"class":"MyClass","args":["hi","there"]}'
will insert 100 jobs for the MyClass
worker onto the myqueue
queue. It is equivalent to:
class MyClass
@queue = :myqueue
end
100.times do
Resque.enqueue MyClass, ['hi', 'there']
end
There are several options which control the operation of the goworker client.
-queues="comma,delimited,queues"
— This is the only required flag. The recommended practice is to separate your Resque workers from your goworkers with different queues. Otherwise, Resque worker classes that have no goworker analog will cause the goworker process to fail the jobs. Because of this, there is no default queue, nor is there a way to select all queues (à la Resque's*
queue). If you have multiple queues you can assign them weights. A queue with a weight of 2 will be checked twice as often as a queue with a weight of 1:-queues='high=2,low=1'
.-interval=5.0
— Specifies the wait period between polling if no job was in the queue the last time one was requested.-concurrency=25
— Specifies the number of concurrently executing workers. This number can be as low as 1 or rather comfortably as high as 100,000, and should be tuned to your workflow and the availability of outside resources.-connections=2
— Specifies the maximum number of Redis connections that goworker will consume between the poller and all workers. There is not much performance gain over two and a slight penalty when using only one. This is configurable in case you need to keep connection counts low for cloud Redis providers who limit plans onmaxclients
.-uri=redis://localhost:6379/
— Specifies the URI of the Redis database from which goworker polls for jobs. Accepts URIs of the formatredis://user:pass@host:port/db
orunix:///path/to/redis.sock
.-namespace=resque:
— Specifies the namespace from which goworker retrieves jobs and stores stats on workers.-exit-on-complete=false
— Exits goworker when there are no jobs left in the queue. This is helpful in conjunction with thetime
command to benchmark different configurations.
You can configure parameters using the Configure
method:
goworker.Configure(map[string]string{
"queues": "myqueue1,myqueue2",
"interval": "0.1",
"concurrency": "5",
"connections": "2",
"uri": "redis://localhost:6379/",
"namespace": "yourNamespace:"})
To stop goworker, send a QUIT
, TERM
, or INT
signal to the process. This will immediately stop job polling. There can be up to $CONCURRENCY
jobs currently running, which will continue to run until they are finished.
Like Resque, goworker makes no guarantees about the safety of jobs in the event of process shutdown. Workers must be both idempotent and tolerant to loss of the job in the event of failure.
If the process is killed with a KILL
or by a system failure, there may be one job that is currently in the poller's buffer that will be lost without any representation in either the queue or the worker variable.
If you are running goworker on a system like Heroku, which sends a TERM
to signal a process that it needs to stop, ten seconds later sends a KILL
to force the process to stop, your jobs must finish within 10 seconds or they may be lost. Jobs will be recoverable from the Redis database under
resque:worker:<hostname>:<process-id>-<worker-id>:<queues>
as a JSON object with keys queue
, run_at
, and payload
, but the process is manual. Additionally, there is no guarantee that the job in Redis under the worker key has not finished, if the process is killed before goworker can flush the update to Redis.