2
0
mirror of https://github.com/hibiken/asynq.git synced 2024-11-15 11:58:47 +08:00
asynq/processor_test.go

741 lines
21 KiB
Go
Raw Normal View History

2020-01-03 10:13:16 +08:00
// Copyright 2020 Kentaro Hibino. All rights reserved.
// Use of this source code is governed by a MIT license
// that can be found in the LICENSE file.
2019-11-28 06:03:04 +08:00
package asynq
import (
"context"
2021-03-21 04:42:13 +08:00
"encoding/json"
2019-11-28 06:03:04 +08:00
"fmt"
"sort"
2019-11-30 04:48:54 +08:00
"sync"
2019-11-28 06:03:04 +08:00
"testing"
2019-11-30 04:48:54 +08:00
"time"
"github.com/google/go-cmp/cmp"
h "github.com/hibiken/asynq/internal/asynqtest"
2019-12-22 23:15:45 +08:00
"github.com/hibiken/asynq/internal/base"
2019-12-04 13:01:26 +08:00
"github.com/hibiken/asynq/internal/rdb"
2019-11-28 06:03:04 +08:00
)
2020-05-19 11:47:35 +08:00
// fakeHeartbeater receives from starting and finished channels and do nothing.
2021-01-28 07:55:43 +08:00
func fakeHeartbeater(starting <-chan *workerInfo, finished <-chan *base.TaskMessage, done <-chan struct{}) {
2020-05-19 11:47:35 +08:00
for {
select {
case <-starting:
case <-finished:
case <-done:
return
}
}
}
2020-06-12 11:58:27 +08:00
// fakeSyncer receives from sync channel and do nothing.
func fakeSyncer(syncCh <-chan *syncRequest, done <-chan struct{}) {
for {
select {
case <-syncCh:
case <-done:
return
}
}
}
2020-08-19 12:21:05 +08:00
func TestProcessorSuccessWithSingleQueue(t *testing.T) {
2019-11-30 04:48:54 +08:00
r := setup(t)
2021-06-03 21:33:21 +08:00
defer r.Close()
2019-12-04 13:01:26 +08:00
rdbClient := rdb.NewRDB(r)
2019-11-30 04:48:54 +08:00
2020-08-19 12:21:05 +08:00
m1 := h.NewTaskMessage("task1", nil)
m2 := h.NewTaskMessage("task2", nil)
m3 := h.NewTaskMessage("task3", nil)
m4 := h.NewTaskMessage("task4", nil)
2019-11-30 04:48:54 +08:00
t1 := NewTask(m1.Type, m1.Payload)
t2 := NewTask(m2.Type, m2.Payload)
t3 := NewTask(m3.Type, m3.Payload)
t4 := NewTask(m4.Type, m4.Payload)
2019-11-30 04:48:54 +08:00
tests := []struct {
2020-09-05 22:03:43 +08:00
pending []*base.TaskMessage // initial default queue state
2019-12-22 23:15:45 +08:00
incoming []*base.TaskMessage // tasks to be enqueued during run
wantProcessed []*Task // tasks to be processed at the end
2019-11-30 04:48:54 +08:00
}{
{
2020-09-05 22:03:43 +08:00
pending: []*base.TaskMessage{m1},
2019-12-22 23:15:45 +08:00
incoming: []*base.TaskMessage{m2, m3, m4},
2019-11-30 04:48:54 +08:00
wantProcessed: []*Task{t1, t2, t3, t4},
},
{
2020-09-05 22:03:43 +08:00
pending: []*base.TaskMessage{},
2019-12-22 23:15:45 +08:00
incoming: []*base.TaskMessage{m1},
2019-11-30 04:48:54 +08:00
wantProcessed: []*Task{t1},
},
}
for _, tc := range tests {
2020-09-05 22:03:43 +08:00
h.FlushDB(t, r) // clean up db before each test case.
h.SeedPendingQueue(t, r, tc.pending, base.DefaultQueueName) // initialize default queue.
2019-11-30 04:48:54 +08:00
// instantiate a new processor
var mu sync.Mutex
var processed []*Task
handler := func(ctx context.Context, task *Task) error {
2019-11-30 04:48:54 +08:00
mu.Lock()
defer mu.Unlock()
processed = append(processed, task)
return nil
}
2021-01-28 07:55:43 +08:00
starting := make(chan *workerInfo)
2020-05-19 11:47:35 +08:00
finished := make(chan *base.TaskMessage)
2020-06-12 11:58:27 +08:00
syncCh := make(chan *syncRequest)
2020-05-19 11:47:35 +08:00
done := make(chan struct{})
defer func() { close(done) }()
go fakeHeartbeater(starting, finished, done)
2020-06-12 11:58:27 +08:00
go fakeSyncer(syncCh, done)
p := newProcessor(processorParams{
logger: testLogger,
2020-04-17 21:56:44 +08:00
broker: rdbClient,
2021-01-13 03:40:26 +08:00
retryDelayFunc: DefaultRetryDelayFunc,
2020-06-12 11:58:27 +08:00
syncCh: syncCh,
2020-05-01 22:22:11 +08:00
cancelations: base.NewCancelations(),
2020-05-19 11:47:35 +08:00
concurrency: 10,
queues: defaultQueueConfig,
strictPriority: false,
errHandler: nil,
shutdownTimeout: defaultShutdownTimeout,
2020-05-19 11:47:35 +08:00
starting: starting,
finished: finished,
})
p.handler = HandlerFunc(handler)
2019-11-30 04:48:54 +08:00
2020-05-01 22:22:11 +08:00
p.start(&sync.WaitGroup{})
2019-11-30 04:48:54 +08:00
for _, msg := range tc.incoming {
2019-12-04 13:01:26 +08:00
err := rdbClient.Enqueue(msg)
2019-11-30 04:48:54 +08:00
if err != nil {
p.shutdown()
2019-11-30 04:48:54 +08:00
t.Fatal(err)
}
}
2020-09-05 22:03:43 +08:00
time.Sleep(2 * time.Second) // wait for two second to allow all pending tasks to be processed.
2020-09-06 03:43:15 +08:00
if l := r.LLen(base.ActiveKey(base.DefaultQueueName)).Val(); l != 0 {
t.Errorf("%q has %d tasks, want 0", base.ActiveKey(base.DefaultQueueName), l)
2020-06-12 11:58:27 +08:00
}
p.shutdown()
2019-11-30 04:48:54 +08:00
2020-06-06 12:23:28 +08:00
mu.Lock()
2021-03-21 04:42:13 +08:00
if diff := cmp.Diff(tc.wantProcessed, processed, sortTaskOpt, cmp.AllowUnexported(Task{})); diff != "" {
2019-11-30 04:48:54 +08:00
t.Errorf("mismatch found in processed tasks; (-want, +got)\n%s", diff)
}
2020-06-06 12:23:28 +08:00
mu.Unlock()
2020-06-12 11:58:27 +08:00
}
}
2020-08-19 12:21:05 +08:00
func TestProcessorSuccessWithMultipleQueues(t *testing.T) {
var (
r = setup(t)
rdbClient = rdb.NewRDB(r)
m1 = h.NewTaskMessage("task1", nil)
m2 = h.NewTaskMessage("task2", nil)
m3 = h.NewTaskMessageWithQueue("task3", nil, "high")
m4 = h.NewTaskMessageWithQueue("task4", nil, "low")
t1 = NewTask(m1.Type, m1.Payload)
t2 = NewTask(m2.Type, m2.Payload)
t3 = NewTask(m3.Type, m3.Payload)
t4 = NewTask(m4.Type, m4.Payload)
)
2021-06-03 21:33:21 +08:00
defer r.Close()
2020-08-19 12:21:05 +08:00
tests := []struct {
2020-09-05 22:03:43 +08:00
pending map[string][]*base.TaskMessage
2020-08-19 12:21:05 +08:00
queues []string // list of queues to consume the tasks from
wantProcessed []*Task // tasks to be processed at the end
}{
{
2020-09-05 22:03:43 +08:00
pending: map[string][]*base.TaskMessage{
2020-08-19 12:21:05 +08:00
"default": {m1, m2},
"high": {m3},
"low": {m4},
},
queues: []string{"default", "high", "low"},
wantProcessed: []*Task{t1, t2, t3, t4},
},
}
for _, tc := range tests {
// Set up test case.
h.FlushDB(t, r)
2020-09-05 22:03:43 +08:00
h.SeedAllPendingQueues(t, r, tc.pending)
2020-08-19 12:21:05 +08:00
// Instantiate a new processor.
var mu sync.Mutex
var processed []*Task
handler := func(ctx context.Context, task *Task) error {
mu.Lock()
defer mu.Unlock()
processed = append(processed, task)
return nil
}
2021-01-28 07:55:43 +08:00
starting := make(chan *workerInfo)
2020-08-19 12:21:05 +08:00
finished := make(chan *base.TaskMessage)
syncCh := make(chan *syncRequest)
done := make(chan struct{})
defer func() { close(done) }()
go fakeHeartbeater(starting, finished, done)
go fakeSyncer(syncCh, done)
p := newProcessor(processorParams{
logger: testLogger,
broker: rdbClient,
2021-01-13 03:40:26 +08:00
retryDelayFunc: DefaultRetryDelayFunc,
2020-08-19 12:21:05 +08:00
syncCh: syncCh,
cancelations: base.NewCancelations(),
concurrency: 10,
queues: map[string]int{
"default": 2,
"high": 3,
"low": 1,
},
strictPriority: false,
errHandler: nil,
shutdownTimeout: defaultShutdownTimeout,
starting: starting,
finished: finished,
})
p.handler = HandlerFunc(handler)
p.start(&sync.WaitGroup{})
2020-09-05 22:03:43 +08:00
// Wait for two second to allow all pending tasks to be processed.
2020-08-19 12:21:05 +08:00
time.Sleep(2 * time.Second)
2020-09-06 03:43:15 +08:00
// Make sure no messages are stuck in active list.
2020-08-19 12:21:05 +08:00
for _, qname := range tc.queues {
2020-09-06 03:43:15 +08:00
if l := r.LLen(base.ActiveKey(qname)).Val(); l != 0 {
t.Errorf("%q has %d tasks, want 0", base.ActiveKey(qname), l)
2020-08-19 12:21:05 +08:00
}
}
p.shutdown()
2020-08-19 12:21:05 +08:00
mu.Lock()
2021-03-21 04:42:13 +08:00
if diff := cmp.Diff(tc.wantProcessed, processed, sortTaskOpt, cmp.AllowUnexported(Task{})); diff != "" {
2020-08-19 12:21:05 +08:00
t.Errorf("mismatch found in processed tasks; (-want, +got)\n%s", diff)
}
mu.Unlock()
}
}
2020-06-12 11:58:27 +08:00
// https://github.com/hibiken/asynq/issues/166
func TestProcessTasksWithLargeNumberInPayload(t *testing.T) {
r := setup(t)
2021-06-03 21:33:21 +08:00
defer r.Close()
2020-06-12 11:58:27 +08:00
rdbClient := rdb.NewRDB(r)
2021-03-21 04:42:13 +08:00
m1 := h.NewTaskMessage("large_number", h.JSON(map[string]interface{}{"data": 111111111111111111}))
2020-06-12 11:58:27 +08:00
t1 := NewTask(m1.Type, m1.Payload)
2019-11-30 04:48:54 +08:00
2020-06-12 11:58:27 +08:00
tests := []struct {
2020-09-05 22:03:43 +08:00
pending []*base.TaskMessage // initial default queue state
2020-06-12 11:58:27 +08:00
wantProcessed []*Task // tasks to be processed at the end
}{
{
2020-09-05 22:03:43 +08:00
pending: []*base.TaskMessage{m1},
2020-06-12 11:58:27 +08:00
wantProcessed: []*Task{t1},
},
}
for _, tc := range tests {
2020-09-05 22:03:43 +08:00
h.FlushDB(t, r) // clean up db before each test case.
h.SeedPendingQueue(t, r, tc.pending, base.DefaultQueueName) // initialize default queue.
2020-06-12 11:58:27 +08:00
var mu sync.Mutex
var processed []*Task
handler := func(ctx context.Context, task *Task) error {
mu.Lock()
defer mu.Unlock()
2021-03-21 04:42:13 +08:00
var payload map[string]int
if err := json.Unmarshal(task.Payload(), &payload); err != nil {
t.Errorf("coult not decode payload: %v", err)
}
if data, ok := payload["data"]; ok {
2020-06-12 11:58:27 +08:00
t.Logf("data == %d", data)
2021-03-21 04:42:13 +08:00
} else {
t.Errorf("could not get data from payload")
2020-06-12 11:58:27 +08:00
}
processed = append(processed, task)
return nil
}
2021-01-28 07:55:43 +08:00
starting := make(chan *workerInfo)
2020-06-12 11:58:27 +08:00
finished := make(chan *base.TaskMessage)
syncCh := make(chan *syncRequest)
done := make(chan struct{})
defer func() { close(done) }()
go fakeHeartbeater(starting, finished, done)
go fakeSyncer(syncCh, done)
p := newProcessor(processorParams{
logger: testLogger,
broker: rdbClient,
2021-01-13 03:40:26 +08:00
retryDelayFunc: DefaultRetryDelayFunc,
2020-06-12 11:58:27 +08:00
syncCh: syncCh,
cancelations: base.NewCancelations(),
concurrency: 10,
queues: defaultQueueConfig,
strictPriority: false,
errHandler: nil,
shutdownTimeout: defaultShutdownTimeout,
starting: starting,
finished: finished,
})
p.handler = HandlerFunc(handler)
p.start(&sync.WaitGroup{})
2020-09-05 22:03:43 +08:00
time.Sleep(2 * time.Second) // wait for two second to allow all pending tasks to be processed.
2020-09-06 03:43:15 +08:00
if l := r.LLen(base.ActiveKey(base.DefaultQueueName)).Val(); l != 0 {
t.Errorf("%q has %d tasks, want 0", base.ActiveKey(base.DefaultQueueName), l)
2019-11-30 04:48:54 +08:00
}
p.shutdown()
2020-06-12 11:58:27 +08:00
mu.Lock()
2021-03-21 04:42:13 +08:00
if diff := cmp.Diff(tc.wantProcessed, processed, sortTaskOpt, cmp.AllowUnexported(Task{})); diff != "" {
2020-06-12 11:58:27 +08:00
t.Errorf("mismatch found in processed tasks; (-want, +got)\n%s", diff)
}
mu.Unlock()
2019-11-30 04:48:54 +08:00
}
}
func TestProcessorRetry(t *testing.T) {
r := setup(t)
2021-06-03 21:33:21 +08:00
defer r.Close()
2019-12-04 13:01:26 +08:00
rdbClient := rdb.NewRDB(r)
2019-11-30 04:48:54 +08:00
m1 := h.NewTaskMessage("send_email", nil)
2019-11-30 04:48:54 +08:00
m1.Retried = m1.Retry // m1 has reached its max retry count
m2 := h.NewTaskMessage("gen_thumbnail", nil)
m3 := h.NewTaskMessage("reindex", nil)
m4 := h.NewTaskMessage("sync", nil)
2019-11-30 04:48:54 +08:00
errMsg := "something went wrong"
wrappedSkipRetry := fmt.Errorf("%s:%w", errMsg, SkipRetry)
2019-11-30 04:48:54 +08:00
tests := []struct {
desc string // test description
2020-09-05 22:03:43 +08:00
pending []*base.TaskMessage // initial default queue state
delay time.Duration // retry delay duration
handler Handler // task handler
wait time.Duration // wait duration between starting and stopping processor for this test case
2021-06-03 21:33:21 +08:00
wantErrMsg string // error message the task should record
wantRetry []*base.TaskMessage // tasks in retry queue at the end
wantArchived []*base.TaskMessage // tasks in archived queue at the end
wantErrCount int // number of times error handler should be called
2019-11-30 04:48:54 +08:00
}{
{
2021-06-03 21:33:21 +08:00
desc: "Should automatically retry errored tasks",
pending: []*base.TaskMessage{m1, m2, m3, m4},
delay: time.Minute,
handler: HandlerFunc(func(ctx context.Context, task *Task) error {
return fmt.Errorf(errMsg)
}),
2021-06-03 21:33:21 +08:00
wait: 2 * time.Second,
wantErrMsg: errMsg,
wantRetry: []*base.TaskMessage{m2, m3, m4},
wantArchived: []*base.TaskMessage{m1},
wantErrCount: 4,
2019-11-30 04:48:54 +08:00
},
{
2021-06-03 21:33:21 +08:00
desc: "Should skip retry errored tasks",
pending: []*base.TaskMessage{m1, m2},
delay: time.Minute,
handler: HandlerFunc(func(ctx context.Context, task *Task) error {
return SkipRetry // return SkipRetry without wrapping
}),
2021-06-03 21:33:21 +08:00
wait: 2 * time.Second,
wantErrMsg: SkipRetry.Error(),
wantRetry: []*base.TaskMessage{},
wantArchived: []*base.TaskMessage{m1, m2},
wantErrCount: 2, // ErrorHandler should still be called with SkipRetry error
},
{
2021-06-03 21:33:21 +08:00
desc: "Should skip retry errored tasks (with error wrapping)",
pending: []*base.TaskMessage{m1, m2},
delay: time.Minute,
handler: HandlerFunc(func(ctx context.Context, task *Task) error {
return wrappedSkipRetry
}),
2021-06-03 21:33:21 +08:00
wait: 2 * time.Second,
wantErrMsg: wrappedSkipRetry.Error(),
wantRetry: []*base.TaskMessage{},
wantArchived: []*base.TaskMessage{m1, m2},
wantErrCount: 2, // ErrorHandler should still be called with SkipRetry error
},
2019-11-30 04:48:54 +08:00
}
for _, tc := range tests {
2020-09-05 22:03:43 +08:00
h.FlushDB(t, r) // clean up db before each test case.
h.SeedPendingQueue(t, r, tc.pending, base.DefaultQueueName) // initialize default queue.
2019-11-30 04:48:54 +08:00
// instantiate a new processor
delayFunc := func(n int, e error, t *Task) time.Duration {
return tc.delay
}
var (
mu sync.Mutex // guards n
n int // number of times error handler is called
)
2020-07-04 20:24:47 +08:00
errHandler := func(ctx context.Context, t *Task, err error) {
mu.Lock()
defer mu.Unlock()
n++
2019-11-30 04:48:54 +08:00
}
2021-01-28 07:55:43 +08:00
starting := make(chan *workerInfo)
2020-05-19 11:47:35 +08:00
finished := make(chan *base.TaskMessage)
done := make(chan struct{})
defer func() { close(done) }()
go fakeHeartbeater(starting, finished, done)
p := newProcessor(processorParams{
logger: testLogger,
2020-04-17 21:56:44 +08:00
broker: rdbClient,
retryDelayFunc: delayFunc,
syncCh: nil,
2020-05-01 22:22:11 +08:00
cancelations: base.NewCancelations(),
2020-05-19 11:47:35 +08:00
concurrency: 10,
queues: defaultQueueConfig,
strictPriority: false,
errHandler: ErrorHandlerFunc(errHandler),
shutdownTimeout: defaultShutdownTimeout,
2020-05-19 11:47:35 +08:00
starting: starting,
finished: finished,
})
p.handler = tc.handler
2019-11-30 04:48:54 +08:00
2020-05-01 22:22:11 +08:00
p.start(&sync.WaitGroup{})
2021-06-03 21:58:07 +08:00
runTime := time.Now() // time when processor is running
time.Sleep(tc.wait) // FIXME: This makes test flaky.
p.shutdown()
2019-11-30 04:48:54 +08:00
2021-06-03 21:33:21 +08:00
cmpOpt := h.EquateInt64Approx(int64(tc.wait.Seconds())) // allow up to a wait-second difference in zset score
2020-08-18 21:01:38 +08:00
gotRetry := h.GetRetryEntries(t, r, base.DefaultQueueName)
2021-06-03 21:33:21 +08:00
var wantRetry []base.Z // Note: construct wantRetry here since `LastFailedAt` and ZSCORE is relative to each test run.
for _, msg := range tc.wantRetry {
wantRetry = append(wantRetry,
base.Z{
2021-06-03 21:58:07 +08:00
Message: h.TaskMessageAfterRetry(*msg, tc.wantErrMsg, runTime),
Score: runTime.Add(tc.delay).Unix(),
2021-06-03 21:33:21 +08:00
})
}
if diff := cmp.Diff(wantRetry, gotRetry, h.SortZSetEntryOpt, cmpOpt); diff != "" {
t.Errorf("%s: mismatch found in %q after running processor; (-want, +got)\n%s", tc.desc, base.RetryKey(base.DefaultQueueName), diff)
2019-11-30 04:48:54 +08:00
}
2021-06-03 21:33:21 +08:00
gotArchived := h.GetArchivedEntries(t, r, base.DefaultQueueName)
var wantArchived []base.Z // Note: construct wantArchived here since `LastFailedAt` and ZSCORE is relative to each test run.
for _, msg := range tc.wantArchived {
wantArchived = append(wantArchived,
base.Z{
2021-06-03 21:58:07 +08:00
Message: h.TaskMessageWithError(*msg, tc.wantErrMsg, runTime),
Score: runTime.Unix(),
2021-06-03 21:33:21 +08:00
})
}
if diff := cmp.Diff(wantArchived, gotArchived, h.SortZSetEntryOpt, cmpOpt); diff != "" {
t.Errorf("%s: mismatch found in %q after running processor; (-want, +got)\n%s", tc.desc, base.ArchivedKey(base.DefaultQueueName), diff)
2019-11-30 04:48:54 +08:00
}
2020-09-06 03:43:15 +08:00
if l := r.LLen(base.ActiveKey(base.DefaultQueueName)).Val(); l != 0 {
t.Errorf("%s: %q has %d tasks, want 0", base.ActiveKey(base.DefaultQueueName), tc.desc, l)
2019-11-30 04:48:54 +08:00
}
if n != tc.wantErrCount {
2020-03-01 13:34:12 +08:00
t.Errorf("error handler was called %d times, want %d", n, tc.wantErrCount)
}
2019-11-30 04:48:54 +08:00
}
}
func TestProcessorQueues(t *testing.T) {
sortOpt := cmp.Transformer("SortStrings", func(in []string) []string {
out := append([]string(nil), in...) // Copy input to avoid mutating it
sort.Strings(out)
return out
})
tests := []struct {
queueCfg map[string]int
want []string
}{
{
queueCfg: map[string]int{
"high": 6,
"default": 3,
"low": 1,
},
want: []string{"high", "default", "low"},
},
{
queueCfg: map[string]int{
"default": 1,
},
want: []string{"default"},
},
}
for _, tc := range tests {
2021-01-28 07:55:43 +08:00
starting := make(chan *workerInfo)
2020-05-19 11:47:35 +08:00
finished := make(chan *base.TaskMessage)
done := make(chan struct{})
defer func() { close(done) }()
go fakeHeartbeater(starting, finished, done)
p := newProcessor(processorParams{
logger: testLogger,
2020-04-17 21:56:44 +08:00
broker: nil,
2021-01-13 03:40:26 +08:00
retryDelayFunc: DefaultRetryDelayFunc,
syncCh: nil,
2020-05-01 22:22:11 +08:00
cancelations: base.NewCancelations(),
2020-05-19 11:47:35 +08:00
concurrency: 10,
queues: tc.queueCfg,
strictPriority: false,
errHandler: nil,
shutdownTimeout: defaultShutdownTimeout,
2020-05-19 11:47:35 +08:00
starting: starting,
finished: finished,
})
got := p.queues()
if diff := cmp.Diff(tc.want, got, sortOpt); diff != "" {
t.Errorf("with queue config: %v\n(*processor).queues() = %v, want %v\n(-want,+got):\n%s",
tc.queueCfg, got, tc.want, diff)
}
}
}
2020-01-12 23:46:51 +08:00
func TestProcessorWithStrictPriority(t *testing.T) {
2020-08-19 12:21:05 +08:00
var (
r = setup(t)
rdbClient = rdb.NewRDB(r)
m1 = h.NewTaskMessageWithQueue("task1", nil, "critical")
m2 = h.NewTaskMessageWithQueue("task2", nil, "critical")
m3 = h.NewTaskMessageWithQueue("task3", nil, "critical")
m4 = h.NewTaskMessageWithQueue("task4", nil, base.DefaultQueueName)
m5 = h.NewTaskMessageWithQueue("task5", nil, base.DefaultQueueName)
m6 = h.NewTaskMessageWithQueue("task6", nil, "low")
m7 = h.NewTaskMessageWithQueue("task7", nil, "low")
t1 = NewTask(m1.Type, m1.Payload)
t2 = NewTask(m2.Type, m2.Payload)
t3 = NewTask(m3.Type, m3.Payload)
t4 = NewTask(m4.Type, m4.Payload)
t5 = NewTask(m5.Type, m5.Payload)
t6 = NewTask(m6.Type, m6.Payload)
t7 = NewTask(m7.Type, m7.Payload)
)
2020-09-08 21:51:01 +08:00
defer r.Close()
2020-01-12 23:46:51 +08:00
tests := []struct {
2020-09-05 22:03:43 +08:00
pending map[string][]*base.TaskMessage // initial queues state
2020-08-19 12:21:05 +08:00
queues []string // list of queues to consume tasks from
2020-01-12 23:46:51 +08:00
wait time.Duration // wait duration between starting and stopping processor for this test case
wantProcessed []*Task // tasks to be processed at the end
}{
{
2020-09-05 22:03:43 +08:00
pending: map[string][]*base.TaskMessage{
2020-01-12 23:46:51 +08:00
base.DefaultQueueName: {m4, m5},
"critical": {m1, m2, m3},
"low": {m6, m7},
},
2020-08-19 12:21:05 +08:00
queues: []string{base.DefaultQueueName, "critical", "low"},
2020-01-12 23:46:51 +08:00
wait: time.Second,
wantProcessed: []*Task{t1, t2, t3, t4, t5, t6, t7},
},
}
for _, tc := range tests {
h.FlushDB(t, r) // clean up db before each test case.
2020-09-05 22:03:43 +08:00
for qname, msgs := range tc.pending {
h.SeedPendingQueue(t, r, msgs, qname)
2020-01-12 23:46:51 +08:00
}
// instantiate a new processor
var mu sync.Mutex
var processed []*Task
handler := func(ctx context.Context, task *Task) error {
2020-01-12 23:46:51 +08:00
mu.Lock()
defer mu.Unlock()
processed = append(processed, task)
return nil
}
queueCfg := map[string]int{
2020-01-12 23:46:51 +08:00
base.DefaultQueueName: 2,
2020-08-19 12:21:05 +08:00
"critical": 3,
2020-01-12 23:46:51 +08:00
"low": 1,
}
2021-01-28 07:55:43 +08:00
starting := make(chan *workerInfo)
2020-05-19 11:47:35 +08:00
finished := make(chan *base.TaskMessage)
2020-08-19 12:21:05 +08:00
syncCh := make(chan *syncRequest)
2020-05-19 11:47:35 +08:00
done := make(chan struct{})
defer func() { close(done) }()
go fakeHeartbeater(starting, finished, done)
2020-08-19 12:21:05 +08:00
go fakeSyncer(syncCh, done)
p := newProcessor(processorParams{
logger: testLogger,
2020-04-17 21:56:44 +08:00
broker: rdbClient,
2021-01-13 03:40:26 +08:00
retryDelayFunc: DefaultRetryDelayFunc,
2020-08-19 12:21:05 +08:00
syncCh: syncCh,
2020-05-01 22:22:11 +08:00
cancelations: base.NewCancelations(),
2020-05-19 11:47:35 +08:00
concurrency: 1, // Set concurrency to 1 to make sure tasks are processed one at a time.
queues: queueCfg,
strictPriority: true,
errHandler: nil,
shutdownTimeout: defaultShutdownTimeout,
2020-05-19 11:47:35 +08:00
starting: starting,
finished: finished,
})
2020-01-12 23:46:51 +08:00
p.handler = HandlerFunc(handler)
2020-05-01 22:22:11 +08:00
p.start(&sync.WaitGroup{})
2020-01-12 23:46:51 +08:00
time.Sleep(tc.wait)
2020-09-06 03:43:15 +08:00
// Make sure no tasks are stuck in active list.
2020-08-19 12:21:05 +08:00
for _, qname := range tc.queues {
2020-09-06 03:43:15 +08:00
if l := r.LLen(base.ActiveKey(qname)).Val(); l != 0 {
t.Errorf("%q has %d tasks, want 0", base.ActiveKey(qname), l)
2020-08-19 12:21:05 +08:00
}
}
p.shutdown()
2020-01-12 23:46:51 +08:00
2021-03-21 04:42:13 +08:00
if diff := cmp.Diff(tc.wantProcessed, processed, sortTaskOpt, cmp.AllowUnexported(Task{})); diff != "" {
2020-01-12 23:46:51 +08:00
t.Errorf("mismatch found in processed tasks; (-want, +got)\n%s", diff)
}
}
}
func TestProcessorPerform(t *testing.T) {
2019-11-28 06:03:04 +08:00
tests := []struct {
desc string
handler HandlerFunc
2019-11-28 06:03:04 +08:00
task *Task
wantErr bool
}{
{
desc: "handler returns nil",
handler: func(ctx context.Context, t *Task) error {
2019-11-28 06:03:04 +08:00
return nil
},
2021-03-21 04:42:13 +08:00
task: NewTask("gen_thumbnail", h.JSON(map[string]interface{}{"src": "some/img/path"})),
2019-11-28 06:03:04 +08:00
wantErr: false,
},
{
desc: "handler returns error",
handler: func(ctx context.Context, t *Task) error {
2019-11-28 06:03:04 +08:00
return fmt.Errorf("something went wrong")
},
2021-03-21 04:42:13 +08:00
task: NewTask("gen_thumbnail", h.JSON(map[string]interface{}{"src": "some/img/path"})),
2019-11-28 06:03:04 +08:00
wantErr: true,
},
{
desc: "handler panics",
handler: func(ctx context.Context, t *Task) error {
2019-11-28 06:03:04 +08:00
panic("something went terribly wrong")
},
2021-03-21 04:42:13 +08:00
task: NewTask("gen_thumbnail", h.JSON(map[string]interface{}{"src": "some/img/path"})),
2019-11-28 06:03:04 +08:00
wantErr: true,
},
}
// Note: We don't need to fully initialize the processor since we are only testing
// perform method.
p := newProcessor(processorParams{
logger: testLogger,
queues: defaultQueueConfig,
})
2019-11-28 06:03:04 +08:00
for _, tc := range tests {
p.handler = tc.handler
got := p.perform(context.Background(), tc.task)
2019-11-28 06:03:04 +08:00
if !tc.wantErr && got != nil {
t.Errorf("%s: perform() = %v, want nil", tc.desc, got)
continue
}
if tc.wantErr && got == nil {
t.Errorf("%s: perform() = nil, want non-nil error", tc.desc)
continue
}
}
}
2020-04-16 00:02:28 +08:00
func TestGCD(t *testing.T) {
tests := []struct {
input []int
want int
}{
{[]int{6, 2, 12}, 2},
{[]int{3, 3, 3}, 3},
{[]int{6, 3, 1}, 1},
{[]int{1}, 1},
{[]int{1, 0, 2}, 1},
{[]int{8, 0, 4}, 4},
{[]int{9, 12, 18, 30}, 3},
}
for _, tc := range tests {
got := gcd(tc.input...)
if got != tc.want {
t.Errorf("gcd(%v) = %d, want %d", tc.input, got, tc.want)
}
}
}
2020-05-19 11:47:35 +08:00
func TestNormalizeQueues(t *testing.T) {
2020-04-16 00:02:28 +08:00
tests := []struct {
input map[string]int
want map[string]int
}{
{
input: map[string]int{
"high": 100,
"default": 20,
"low": 5,
},
want: map[string]int{
"high": 20,
"default": 4,
"low": 1,
},
},
{
input: map[string]int{
"default": 10,
},
want: map[string]int{
"default": 1,
},
},
{
input: map[string]int{
"critical": 5,
"default": 1,
},
want: map[string]int{
"critical": 5,
"default": 1,
},
},
{
input: map[string]int{
"critical": 6,
"default": 3,
"low": 0,
},
want: map[string]int{
"critical": 2,
"default": 1,
"low": 0,
},
},
}
for _, tc := range tests {
2020-05-19 11:47:35 +08:00
got := normalizeQueues(tc.input)
2020-04-16 00:02:28 +08:00
if diff := cmp.Diff(tc.want, got); diff != "" {
2020-05-19 11:47:35 +08:00
t.Errorf("normalizeQueues(%v) = %v, want %v; (-want, +got):\n%s",
2020-04-16 00:02:28 +08:00
tc.input, got, tc.want, diff)
}
}
}