mirror of
https://github.com/hibiken/asynq.git
synced 2024-11-14 19:38:49 +08:00
50e7f38365
- Renamed previously called scheduler to forwarder to resolve name conflicts
374 lines
13 KiB
Go
374 lines
13 KiB
Go
// Copyright 2020 Kentaro Hibino. All rights reserved.
|
|
// Use of this source code is governed by a MIT license
|
|
// that can be found in the LICENSE file.
|
|
|
|
// Package asynqtest defines test helpers for asynq and its internal packages.
|
|
package asynqtest
|
|
|
|
import (
|
|
"encoding/json"
|
|
"math"
|
|
"sort"
|
|
"testing"
|
|
|
|
"github.com/go-redis/redis/v7"
|
|
"github.com/google/go-cmp/cmp"
|
|
"github.com/google/go-cmp/cmp/cmpopts"
|
|
"github.com/google/uuid"
|
|
"github.com/hibiken/asynq/internal/base"
|
|
)
|
|
|
|
// EquateInt64Approx returns a Comparer option that treats int64 values
|
|
// to be equal if they are within the given margin.
|
|
func EquateInt64Approx(margin int64) cmp.Option {
|
|
return cmp.Comparer(func(a, b int64) bool {
|
|
return math.Abs(float64(a-b)) <= float64(margin)
|
|
})
|
|
}
|
|
|
|
// SortMsgOpt is a cmp.Option to sort base.TaskMessage for comparing slice of task messages.
|
|
var SortMsgOpt = cmp.Transformer("SortTaskMessages", func(in []*base.TaskMessage) []*base.TaskMessage {
|
|
out := append([]*base.TaskMessage(nil), in...) // Copy input to avoid mutating it
|
|
sort.Slice(out, func(i, j int) bool {
|
|
return out[i].ID.String() < out[j].ID.String()
|
|
})
|
|
return out
|
|
})
|
|
|
|
// SortZSetEntryOpt is an cmp.Option to sort ZSetEntry for comparing slice of zset entries.
|
|
var SortZSetEntryOpt = cmp.Transformer("SortZSetEntries", func(in []base.Z) []base.Z {
|
|
out := append([]base.Z(nil), in...) // Copy input to avoid mutating it
|
|
sort.Slice(out, func(i, j int) bool {
|
|
return out[i].Message.ID.String() < out[j].Message.ID.String()
|
|
})
|
|
return out
|
|
})
|
|
|
|
// SortServerInfoOpt is a cmp.Option to sort base.ServerInfo for comparing slice of process info.
|
|
var SortServerInfoOpt = cmp.Transformer("SortServerInfo", func(in []*base.ServerInfo) []*base.ServerInfo {
|
|
out := append([]*base.ServerInfo(nil), in...) // Copy input to avoid mutating it
|
|
sort.Slice(out, func(i, j int) bool {
|
|
if out[i].Host != out[j].Host {
|
|
return out[i].Host < out[j].Host
|
|
}
|
|
return out[i].PID < out[j].PID
|
|
})
|
|
return out
|
|
})
|
|
|
|
// SortWorkerInfoOpt is a cmp.Option to sort base.WorkerInfo for comparing slice of worker info.
|
|
var SortWorkerInfoOpt = cmp.Transformer("SortWorkerInfo", func(in []*base.WorkerInfo) []*base.WorkerInfo {
|
|
out := append([]*base.WorkerInfo(nil), in...) // Copy input to avoid mutating it
|
|
sort.Slice(out, func(i, j int) bool {
|
|
return out[i].ID < out[j].ID
|
|
})
|
|
return out
|
|
})
|
|
|
|
// SortSchedulerEntryOpt is a cmp.Option to sort base.SchedulerEntry for comparing slice of entries.
|
|
var SortSchedulerEntryOpt = cmp.Transformer("SortSchedulerEntry", func(in []*base.SchedulerEntry) []*base.SchedulerEntry {
|
|
out := append([]*base.SchedulerEntry(nil), in...) // Copy input to avoid mutating it
|
|
sort.Slice(out, func(i, j int) bool {
|
|
return out[i].Spec < out[j].Spec
|
|
})
|
|
return out
|
|
})
|
|
|
|
// SortSchedulerEnqueueEventOpt is a cmp.Option to sort base.SchedulerEnqueueEvent for comparing slice of events.
|
|
var SortSchedulerEnqueueEventOpt = cmp.Transformer("SortSchedulerEnqueueEvent", func(in []*base.SchedulerEnqueueEvent) []*base.SchedulerEnqueueEvent {
|
|
out := append([]*base.SchedulerEnqueueEvent(nil), in...)
|
|
sort.Slice(out, func(i, j int) bool {
|
|
return out[i].EnqueuedAt.Unix() < out[j].EnqueuedAt.Unix()
|
|
})
|
|
return out
|
|
})
|
|
|
|
// SortStringSliceOpt is a cmp.Option to sort string slice.
|
|
var SortStringSliceOpt = cmp.Transformer("SortStringSlice", func(in []string) []string {
|
|
out := append([]string(nil), in...)
|
|
sort.Strings(out)
|
|
return out
|
|
})
|
|
|
|
// IgnoreIDOpt is an cmp.Option to ignore ID field in task messages when comparing.
|
|
var IgnoreIDOpt = cmpopts.IgnoreFields(base.TaskMessage{}, "ID")
|
|
|
|
// NewTaskMessage returns a new instance of TaskMessage given a task type and payload.
|
|
func NewTaskMessage(taskType string, payload map[string]interface{}) *base.TaskMessage {
|
|
return NewTaskMessageWithQueue(taskType, payload, base.DefaultQueueName)
|
|
}
|
|
|
|
// NewTaskMessageWithQueue returns a new instance of TaskMessage given a
|
|
// task type, payload and queue name.
|
|
func NewTaskMessageWithQueue(taskType string, payload map[string]interface{}, qname string) *base.TaskMessage {
|
|
return &base.TaskMessage{
|
|
ID: uuid.New(),
|
|
Type: taskType,
|
|
Queue: qname,
|
|
Retry: 25,
|
|
Payload: payload,
|
|
Timeout: 1800, // default timeout of 30 mins
|
|
Deadline: 0, // no deadline
|
|
}
|
|
}
|
|
|
|
// TaskMessageAfterRetry returns an updated copy of t after retry.
|
|
// It increments retry count and sets the error message.
|
|
func TaskMessageAfterRetry(t base.TaskMessage, errMsg string) *base.TaskMessage {
|
|
t.Retried = t.Retried + 1
|
|
t.ErrorMsg = errMsg
|
|
return &t
|
|
}
|
|
|
|
// TaskMessageWithError returns an updated copy of t with the given error message.
|
|
func TaskMessageWithError(t base.TaskMessage, errMsg string) *base.TaskMessage {
|
|
t.ErrorMsg = errMsg
|
|
return &t
|
|
}
|
|
|
|
// MustMarshal marshals given task message and returns a json string.
|
|
// Calling test will fail if marshaling errors out.
|
|
func MustMarshal(tb testing.TB, msg *base.TaskMessage) string {
|
|
tb.Helper()
|
|
data, err := json.Marshal(msg)
|
|
if err != nil {
|
|
tb.Fatal(err)
|
|
}
|
|
return string(data)
|
|
}
|
|
|
|
// MustUnmarshal unmarshals given string into task message struct.
|
|
// Calling test will fail if unmarshaling errors out.
|
|
func MustUnmarshal(tb testing.TB, data string) *base.TaskMessage {
|
|
tb.Helper()
|
|
var msg base.TaskMessage
|
|
err := json.Unmarshal([]byte(data), &msg)
|
|
if err != nil {
|
|
tb.Fatal(err)
|
|
}
|
|
return &msg
|
|
}
|
|
|
|
// MustMarshalSlice marshals a slice of task messages and return a slice of
|
|
// json strings. Calling test will fail if marshaling errors out.
|
|
func MustMarshalSlice(tb testing.TB, msgs []*base.TaskMessage) []string {
|
|
tb.Helper()
|
|
var data []string
|
|
for _, m := range msgs {
|
|
data = append(data, MustMarshal(tb, m))
|
|
}
|
|
return data
|
|
}
|
|
|
|
// MustUnmarshalSlice unmarshals a slice of strings into a slice of task message structs.
|
|
// Calling test will fail if marshaling errors out.
|
|
func MustUnmarshalSlice(tb testing.TB, data []string) []*base.TaskMessage {
|
|
tb.Helper()
|
|
var msgs []*base.TaskMessage
|
|
for _, s := range data {
|
|
msgs = append(msgs, MustUnmarshal(tb, s))
|
|
}
|
|
return msgs
|
|
}
|
|
|
|
// FlushDB deletes all the keys of the currently selected DB.
|
|
func FlushDB(tb testing.TB, r redis.UniversalClient) {
|
|
tb.Helper()
|
|
switch r := r.(type) {
|
|
case *redis.Client:
|
|
if err := r.FlushDB().Err(); err != nil {
|
|
tb.Fatal(err)
|
|
}
|
|
case *redis.ClusterClient:
|
|
err := r.ForEachMaster(func(c *redis.Client) error {
|
|
if err := c.FlushAll().Err(); err != nil {
|
|
return err
|
|
}
|
|
return nil
|
|
})
|
|
if err != nil {
|
|
tb.Fatal(err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// SeedPendingQueue initializes the specified queue with the given messages.
|
|
func SeedPendingQueue(tb testing.TB, r redis.UniversalClient, msgs []*base.TaskMessage, qname string) {
|
|
tb.Helper()
|
|
r.SAdd(base.AllQueues, qname)
|
|
seedRedisList(tb, r, base.QueueKey(qname), msgs)
|
|
}
|
|
|
|
// SeedActiveQueue initializes the active queue with the given messages.
|
|
func SeedActiveQueue(tb testing.TB, r redis.UniversalClient, msgs []*base.TaskMessage, qname string) {
|
|
tb.Helper()
|
|
r.SAdd(base.AllQueues, qname)
|
|
seedRedisList(tb, r, base.ActiveKey(qname), msgs)
|
|
}
|
|
|
|
// SeedScheduledQueue initializes the scheduled queue with the given messages.
|
|
func SeedScheduledQueue(tb testing.TB, r redis.UniversalClient, entries []base.Z, qname string) {
|
|
tb.Helper()
|
|
r.SAdd(base.AllQueues, qname)
|
|
seedRedisZSet(tb, r, base.ScheduledKey(qname), entries)
|
|
}
|
|
|
|
// SeedRetryQueue initializes the retry queue with the given messages.
|
|
func SeedRetryQueue(tb testing.TB, r redis.UniversalClient, entries []base.Z, qname string) {
|
|
tb.Helper()
|
|
r.SAdd(base.AllQueues, qname)
|
|
seedRedisZSet(tb, r, base.RetryKey(qname), entries)
|
|
}
|
|
|
|
// SeedDeadQueue initializes the dead queue with the given messages.
|
|
func SeedDeadQueue(tb testing.TB, r redis.UniversalClient, entries []base.Z, qname string) {
|
|
tb.Helper()
|
|
r.SAdd(base.AllQueues, qname)
|
|
seedRedisZSet(tb, r, base.DeadKey(qname), entries)
|
|
}
|
|
|
|
// SeedDeadlines initializes the deadlines set with the given entries.
|
|
func SeedDeadlines(tb testing.TB, r redis.UniversalClient, entries []base.Z, qname string) {
|
|
tb.Helper()
|
|
r.SAdd(base.AllQueues, qname)
|
|
seedRedisZSet(tb, r, base.DeadlinesKey(qname), entries)
|
|
}
|
|
|
|
// SeedAllPendingQueues initializes all of the specified queues with the given messages.
|
|
//
|
|
// pending maps a queue name to a list of messages.
|
|
func SeedAllPendingQueues(tb testing.TB, r redis.UniversalClient, pending map[string][]*base.TaskMessage) {
|
|
for q, msgs := range pending {
|
|
SeedPendingQueue(tb, r, msgs, q)
|
|
}
|
|
}
|
|
|
|
// SeedAllActiveQueues initializes all of the specified active queues with the given messages.
|
|
func SeedAllActiveQueues(tb testing.TB, r redis.UniversalClient, active map[string][]*base.TaskMessage) {
|
|
for q, msgs := range active {
|
|
SeedActiveQueue(tb, r, msgs, q)
|
|
}
|
|
}
|
|
|
|
// SeedAllScheduledQueues initializes all of the specified scheduled queues with the given entries.
|
|
func SeedAllScheduledQueues(tb testing.TB, r redis.UniversalClient, scheduled map[string][]base.Z) {
|
|
for q, entries := range scheduled {
|
|
SeedScheduledQueue(tb, r, entries, q)
|
|
}
|
|
}
|
|
|
|
// SeedAllRetryQueues initializes all of the specified retry queues with the given entries.
|
|
func SeedAllRetryQueues(tb testing.TB, r redis.UniversalClient, retry map[string][]base.Z) {
|
|
for q, entries := range retry {
|
|
SeedRetryQueue(tb, r, entries, q)
|
|
}
|
|
}
|
|
|
|
// SeedAllDeadQueues initializes all of the specified dead queues with the given entries.
|
|
func SeedAllDeadQueues(tb testing.TB, r redis.UniversalClient, dead map[string][]base.Z) {
|
|
for q, entries := range dead {
|
|
SeedDeadQueue(tb, r, entries, q)
|
|
}
|
|
}
|
|
|
|
// SeedAllDeadlines initializes all of the deadlines with the given entries.
|
|
func SeedAllDeadlines(tb testing.TB, r redis.UniversalClient, deadlines map[string][]base.Z) {
|
|
for q, entries := range deadlines {
|
|
SeedDeadlines(tb, r, entries, q)
|
|
}
|
|
}
|
|
|
|
func seedRedisList(tb testing.TB, c redis.UniversalClient, key string, msgs []*base.TaskMessage) {
|
|
data := MustMarshalSlice(tb, msgs)
|
|
for _, s := range data {
|
|
if err := c.LPush(key, s).Err(); err != nil {
|
|
tb.Fatal(err)
|
|
}
|
|
}
|
|
}
|
|
|
|
func seedRedisZSet(tb testing.TB, c redis.UniversalClient, key string, items []base.Z) {
|
|
for _, item := range items {
|
|
z := &redis.Z{Member: MustMarshal(tb, item.Message), Score: float64(item.Score)}
|
|
if err := c.ZAdd(key, z).Err(); err != nil {
|
|
tb.Fatal(err)
|
|
}
|
|
}
|
|
}
|
|
|
|
// GetPendingMessages returns all pending messages in the given queue.
|
|
func GetPendingMessages(tb testing.TB, r redis.UniversalClient, qname string) []*base.TaskMessage {
|
|
tb.Helper()
|
|
return getListMessages(tb, r, base.QueueKey(qname))
|
|
}
|
|
|
|
// GetActiveMessages returns all active messages in the given queue.
|
|
func GetActiveMessages(tb testing.TB, r redis.UniversalClient, qname string) []*base.TaskMessage {
|
|
tb.Helper()
|
|
return getListMessages(tb, r, base.ActiveKey(qname))
|
|
}
|
|
|
|
// GetScheduledMessages returns all scheduled task messages in the given queue.
|
|
func GetScheduledMessages(tb testing.TB, r redis.UniversalClient, qname string) []*base.TaskMessage {
|
|
tb.Helper()
|
|
return getZSetMessages(tb, r, base.ScheduledKey(qname))
|
|
}
|
|
|
|
// GetRetryMessages returns all retry messages in the given queue.
|
|
func GetRetryMessages(tb testing.TB, r redis.UniversalClient, qname string) []*base.TaskMessage {
|
|
tb.Helper()
|
|
return getZSetMessages(tb, r, base.RetryKey(qname))
|
|
}
|
|
|
|
// GetDeadMessages returns all dead messages in the given queue.
|
|
func GetDeadMessages(tb testing.TB, r redis.UniversalClient, qname string) []*base.TaskMessage {
|
|
tb.Helper()
|
|
return getZSetMessages(tb, r, base.DeadKey(qname))
|
|
}
|
|
|
|
// GetScheduledEntries returns all scheduled messages and its score in the given queue.
|
|
func GetScheduledEntries(tb testing.TB, r redis.UniversalClient, qname string) []base.Z {
|
|
tb.Helper()
|
|
return getZSetEntries(tb, r, base.ScheduledKey(qname))
|
|
}
|
|
|
|
// GetRetryEntries returns all retry messages and its score in the given queue.
|
|
func GetRetryEntries(tb testing.TB, r redis.UniversalClient, qname string) []base.Z {
|
|
tb.Helper()
|
|
return getZSetEntries(tb, r, base.RetryKey(qname))
|
|
}
|
|
|
|
// GetDeadEntries returns all dead messages and its score in the given queue.
|
|
func GetDeadEntries(tb testing.TB, r redis.UniversalClient, qname string) []base.Z {
|
|
tb.Helper()
|
|
return getZSetEntries(tb, r, base.DeadKey(qname))
|
|
}
|
|
|
|
// GetDeadlinesEntries returns all task messages and its score in the deadlines set for the given queue.
|
|
func GetDeadlinesEntries(tb testing.TB, r redis.UniversalClient, qname string) []base.Z {
|
|
tb.Helper()
|
|
return getZSetEntries(tb, r, base.DeadlinesKey(qname))
|
|
}
|
|
|
|
func getListMessages(tb testing.TB, r redis.UniversalClient, list string) []*base.TaskMessage {
|
|
data := r.LRange(list, 0, -1).Val()
|
|
return MustUnmarshalSlice(tb, data)
|
|
}
|
|
|
|
func getZSetMessages(tb testing.TB, r redis.UniversalClient, zset string) []*base.TaskMessage {
|
|
data := r.ZRange(zset, 0, -1).Val()
|
|
return MustUnmarshalSlice(tb, data)
|
|
}
|
|
|
|
func getZSetEntries(tb testing.TB, r redis.UniversalClient, zset string) []base.Z {
|
|
data := r.ZRangeWithScores(zset, 0, -1).Val()
|
|
var entries []base.Z
|
|
for _, z := range data {
|
|
entries = append(entries, base.Z{
|
|
Message: MustUnmarshal(tb, z.Member.(string)),
|
|
Score: int64(z.Score),
|
|
})
|
|
}
|
|
return entries
|
|
}
|