2020-01-03 10:13:16 +08:00
|
|
|
// Copyright 2020 Kentaro Hibino. All rights reserved.
|
|
|
|
// Use of this source code is governed by a MIT license
|
|
|
|
// that can be found in the LICENSE file.
|
|
|
|
|
2020-01-01 04:36:46 +08:00
|
|
|
package asynq
|
|
|
|
|
|
|
|
import (
|
2020-02-11 23:06:52 +08:00
|
|
|
"context"
|
2020-01-01 04:36:46 +08:00
|
|
|
"fmt"
|
|
|
|
"math/rand"
|
|
|
|
"sync"
|
|
|
|
"testing"
|
|
|
|
"time"
|
|
|
|
)
|
|
|
|
|
2020-02-09 03:06:14 +08:00
|
|
|
// Simple E2E Benchmark testing with no scheduled tasks and retries.
|
2020-01-01 04:36:46 +08:00
|
|
|
func BenchmarkEndToEndSimple(b *testing.B) {
|
|
|
|
const count = 100000
|
|
|
|
for n := 0; n < b.N; n++ {
|
|
|
|
b.StopTimer() // begin setup
|
2020-01-19 02:17:39 +08:00
|
|
|
setup(b)
|
|
|
|
redis := &RedisClientOpt{
|
|
|
|
Addr: redisAddr,
|
|
|
|
DB: redisDB,
|
|
|
|
}
|
|
|
|
client := NewClient(redis)
|
|
|
|
bg := NewBackground(redis, &Config{
|
2020-01-01 04:36:46 +08:00
|
|
|
Concurrency: 10,
|
|
|
|
RetryDelayFunc: func(n int, err error, t *Task) time.Duration {
|
|
|
|
return time.Second
|
|
|
|
},
|
|
|
|
})
|
|
|
|
// Create a bunch of tasks
|
|
|
|
for i := 0; i < count; i++ {
|
2020-01-05 05:13:46 +08:00
|
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
2020-02-24 07:40:04 +08:00
|
|
|
if err := client.Enqueue(t); err != nil {
|
|
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
|
|
}
|
2020-01-01 04:36:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(count)
|
2020-02-11 23:06:52 +08:00
|
|
|
handler := func(ctx context.Context, t *Task) error {
|
2020-01-01 04:36:46 +08:00
|
|
|
wg.Done()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
b.StartTimer() // end setup
|
|
|
|
|
|
|
|
bg.start(HandlerFunc(handler))
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
b.StopTimer() // begin teardown
|
|
|
|
bg.stop()
|
|
|
|
b.StartTimer() // end teardown
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
// E2E benchmark with scheduled tasks and retries.
|
|
|
|
func BenchmarkEndToEnd(b *testing.B) {
|
|
|
|
const count = 100000
|
|
|
|
for n := 0; n < b.N; n++ {
|
|
|
|
b.StopTimer() // begin setup
|
|
|
|
rand.Seed(time.Now().UnixNano())
|
2020-01-19 02:17:39 +08:00
|
|
|
setup(b)
|
|
|
|
redis := &RedisClientOpt{
|
|
|
|
Addr: redisAddr,
|
|
|
|
DB: redisDB,
|
|
|
|
}
|
|
|
|
client := NewClient(redis)
|
|
|
|
bg := NewBackground(redis, &Config{
|
2020-01-01 04:36:46 +08:00
|
|
|
Concurrency: 10,
|
|
|
|
RetryDelayFunc: func(n int, err error, t *Task) time.Duration {
|
|
|
|
return time.Second
|
|
|
|
},
|
|
|
|
})
|
|
|
|
// Create a bunch of tasks
|
|
|
|
for i := 0; i < count; i++ {
|
2020-01-05 05:13:46 +08:00
|
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
2020-02-24 07:40:04 +08:00
|
|
|
if err := client.Enqueue(t); err != nil {
|
|
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
|
|
}
|
2020-01-01 04:36:46 +08:00
|
|
|
}
|
|
|
|
for i := 0; i < count; i++ {
|
2020-01-05 05:13:46 +08:00
|
|
|
t := NewTask(fmt.Sprintf("scheduled%d", i), map[string]interface{}{"data": i})
|
2020-02-24 07:40:04 +08:00
|
|
|
if err := client.EnqueueAt(time.Now().Add(time.Second), t); err != nil {
|
|
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
|
|
}
|
2020-01-01 04:36:46 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(count * 2)
|
2020-02-11 23:06:52 +08:00
|
|
|
handler := func(ctx context.Context, t *Task) error {
|
2020-01-01 04:36:46 +08:00
|
|
|
// randomly fail 1% of tasks
|
|
|
|
if rand.Intn(100) == 1 {
|
|
|
|
return fmt.Errorf(":(")
|
|
|
|
}
|
|
|
|
wg.Done()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
b.StartTimer() // end setup
|
|
|
|
|
|
|
|
bg.start(HandlerFunc(handler))
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
b.StopTimer() // begin teardown
|
|
|
|
bg.stop()
|
|
|
|
b.StartTimer() // end teardown
|
|
|
|
}
|
|
|
|
}
|
2020-02-09 03:06:14 +08:00
|
|
|
|
|
|
|
// Simple E2E Benchmark testing with no scheduled tasks and retries with multiple queues.
|
|
|
|
func BenchmarkEndToEndMultipleQueues(b *testing.B) {
|
|
|
|
// number of tasks to create for each queue
|
|
|
|
const (
|
|
|
|
highCount = 20000
|
|
|
|
defaultCount = 20000
|
|
|
|
lowCount = 20000
|
|
|
|
)
|
|
|
|
for n := 0; n < b.N; n++ {
|
|
|
|
b.StopTimer() // begin setup
|
|
|
|
setup(b)
|
|
|
|
redis := &RedisClientOpt{
|
|
|
|
Addr: redisAddr,
|
|
|
|
DB: redisDB,
|
|
|
|
}
|
|
|
|
client := NewClient(redis)
|
|
|
|
bg := NewBackground(redis, &Config{
|
|
|
|
Concurrency: 10,
|
2020-02-13 14:23:25 +08:00
|
|
|
Queues: map[string]int{
|
2020-02-09 03:06:14 +08:00
|
|
|
"high": 6,
|
|
|
|
"default": 3,
|
|
|
|
"low": 1,
|
|
|
|
},
|
|
|
|
})
|
|
|
|
// Create a bunch of tasks
|
|
|
|
for i := 0; i < highCount; i++ {
|
|
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
2020-02-24 07:40:04 +08:00
|
|
|
if err := client.Enqueue(t, Queue("high")); err != nil {
|
|
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
|
|
}
|
2020-02-09 03:06:14 +08:00
|
|
|
}
|
|
|
|
for i := 0; i < defaultCount; i++ {
|
|
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
2020-02-24 07:40:04 +08:00
|
|
|
if err := client.Enqueue(t); err != nil {
|
|
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
|
|
}
|
2020-02-09 03:06:14 +08:00
|
|
|
}
|
|
|
|
for i := 0; i < lowCount; i++ {
|
|
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
2020-02-24 07:40:04 +08:00
|
|
|
if err := client.Enqueue(t, Queue("low")); err != nil {
|
|
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
|
|
}
|
2020-02-09 03:06:14 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
var wg sync.WaitGroup
|
|
|
|
wg.Add(highCount + defaultCount + lowCount)
|
2020-02-11 23:06:52 +08:00
|
|
|
handler := func(ctx context.Context, t *Task) error {
|
2020-02-09 03:06:14 +08:00
|
|
|
wg.Done()
|
|
|
|
return nil
|
|
|
|
}
|
|
|
|
b.StartTimer() // end setup
|
|
|
|
|
|
|
|
bg.start(HandlerFunc(handler))
|
|
|
|
wg.Wait()
|
|
|
|
|
|
|
|
b.StopTimer() // begin teardown
|
|
|
|
bg.stop()
|
|
|
|
b.StartTimer() // end teardown
|
|
|
|
}
|
|
|
|
}
|