mirror of
https://github.com/hibiken/asynq.git
synced 2024-11-15 11:58:47 +08:00
231 lines
5.9 KiB
Go
231 lines
5.9 KiB
Go
// Copyright 2020 Kentaro Hibino. All rights reserved.
|
|
// Use of this source code is governed by a MIT license
|
|
// that can be found in the LICENSE file.
|
|
|
|
package asynq
|
|
|
|
import (
|
|
"context"
|
|
"fmt"
|
|
"sync"
|
|
"testing"
|
|
"time"
|
|
)
|
|
|
|
// Simple E2E Benchmark testing with no scheduled tasks and retries.
|
|
func BenchmarkEndToEndSimple(b *testing.B) {
|
|
const count = 100000
|
|
for n := 0; n < b.N; n++ {
|
|
b.StopTimer() // begin setup
|
|
setup(b)
|
|
redis := getRedisConnOpt(b)
|
|
client := NewClient(redis)
|
|
srv := NewServer(redis, Config{
|
|
Concurrency: 10,
|
|
RetryDelayFunc: func(n int, err error, t *Task) time.Duration {
|
|
return time.Second
|
|
},
|
|
LogLevel: testLogLevel,
|
|
})
|
|
// Create a bunch of tasks
|
|
for i := 0; i < count; i++ {
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
client.Close()
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(count)
|
|
handler := func(ctx context.Context, t *Task) error {
|
|
wg.Done()
|
|
return nil
|
|
}
|
|
b.StartTimer() // end setup
|
|
|
|
srv.Start(HandlerFunc(handler))
|
|
wg.Wait()
|
|
|
|
b.StopTimer() // begin teardown
|
|
srv.Stop()
|
|
b.StartTimer() // end teardown
|
|
}
|
|
}
|
|
|
|
// E2E benchmark with scheduled tasks and retries.
|
|
func BenchmarkEndToEnd(b *testing.B) {
|
|
const count = 100000
|
|
for n := 0; n < b.N; n++ {
|
|
b.StopTimer() // begin setup
|
|
setup(b)
|
|
redis := getRedisConnOpt(b)
|
|
client := NewClient(redis)
|
|
srv := NewServer(redis, Config{
|
|
Concurrency: 10,
|
|
RetryDelayFunc: func(n int, err error, t *Task) time.Duration {
|
|
return time.Second
|
|
},
|
|
LogLevel: testLogLevel,
|
|
})
|
|
// Create a bunch of tasks
|
|
for i := 0; i < count; i++ {
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
for i := 0; i < count; i++ {
|
|
t := NewTask(fmt.Sprintf("scheduled%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t, ProcessIn(1*time.Second)); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
client.Close()
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(count * 2)
|
|
handler := func(ctx context.Context, t *Task) error {
|
|
n, err := t.Payload.GetInt("data")
|
|
if err != nil {
|
|
b.Logf("internal error: %v", err)
|
|
}
|
|
retried, ok := GetRetryCount(ctx)
|
|
if !ok {
|
|
b.Logf("internal error: %v", err)
|
|
}
|
|
// Fail 1% of tasks for the first attempt.
|
|
if retried == 0 && n%100 == 0 {
|
|
return fmt.Errorf(":(")
|
|
}
|
|
wg.Done()
|
|
return nil
|
|
}
|
|
b.StartTimer() // end setup
|
|
|
|
srv.Start(HandlerFunc(handler))
|
|
wg.Wait()
|
|
|
|
b.StopTimer() // begin teardown
|
|
srv.Stop()
|
|
b.StartTimer() // end teardown
|
|
}
|
|
}
|
|
|
|
// Simple E2E Benchmark testing with no scheduled tasks and retries with multiple queues.
|
|
func BenchmarkEndToEndMultipleQueues(b *testing.B) {
|
|
// number of tasks to create for each queue
|
|
const (
|
|
highCount = 20000
|
|
defaultCount = 20000
|
|
lowCount = 20000
|
|
)
|
|
for n := 0; n < b.N; n++ {
|
|
b.StopTimer() // begin setup
|
|
setup(b)
|
|
redis := getRedisConnOpt(b)
|
|
client := NewClient(redis)
|
|
srv := NewServer(redis, Config{
|
|
Concurrency: 10,
|
|
Queues: map[string]int{
|
|
"high": 6,
|
|
"default": 3,
|
|
"low": 1,
|
|
},
|
|
LogLevel: testLogLevel,
|
|
})
|
|
// Create a bunch of tasks
|
|
for i := 0; i < highCount; i++ {
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t, Queue("high")); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
for i := 0; i < defaultCount; i++ {
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
for i := 0; i < lowCount; i++ {
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t, Queue("low")); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
client.Close()
|
|
|
|
var wg sync.WaitGroup
|
|
wg.Add(highCount + defaultCount + lowCount)
|
|
handler := func(ctx context.Context, t *Task) error {
|
|
wg.Done()
|
|
return nil
|
|
}
|
|
b.StartTimer() // end setup
|
|
|
|
srv.Start(HandlerFunc(handler))
|
|
wg.Wait()
|
|
|
|
b.StopTimer() // begin teardown
|
|
srv.Stop()
|
|
b.StartTimer() // end teardown
|
|
}
|
|
}
|
|
|
|
// E2E benchmark to check client enqueue operation performs correctly,
|
|
// while server is busy processing tasks.
|
|
func BenchmarkClientWhileServerRunning(b *testing.B) {
|
|
const count = 10000
|
|
for n := 0; n < b.N; n++ {
|
|
b.StopTimer() // begin setup
|
|
setup(b)
|
|
redis := getRedisConnOpt(b)
|
|
client := NewClient(redis)
|
|
srv := NewServer(redis, Config{
|
|
Concurrency: 10,
|
|
RetryDelayFunc: func(n int, err error, t *Task) time.Duration {
|
|
return time.Second
|
|
},
|
|
LogLevel: testLogLevel,
|
|
})
|
|
// Enqueue 10,000 tasks.
|
|
for i := 0; i < count; i++ {
|
|
t := NewTask(fmt.Sprintf("task%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
// Schedule 10,000 tasks.
|
|
for i := 0; i < count; i++ {
|
|
t := NewTask(fmt.Sprintf("scheduled%d", i), map[string]interface{}{"data": i})
|
|
if _, err := client.Enqueue(t, ProcessIn(1*time.Second)); err != nil {
|
|
b.Fatalf("could not enqueue a task: %v", err)
|
|
}
|
|
}
|
|
|
|
handler := func(ctx context.Context, t *Task) error {
|
|
return nil
|
|
}
|
|
srv.Start(HandlerFunc(handler))
|
|
|
|
b.StartTimer() // end setup
|
|
|
|
b.Log("Starting enqueueing")
|
|
enqueued := 0
|
|
for enqueued < 100000 {
|
|
t := NewTask(fmt.Sprintf("enqueued%d", enqueued), map[string]interface{}{"data": enqueued})
|
|
if _, err := client.Enqueue(t); err != nil {
|
|
b.Logf("could not enqueue task %d: %v", enqueued, err)
|
|
continue
|
|
}
|
|
enqueued++
|
|
}
|
|
b.Logf("Finished enqueueing %d tasks", enqueued)
|
|
|
|
b.StopTimer() // begin teardown
|
|
srv.Stop()
|
|
client.Close()
|
|
b.StartTimer() // end teardown
|
|
}
|
|
}
|