mirror of
https://codeberg.org/forgejo/forgejo.git
synced 2024-11-24 03:21:49 +01:00
62eb1b0f25
* Queue: Add generic graceful queues with settings
* Queue & Setting: Add worker pool implementation
* Queue: Add worker settings
* Queue: Make resizing worker pools
* Queue: Add name variable to queues
* Queue: Add monitoring
* Queue: Improve logging
* Issues: Gracefulise the issues indexer
Remove the old now unused specific queues
* Task: Move to generic queue and gracefulise
* Issues: Standardise the issues indexer queue settings
* Fix test
* Queue: Allow Redis to connect to unix
* Prevent deadlock during early shutdown of issue indexer
* Add MaxWorker settings to queues
* Merge branch 'master' into graceful-queues
* Update modules/indexer/issues/indexer.go
Co-Authored-By: guillep2k <18600385+guillep2k@users.noreply.github.com>
* Update modules/indexer/issues/indexer.go
Co-Authored-By: guillep2k <18600385+guillep2k@users.noreply.github.com>
* Update modules/queue/queue_channel.go
Co-Authored-By: guillep2k <18600385+guillep2k@users.noreply.github.com>
* Update modules/queue/queue_disk.go
* Update modules/queue/queue_disk_channel.go
Co-Authored-By: guillep2k <18600385+guillep2k@users.noreply.github.com>
* Rename queue.Description to queue.ManagedQueue as per @guillep2k
* Cancel pool workers when removed
* Remove dependency on queue from setting
* Update modules/queue/queue_redis.go
Co-Authored-By: guillep2k <18600385+guillep2k@users.noreply.github.com>
* As per @guillep2k add mutex locks on shutdown/terminate
* move unlocking out of setInternal
* Add warning if number of workers < 0
* Small changes as per @guillep2k
* No redis host specified not found
* Clean up documentation for queues
* Update docs/content/doc/advanced/config-cheat-sheet.en-us.md
* Update modules/indexer/issues/indexer_test.go
* Ensure that persistable channel queue is added to manager
* Rename QUEUE_NAME REDIS_QUEUE_NAME
* Revert "Rename QUEUE_NAME REDIS_QUEUE_NAME"
This reverts commit 1f83b4fc9b
.
Co-authored-by: guillep2k <18600385+guillep2k@users.noreply.github.com>
Co-authored-by: Lauris BH <lauris@nix.lv>
Co-authored-by: techknowlogick <matti@mdranta.net>
Co-authored-by: Lunny Xiao <xiaolunwen@gmail.com>
213 lines
4.9 KiB
Go
213 lines
4.9 KiB
Go
// Copyright 2019 The Gitea Authors. All rights reserved.
|
|
// Use of this source code is governed by a MIT-style
|
|
// license that can be found in the LICENSE file.
|
|
|
|
package queue
|
|
|
|
import (
|
|
"context"
|
|
"encoding/json"
|
|
"fmt"
|
|
"reflect"
|
|
"sync"
|
|
"time"
|
|
|
|
"code.gitea.io/gitea/modules/log"
|
|
|
|
"gitea.com/lunny/levelqueue"
|
|
)
|
|
|
|
// LevelQueueType is the type for level queue
|
|
const LevelQueueType Type = "level"
|
|
|
|
// LevelQueueConfiguration is the configuration for a LevelQueue
|
|
type LevelQueueConfiguration struct {
|
|
DataDir string
|
|
QueueLength int
|
|
BatchLength int
|
|
Workers int
|
|
MaxWorkers int
|
|
BlockTimeout time.Duration
|
|
BoostTimeout time.Duration
|
|
BoostWorkers int
|
|
Name string
|
|
}
|
|
|
|
// LevelQueue implements a disk library queue
|
|
type LevelQueue struct {
|
|
pool *WorkerPool
|
|
queue *levelqueue.Queue
|
|
closed chan struct{}
|
|
terminated chan struct{}
|
|
lock sync.Mutex
|
|
exemplar interface{}
|
|
workers int
|
|
name string
|
|
}
|
|
|
|
// NewLevelQueue creates a ledis local queue
|
|
func NewLevelQueue(handle HandlerFunc, cfg, exemplar interface{}) (Queue, error) {
|
|
configInterface, err := toConfig(LevelQueueConfiguration{}, cfg)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
config := configInterface.(LevelQueueConfiguration)
|
|
|
|
internal, err := levelqueue.Open(config.DataDir)
|
|
if err != nil {
|
|
return nil, err
|
|
}
|
|
|
|
dataChan := make(chan Data, config.QueueLength)
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
|
|
queue := &LevelQueue{
|
|
pool: &WorkerPool{
|
|
baseCtx: ctx,
|
|
cancel: cancel,
|
|
batchLength: config.BatchLength,
|
|
handle: handle,
|
|
dataChan: dataChan,
|
|
blockTimeout: config.BlockTimeout,
|
|
boostTimeout: config.BoostTimeout,
|
|
boostWorkers: config.BoostWorkers,
|
|
maxNumberOfWorkers: config.MaxWorkers,
|
|
},
|
|
queue: internal,
|
|
exemplar: exemplar,
|
|
closed: make(chan struct{}),
|
|
terminated: make(chan struct{}),
|
|
workers: config.Workers,
|
|
name: config.Name,
|
|
}
|
|
queue.pool.qid = GetManager().Add(queue, LevelQueueType, config, exemplar, queue.pool)
|
|
return queue, nil
|
|
}
|
|
|
|
// Run starts to run the queue
|
|
func (l *LevelQueue) Run(atShutdown, atTerminate func(context.Context, func())) {
|
|
atShutdown(context.Background(), l.Shutdown)
|
|
atTerminate(context.Background(), l.Terminate)
|
|
|
|
go func() {
|
|
_ = l.pool.AddWorkers(l.workers, 0)
|
|
}()
|
|
|
|
go l.readToChan()
|
|
|
|
log.Trace("LevelQueue: %s Waiting til closed", l.name)
|
|
<-l.closed
|
|
|
|
log.Trace("LevelQueue: %s Waiting til done", l.name)
|
|
l.pool.Wait()
|
|
|
|
log.Trace("LevelQueue: %s Waiting til cleaned", l.name)
|
|
ctx, cancel := context.WithCancel(context.Background())
|
|
atTerminate(ctx, cancel)
|
|
l.pool.CleanUp(ctx)
|
|
cancel()
|
|
log.Trace("LevelQueue: %s Cleaned", l.name)
|
|
|
|
}
|
|
|
|
func (l *LevelQueue) readToChan() {
|
|
for {
|
|
select {
|
|
case <-l.closed:
|
|
// tell the pool to shutdown.
|
|
l.pool.cancel()
|
|
return
|
|
default:
|
|
bs, err := l.queue.RPop()
|
|
if err != nil {
|
|
if err != levelqueue.ErrNotFound {
|
|
log.Error("LevelQueue: %s Error on RPop: %v", l.name, err)
|
|
}
|
|
time.Sleep(time.Millisecond * 100)
|
|
continue
|
|
}
|
|
|
|
if len(bs) == 0 {
|
|
time.Sleep(time.Millisecond * 100)
|
|
continue
|
|
}
|
|
|
|
var data Data
|
|
if l.exemplar != nil {
|
|
t := reflect.TypeOf(l.exemplar)
|
|
n := reflect.New(t)
|
|
ne := n.Elem()
|
|
err = json.Unmarshal(bs, ne.Addr().Interface())
|
|
data = ne.Interface().(Data)
|
|
} else {
|
|
err = json.Unmarshal(bs, &data)
|
|
}
|
|
if err != nil {
|
|
log.Error("LevelQueue: %s Failed to unmarshal with error: %v", l.name, err)
|
|
time.Sleep(time.Millisecond * 100)
|
|
continue
|
|
}
|
|
|
|
log.Trace("LevelQueue %s: Task found: %#v", l.name, data)
|
|
l.pool.Push(data)
|
|
|
|
}
|
|
}
|
|
}
|
|
|
|
// Push will push the indexer data to queue
|
|
func (l *LevelQueue) Push(data Data) error {
|
|
if l.exemplar != nil {
|
|
// Assert data is of same type as r.exemplar
|
|
value := reflect.ValueOf(data)
|
|
t := value.Type()
|
|
exemplarType := reflect.ValueOf(l.exemplar).Type()
|
|
if !t.AssignableTo(exemplarType) || data == nil {
|
|
return fmt.Errorf("Unable to assign data: %v to same type as exemplar: %v in %s", data, l.exemplar, l.name)
|
|
}
|
|
}
|
|
bs, err := json.Marshal(data)
|
|
if err != nil {
|
|
return err
|
|
}
|
|
return l.queue.LPush(bs)
|
|
}
|
|
|
|
// Shutdown this queue and stop processing
|
|
func (l *LevelQueue) Shutdown() {
|
|
l.lock.Lock()
|
|
defer l.lock.Unlock()
|
|
log.Trace("LevelQueue: %s Shutdown", l.name)
|
|
select {
|
|
case <-l.closed:
|
|
default:
|
|
close(l.closed)
|
|
}
|
|
}
|
|
|
|
// Terminate this queue and close the queue
|
|
func (l *LevelQueue) Terminate() {
|
|
log.Trace("LevelQueue: %s Terminating", l.name)
|
|
l.Shutdown()
|
|
l.lock.Lock()
|
|
select {
|
|
case <-l.terminated:
|
|
l.lock.Unlock()
|
|
default:
|
|
close(l.terminated)
|
|
l.lock.Unlock()
|
|
if err := l.queue.Close(); err != nil && err.Error() != "leveldb: closed" {
|
|
log.Error("Error whilst closing internal queue in %s: %v", l.name, err)
|
|
}
|
|
|
|
}
|
|
}
|
|
|
|
// Name returns the name of this queue
|
|
func (l *LevelQueue) Name() string {
|
|
return l.name
|
|
}
|
|
|
|
func init() {
|
|
queuesMap[LevelQueueType] = NewLevelQueue
|
|
}
|