mirror of
https://github.com/Motorhead1991/qemu.git
synced 2025-07-27 04:13:53 -06:00

Migration code wants to manage device data sending threads in one place. QEMU has an existing thread pool implementation, however it is limited to queuing AIO operations only and essentially has a 1:1 mapping between the current AioContext and the AIO ThreadPool in use. Implement generic (non-AIO) ThreadPool by essentially wrapping Glib's GThreadPool. This brings a few new operations on a pool: * thread_pool_wait() operation waits until all the submitted work requests have finished. * thread_pool_set_max_threads() explicitly sets the maximum thread count in the pool. * thread_pool_adjust_max_threads_to_work() adjusts the maximum thread count in the pool to equal the number of still waiting in queue or unfinished work. Reviewed-by: Fabiano Rosas <farosas@suse.de> Reviewed-by: Peter Xu <peterx@redhat.com> Signed-off-by: Maciej S. Szmigiero <maciej.szmigiero@oracle.com> Link: https://lore.kernel.org/qemu-devel/b1efaebdbea7cb7068b8fb74148777012383e12b.1741124640.git.maciej.szmigiero@oracle.com Signed-off-by: Cédric Le Goater <clg@redhat.com>
495 lines
13 KiB
C
495 lines
13 KiB
C
/*
|
|
* QEMU block layer thread pool
|
|
*
|
|
* Copyright IBM, Corp. 2008
|
|
* Copyright Red Hat, Inc. 2012
|
|
*
|
|
* Authors:
|
|
* Anthony Liguori <aliguori@us.ibm.com>
|
|
* Paolo Bonzini <pbonzini@redhat.com>
|
|
*
|
|
* This work is licensed under the terms of the GNU GPL, version 2. See
|
|
* the COPYING file in the top-level directory.
|
|
*
|
|
* Contributions after 2012-01-13 are licensed under the terms of the
|
|
* GNU GPL, version 2 or (at your option) any later version.
|
|
*/
|
|
#include "qemu/osdep.h"
|
|
#include "qemu/defer-call.h"
|
|
#include "qemu/queue.h"
|
|
#include "qemu/thread.h"
|
|
#include "qemu/coroutine.h"
|
|
#include "trace.h"
|
|
#include "block/thread-pool.h"
|
|
#include "qemu/main-loop.h"
|
|
|
|
static void do_spawn_thread(ThreadPoolAio *pool);
|
|
|
|
typedef struct ThreadPoolElementAio ThreadPoolElementAio;
|
|
|
|
enum ThreadState {
|
|
THREAD_QUEUED,
|
|
THREAD_ACTIVE,
|
|
THREAD_DONE,
|
|
};
|
|
|
|
struct ThreadPoolElementAio {
|
|
BlockAIOCB common;
|
|
ThreadPoolAio *pool;
|
|
ThreadPoolFunc *func;
|
|
void *arg;
|
|
|
|
/* Moving state out of THREAD_QUEUED is protected by lock. After
|
|
* that, only the worker thread can write to it. Reads and writes
|
|
* of state and ret are ordered with memory barriers.
|
|
*/
|
|
enum ThreadState state;
|
|
int ret;
|
|
|
|
/* Access to this list is protected by lock. */
|
|
QTAILQ_ENTRY(ThreadPoolElementAio) reqs;
|
|
|
|
/* This list is only written by the thread pool's mother thread. */
|
|
QLIST_ENTRY(ThreadPoolElementAio) all;
|
|
};
|
|
|
|
struct ThreadPoolAio {
|
|
AioContext *ctx;
|
|
QEMUBH *completion_bh;
|
|
QemuMutex lock;
|
|
QemuCond worker_stopped;
|
|
QemuCond request_cond;
|
|
QEMUBH *new_thread_bh;
|
|
|
|
/* The following variables are only accessed from one AioContext. */
|
|
QLIST_HEAD(, ThreadPoolElementAio) head;
|
|
|
|
/* The following variables are protected by lock. */
|
|
QTAILQ_HEAD(, ThreadPoolElementAio) request_list;
|
|
int cur_threads;
|
|
int idle_threads;
|
|
int new_threads; /* backlog of threads we need to create */
|
|
int pending_threads; /* threads created but not running yet */
|
|
int min_threads;
|
|
int max_threads;
|
|
};
|
|
|
|
static void *worker_thread(void *opaque)
|
|
{
|
|
ThreadPoolAio *pool = opaque;
|
|
|
|
qemu_mutex_lock(&pool->lock);
|
|
pool->pending_threads--;
|
|
do_spawn_thread(pool);
|
|
|
|
while (pool->cur_threads <= pool->max_threads) {
|
|
ThreadPoolElementAio *req;
|
|
int ret;
|
|
|
|
if (QTAILQ_EMPTY(&pool->request_list)) {
|
|
pool->idle_threads++;
|
|
ret = qemu_cond_timedwait(&pool->request_cond, &pool->lock, 10000);
|
|
pool->idle_threads--;
|
|
if (ret == 0 &&
|
|
QTAILQ_EMPTY(&pool->request_list) &&
|
|
pool->cur_threads > pool->min_threads) {
|
|
/* Timed out + no work to do + no need for warm threads = exit. */
|
|
break;
|
|
}
|
|
/*
|
|
* Even if there was some work to do, check if there aren't
|
|
* too many worker threads before picking it up.
|
|
*/
|
|
continue;
|
|
}
|
|
|
|
req = QTAILQ_FIRST(&pool->request_list);
|
|
QTAILQ_REMOVE(&pool->request_list, req, reqs);
|
|
req->state = THREAD_ACTIVE;
|
|
qemu_mutex_unlock(&pool->lock);
|
|
|
|
ret = req->func(req->arg);
|
|
|
|
req->ret = ret;
|
|
/* Write ret before state. */
|
|
smp_wmb();
|
|
req->state = THREAD_DONE;
|
|
|
|
qemu_bh_schedule(pool->completion_bh);
|
|
qemu_mutex_lock(&pool->lock);
|
|
}
|
|
|
|
pool->cur_threads--;
|
|
qemu_cond_signal(&pool->worker_stopped);
|
|
|
|
/*
|
|
* Wake up another thread, in case we got a wakeup but decided
|
|
* to exit due to pool->cur_threads > pool->max_threads.
|
|
*/
|
|
qemu_cond_signal(&pool->request_cond);
|
|
qemu_mutex_unlock(&pool->lock);
|
|
return NULL;
|
|
}
|
|
|
|
static void do_spawn_thread(ThreadPoolAio *pool)
|
|
{
|
|
QemuThread t;
|
|
|
|
/* Runs with lock taken. */
|
|
if (!pool->new_threads) {
|
|
return;
|
|
}
|
|
|
|
pool->new_threads--;
|
|
pool->pending_threads++;
|
|
|
|
qemu_thread_create(&t, "worker", worker_thread, pool, QEMU_THREAD_DETACHED);
|
|
}
|
|
|
|
static void spawn_thread_bh_fn(void *opaque)
|
|
{
|
|
ThreadPoolAio *pool = opaque;
|
|
|
|
qemu_mutex_lock(&pool->lock);
|
|
do_spawn_thread(pool);
|
|
qemu_mutex_unlock(&pool->lock);
|
|
}
|
|
|
|
static void spawn_thread(ThreadPoolAio *pool)
|
|
{
|
|
pool->cur_threads++;
|
|
pool->new_threads++;
|
|
/* If there are threads being created, they will spawn new workers, so
|
|
* we don't spend time creating many threads in a loop holding a mutex or
|
|
* starving the current vcpu.
|
|
*
|
|
* If there are no idle threads, ask the main thread to create one, so we
|
|
* inherit the correct affinity instead of the vcpu affinity.
|
|
*/
|
|
if (!pool->pending_threads) {
|
|
qemu_bh_schedule(pool->new_thread_bh);
|
|
}
|
|
}
|
|
|
|
static void thread_pool_completion_bh(void *opaque)
|
|
{
|
|
ThreadPoolAio *pool = opaque;
|
|
ThreadPoolElementAio *elem, *next;
|
|
|
|
defer_call_begin(); /* cb() may use defer_call() to coalesce work */
|
|
|
|
restart:
|
|
QLIST_FOREACH_SAFE(elem, &pool->head, all, next) {
|
|
if (elem->state != THREAD_DONE) {
|
|
continue;
|
|
}
|
|
|
|
trace_thread_pool_complete_aio(pool, elem, elem->common.opaque,
|
|
elem->ret);
|
|
QLIST_REMOVE(elem, all);
|
|
|
|
if (elem->common.cb) {
|
|
/* Read state before ret. */
|
|
smp_rmb();
|
|
|
|
/* Schedule ourselves in case elem->common.cb() calls aio_poll() to
|
|
* wait for another request that completed at the same time.
|
|
*/
|
|
qemu_bh_schedule(pool->completion_bh);
|
|
|
|
elem->common.cb(elem->common.opaque, elem->ret);
|
|
|
|
/* We can safely cancel the completion_bh here regardless of someone
|
|
* else having scheduled it meanwhile because we reenter the
|
|
* completion function anyway (goto restart).
|
|
*/
|
|
qemu_bh_cancel(pool->completion_bh);
|
|
|
|
qemu_aio_unref(elem);
|
|
goto restart;
|
|
} else {
|
|
qemu_aio_unref(elem);
|
|
}
|
|
}
|
|
|
|
defer_call_end();
|
|
}
|
|
|
|
static void thread_pool_cancel(BlockAIOCB *acb)
|
|
{
|
|
ThreadPoolElementAio *elem = (ThreadPoolElementAio *)acb;
|
|
ThreadPoolAio *pool = elem->pool;
|
|
|
|
trace_thread_pool_cancel_aio(elem, elem->common.opaque);
|
|
|
|
QEMU_LOCK_GUARD(&pool->lock);
|
|
if (elem->state == THREAD_QUEUED) {
|
|
QTAILQ_REMOVE(&pool->request_list, elem, reqs);
|
|
qemu_bh_schedule(pool->completion_bh);
|
|
|
|
elem->state = THREAD_DONE;
|
|
elem->ret = -ECANCELED;
|
|
}
|
|
|
|
}
|
|
|
|
static const AIOCBInfo thread_pool_aiocb_info = {
|
|
.aiocb_size = sizeof(ThreadPoolElementAio),
|
|
.cancel_async = thread_pool_cancel,
|
|
};
|
|
|
|
BlockAIOCB *thread_pool_submit_aio(ThreadPoolFunc *func, void *arg,
|
|
BlockCompletionFunc *cb, void *opaque)
|
|
{
|
|
ThreadPoolElementAio *req;
|
|
AioContext *ctx = qemu_get_current_aio_context();
|
|
ThreadPoolAio *pool = aio_get_thread_pool(ctx);
|
|
|
|
/* Assert that the thread submitting work is the same running the pool */
|
|
assert(pool->ctx == qemu_get_current_aio_context());
|
|
|
|
req = qemu_aio_get(&thread_pool_aiocb_info, NULL, cb, opaque);
|
|
req->func = func;
|
|
req->arg = arg;
|
|
req->state = THREAD_QUEUED;
|
|
req->pool = pool;
|
|
|
|
QLIST_INSERT_HEAD(&pool->head, req, all);
|
|
|
|
trace_thread_pool_submit_aio(pool, req, arg);
|
|
|
|
qemu_mutex_lock(&pool->lock);
|
|
if (pool->idle_threads == 0 && pool->cur_threads < pool->max_threads) {
|
|
spawn_thread(pool);
|
|
}
|
|
QTAILQ_INSERT_TAIL(&pool->request_list, req, reqs);
|
|
qemu_mutex_unlock(&pool->lock);
|
|
qemu_cond_signal(&pool->request_cond);
|
|
return &req->common;
|
|
}
|
|
|
|
typedef struct ThreadPoolCo {
|
|
Coroutine *co;
|
|
int ret;
|
|
} ThreadPoolCo;
|
|
|
|
static void thread_pool_co_cb(void *opaque, int ret)
|
|
{
|
|
ThreadPoolCo *co = opaque;
|
|
|
|
co->ret = ret;
|
|
aio_co_wake(co->co);
|
|
}
|
|
|
|
int coroutine_fn thread_pool_submit_co(ThreadPoolFunc *func, void *arg)
|
|
{
|
|
ThreadPoolCo tpc = { .co = qemu_coroutine_self(), .ret = -EINPROGRESS };
|
|
assert(qemu_in_coroutine());
|
|
thread_pool_submit_aio(func, arg, thread_pool_co_cb, &tpc);
|
|
qemu_coroutine_yield();
|
|
return tpc.ret;
|
|
}
|
|
|
|
void thread_pool_update_params(ThreadPoolAio *pool, AioContext *ctx)
|
|
{
|
|
qemu_mutex_lock(&pool->lock);
|
|
|
|
pool->min_threads = ctx->thread_pool_min;
|
|
pool->max_threads = ctx->thread_pool_max;
|
|
|
|
/*
|
|
* We either have to:
|
|
* - Increase the number available of threads until over the min_threads
|
|
* threshold.
|
|
* - Bump the worker threads so that they exit, until under the max_threads
|
|
* threshold.
|
|
* - Do nothing. The current number of threads fall in between the min and
|
|
* max thresholds. We'll let the pool manage itself.
|
|
*/
|
|
for (int i = pool->cur_threads; i < pool->min_threads; i++) {
|
|
spawn_thread(pool);
|
|
}
|
|
|
|
for (int i = pool->cur_threads; i > pool->max_threads; i--) {
|
|
qemu_cond_signal(&pool->request_cond);
|
|
}
|
|
|
|
qemu_mutex_unlock(&pool->lock);
|
|
}
|
|
|
|
static void thread_pool_init_one(ThreadPoolAio *pool, AioContext *ctx)
|
|
{
|
|
if (!ctx) {
|
|
ctx = qemu_get_aio_context();
|
|
}
|
|
|
|
memset(pool, 0, sizeof(*pool));
|
|
pool->ctx = ctx;
|
|
pool->completion_bh = aio_bh_new(ctx, thread_pool_completion_bh, pool);
|
|
qemu_mutex_init(&pool->lock);
|
|
qemu_cond_init(&pool->worker_stopped);
|
|
qemu_cond_init(&pool->request_cond);
|
|
pool->new_thread_bh = aio_bh_new(ctx, spawn_thread_bh_fn, pool);
|
|
|
|
QLIST_INIT(&pool->head);
|
|
QTAILQ_INIT(&pool->request_list);
|
|
|
|
thread_pool_update_params(pool, ctx);
|
|
}
|
|
|
|
ThreadPoolAio *thread_pool_new_aio(AioContext *ctx)
|
|
{
|
|
ThreadPoolAio *pool = g_new(ThreadPoolAio, 1);
|
|
thread_pool_init_one(pool, ctx);
|
|
return pool;
|
|
}
|
|
|
|
void thread_pool_free_aio(ThreadPoolAio *pool)
|
|
{
|
|
if (!pool) {
|
|
return;
|
|
}
|
|
|
|
assert(QLIST_EMPTY(&pool->head));
|
|
|
|
qemu_mutex_lock(&pool->lock);
|
|
|
|
/* Stop new threads from spawning */
|
|
qemu_bh_delete(pool->new_thread_bh);
|
|
pool->cur_threads -= pool->new_threads;
|
|
pool->new_threads = 0;
|
|
|
|
/* Wait for worker threads to terminate */
|
|
pool->max_threads = 0;
|
|
qemu_cond_broadcast(&pool->request_cond);
|
|
while (pool->cur_threads > 0) {
|
|
qemu_cond_wait(&pool->worker_stopped, &pool->lock);
|
|
}
|
|
|
|
qemu_mutex_unlock(&pool->lock);
|
|
|
|
qemu_bh_delete(pool->completion_bh);
|
|
qemu_cond_destroy(&pool->request_cond);
|
|
qemu_cond_destroy(&pool->worker_stopped);
|
|
qemu_mutex_destroy(&pool->lock);
|
|
g_free(pool);
|
|
}
|
|
|
|
struct ThreadPool {
|
|
GThreadPool *t;
|
|
size_t cur_work;
|
|
QemuMutex cur_work_lock;
|
|
QemuCond all_finished_cond;
|
|
};
|
|
|
|
typedef struct {
|
|
ThreadPoolFunc *func;
|
|
void *opaque;
|
|
GDestroyNotify opaque_destroy;
|
|
} ThreadPoolElement;
|
|
|
|
static void thread_pool_func(gpointer data, gpointer user_data)
|
|
{
|
|
ThreadPool *pool = user_data;
|
|
g_autofree ThreadPoolElement *el = data;
|
|
|
|
el->func(el->opaque);
|
|
|
|
if (el->opaque_destroy) {
|
|
el->opaque_destroy(el->opaque);
|
|
}
|
|
|
|
QEMU_LOCK_GUARD(&pool->cur_work_lock);
|
|
|
|
assert(pool->cur_work > 0);
|
|
pool->cur_work--;
|
|
|
|
if (pool->cur_work == 0) {
|
|
qemu_cond_signal(&pool->all_finished_cond);
|
|
}
|
|
}
|
|
|
|
ThreadPool *thread_pool_new(void)
|
|
{
|
|
ThreadPool *pool = g_new(ThreadPool, 1);
|
|
|
|
pool->cur_work = 0;
|
|
qemu_mutex_init(&pool->cur_work_lock);
|
|
qemu_cond_init(&pool->all_finished_cond);
|
|
|
|
pool->t = g_thread_pool_new(thread_pool_func, pool, 0, TRUE, NULL);
|
|
/*
|
|
* g_thread_pool_new() can only return errors if initial thread(s)
|
|
* creation fails but we ask for 0 initial threads above.
|
|
*/
|
|
assert(pool->t);
|
|
|
|
return pool;
|
|
}
|
|
|
|
void thread_pool_free(ThreadPool *pool)
|
|
{
|
|
/*
|
|
* With _wait = TRUE this effectively waits for all
|
|
* previously submitted work to complete first.
|
|
*/
|
|
g_thread_pool_free(pool->t, FALSE, TRUE);
|
|
|
|
qemu_cond_destroy(&pool->all_finished_cond);
|
|
qemu_mutex_destroy(&pool->cur_work_lock);
|
|
|
|
g_free(pool);
|
|
}
|
|
|
|
void thread_pool_submit(ThreadPool *pool, ThreadPoolFunc *func,
|
|
void *opaque, GDestroyNotify opaque_destroy)
|
|
{
|
|
ThreadPoolElement *el = g_new(ThreadPoolElement, 1);
|
|
|
|
el->func = func;
|
|
el->opaque = opaque;
|
|
el->opaque_destroy = opaque_destroy;
|
|
|
|
WITH_QEMU_LOCK_GUARD(&pool->cur_work_lock) {
|
|
pool->cur_work++;
|
|
}
|
|
|
|
/*
|
|
* Ignore the return value since this function can only return errors
|
|
* if creation of an additional thread fails but even in this case the
|
|
* provided work is still getting queued (just for the existing threads).
|
|
*/
|
|
g_thread_pool_push(pool->t, el, NULL);
|
|
}
|
|
|
|
void thread_pool_submit_immediate(ThreadPool *pool, ThreadPoolFunc *func,
|
|
void *opaque, GDestroyNotify opaque_destroy)
|
|
{
|
|
thread_pool_submit(pool, func, opaque, opaque_destroy);
|
|
thread_pool_adjust_max_threads_to_work(pool);
|
|
}
|
|
|
|
void thread_pool_wait(ThreadPool *pool)
|
|
{
|
|
QEMU_LOCK_GUARD(&pool->cur_work_lock);
|
|
|
|
while (pool->cur_work > 0) {
|
|
qemu_cond_wait(&pool->all_finished_cond,
|
|
&pool->cur_work_lock);
|
|
}
|
|
}
|
|
|
|
bool thread_pool_set_max_threads(ThreadPool *pool,
|
|
int max_threads)
|
|
{
|
|
assert(max_threads > 0);
|
|
|
|
return g_thread_pool_set_max_threads(pool->t, max_threads, NULL);
|
|
}
|
|
|
|
bool thread_pool_adjust_max_threads_to_work(ThreadPool *pool)
|
|
{
|
|
QEMU_LOCK_GUARD(&pool->cur_work_lock);
|
|
|
|
return thread_pool_set_max_threads(pool, pool->cur_work);
|
|
}
|