Blame SOURCES/kvm-util-event-loop-base-Introduce-options-to-set-the-th.patch

586cba
From 7a6fa42d4a4263c94b9bf18290f9e7680ea9e7f4 Mon Sep 17 00:00:00 2001
586cba
From: Nicolas Saenz Julienne <nsaenzju@redhat.com>
586cba
Date: Mon, 25 Apr 2022 09:57:23 +0200
586cba
Subject: [PATCH 03/16] util/event-loop-base: Introduce options to set the
586cba
 thread pool size
586cba
586cba
RH-Author: Nicolas Saenz Julienne <nsaenzju@redhat.com>
586cba
RH-MergeRequest: 93: util/thread-pool: Expose minimum and maximum size
586cba
RH-Commit: [3/3] af78a88ff3c69701cbb5f9e980c3d6ebbd13ff98
586cba
RH-Bugzilla: 2031024
586cba
RH-Acked-by: Miroslav Rezanina <mrezanin@redhat.com>
586cba
RH-Acked-by: Stefano Garzarella <sgarzare@redhat.com>
586cba
RH-Acked-by: Stefan Hajnoczi <stefanha@redhat.com>
586cba
586cba
The thread pool regulates itself: when idle, it kills threads until
586cba
empty, when in demand, it creates new threads until full. This behaviour
586cba
doesn't play well with latency sensitive workloads where the price of
586cba
creating a new thread is too high. For example, when paired with qemu's
586cba
'-mlock', or using safety features like SafeStack, creating a new thread
586cba
has been measured take multiple milliseconds.
586cba
586cba
In order to mitigate this let's introduce a new 'EventLoopBase'
586cba
property to set the thread pool size. The threads will be created during
586cba
the pool's initialization or upon updating the property's value, remain
586cba
available during its lifetime regardless of demand, and destroyed upon
586cba
freeing it. A properly characterized workload will then be able to
586cba
configure the pool to avoid any latency spikes.
586cba
586cba
Signed-off-by: Nicolas Saenz Julienne <nsaenzju@redhat.com>
586cba
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
586cba
Acked-by: Markus Armbruster <armbru@redhat.com>
586cba
Message-id: 20220425075723.20019-4-nsaenzju@redhat.com
586cba
Signed-off-by: Stefan Hajnoczi <stefanha@redhat.com>
586cba
(cherry picked from commit 71ad4713cc1d7fca24388b828ef31ae6cb38a31c)
586cba
---
586cba
 event-loop-base.c                | 23 +++++++++++++
586cba
 include/block/aio.h              | 10 ++++++
586cba
 include/block/thread-pool.h      |  3 ++
586cba
 include/sysemu/event-loop-base.h |  4 +++
586cba
 iothread.c                       |  3 ++
586cba
 qapi/qom.json                    | 10 +++++-
586cba
 util/aio-posix.c                 |  1 +
586cba
 util/async.c                     | 20 ++++++++++++
586cba
 util/main-loop.c                 |  9 ++++++
586cba
 util/thread-pool.c               | 55 +++++++++++++++++++++++++++++---
586cba
 10 files changed, 133 insertions(+), 5 deletions(-)
586cba
586cba
diff --git a/event-loop-base.c b/event-loop-base.c
586cba
index e7f99a6ec8..d5be4dc6fc 100644
586cba
--- a/event-loop-base.c
586cba
+++ b/event-loop-base.c
586cba
@@ -14,6 +14,7 @@
586cba
 #include "qemu/osdep.h"
586cba
 #include "qom/object_interfaces.h"
586cba
 #include "qapi/error.h"
586cba
+#include "block/thread-pool.h"
586cba
 #include "sysemu/event-loop-base.h"
586cba
 
586cba
 typedef struct {
586cba
@@ -21,9 +22,22 @@ typedef struct {
586cba
     ptrdiff_t offset; /* field's byte offset in EventLoopBase struct */
586cba
 } EventLoopBaseParamInfo;
586cba
 
586cba
+static void event_loop_base_instance_init(Object *obj)
586cba
+{
586cba
+    EventLoopBase *base = EVENT_LOOP_BASE(obj);
586cba
+
586cba
+    base->thread_pool_max = THREAD_POOL_MAX_THREADS_DEFAULT;
586cba
+}
586cba
+
586cba
 static EventLoopBaseParamInfo aio_max_batch_info = {
586cba
     "aio-max-batch", offsetof(EventLoopBase, aio_max_batch),
586cba
 };
586cba
+static EventLoopBaseParamInfo thread_pool_min_info = {
586cba
+    "thread-pool-min", offsetof(EventLoopBase, thread_pool_min),
586cba
+};
586cba
+static EventLoopBaseParamInfo thread_pool_max_info = {
586cba
+    "thread-pool-max", offsetof(EventLoopBase, thread_pool_max),
586cba
+};
586cba
 
586cba
 static void event_loop_base_get_param(Object *obj, Visitor *v,
586cba
         const char *name, void *opaque, Error **errp)
586cba
@@ -95,12 +109,21 @@ static void event_loop_base_class_init(ObjectClass *klass, void *class_data)
586cba
                               event_loop_base_get_param,
586cba
                               event_loop_base_set_param,
586cba
                               NULL, &aio_max_batch_info);
586cba
+    object_class_property_add(klass, "thread-pool-min", "int",
586cba
+                              event_loop_base_get_param,
586cba
+                              event_loop_base_set_param,
586cba
+                              NULL, &thread_pool_min_info);
586cba
+    object_class_property_add(klass, "thread-pool-max", "int",
586cba
+                              event_loop_base_get_param,
586cba
+                              event_loop_base_set_param,
586cba
+                              NULL, &thread_pool_max_info);
586cba
 }
586cba
 
586cba
 static const TypeInfo event_loop_base_info = {
586cba
     .name = TYPE_EVENT_LOOP_BASE,
586cba
     .parent = TYPE_OBJECT,
586cba
     .instance_size = sizeof(EventLoopBase),
586cba
+    .instance_init = event_loop_base_instance_init,
586cba
     .class_size = sizeof(EventLoopBaseClass),
586cba
     .class_init = event_loop_base_class_init,
586cba
     .abstract = true,
586cba
diff --git a/include/block/aio.h b/include/block/aio.h
586cba
index 5634173b12..d128558f1d 100644
586cba
--- a/include/block/aio.h
586cba
+++ b/include/block/aio.h
586cba
@@ -192,6 +192,8 @@ struct AioContext {
586cba
     QSLIST_HEAD(, Coroutine) scheduled_coroutines;
586cba
     QEMUBH *co_schedule_bh;
586cba
 
586cba
+    int thread_pool_min;
586cba
+    int thread_pool_max;
586cba
     /* Thread pool for performing work and receiving completion callbacks.
586cba
      * Has its own locking.
586cba
      */
586cba
@@ -769,4 +771,12 @@ void aio_context_set_poll_params(AioContext *ctx, int64_t max_ns,
586cba
 void aio_context_set_aio_params(AioContext *ctx, int64_t max_batch,
586cba
                                 Error **errp);
586cba
 
586cba
+/**
586cba
+ * aio_context_set_thread_pool_params:
586cba
+ * @ctx: the aio context
586cba
+ * @min: min number of threads to have readily available in the thread pool
586cba
+ * @min: max number of threads the thread pool can contain
586cba
+ */
586cba
+void aio_context_set_thread_pool_params(AioContext *ctx, int64_t min,
586cba
+                                        int64_t max, Error **errp);
586cba
 #endif
586cba
diff --git a/include/block/thread-pool.h b/include/block/thread-pool.h
586cba
index 7dd7d730a0..2020bcc92d 100644
586cba
--- a/include/block/thread-pool.h
586cba
+++ b/include/block/thread-pool.h
586cba
@@ -20,6 +20,8 @@
586cba
 
586cba
 #include "block/block.h"
586cba
 
586cba
+#define THREAD_POOL_MAX_THREADS_DEFAULT         64
586cba
+
586cba
 typedef int ThreadPoolFunc(void *opaque);
586cba
 
586cba
 typedef struct ThreadPool ThreadPool;
586cba
@@ -33,5 +35,6 @@ BlockAIOCB *thread_pool_submit_aio(ThreadPool *pool,
586cba
 int coroutine_fn thread_pool_submit_co(ThreadPool *pool,
586cba
         ThreadPoolFunc *func, void *arg);
586cba
 void thread_pool_submit(ThreadPool *pool, ThreadPoolFunc *func, void *arg);
586cba
+void thread_pool_update_params(ThreadPool *pool, struct AioContext *ctx);
586cba
 
586cba
 #endif
586cba
diff --git a/include/sysemu/event-loop-base.h b/include/sysemu/event-loop-base.h
586cba
index fced4c9fea..2748bf6ae1 100644
586cba
--- a/include/sysemu/event-loop-base.h
586cba
+++ b/include/sysemu/event-loop-base.h
586cba
@@ -33,5 +33,9 @@ struct EventLoopBase {
586cba
 
586cba
     /* AioContext AIO engine parameters */
586cba
     int64_t aio_max_batch;
586cba
+
586cba
+    /* AioContext thread pool parameters */
586cba
+    int64_t thread_pool_min;
586cba
+    int64_t thread_pool_max;
586cba
 };
586cba
 #endif
586cba
diff --git a/iothread.c b/iothread.c
586cba
index 8fa2f3bfb8..529194a566 100644
586cba
--- a/iothread.c
586cba
+++ b/iothread.c
586cba
@@ -174,6 +174,9 @@ static void iothread_set_aio_context_params(EventLoopBase *base, Error **errp)
586cba
     aio_context_set_aio_params(iothread->ctx,
586cba
                                iothread->parent_obj.aio_max_batch,
586cba
                                errp);
586cba
+
586cba
+    aio_context_set_thread_pool_params(iothread->ctx, base->thread_pool_min,
586cba
+                                       base->thread_pool_max, errp);
586cba
 }
586cba
 
586cba
 
586cba
diff --git a/qapi/qom.json b/qapi/qom.json
586cba
index 7d4a2ac1b9..6a653c6636 100644
586cba
--- a/qapi/qom.json
586cba
+++ b/qapi/qom.json
586cba
@@ -508,10 +508,18 @@
586cba
 #                 0 means that the engine will use its default.
586cba
 #                 (default: 0)
586cba
 #
586cba
+# @thread-pool-min: minimum number of threads reserved in the thread pool
586cba
+#                   (default:0)
586cba
+#
586cba
+# @thread-pool-max: maximum number of threads the thread pool can contain
586cba
+#                   (default:64)
586cba
+#
586cba
 # Since: 7.1
586cba
 ##
586cba
 { 'struct': 'EventLoopBaseProperties',
586cba
-  'data': { '*aio-max-batch': 'int' } }
586cba
+  'data': { '*aio-max-batch': 'int',
586cba
+            '*thread-pool-min': 'int',
586cba
+            '*thread-pool-max': 'int' } }
586cba
 
586cba
 ##
586cba
 # @IothreadProperties:
586cba
diff --git a/util/aio-posix.c b/util/aio-posix.c
586cba
index be0182a3c6..731f3826c0 100644
586cba
--- a/util/aio-posix.c
586cba
+++ b/util/aio-posix.c
586cba
@@ -15,6 +15,7 @@
586cba
 
586cba
 #include "qemu/osdep.h"
586cba
 #include "block/block.h"
586cba
+#include "block/thread-pool.h"
586cba
 #include "qemu/main-loop.h"
586cba
 #include "qemu/rcu.h"
586cba
 #include "qemu/rcu_queue.h"
586cba
diff --git a/util/async.c b/util/async.c
586cba
index 2ea1172f3e..554ba70cca 100644
586cba
--- a/util/async.c
586cba
+++ b/util/async.c
586cba
@@ -563,6 +563,9 @@ AioContext *aio_context_new(Error **errp)
586cba
 
586cba
     ctx->aio_max_batch = 0;
586cba
 
586cba
+    ctx->thread_pool_min = 0;
586cba
+    ctx->thread_pool_max = THREAD_POOL_MAX_THREADS_DEFAULT;
586cba
+
586cba
     return ctx;
586cba
 fail:
586cba
     g_source_destroy(&ctx->source);
586cba
@@ -696,3 +699,20 @@ void qemu_set_current_aio_context(AioContext *ctx)
586cba
     assert(!get_my_aiocontext());
586cba
     set_my_aiocontext(ctx);
586cba
 }
586cba
+
586cba
+void aio_context_set_thread_pool_params(AioContext *ctx, int64_t min,
586cba
+                                        int64_t max, Error **errp)
586cba
+{
586cba
+
586cba
+    if (min > max || !max || min > INT_MAX || max > INT_MAX) {
586cba
+        error_setg(errp, "bad thread-pool-min/thread-pool-max values");
586cba
+        return;
586cba
+    }
586cba
+
586cba
+    ctx->thread_pool_min = min;
586cba
+    ctx->thread_pool_max = max;
586cba
+
586cba
+    if (ctx->thread_pool) {
586cba
+        thread_pool_update_params(ctx->thread_pool, ctx);
586cba
+    }
586cba
+}
586cba
diff --git a/util/main-loop.c b/util/main-loop.c
586cba
index 5b13f456fa..a0f48186ab 100644
586cba
--- a/util/main-loop.c
586cba
+++ b/util/main-loop.c
586cba
@@ -30,6 +30,7 @@
586cba
 #include "sysemu/replay.h"
586cba
 #include "qemu/main-loop.h"
586cba
 #include "block/aio.h"
586cba
+#include "block/thread-pool.h"
586cba
 #include "qemu/error-report.h"
586cba
 #include "qemu/queue.h"
586cba
 #include "qemu/compiler.h"
586cba
@@ -187,12 +188,20 @@ int qemu_init_main_loop(Error **errp)
586cba
 
586cba
 static void main_loop_update_params(EventLoopBase *base, Error **errp)
586cba
 {
586cba
+    ERRP_GUARD();
586cba
+
586cba
     if (!qemu_aio_context) {
586cba
         error_setg(errp, "qemu aio context not ready");
586cba
         return;
586cba
     }
586cba
 
586cba
     aio_context_set_aio_params(qemu_aio_context, base->aio_max_batch, errp);
586cba
+    if (*errp) {
586cba
+        return;
586cba
+    }
586cba
+
586cba
+    aio_context_set_thread_pool_params(qemu_aio_context, base->thread_pool_min,
586cba
+                                       base->thread_pool_max, errp);
586cba
 }
586cba
 
586cba
 MainLoop *mloop;
586cba
diff --git a/util/thread-pool.c b/util/thread-pool.c
586cba
index d763cea505..196835b4d3 100644
586cba
--- a/util/thread-pool.c
586cba
+++ b/util/thread-pool.c
586cba
@@ -58,7 +58,6 @@ struct ThreadPool {
586cba
     QemuMutex lock;
586cba
     QemuCond worker_stopped;
586cba
     QemuSemaphore sem;
586cba
-    int max_threads;
586cba
     QEMUBH *new_thread_bh;
586cba
 
586cba
     /* The following variables are only accessed from one AioContext. */
586cba
@@ -71,8 +70,27 @@ struct ThreadPool {
586cba
     int new_threads;     /* backlog of threads we need to create */
586cba
     int pending_threads; /* threads created but not running yet */
586cba
     bool stopping;
586cba
+    int min_threads;
586cba
+    int max_threads;
586cba
 };
586cba
 
586cba
+static inline bool back_to_sleep(ThreadPool *pool, int ret)
586cba
+{
586cba
+    /*
586cba
+     * The semaphore timed out, we should exit the loop except when:
586cba
+     *  - There is work to do, we raced with the signal.
586cba
+     *  - The max threads threshold just changed, we raced with the signal.
586cba
+     *  - The thread pool forces a minimum number of readily available threads.
586cba
+     */
586cba
+    if (ret == -1 && (!QTAILQ_EMPTY(&pool->request_list) ||
586cba
+            pool->cur_threads > pool->max_threads ||
586cba
+            pool->cur_threads <= pool->min_threads)) {
586cba
+            return true;
586cba
+    }
586cba
+
586cba
+    return false;
586cba
+}
586cba
+
586cba
 static void *worker_thread(void *opaque)
586cba
 {
586cba
     ThreadPool *pool = opaque;
586cba
@@ -91,8 +109,9 @@ static void *worker_thread(void *opaque)
586cba
             ret = qemu_sem_timedwait(&pool->sem, 10000);
586cba
             qemu_mutex_lock(&pool->lock);
586cba
             pool->idle_threads--;
586cba
-        } while (ret == -1 && !QTAILQ_EMPTY(&pool->request_list));
586cba
-        if (ret == -1 || pool->stopping) {
586cba
+        } while (back_to_sleep(pool, ret));
586cba
+        if (ret == -1 || pool->stopping ||
586cba
+            pool->cur_threads > pool->max_threads) {
586cba
             break;
586cba
         }
586cba
 
586cba
@@ -294,6 +313,33 @@ void thread_pool_submit(ThreadPool *pool, ThreadPoolFunc *func, void *arg)
586cba
     thread_pool_submit_aio(pool, func, arg, NULL, NULL);
586cba
 }
586cba
 
586cba
+void thread_pool_update_params(ThreadPool *pool, AioContext *ctx)
586cba
+{
586cba
+    qemu_mutex_lock(&pool->lock);
586cba
+
586cba
+    pool->min_threads = ctx->thread_pool_min;
586cba
+    pool->max_threads = ctx->thread_pool_max;
586cba
+
586cba
+    /*
586cba
+     * We either have to:
586cba
+     *  - Increase the number available of threads until over the min_threads
586cba
+     *    threshold.
586cba
+     *  - Decrease the number of available threads until under the max_threads
586cba
+     *    threshold.
586cba
+     *  - Do nothing. The current number of threads fall in between the min and
586cba
+     *    max thresholds. We'll let the pool manage itself.
586cba
+     */
586cba
+    for (int i = pool->cur_threads; i < pool->min_threads; i++) {
586cba
+        spawn_thread(pool);
586cba
+    }
586cba
+
586cba
+    for (int i = pool->cur_threads; i > pool->max_threads; i--) {
586cba
+        qemu_sem_post(&pool->sem);
586cba
+    }
586cba
+
586cba
+    qemu_mutex_unlock(&pool->lock);
586cba
+}
586cba
+
586cba
 static void thread_pool_init_one(ThreadPool *pool, AioContext *ctx)
586cba
 {
586cba
     if (!ctx) {
586cba
@@ -306,11 +352,12 @@ static void thread_pool_init_one(ThreadPool *pool, AioContext *ctx)
586cba
     qemu_mutex_init(&pool->lock);
586cba
     qemu_cond_init(&pool->worker_stopped);
586cba
     qemu_sem_init(&pool->sem, 0);
586cba
-    pool->max_threads = 64;
586cba
     pool->new_thread_bh = aio_bh_new(ctx, spawn_thread_bh_fn, pool);
586cba
 
586cba
     QLIST_INIT(&pool->head);
586cba
     QTAILQ_INIT(&pool->request_list);
586cba
+
586cba
+    thread_pool_update_params(pool, ctx);
586cba
 }
586cba
 
586cba
 ThreadPool *thread_pool_new(AioContext *ctx)
586cba
-- 
586cba
2.31.1
586cba