thebeanogamer / rpms / qemu-kvm

Forked from rpms/qemu-kvm 5 months ago
Clone
26ba25
From 6fe050124f9433521227b5da8f560c7a3c248513 Mon Sep 17 00:00:00 2001
26ba25
From: Kevin Wolf <kwolf@redhat.com>
26ba25
Date: Wed, 10 Oct 2018 20:08:38 +0100
26ba25
Subject: [PATCH 07/49] block: Don't manually poll in bdrv_drain_all()
26ba25
26ba25
RH-Author: Kevin Wolf <kwolf@redhat.com>
26ba25
Message-id: <20181010200843.6710-5-kwolf@redhat.com>
26ba25
Patchwork-id: 82584
26ba25
O-Subject: [RHEL-8 qemu-kvm PATCH 04/44] block: Don't manually poll in bdrv_drain_all()
26ba25
Bugzilla: 1637976
26ba25
RH-Acked-by: Max Reitz <mreitz@redhat.com>
26ba25
RH-Acked-by: John Snow <jsnow@redhat.com>
26ba25
RH-Acked-by: Thomas Huth <thuth@redhat.com>
26ba25
26ba25
All involved nodes are already idle, we called bdrv_do_drain_begin() on
26ba25
them.
26ba25
26ba25
The comment in the code suggested that this was not correct because the
26ba25
completion of a request on one node could spawn a new request on a
26ba25
different node (which might have been drained before, so we wouldn't
26ba25
drain the new request). In reality, new requests to different nodes
26ba25
aren't spawned out of nothing, but only in the context of a parent
26ba25
request, and they aren't submitted to random nodes, but only to child
26ba25
nodes. As long as we still poll for the completion of the parent request
26ba25
(which we do), draining each root node separately is good enough.
26ba25
26ba25
Remove the additional polling code from bdrv_drain_all_begin() and
26ba25
replace it with an assertion that all nodes are already idle after we
26ba25
drained them separately.
26ba25
26ba25
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
26ba25
Reviewed-by: Stefan Hajnoczi <stefanha@redhat.com>
26ba25
(cherry picked from commit c13ad59f012cbbccb866a10477458e69bc868dbb)
26ba25
Signed-off-by: Kevin Wolf <kwolf@redhat.com>
26ba25
Signed-off-by: Danilo C. L. de Paula <ddepaula@redhat.com>
26ba25
---
26ba25
 block/io.c | 41 ++++++++++++-----------------------------
26ba25
 1 file changed, 12 insertions(+), 29 deletions(-)
26ba25
26ba25
diff --git a/block/io.c b/block/io.c
26ba25
index aa41f1e..e5fc42c 100644
26ba25
--- a/block/io.c
26ba25
+++ b/block/io.c
26ba25
@@ -376,6 +376,16 @@ void bdrv_drain(BlockDriverState *bs)
26ba25
     bdrv_drained_end(bs);
26ba25
 }
26ba25
 
26ba25
+static void bdrv_drain_assert_idle(BlockDriverState *bs)
26ba25
+{
26ba25
+    BdrvChild *child, *next;
26ba25
+
26ba25
+    assert(atomic_read(&bs->in_flight) == 0);
26ba25
+    QLIST_FOREACH_SAFE(child, &bs->children, next, next) {
26ba25
+        bdrv_drain_assert_idle(child->bs);
26ba25
+    }
26ba25
+}
26ba25
+
26ba25
 /*
26ba25
  * Wait for pending requests to complete across all BlockDriverStates
26ba25
  *
26ba25
@@ -390,11 +400,8 @@ void bdrv_drain(BlockDriverState *bs)
26ba25
  */
26ba25
 void bdrv_drain_all_begin(void)
26ba25
 {
26ba25
-    /* Always run first iteration so any pending completion BHs run */
26ba25
-    bool waited = true;
26ba25
     BlockDriverState *bs;
26ba25
     BdrvNextIterator it;
26ba25
-    GSList *aio_ctxs = NULL, *ctx;
26ba25
 
26ba25
     /* BDRV_POLL_WHILE() for a node can only be called from its own I/O thread
26ba25
      * or the main loop AioContext. We potentially use BDRV_POLL_WHILE() on
26ba25
@@ -408,35 +415,11 @@ void bdrv_drain_all_begin(void)
26ba25
         aio_context_acquire(aio_context);
26ba25
         bdrv_do_drained_begin(bs, true, NULL);
26ba25
         aio_context_release(aio_context);
26ba25
-
26ba25
-        if (!g_slist_find(aio_ctxs, aio_context)) {
26ba25
-            aio_ctxs = g_slist_prepend(aio_ctxs, aio_context);
26ba25
-        }
26ba25
     }
26ba25
 
26ba25
-    /* Note that completion of an asynchronous I/O operation can trigger any
26ba25
-     * number of other I/O operations on other devices---for example a
26ba25
-     * coroutine can submit an I/O request to another device in response to
26ba25
-     * request completion.  Therefore we must keep looping until there was no
26ba25
-     * more activity rather than simply draining each device independently.
26ba25
-     */
26ba25
-    while (waited) {
26ba25
-        waited = false;
26ba25
-
26ba25
-        for (ctx = aio_ctxs; ctx != NULL; ctx = ctx->next) {
26ba25
-            AioContext *aio_context = ctx->data;
26ba25
-
26ba25
-            aio_context_acquire(aio_context);
26ba25
-            for (bs = bdrv_first(&it); bs; bs = bdrv_next(&it)) {
26ba25
-                if (aio_context == bdrv_get_aio_context(bs)) {
26ba25
-                    waited |= bdrv_drain_recurse(bs);
26ba25
-                }
26ba25
-            }
26ba25
-            aio_context_release(aio_context);
26ba25
-        }
26ba25
+    for (bs = bdrv_first(&it); bs; bs = bdrv_next(&it)) {
26ba25
+        bdrv_drain_assert_idle(bs);
26ba25
     }
26ba25
-
26ba25
-    g_slist_free(aio_ctxs);
26ba25
 }
26ba25
 
26ba25
 void bdrv_drain_all_end(void)
26ba25
-- 
26ba25
1.8.3.1
26ba25