|
|
383d26 |
From cb638b3d14fb19db26f6108a08f1a3b62a96b6c3 Mon Sep 17 00:00:00 2001
|
|
|
383d26 |
From: "Dr. David Alan Gilbert" <dgilbert@redhat.com>
|
|
|
383d26 |
Date: Fri, 22 Jun 2018 18:59:53 +0200
|
|
|
383d26 |
Subject: [PATCH 14/57] migration: move some code to ram_save_host_page
|
|
|
383d26 |
|
|
|
383d26 |
RH-Author: Dr. David Alan Gilbert <dgilbert@redhat.com>
|
|
|
383d26 |
Message-id: <20180622190005.21297-7-dgilbert@redhat.com>
|
|
|
383d26 |
Patchwork-id: 81007
|
|
|
383d26 |
O-Subject: [RHEL7.6 qemu-kvm-rhev PATCH 06/18] migration: move some code to ram_save_host_page
|
|
|
383d26 |
Bugzilla: 1584139
|
|
|
383d26 |
RH-Acked-by: Peter Xu <peterx@redhat.com>
|
|
|
383d26 |
RH-Acked-by: Juan Quintela <quintela@redhat.com>
|
|
|
383d26 |
RH-Acked-by: Laurent Vivier <lvivier@redhat.com>
|
|
|
383d26 |
|
|
|
383d26 |
From: Xiao Guangrong <xiaoguangrong@tencent.com>
|
|
|
383d26 |
|
|
|
383d26 |
Move some code from ram_save_target_page() to ram_save_host_page()
|
|
|
383d26 |
to make it be more readable for latter patches that dramatically
|
|
|
383d26 |
clean ram_save_target_page() up
|
|
|
383d26 |
|
|
|
383d26 |
Reviewed-by: Peter Xu <peterx@redhat.com>
|
|
|
383d26 |
Signed-off-by: Xiao Guangrong <xiaoguangrong@tencent.com>
|
|
|
383d26 |
Message-Id: <20180330075128.26919-7-xiaoguangrong@tencent.com>
|
|
|
383d26 |
Signed-off-by: Dr. David Alan Gilbert <dgilbert@redhat.com>
|
|
|
383d26 |
(cherry picked from commit 1faa5665c0f1df2eff291454a3a85625a3bc93dd)
|
|
|
383d26 |
Signed-off-by: Miroslav Rezanina <mrezanin@redhat.com>
|
|
|
383d26 |
---
|
|
|
383d26 |
migration/ram.c | 43 +++++++++++++++++++------------------------
|
|
|
383d26 |
1 file changed, 19 insertions(+), 24 deletions(-)
|
|
|
383d26 |
|
|
|
383d26 |
diff --git a/migration/ram.c b/migration/ram.c
|
|
|
383d26 |
index 8dc98a5..106fcf1 100644
|
|
|
383d26 |
--- a/migration/ram.c
|
|
|
383d26 |
+++ b/migration/ram.c
|
|
|
383d26 |
@@ -1484,38 +1484,23 @@ err:
|
|
|
383d26 |
* Returns the number of pages written
|
|
|
383d26 |
*
|
|
|
383d26 |
* @rs: current RAM state
|
|
|
383d26 |
- * @ms: current migration state
|
|
|
383d26 |
* @pss: data about the page we want to send
|
|
|
383d26 |
* @last_stage: if we are at the completion stage
|
|
|
383d26 |
*/
|
|
|
383d26 |
static int ram_save_target_page(RAMState *rs, PageSearchStatus *pss,
|
|
|
383d26 |
bool last_stage)
|
|
|
383d26 |
{
|
|
|
383d26 |
- int res = 0;
|
|
|
383d26 |
-
|
|
|
383d26 |
- /* Check the pages is dirty and if it is send it */
|
|
|
383d26 |
- if (migration_bitmap_clear_dirty(rs, pss->block, pss->page)) {
|
|
|
383d26 |
- /*
|
|
|
383d26 |
- * If xbzrle is on, stop using the data compression after first
|
|
|
383d26 |
- * round of migration even if compression is enabled. In theory,
|
|
|
383d26 |
- * xbzrle can do better than compression.
|
|
|
383d26 |
- */
|
|
|
383d26 |
- if (migrate_use_compression() &&
|
|
|
383d26 |
- (rs->ram_bulk_stage || !migrate_use_xbzrle())) {
|
|
|
383d26 |
- res = ram_save_compressed_page(rs, pss, last_stage);
|
|
|
383d26 |
- } else {
|
|
|
383d26 |
- res = ram_save_page(rs, pss, last_stage);
|
|
|
383d26 |
- }
|
|
|
383d26 |
-
|
|
|
383d26 |
- if (res < 0) {
|
|
|
383d26 |
- return res;
|
|
|
383d26 |
- }
|
|
|
383d26 |
- if (pss->block->unsentmap) {
|
|
|
383d26 |
- clear_bit(pss->page, pss->block->unsentmap);
|
|
|
383d26 |
- }
|
|
|
383d26 |
+ /*
|
|
|
383d26 |
+ * If xbzrle is on, stop using the data compression after first
|
|
|
383d26 |
+ * round of migration even if compression is enabled. In theory,
|
|
|
383d26 |
+ * xbzrle can do better than compression.
|
|
|
383d26 |
+ */
|
|
|
383d26 |
+ if (migrate_use_compression() &&
|
|
|
383d26 |
+ (rs->ram_bulk_stage || !migrate_use_xbzrle())) {
|
|
|
383d26 |
+ return ram_save_compressed_page(rs, pss, last_stage);
|
|
|
383d26 |
}
|
|
|
383d26 |
|
|
|
383d26 |
- return res;
|
|
|
383d26 |
+ return ram_save_page(rs, pss, last_stage);
|
|
|
383d26 |
}
|
|
|
383d26 |
|
|
|
383d26 |
/**
|
|
|
383d26 |
@@ -1544,12 +1529,22 @@ static int ram_save_host_page(RAMState *rs, PageSearchStatus *pss,
|
|
|
383d26 |
qemu_ram_pagesize(pss->block) >> TARGET_PAGE_BITS;
|
|
|
383d26 |
|
|
|
383d26 |
do {
|
|
|
383d26 |
+ /* Check the pages is dirty and if it is send it */
|
|
|
383d26 |
+ if (!migration_bitmap_clear_dirty(rs, pss->block, pss->page)) {
|
|
|
383d26 |
+ pss->page++;
|
|
|
383d26 |
+ continue;
|
|
|
383d26 |
+ }
|
|
|
383d26 |
+
|
|
|
383d26 |
tmppages = ram_save_target_page(rs, pss, last_stage);
|
|
|
383d26 |
if (tmppages < 0) {
|
|
|
383d26 |
return tmppages;
|
|
|
383d26 |
}
|
|
|
383d26 |
|
|
|
383d26 |
pages += tmppages;
|
|
|
383d26 |
+ if (pss->block->unsentmap) {
|
|
|
383d26 |
+ clear_bit(pss->page, pss->block->unsentmap);
|
|
|
383d26 |
+ }
|
|
|
383d26 |
+
|
|
|
383d26 |
pss->page++;
|
|
|
383d26 |
} while ((pss->page & (pagesize_bits - 1)) &&
|
|
|
383d26 |
offset_in_ramblock(pss->block, pss->page << TARGET_PAGE_BITS));
|
|
|
383d26 |
--
|
|
|
383d26 |
1.8.3.1
|
|
|
383d26 |
|