e3c68b
From 998d9b8b5e271f407e1c654c34f45f0db36abc71 Mon Sep 17 00:00:00 2001
e3c68b
From: Mohammed Rafi KC <rkavunga@redhat.com>
e3c68b
Date: Tue, 21 May 2019 17:15:07 +0530
e3c68b
Subject: [PATCH 172/178] ec/fini: Fix race with ec_fini and ec_notify
e3c68b
e3c68b
During a graph cleanup, we first sent a PARENT_DOWN and wait for
e3c68b
a child down to ultimately free the xlator and the graph.
e3c68b
e3c68b
In the ec xlator, we cleanup the threads when we get a PARENT_DOWN event.
e3c68b
But a racing event like CHILD_UP or event xl_op may trigger healing threads
e3c68b
after threads cleanup.
e3c68b
e3c68b
So there is a chance that the threads might access a freed private variabe
e3c68b
e3c68b
Upstream patch: https://review.gluster.org/#/c/glusterfs/+/22758/
e3c68b
e3c68b
>Change-Id: I252d10181bb67b95900c903d479de707a8489532
e3c68b
>fixes: bz#1703948
e3c68b
>Signed-off-by: Mohammed Rafi KC <rkavunga@redhat.com>
e3c68b
e3c68b
Change-Id: I84a10352d9fb3e68d4147b3791e3af45ab79050e
e3c68b
BUG: 1703434
e3c68b
Signed-off-by: Mohammed Rafi KC <rkavunga@redhat.com>
e3c68b
Reviewed-on: https://code.engineering.redhat.com/gerrit/172285
e3c68b
Tested-by: RHGS Build Bot <nigelb@redhat.com>
e3c68b
Reviewed-by: Atin Mukherjee <amukherj@redhat.com>
e3c68b
---
e3c68b
 libglusterfs/src/glusterfs/xlator.h |  3 +++
e3c68b
 libglusterfs/src/libglusterfs.sym   |  1 +
e3c68b
 libglusterfs/src/xlator.c           | 21 +++++++++++++++++++++
e3c68b
 xlators/cluster/ec/src/ec-heal.c    |  4 ++++
e3c68b
 xlators/cluster/ec/src/ec-heald.c   |  6 ++++++
e3c68b
 xlators/cluster/ec/src/ec.c         |  3 +++
e3c68b
 6 files changed, 38 insertions(+)
e3c68b
e3c68b
diff --git a/libglusterfs/src/glusterfs/xlator.h b/libglusterfs/src/glusterfs/xlator.h
e3c68b
index 8998976..09e463e 100644
e3c68b
--- a/libglusterfs/src/glusterfs/xlator.h
e3c68b
+++ b/libglusterfs/src/glusterfs/xlator.h
e3c68b
@@ -1092,4 +1092,7 @@ gluster_graph_take_reference(xlator_t *tree);
e3c68b
 
e3c68b
 gf_boolean_t
e3c68b
 mgmt_is_multiplexed_daemon(char *name);
e3c68b
+
e3c68b
+gf_boolean_t
e3c68b
+xlator_is_cleanup_starting(xlator_t *this);
e3c68b
 #endif /* _XLATOR_H */
e3c68b
diff --git a/libglusterfs/src/libglusterfs.sym b/libglusterfs/src/libglusterfs.sym
e3c68b
index ec474e7..7a2edef 100644
e3c68b
--- a/libglusterfs/src/libglusterfs.sym
e3c68b
+++ b/libglusterfs/src/libglusterfs.sym
e3c68b
@@ -1161,3 +1161,4 @@ glusterfs_process_svc_attach_volfp
e3c68b
 glusterfs_mux_volfile_reconfigure
e3c68b
 glusterfs_process_svc_detach
e3c68b
 mgmt_is_multiplexed_daemon
e3c68b
+xlator_is_cleanup_starting
e3c68b
diff --git a/libglusterfs/src/xlator.c b/libglusterfs/src/xlator.c
e3c68b
index 022c3ed..fbfbbe2 100644
e3c68b
--- a/libglusterfs/src/xlator.c
e3c68b
+++ b/libglusterfs/src/xlator.c
e3c68b
@@ -1486,3 +1486,24 @@ mgmt_is_multiplexed_daemon(char *name)
e3c68b
     }
e3c68b
     return _gf_false;
e3c68b
 }
e3c68b
+
e3c68b
+gf_boolean_t
e3c68b
+xlator_is_cleanup_starting(xlator_t *this)
e3c68b
+{
e3c68b
+    gf_boolean_t cleanup = _gf_false;
e3c68b
+    glusterfs_graph_t *graph = NULL;
e3c68b
+    xlator_t *xl = NULL;
e3c68b
+
e3c68b
+    if (!this)
e3c68b
+        goto out;
e3c68b
+    graph = this->graph;
e3c68b
+
e3c68b
+    if (!graph)
e3c68b
+        goto out;
e3c68b
+
e3c68b
+    xl = graph->first;
e3c68b
+    if (xl && xl->cleanup_starting)
e3c68b
+        cleanup = _gf_true;
e3c68b
+out:
e3c68b
+    return cleanup;
e3c68b
+}
e3c68b
diff --git a/xlators/cluster/ec/src/ec-heal.c b/xlators/cluster/ec/src/ec-heal.c
e3c68b
index 2fa1f11..8844c29 100644
e3c68b
--- a/xlators/cluster/ec/src/ec-heal.c
e3c68b
+++ b/xlators/cluster/ec/src/ec-heal.c
e3c68b
@@ -2855,6 +2855,10 @@ ec_replace_brick_heal_wrap(void *opaque)
e3c68b
         itable = ec->xl->itable;
e3c68b
     else
e3c68b
         goto out;
e3c68b
+
e3c68b
+    if (xlator_is_cleanup_starting(ec->xl))
e3c68b
+        goto out;
e3c68b
+
e3c68b
     ret = ec_replace_heal(ec, itable->root);
e3c68b
 out:
e3c68b
     return ret;
e3c68b
diff --git a/xlators/cluster/ec/src/ec-heald.c b/xlators/cluster/ec/src/ec-heald.c
e3c68b
index edf5e11..91512d7 100644
e3c68b
--- a/xlators/cluster/ec/src/ec-heald.c
e3c68b
+++ b/xlators/cluster/ec/src/ec-heald.c
e3c68b
@@ -444,6 +444,9 @@ unlock:
e3c68b
 int
e3c68b
 ec_shd_full_healer_spawn(xlator_t *this, int subvol)
e3c68b
 {
e3c68b
+    if (xlator_is_cleanup_starting(this))
e3c68b
+        return -1;
e3c68b
+
e3c68b
     return ec_shd_healer_spawn(this, NTH_FULL_HEALER(this, subvol),
e3c68b
                                ec_shd_full_healer);
e3c68b
 }
e3c68b
@@ -451,6 +454,9 @@ ec_shd_full_healer_spawn(xlator_t *this, int subvol)
e3c68b
 int
e3c68b
 ec_shd_index_healer_spawn(xlator_t *this, int subvol)
e3c68b
 {
e3c68b
+    if (xlator_is_cleanup_starting(this))
e3c68b
+        return -1;
e3c68b
+
e3c68b
     return ec_shd_healer_spawn(this, NTH_INDEX_HEALER(this, subvol),
e3c68b
                                ec_shd_index_healer);
e3c68b
 }
e3c68b
diff --git a/xlators/cluster/ec/src/ec.c b/xlators/cluster/ec/src/ec.c
e3c68b
index 264582a..df5912c 100644
e3c68b
--- a/xlators/cluster/ec/src/ec.c
e3c68b
+++ b/xlators/cluster/ec/src/ec.c
e3c68b
@@ -486,6 +486,9 @@ ec_set_up_state(ec_t *ec, uintptr_t index_mask, uintptr_t new_state)
e3c68b
 {
e3c68b
     uintptr_t current_state = 0;
e3c68b
 
e3c68b
+    if (xlator_is_cleanup_starting(ec->xl))
e3c68b
+        return _gf_false;
e3c68b
+
e3c68b
     if ((ec->xl_notify & index_mask) == 0) {
e3c68b
         ec->xl_notify |= index_mask;
e3c68b
         ec->xl_notify_count++;
e3c68b
-- 
e3c68b
1.8.3.1
e3c68b