From 53042d3aa2a619c0e5e56d0707ea9d503429d28a Mon Sep 17 00:00:00 2001 From: Mohit Agrawal Date: Mon, 12 Mar 2018 19:43:15 +0530 Subject: [PATCH 226/236] gluster: Sometimes Brick process is crashed at the time of stopping brick Problem: Sometimes brick process is getting crashed at the time of stop brick while brick mux is enabled. Solution: Brick process was getting crashed because of rpc connection was not cleaning properly while brick mux is enabled.In this patch after sending GF_EVENT_CLEANUP notification to xlator(server) waits for all rpc client connection destroy for specific xlator.Once rpc connections are destroyed in server_rpc_notify for all associated client for that brick then call xlator_mem_cleanup for for brick xlator as well as all child xlators.To avoid races at the time of cleanup introduce two new flags at each xlator cleanup_starting, call_cleanup. > BUG: 1544090 > Signed-off-by: Mohit Agrawal > (cherry picked from commit 0043c63f70776444f69667a4ef9596217ecb42b7) > (Upstream link for review https://review.gluster.org/#/c/19734/) > Note: Run all test-cases in separate build (https://review.gluster.org/#/c/19700/) > with same patch after enable brick mux forcefully, all test cases are > passed. BUG: 1535281 Signed-off-by: Mohit Agrawal Change-Id: Ic4ab9c128df282d146cf1135640281fcb31997bf Reviewed-on: https://code.engineering.redhat.com/gerrit/136229 Tested-by: RHGS Build Bot Reviewed-by: Atin Mukherjee --- glusterfsd/src/glusterfsd-mgmt.c | 96 +++++++++++--- libglusterfs/src/client_t.c | 4 +- libglusterfs/src/fd.c | 4 +- libglusterfs/src/graph.c | 6 +- libglusterfs/src/inode.c | 2 +- libglusterfs/src/statedump.c | 61 ++++++--- libglusterfs/src/xlator.c | 8 +- libglusterfs/src/xlator.h | 11 +- tests/volume.rc | 31 ++++- xlators/features/changelog/src/changelog-rpc.c | 9 ++ xlators/features/changelog/src/changelog.c | 5 + xlators/features/trash/src/trash.c | 18 ++- xlators/protocol/server/src/server-handshake.c | 25 +++- xlators/protocol/server/src/server.c | 170 ++++++++++++++++++------- xlators/protocol/server/src/server.h | 2 +- xlators/storage/posix/src/posix-helpers.c | 14 +- xlators/storage/posix/src/posix.c | 6 +- 17 files changed, 362 insertions(+), 110 deletions(-) diff --git a/glusterfsd/src/glusterfsd-mgmt.c b/glusterfsd/src/glusterfsd-mgmt.c index 8f4450b..fdf403c 100644 --- a/glusterfsd/src/glusterfsd-mgmt.c +++ b/glusterfsd/src/glusterfsd-mgmt.c @@ -206,11 +206,6 @@ xlator_mem_free (xlator_t *xl) if (!xl) return 0; - GF_FREE (xl->name); - GF_FREE (xl->type); - xl->name = NULL; - xl->type = NULL; - if (xl->options) { dict_ref (xl->options); dict_unref (xl->options); @@ -222,25 +217,38 @@ xlator_mem_free (xlator_t *xl) GF_FREE (vol_opt); } + xlator_memrec_free (xl); + return 0; } void xlator_call_fini (xlator_t *this) { - if (!this) + if (!this || this->cleanup_starting) return; + this->cleanup_starting = 1; + this->call_cleanup = 1; xlator_call_fini (this->next); this->fini (this); } void xlator_mem_cleanup (xlator_t *this) { - xlator_list_t *list = this->children; - xlator_t *trav = list->xlator; - inode_table_t *inode_table = NULL; - xlator_t *prev = trav; + xlator_list_t *list = this->children; + xlator_t *trav = list->xlator; + inode_table_t *inode_table = NULL; + xlator_t *prev = trav; + glusterfs_ctx_t *ctx = NULL; + xlator_list_t **trav_p = NULL; + xlator_t *top = NULL; + xlator_t *victim = NULL; - inode_table = this->itable; + + if (this->call_cleanup) + return; + + this->call_cleanup = 1; + ctx = glusterfsd_ctx; xlator_call_fini (trav); @@ -250,6 +258,7 @@ xlator_mem_cleanup (xlator_t *this) { prev = trav; } + inode_table = this->itable; if (inode_table) { inode_table_destroy (inode_table); this->itable = NULL; @@ -260,6 +269,33 @@ xlator_mem_cleanup (xlator_t *this) { } xlator_mem_free (this); + + if (glusterfsd_ctx->active) { + top = glusterfsd_ctx->active->first; + LOCK (&ctx->volfile_lock); + /* TODO here we have leak for xlator node in a graph */ + for (trav_p = &top->children; *trav_p; trav_p = &(*trav_p)->next) { + victim = (*trav_p)->xlator; + if (victim->call_cleanup && !strcmp (victim->name, this->name)) { + (*trav_p) = (*trav_p)->next; + break; + } + } + /* TODO Sometime brick xlator is not moved from graph so followed below + approach to move brick xlator from a graph, will move specific brick + xlator from graph only while inode table and mem_acct are cleaned up + */ + trav_p = &top->children; + while (*trav_p) { + victim = (*trav_p)->xlator; + if (victim->call_cleanup && !victim->itable && !victim->mem_acct) { + (*trav_p) = (*trav_p)->next; + } else { + trav_p = &(*trav_p)->next; + } + } + UNLOCK (&ctx->volfile_lock); + } } @@ -271,8 +307,10 @@ glusterfs_handle_terminate (rpcsvc_request_t *req) glusterfs_ctx_t *ctx = NULL; xlator_t *top = NULL; xlator_t *victim = NULL; + xlator_t *tvictim = NULL; xlator_list_t **trav_p = NULL; gf_boolean_t lockflag = _gf_false; + gf_boolean_t last_brick = _gf_false; ret = xdr_to_generic (req->msg[0], &xlator_req, (xdrproc_t)xdr_gd1_mgmt_brick_op_req); @@ -290,11 +328,15 @@ glusterfs_handle_terminate (rpcsvc_request_t *req) for (trav_p = &top->children; *trav_p; trav_p = &(*trav_p)->next) { victim = (*trav_p)->xlator; - if (strcmp (victim->name, xlator_req.name) == 0) { + if (!victim->cleanup_starting && strcmp (victim->name, xlator_req.name) == 0) { break; } } } + + if (!top) + goto err; + } if (!*trav_p) { gf_log (THIS->name, GF_LOG_ERROR, @@ -310,26 +352,38 @@ glusterfs_handle_terminate (rpcsvc_request_t *req) } glusterfs_terminate_response_send (req, 0); - if ((trav_p == &top->children) && !(*trav_p)->next) { + for (trav_p = &top->children; *trav_p; trav_p = &(*trav_p)->next) { + tvictim = (*trav_p)->xlator; + if (!tvictim->cleanup_starting && !strcmp (tvictim->name, xlator_req.name)) { + continue; + } + if (!tvictim->cleanup_starting) { + last_brick = _gf_true; + break; + } + } + if (!last_brick) { gf_log (THIS->name, GF_LOG_INFO, "terminating after loss of last child %s", xlator_req.name); glusterfs_mgmt_pmap_signout (glusterfsd_ctx, xlator_req.name); kill (getpid(), SIGTERM); } else { - /* - * This is terribly unsafe without quiescing or shutting - * things down properly but it gets us to the point - * where we can test other stuff. - * - * TBD: finish implementing this "detach" code properly - */ + /* TODO cleanup sequence needs to be done properly for + Quota and Changelog + */ + if (victim->cleanup_starting) + goto err; + + victim->cleanup_starting = 1; + UNLOCK (&ctx->volfile_lock); lockflag = _gf_true; + gf_log (THIS->name, GF_LOG_INFO, "detaching not-only" " child %s", xlator_req.name); top->notify (top, GF_EVENT_CLEANUP, victim); - xlator_mem_cleanup (victim); + } err: if (!lockflag) diff --git a/libglusterfs/src/client_t.c b/libglusterfs/src/client_t.c index eda1c46..55d891f 100644 --- a/libglusterfs/src/client_t.c +++ b/libglusterfs/src/client_t.c @@ -338,7 +338,7 @@ gf_client_destroy_recursive (xlator_t *xl, client_t *client) { xlator_list_t *trav; - if (xl->cbks->client_destroy) { + if (!xl->call_cleanup && xl->cbks->client_destroy) { xl->cbks->client_destroy (xl, client); } @@ -397,7 +397,7 @@ gf_client_disconnect_recursive (xlator_t *xl, client_t *client) int ret = 0; xlator_list_t *trav; - if (xl->cbks->client_disconnect) { + if (!xl->call_cleanup && xl->cbks->client_disconnect) { ret = xl->cbks->client_disconnect (xl, client); } diff --git a/libglusterfs/src/fd.c b/libglusterfs/src/fd.c index 45b0d32..2dc52ba 100644 --- a/libglusterfs/src/fd.c +++ b/libglusterfs/src/fd.c @@ -500,7 +500,7 @@ fd_destroy (fd_t *fd, gf_boolean_t bound) xl = fd->_ctx[i].xl_key; old_THIS = THIS; THIS = xl; - if (xl->cbks->releasedir) + if (!xl->call_cleanup && xl->cbks->releasedir) xl->cbks->releasedir (xl, fd); THIS = old_THIS; } @@ -511,7 +511,7 @@ fd_destroy (fd_t *fd, gf_boolean_t bound) xl = fd->_ctx[i].xl_key; old_THIS = THIS; THIS = xl; - if (xl->cbks->release) + if (!xl->call_cleanup && xl->cbks->release) xl->cbks->release (xl, fd); THIS = old_THIS; } diff --git a/libglusterfs/src/graph.c b/libglusterfs/src/graph.c index cdd7123..c0e67e4 100644 --- a/libglusterfs/src/graph.c +++ b/libglusterfs/src/graph.c @@ -841,7 +841,7 @@ is_graph_topology_equal (glusterfs_graph_t *graph1, glusterfs_graph_t *graph2) trav2 = trav2->children->xlator; for (ltrav = trav1->children; ltrav; ltrav = ltrav->next) { trav1 = ltrav->xlator; - if (strcmp (trav1->name, trav2->name) == 0) { + if (!trav1->cleanup_starting && !strcmp (trav1->name, trav2->name)) { break; } } @@ -1088,7 +1088,7 @@ glusterfs_graph_reconfigure (glusterfs_graph_t *oldgraph, new_xl = FIRST_CHILD (new_xl); for (trav = old_xl->children; trav; trav = trav->next) { - if (strcmp (trav->xlator->name, new_xl->name) == 0) { + if (!trav->xlator->cleanup_starting && !strcmp (trav->xlator->name, new_xl->name)) { return xlator_tree_reconfigure (trav->xlator, new_xl); } } @@ -1240,7 +1240,7 @@ glusterfs_graph_attach (glusterfs_graph_t *orig_graph, char *path, xl->volfile_id[strlen(xl->volfile_id)-4] = '\0'; } - /* TBD: memory leaks everywhere */ + /* TODO memory leaks everywhere need to free graph in case of error */ if (glusterfs_graph_prepare (graph, this->ctx, xl->name)) { gf_log (this->name, GF_LOG_WARNING, "failed to prepare graph for xlator %s", xl->name); diff --git a/libglusterfs/src/inode.c b/libglusterfs/src/inode.c index ffba1bf..cf264d8 100644 --- a/libglusterfs/src/inode.c +++ b/libglusterfs/src/inode.c @@ -328,7 +328,7 @@ __inode_ctx_free (inode_t *inode) xl = (xlator_t *)(long)inode->_ctx[index].xl_key; old_THIS = THIS; THIS = xl; - if (xl->cbks->forget) + if (!xl->call_cleanup && xl->cbks->forget) xl->cbks->forget (xl, inode); THIS = old_THIS; } diff --git a/libglusterfs/src/statedump.c b/libglusterfs/src/statedump.c index 697ddc3..e9ecef5 100644 --- a/libglusterfs/src/statedump.c +++ b/libglusterfs/src/statedump.c @@ -483,21 +483,14 @@ gf_proc_dump_dict_info (glusterfs_ctx_t *ctx) (total_pairs / total_dicts)); } -void -gf_proc_dump_xlator_info (xlator_t *top) +static void +gf_proc_dump_per_xlator_info (xlator_t *top) { - xlator_t *trav = NULL; - glusterfs_ctx_t *ctx = NULL; + xlator_t *trav = top; + glusterfs_ctx_t *ctx = top->ctx; char itable_key[1024] = {0,}; - if (!top) - return; - - ctx = top->ctx; - - trav = top; - while (trav) { - + while (trav && !trav->cleanup_starting) { if (ctx->measure_latency) gf_proc_dump_latency_info (trav); @@ -521,7 +514,6 @@ gf_proc_dump_xlator_info (xlator_t *top) if (GF_PROC_DUMP_IS_XL_OPTION_ENABLED (inode) && (trav->dumpops->inode)) trav->dumpops->inode (trav); - if (trav->dumpops->fd && GF_PROC_DUMP_IS_XL_OPTION_ENABLED (fd)) trav->dumpops->fd (trav); @@ -532,6 +524,30 @@ gf_proc_dump_xlator_info (xlator_t *top) trav = trav->next; } +} + + + +void +gf_proc_dump_xlator_info (xlator_t *top, gf_boolean_t brick_mux) +{ + xlator_t *trav = NULL; + xlator_list_t **trav_p = NULL; + + if (!top) + return; + + trav = top; + gf_proc_dump_per_xlator_info (trav); + + if (brick_mux) { + trav_p = &top->children; + while (*trav_p) { + trav = (*trav_p)->xlator; + gf_proc_dump_per_xlator_info (trav); + trav_p = &(*trav_p)->next; + } + } return; } @@ -785,12 +801,27 @@ gf_proc_dump_info (int signum, glusterfs_ctx_t *ctx) char tmp_dump_name[PATH_MAX] = {0,}; char path[PATH_MAX] = {0,}; struct timeval tv = {0,}; + gf_boolean_t is_brick_mux = _gf_false; + xlator_t *top = NULL; + xlator_list_t **trav_p = NULL; + int brick_count = 0; gf_proc_dump_lock (); if (!ctx) goto out; + if (ctx) { + top = ctx->active->first; + for (trav_p = &top->children; *trav_p; + trav_p = &(*trav_p)->next) { + brick_count++; + } + + if (brick_count > 1) + is_brick_mux = _gf_true; + } + if (ctx->cmd_args.brick_name) { GF_REMOVE_SLASH_FROM_PATH (ctx->cmd_args.brick_name, brick_name); } else @@ -850,12 +881,12 @@ gf_proc_dump_info (int signum, glusterfs_ctx_t *ctx) if (ctx->master) { gf_proc_dump_add_section ("fuse"); - gf_proc_dump_xlator_info (ctx->master); + gf_proc_dump_xlator_info (ctx->master, _gf_false); } if (ctx->active) { gf_proc_dump_add_section ("active graph - %d", ctx->graph_id); - gf_proc_dump_xlator_info (ctx->active->top); + gf_proc_dump_xlator_info (ctx->active->top, is_brick_mux); } i = 0; diff --git a/libglusterfs/src/xlator.c b/libglusterfs/src/xlator.c index de97dff..4722800 100644 --- a/libglusterfs/src/xlator.c +++ b/libglusterfs/src/xlator.c @@ -424,7 +424,7 @@ get_xlator_by_name_or_type (xlator_t *this, char *target, int is_name) for (trav = this->children; trav; trav = trav->next) { value = is_name ? trav->xlator->name : trav->xlator->type; - if (strcmp(value, target) == 0) { + if (!strcmp(value, target)) { return trav->xlator; } child_xl = get_xlator_by_name_or_type (trav->xlator, target, @@ -505,7 +505,9 @@ xlator_init (xlator_t *xl) } xl->init_succeeded = 1; - + /*xl->cleanup_starting = 0; + xl->call_cleanup = 0; + */ ret = 0; out: return ret; @@ -638,7 +640,7 @@ xlator_list_destroy (xlator_list_t *list) return 0; } -static int +int xlator_memrec_free (xlator_t *xl) { uint32_t i = 0; diff --git a/libglusterfs/src/xlator.h b/libglusterfs/src/xlator.h index 26d2cc5..5ed8646 100644 --- a/libglusterfs/src/xlator.h +++ b/libglusterfs/src/xlator.h @@ -957,6 +957,14 @@ struct _xlator { /* Its used as an index to inode_ctx*/ uint32_t xl_id; + + /* cleanup flag to avoid races during xlator cleanup */ + uint32_t cleanup_starting; + + /* flag to avoid recall of xlator_mem_cleanup for xame xlator */ + uint32_t call_cleanup; + + }; typedef struct { @@ -1069,5 +1077,6 @@ copy_opts_to_child (xlator_t *src, xlator_t *dst, char *glob); int glusterfs_delete_volfile_checksum (glusterfs_ctx_t *ctx, const char *volfile_id); - +int +xlator_memrec_free (xlator_t *xl); #endif /* _XLATOR_H */ diff --git a/tests/volume.rc b/tests/volume.rc index d57aa93..f9e16c5 100644 --- a/tests/volume.rc +++ b/tests/volume.rc @@ -797,26 +797,53 @@ function count_sh_entries() ls $1/.glusterfs/indices/xattrop | grep -v "xattrop-" | wc -l } +function check_brick_multiplex() { + cnt="$(ls /var/log/glusterfs/bricks|wc -l)" + local ret=$($CLI volume info|grep "cluster.brick-multiplex"|cut -d" " -f2) + local bcnt="$(brick_count)" + + if [ $bcnt -ne 1 ]; then + if [ "$ret" = "on" ] || [ $cnt -eq 1 ]; then + echo "Y" + else + echo "N" + fi + else + echo "N" + fi +} + function get_fd_count { local vol=$1 local host=$2 local brick=$3 local fname=$4 + local val="$(check_brick_multiplex)" local gfid_str=$(gf_gfid_xattr_to_str $(gf_get_gfid_xattr $brick/$fname)) local statedump=$(generate_brick_statedump $vol $host $brick) - local count=$(grep "gfid=$gfid_str" $statedump -A2 -B1 | grep $brick -A3 | grep -w fd-count | cut -f2 -d'=' | tail -1) + if [ $val == "N" ]; then + count=$(grep "gfid=$gfid_str" $statedump -A2 | grep fd-count | cut -f2 -d'=' | tail -1) + else + count=$(grep "${brick}.active.1" -A3 $statedump | grep "gfid=$gfid_str" -A2 | grep fd-count | cut -f2 -d'=' | tail -1) + fi rm -f $statedump echo $count } + function get_active_fd_count { local vol=$1 local host=$2 local brick=$3 local fname=$4 + local val="$(check_brick_multiplex)" local gfid_str=$(gf_gfid_xattr_to_str $(gf_get_gfid_xattr $brick/$fname)) local statedump=$(generate_brick_statedump $vol $host $brick) - local count=$(grep "gfid=$gfid_str" $statedump -A2 -B1 | grep $brick -A3 | grep -w active-fd-count | cut -f2 -d'=' | tail -1) + if [ $val == "N" ]; then + count=$(grep "gfid=$gfid_str" $statedump -A2 | grep fd-count | cut -f2 -d'=' | tail -1) + else + count=$(grep "${brick}.active.1" -A3 $statedump | grep "gfid=$gfid_str" -A2 | grep fd-count | cut -f2 -d'=' | tail -1) + fi rm -f $statedump echo $count } diff --git a/xlators/features/changelog/src/changelog-rpc.c b/xlators/features/changelog/src/changelog-rpc.c index 5524e43..ccb22b5 100644 --- a/xlators/features/changelog/src/changelog-rpc.c +++ b/xlators/features/changelog/src/changelog-rpc.c @@ -255,6 +255,15 @@ changelog_handle_probe (rpcsvc_request_t *req) changelog_probe_req rpc_req = {0,}; changelog_probe_rsp rpc_rsp = {0,}; + + this = req->trans->xl; + if (this->cleanup_starting) { + gf_msg (this->name, GF_LOG_DEBUG, 0, + CHANGELOG_MSG_HANDLE_PROBE_ERROR, + "cleanup_starting flag is already set for xl"); + return 0; + } + ret = xdr_to_generic (req->msg[0], &rpc_req, (xdrproc_t)xdr_changelog_probe_req); if (ret < 0) { diff --git a/xlators/features/changelog/src/changelog.c b/xlators/features/changelog/src/changelog.c index a472208..121ab04 100644 --- a/xlators/features/changelog/src/changelog.c +++ b/xlators/features/changelog/src/changelog.c @@ -2877,6 +2877,7 @@ void fini (xlator_t *this) { changelog_priv_t *priv = NULL; + struct list_head queue = {0, }; priv = this->private; @@ -2884,6 +2885,10 @@ fini (xlator_t *this) /* terminate RPC server/threads */ changelog_cleanup_rpc (this, priv); + /* call barrier_disable to cancel timer */ + if (priv->barrier_enabled) + __chlog_barrier_disable (this, &queue); + /* cleanup barrier related objects */ changelog_barrier_pthread_destroy (priv); diff --git a/xlators/features/trash/src/trash.c b/xlators/features/trash/src/trash.c index d114858..35db245 100644 --- a/xlators/features/trash/src/trash.c +++ b/xlators/features/trash/src/trash.c @@ -2591,16 +2591,24 @@ fini (xlator_t *this) GF_VALIDATE_OR_GOTO ("trash", this, out); priv = this->private; - inode_table = priv->trash_itable; if (priv) { - if (priv->newtrash_dir) + inode_table = priv->trash_itable; + if (priv->newtrash_dir) { GF_FREE (priv->newtrash_dir); - if (priv->oldtrash_dir) + priv->newtrash_dir = NULL; + } + if (priv->oldtrash_dir) { GF_FREE (priv->oldtrash_dir); - if (priv->brick_path) + priv->oldtrash_dir = NULL; + } + if (priv->brick_path) { GF_FREE (priv->brick_path); - if (priv->eliminate) + priv->brick_path = NULL; + } + if (priv->eliminate) { wipe_eliminate_path (&priv->eliminate); + priv->eliminate = NULL; + } if (inode_table) { inode_table_destroy (inode_table); priv->trash_itable = NULL; diff --git a/xlators/protocol/server/src/server-handshake.c b/xlators/protocol/server/src/server-handshake.c index 50796d5..cbc93a3 100644 --- a/xlators/protocol/server/src/server-handshake.c +++ b/xlators/protocol/server/src/server-handshake.c @@ -475,6 +475,7 @@ server_setvolume (rpcsvc_request_t *req) glusterfs_ctx_t *ctx = NULL; struct _child_status *tmp = NULL; char *subdir_mount = NULL; + gf_boolean_t cleanup_starting = _gf_false; params = dict_new (); reply = dict_new (); @@ -574,11 +575,13 @@ server_setvolume (rpcsvc_request_t *req) "initialised yet. Try again later"); goto fail; } + list_for_each_entry (tmp, &conf->child_status->status_list, status_list) { if (strcmp (tmp->name, name) == 0) break; } + if (!tmp->name) { gf_msg (this->name, GF_LOG_ERROR, 0, PS_MSG_CHILD_STATUS_FAILED, @@ -592,6 +595,7 @@ server_setvolume (rpcsvc_request_t *req) "Failed to set 'child_up' for xlator %s " "in the reply dict", tmp->name); } + ret = dict_get_str (params, "process-uuid", &client_uid); if (ret < 0) { ret = dict_set_str (reply, "ERROR", @@ -665,8 +669,27 @@ server_setvolume (rpcsvc_request_t *req) INTERNAL_LOCKS | POSIX_LOCKS); } - if (req->trans->xl_private != client) + pthread_mutex_lock (&conf->mutex); + if (xl->cleanup_starting) { + cleanup_starting = _gf_true; + } else if (req->trans->xl_private != client) { req->trans->xl_private = client; + } + pthread_mutex_unlock (&conf->mutex); + + if (cleanup_starting) { + op_ret = -1; + op_errno = EAGAIN; + + ret = dict_set_str (reply, "ERROR", + "cleanup flag is set for xlator. " + " Try again later"); + if (ret < 0) + gf_msg_debug (this->name, 0, "failed to set error: " + "cleanup flag is set for xlator. " + "Try again later"); + goto fail; + } auth_set_username_passwd (params, config_params, client); if (req->trans->ssl_name) { diff --git a/xlators/protocol/server/src/server.c b/xlators/protocol/server/src/server.c index 737bb96..792dfb3 100644 --- a/xlators/protocol/server/src/server.c +++ b/xlators/protocol/server/src/server.c @@ -467,7 +467,16 @@ server_rpc_notify (rpcsvc_t *rpc, void *xl, rpcsvc_event_t event, server_ctx_t *serv_ctx = NULL; struct timespec grace_ts = {0, }; char *auth_path = NULL; - int ret = -1; + int ret = -1; + gf_boolean_t victim_found = _gf_false; + char *xlator_name = NULL; + glusterfs_ctx_t *ctx = NULL; + xlator_t *top = NULL; + xlator_list_t **trav_p = NULL; + xlator_t *travxl = NULL; + uint64_t xprtrefcount = 0; + struct _child_status *tmp = NULL; + if (!xl || !data) { gf_msg_callingfn ("server", GF_LOG_WARNING, 0, @@ -479,6 +488,7 @@ server_rpc_notify (rpcsvc_t *rpc, void *xl, rpcsvc_event_t event, this = xl; trans = data; conf = this->private; + ctx = glusterfsd_ctx; switch (event) { case RPCSVC_EVENT_ACCEPT: @@ -626,11 +636,49 @@ unref_transport: client = trans->xl_private; if (!client) break; + pthread_mutex_lock (&conf->mutex); + list_for_each_entry (tmp, &conf->child_status->status_list, + status_list) { + if (tmp->name && client->bound_xl && + client->bound_xl->cleanup_starting && + !strcmp (tmp->name, client->bound_xl->name)) { + xprtrefcount = GF_ATOMIC_GET (tmp->xprtrefcnt); + if (xprtrefcount > 0) { + xprtrefcount = GF_ATOMIC_DEC (tmp->xprtrefcnt); + if (xprtrefcount == 0) + xlator_name = gf_strdup(client->bound_xl->name); + } + break; + } + } + pthread_mutex_unlock (&conf->mutex); /* unref only for if (!client->lk_heal) */ if (!conf->lk_heal) gf_client_unref (client); + if (xlator_name) { + if (this->ctx->active) { + top = this->ctx->active->first; + LOCK (&ctx->volfile_lock); + for (trav_p = &top->children; *trav_p; + trav_p = &(*trav_p)->next) { + travxl = (*trav_p)->xlator; + if (!travxl->call_cleanup && + strcmp (travxl->name, xlator_name) == 0) { + victim_found = _gf_true; + break; + } + } + UNLOCK (&ctx->volfile_lock); + if (victim_found) { + xlator_mem_cleanup (travxl); + glusterfs_autoscale_threads (ctx, -1, this); + } + } + GF_FREE (xlator_name); + } + trans->xl_private = NULL; break; default: @@ -1078,6 +1126,7 @@ init (xlator_t *this) conf->child_status = GF_CALLOC (1, sizeof (struct _child_status), gf_server_mt_child_status); INIT_LIST_HEAD (&conf->child_status->status_list); + GF_ATOMIC_INIT (conf->child_status->xprtrefcnt, 0); /*ret = dict_get_str (this->options, "statedump-path", &statedump_path); if (!ret) { @@ -1392,14 +1441,53 @@ server_process_child_event (xlator_t *this, int32_t event, void *data, int ret = -1; server_conf_t *conf = NULL; rpc_transport_t *xprt = NULL; + xlator_t *victim = NULL; + struct _child_status *tmp = NULL; GF_VALIDATE_OR_GOTO(this->name, data, out); conf = this->private; GF_VALIDATE_OR_GOTO(this->name, conf, out); + victim = data; pthread_mutex_lock (&conf->mutex); { + if (cbk_procnum == GF_CBK_CHILD_UP) { + list_for_each_entry (tmp, &conf->child_status->status_list, + status_list) { + if (tmp->name == NULL) + break; + if (strcmp (tmp->name, victim->name) == 0) + break; + } + if (tmp->name) { + tmp->child_up = _gf_true; + } else { + tmp = GF_CALLOC (1, sizeof (struct _child_status), + gf_server_mt_child_status); + INIT_LIST_HEAD (&tmp->status_list); + tmp->name = gf_strdup (victim->name); + tmp->child_up = _gf_true; + list_add_tail (&tmp->status_list, + &conf->child_status->status_list); + } + } + + if (cbk_procnum == GF_CBK_CHILD_DOWN) { + list_for_each_entry (tmp, &conf->child_status->status_list, + status_list) { + if (strcmp (tmp->name, victim->name) == 0) { + tmp->child_up = _gf_false; + break; + } + } + + if (!tmp->name) + gf_msg (this->name, GF_LOG_ERROR, 0, + PS_MSG_CHILD_STATUS_FAILED, + "No xlator %s is found in " + "child status list", victim->name); + } list_for_each_entry (xprt, &conf->xprt_list, list) { if (!xprt->xl_private) { continue; @@ -1433,6 +1521,8 @@ notify (xlator_t *this, int32_t event, void *data, ...) struct _child_status *tmp = NULL; gf_boolean_t victim_found = _gf_false; glusterfs_ctx_t *ctx = NULL; + gf_boolean_t xprt_found = _gf_false; + uint64_t totxprt = 0; GF_VALIDATE_OR_GOTO (THIS->name, this, out); conf = this->private; @@ -1467,24 +1557,6 @@ notify (xlator_t *this, int32_t event, void *data, ...) case GF_EVENT_CHILD_UP: { - list_for_each_entry (tmp, &conf->child_status->status_list, - status_list) { - if (tmp->name == NULL) - break; - if (strcmp (tmp->name, victim->name) == 0) - break; - } - if (tmp->name) { - tmp->child_up = _gf_true; - } else { - tmp = GF_CALLOC (1, sizeof (struct _child_status), - gf_server_mt_child_status); - INIT_LIST_HEAD (&tmp->status_list); - tmp->name = gf_strdup (victim->name); - tmp->child_up = _gf_true; - list_add_tail (&tmp->status_list, - &conf->child_status->status_list); - } ret = server_process_child_event (this, event, data, GF_CBK_CHILD_UP); if (ret) { @@ -1499,19 +1571,6 @@ notify (xlator_t *this, int32_t event, void *data, ...) case GF_EVENT_CHILD_DOWN: { - list_for_each_entry (tmp, &conf->child_status->status_list, - status_list) { - if (strcmp (tmp->name, victim->name) == 0) { - tmp->child_up = _gf_false; - break; - } - } - if (!tmp->name) - gf_msg (this->name, GF_LOG_ERROR, 0, - PS_MSG_CHILD_STATUS_FAILED, - "No xlator %s is found in " - "child status list", victim->name); - ret = server_process_child_event (this, event, data, GF_CBK_CHILD_DOWN); if (ret) { @@ -1528,6 +1587,28 @@ notify (xlator_t *this, int32_t event, void *data, ...) case GF_EVENT_CLEANUP: conf = this->private; pthread_mutex_lock (&conf->mutex); + /* Calculate total no. of xprt available in list for this + brick xlator + */ + list_for_each_entry_safe (xprt, xp_next, + &conf->xprt_list, list) { + if (!xprt->xl_private) { + continue; + } + if (xprt->xl_private->bound_xl == data) { + totxprt++; + } + } + + list_for_each_entry (tmp, &conf->child_status->status_list, + status_list) { + if (strcmp (tmp->name, victim->name) == 0) { + tmp->child_up = _gf_false; + GF_ATOMIC_INIT (tmp->xprtrefcnt, totxprt); + break; + } + } + /* * Disconnecting will (usually) drop the last ref, which will * cause the transport to be unlinked and freed while we're @@ -1543,18 +1624,11 @@ notify (xlator_t *this, int32_t event, void *data, ...) gf_log (this->name, GF_LOG_INFO, "disconnecting %s", xprt->peerinfo.identifier); + xprt_found = _gf_true; rpc_transport_disconnect (xprt, _gf_false); } } - list_for_each_entry (tmp, &conf->child_status->status_list, - status_list) { - if (strcmp (tmp->name, victim->name) == 0) - break; - } - if (tmp->name && (strcmp (tmp->name, victim->name) == 0)) { - GF_FREE (tmp->name); - list_del (&tmp->status_list); - } + pthread_mutex_unlock (&conf->mutex); if (this->ctx->active) { top = this->ctx->active->first; @@ -1562,8 +1636,8 @@ notify (xlator_t *this, int32_t event, void *data, ...) for (trav_p = &top->children; *trav_p; trav_p = &(*trav_p)->next) { travxl = (*trav_p)->xlator; - if (travxl && - strcmp (travxl->name, victim->name) == 0) { + if (!travxl->call_cleanup && + strcmp (travxl->name, victim->name) == 0) { victim_found = _gf_true; break; } @@ -1572,12 +1646,14 @@ notify (xlator_t *this, int32_t event, void *data, ...) glusterfs_delete_volfile_checksum (ctx, victim->volfile_id); UNLOCK (&ctx->volfile_lock); - if (victim_found) - (*trav_p) = (*trav_p)->next; + glusterfs_mgmt_pmap_signout (ctx, victim->name); - /* we need the protocol/server xlator here as 'this' */ - glusterfs_autoscale_threads (ctx, -1, this); + + if (!xprt_found && victim_found) { + xlator_mem_cleanup (victim); + glusterfs_autoscale_threads (ctx, -1, this); + } } break; diff --git a/xlators/protocol/server/src/server.h b/xlators/protocol/server/src/server.h index ea15b56..af987aa 100644 --- a/xlators/protocol/server/src/server.h +++ b/xlators/protocol/server/src/server.h @@ -77,7 +77,7 @@ struct _child_status { struct list_head status_list; char *name; gf_boolean_t child_up; - + gf_atomic_t xprtrefcnt; }; struct server_conf { rpcsvc_t *rpc; diff --git a/xlators/storage/posix/src/posix-helpers.c b/xlators/storage/posix/src/posix-helpers.c index 334175d..49f9f46 100644 --- a/xlators/storage/posix/src/posix-helpers.c +++ b/xlators/storage/posix/src/posix-helpers.c @@ -1921,6 +1921,12 @@ out: return NULL; abort: + LOCK (&priv->lock); + { + priv->health_check_active = _gf_false; + } + UNLOCK (&priv->lock); + /* health-check failed */ gf_msg (this->name, GF_LOG_EMERG, 0, P_MSG_HEALTHCHECK_FAILED, "health-check failed, going down"); @@ -1961,18 +1967,18 @@ abort: for (trav_p = &top->children; *trav_p; trav_p = &(*trav_p)->next) { victim = (*trav_p)->xlator; - if (victim && - strcmp (victim->name, priv->base_path) == 0) { + if (!victim->call_cleanup && + strcmp (victim->name, priv->base_path) == 0) { victim_found = _gf_true; break; } } UNLOCK (&ctx->volfile_lock); - if (victim_found) { + if (victim_found && !victim->cleanup_starting) { gf_log (THIS->name, GF_LOG_INFO, "detaching not-only " " child %s", priv->base_path); + victim->cleanup_starting = 1; top->notify (top, GF_EVENT_CLEANUP, victim); - xlator_mem_cleanup (victim); } } diff --git a/xlators/storage/posix/src/posix.c b/xlators/storage/posix/src/posix.c index 74ee98f..c680e3f 100644 --- a/xlators/storage/posix/src/posix.c +++ b/xlators/storage/posix/src/posix.c @@ -7899,12 +7899,13 @@ fini (xlator_t *this) struct posix_private *priv = this->private; if (!priv) return; - this->private = NULL; - if (priv->health_check) { + LOCK (&priv->lock); + if (priv->health_check_active) { priv->health_check_active = _gf_false; pthread_cancel (priv->health_check); priv->health_check = 0; } + UNLOCK (&priv->lock); if (priv->disk_space_check) { priv->disk_space_check_active = _gf_false; pthread_cancel (priv->disk_space_check); @@ -7929,6 +7930,7 @@ fini (xlator_t *this) GF_FREE (priv->hostname); GF_FREE (priv->trash_path); GF_FREE (priv); + this->private = NULL; return; } -- 1.8.3.1