diff --git a/SOURCES/011-remote.patch b/SOURCES/011-remote.patch
new file mode 100644
index 0000000..c09a8e2
--- /dev/null
+++ b/SOURCES/011-remote.patch
@@ -0,0 +1,786 @@
+From a81ca9625e8d1ccd7f79fbe464b9f4221e8671f2 Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Thu, 9 May 2019 20:26:08 -0500
+Subject: [PATCH 1/6] Refactor: libpe_status: functionize unfencing digest code
+ more
+
+... for readability, reusability, and avoiding unnecessary function calls or
+memory allocation.
+---
+ lib/pengine/utils.c | 159 ++++++++++++++++++++++++++++++++++++++--------------
+ 1 file changed, 118 insertions(+), 41 deletions(-)
+
+diff --git a/lib/pengine/utils.c b/lib/pengine/utils.c
+index d09b0d8..b6a31d1 100644
+--- a/lib/pengine/utils.c
++++ b/lib/pengine/utils.c
+@@ -2091,57 +2091,134 @@ rsc_action_digest_cmp(resource_t * rsc, xmlNode * xml_op, node_t * node,
+     return data;
+ }
+ 
++/*!
++ * \internal
++ * \brief Create an unfencing summary for use in special node attribute
++ *
++ * Create a string combining a fence device's resource ID, agent type, and
++ * parameter digest (whether for all parameters or just non-private parameters).
++ * This can be stored in a special node attribute, allowing us to detect changes
++ * in either the agent type or parameters, to know whether unfencing must be
++ * redone or can be safely skipped when the device's history is cleaned.
++ *
++ * \param[in] rsc_id        Fence device resource ID
++ * \param[in] agent_type    Fence device agent
++ * \param[in] param_digest  Fence device parameter digest
++ *
++ * \return Newly allocated string with unfencing digest
++ * \note The caller is responsible for freeing the result.
++ */
++static inline char *
++create_unfencing_summary(const char *rsc_id, const char *agent_type,
++                         const char *param_digest)
++{
++    return crm_strdup_printf("%s:%s:%s", rsc_id, agent_type, param_digest);
++}
++
++/*!
++ * \internal
++ * \brief Check whether a node can skip unfencing
++ *
++ * Check whether a fence device's current definition matches a node's
++ * stored summary of when it was last unfenced by the device.
++ *
++ * \param[in] rsc_id        Fence device's resource ID
++ * \param[in] agent         Fence device's agent type
++ * \param[in] digest_calc   Fence device's current parameter digest
++ * \param[in] node_summary  Value of node's special unfencing node attribute
++ *                          (a comma-separated list of unfencing summaries for
++ *                          all devices that have unfenced this node)
++ *
++ * \return TRUE if digest matches, FALSE otherwise
++ */
++static bool
++unfencing_digest_matches(const char *rsc_id, const char *agent,
++                         const char *digest_calc, const char *node_summary)
++{
++    bool matches = FALSE;
++
++    if (rsc_id && agent && digest_calc && node_summary) {
++        char *search_secure = create_unfencing_summary(rsc_id, agent,
++                                                       digest_calc);
++
++        /* The digest was calculated including the device ID and agent,
++         * so there is no risk of collision using strstr().
++         */
++        matches = (strstr(node_summary, search_secure) != NULL);
++        crm_trace("Calculated unfencing digest '%s' %sfound in '%s'",
++                  search_secure, matches? "" : "not ", node_summary);
++        free(search_secure);
++    }
++    return matches;
++}
++
++/* Magic string to use as action name for digest cache entries used for
++ * unfencing checks. This is not a real action name (i.e. "on"), so
++ * check_action_definition() won't confuse these entries with real actions.
++ */
+ #define STONITH_DIGEST_TASK "stonith-on"
+ 
++/*!
++ * \internal
++ * \brief Calculate fence device digests and digest comparison result
++ *
++ * \param[in] rsc       Fence device resource
++ * \param[in] agent     Fence device's agent type
++ * \param[in] node      Node with digest cache to use
++ * \param[in] data_set  Cluster working set
++ *
++ * \return Node's digest cache entry
++ */
+ static op_digest_cache_t *
+-fencing_action_digest_cmp(resource_t * rsc, node_t * node, pe_working_set_t * data_set)
++fencing_action_digest_cmp(pe_resource_t *rsc, const char *agent,
++                          pe_node_t *node, pe_working_set_t *data_set)
+ {
+-    char *key = generate_op_key(rsc->id, STONITH_DIGEST_TASK, 0);
+-    op_digest_cache_t *data = rsc_action_digest(rsc, STONITH_DIGEST_TASK, key, node, NULL, data_set);
++    const char *node_summary = NULL;
+ 
+-    const char *digest_all = pe_node_attribute_raw(node, CRM_ATTR_DIGESTS_ALL);
+-    const char *digest_secure = pe_node_attribute_raw(node, CRM_ATTR_DIGESTS_SECURE);
++    // Calculate device's current parameter digests
++    char *key = generate_op_key(rsc->id, STONITH_DIGEST_TASK, 0);
++    op_digest_cache_t *data = rsc_action_digest(rsc, STONITH_DIGEST_TASK, key,
++                                                node, NULL, data_set);
+ 
+-    /* No 'reloads' for fencing device changes
+-     *
+-     * We use the resource id + agent + digest so that we can detect
+-     * changes to the agent and/or the parameters used
+-     */
+-    char *search_all = crm_strdup_printf("%s:%s:%s", rsc->id, (const char*)g_hash_table_lookup(rsc->meta, XML_ATTR_TYPE), data->digest_all_calc);
+-    char *search_secure = crm_strdup_printf("%s:%s:%s", rsc->id, (const char*)g_hash_table_lookup(rsc->meta, XML_ATTR_TYPE), data->digest_secure_calc);
++    free(key);
+ 
+-    data->rc = RSC_DIGEST_ALL;
+-    if (digest_all == NULL) {
+-        /* it is unknown what the previous op digest was */
++    // Check whether node has special unfencing summary node attribute
++    node_summary = pe_node_attribute_raw(node, CRM_ATTR_DIGESTS_ALL);
++    if (node_summary == NULL) {
+         data->rc = RSC_DIGEST_UNKNOWN;
++        return data;
++    }
+ 
+-    } else if (strstr(digest_all, search_all)) {
++    // Check whether full parameter digest matches
++    if (unfencing_digest_matches(rsc->id, agent, data->digest_all_calc,
++                                 node_summary)) {
+         data->rc = RSC_DIGEST_MATCH;
++        return data;
++    }
+ 
+-    } else if(digest_secure && data->digest_secure_calc) {
+-        if(strstr(digest_secure, search_secure)) {
+-            if (is_set(data_set->flags, pe_flag_stdout)) {
+-                printf("Only 'private' parameters to %s for unfencing %s changed\n",
+-                       rsc->id, node->details->uname);
+-            }
+-            data->rc = RSC_DIGEST_MATCH;
++    // Check whether secure parameter digest matches
++    node_summary = pe_node_attribute_raw(node, CRM_ATTR_DIGESTS_SECURE);
++    if (unfencing_digest_matches(rsc->id, agent, data->digest_secure_calc,
++                                 node_summary)) {
++        data->rc = RSC_DIGEST_MATCH;
++        if (is_set(data_set->flags, pe_flag_stdout)) {
++            printf("Only 'private' parameters to %s for unfencing %s changed\n",
++                   rsc->id, node->details->uname);
+         }
++        return data;
+     }
+ 
+-    if (is_set(data_set->flags, pe_flag_sanitized)
+-        && is_set(data_set->flags, pe_flag_stdout)
+-        && (data->rc == RSC_DIGEST_ALL)
++    // Parameters don't match
++    data->rc = RSC_DIGEST_ALL;
++    if (is_set(data_set->flags, (pe_flag_sanitized|pe_flag_stdout))
+         && data->digest_secure_calc) {
+-        printf("Parameters to %s for unfencing %s changed, try '%s:%s:%s'\n",
+-               rsc->id, node->details->uname, rsc->id,
+-               (const char *) g_hash_table_lookup(rsc->meta, XML_ATTR_TYPE),
+-               data->digest_secure_calc);
+-    }
+-
+-    free(key);
+-    free(search_all);
+-    free(search_secure);
++        char *digest = create_unfencing_summary(rsc->id, agent,
++                                                data->digest_secure_calc);
+ 
++        printf("Parameters to %s for unfencing %s changed, try '%s'\n",
++               rsc->id, node->details->uname, digest);
++        free(digest);
++    }
+     return data;
+ }
+ 
+@@ -2228,9 +2305,6 @@ pe_fence_op(node_t * node, const char *op, bool optional, const char *reason, pe
+              *
+              * We may do this for all nodes in the future, but for now
+              * the check_action_definition() based stuff works fine.
+-             *
+-             * Use "stonith-on" to avoid creating cache entries for
+-             * operations check_action_definition() would look for.
+              */
+             long max = 1024;
+             long digests_all_offset = 0;
+@@ -2242,8 +2316,11 @@ pe_fence_op(node_t * node, const char *op, bool optional, const char *reason, pe
+ 
+             for (GListPtr gIter = matches; gIter != NULL; gIter = gIter->next) {
+                 resource_t *match = gIter->data;
+-                op_digest_cache_t *data = fencing_action_digest_cmp(match, node, data_set);
++                const char *agent = g_hash_table_lookup(match->meta,
++                                                        XML_ATTR_TYPE);
++                op_digest_cache_t *data = NULL;
+ 
++                data = fencing_action_digest_cmp(match, agent, node, data_set);
+                 if(data->rc == RSC_DIGEST_ALL) {
+                     optional = FALSE;
+                     crm_notice("Unfencing %s (remote): because the definition of %s changed", node->details->uname, match->id);
+@@ -2254,11 +2331,11 @@ pe_fence_op(node_t * node, const char *op, bool optional, const char *reason, pe
+ 
+                 digests_all_offset += snprintf(
+                     digests_all+digests_all_offset, max-digests_all_offset,
+-                    "%s:%s:%s,", match->id, (const char*)g_hash_table_lookup(match->meta, XML_ATTR_TYPE), data->digest_all_calc);
++                    "%s:%s:%s,", match->id, agent, data->digest_all_calc);
+ 
+                 digests_secure_offset += snprintf(
+                     digests_secure+digests_secure_offset, max-digests_secure_offset,
+-                    "%s:%s:%s,", match->id, (const char*)g_hash_table_lookup(match->meta, XML_ATTR_TYPE), data->digest_secure_calc);
++                    "%s:%s:%s,", match->id, agent, data->digest_secure_calc);
+             }
+             g_hash_table_insert(stonith_op->meta,
+                                 strdup(XML_OP_ATTR_DIGESTS_ALL),
+-- 
+1.8.3.1
+
+
+From be34a73f9cfb6abdb3e2799593cb0358c01c2521 Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Fri, 10 May 2019 11:57:31 -0500
+Subject: [PATCH 2/6] Fix: libpe_status: calculate secure digests for unfencing
+ ops
+
+The calculation of digests for detection of when unfencing is needed reused
+rsc_action_digest(). However that would only add secure digests when the
+pe_flag_sanitized flag was set, which is only set by crm_simulate, so secure
+digests would never be added in normal cluster operation. This led to
+node attributes like name="#digests-secure"
+value="stonith-fence_compute-fence-nova:fence_compute:(null),".
+
+Now, rsc_action_digest() takes a new argument to select whether secure digests
+are added, which is always set to TRUE when calculating unfencing digests.
+---
+ lib/pengine/utils.c | 27 ++++++++++++++++++++++-----
+ 1 file changed, 22 insertions(+), 5 deletions(-)
+
+diff --git a/lib/pengine/utils.c b/lib/pengine/utils.c
+index b6a31d1..f52f1c7 100644
+--- a/lib/pengine/utils.c
++++ b/lib/pengine/utils.c
+@@ -1948,9 +1948,24 @@ append_versioned_params(xmlNode *versioned_params, const char *ra_version, xmlNo
+ }
+ #endif
+ 
++/*!
++ * \internal
++ * \brief Calculate action digests and store in node's digest cache
++ *
++ * \param[in] rsc          Resource that action was for
++ * \param[in] task         Name of action performed
++ * \param[in] key          Action's task key
++ * \param[in] node         Node action was performed on
++ * \param[in] xml_op       XML of operation in CIB status (if available)
++ * \param[in] calc_secure  Whether to calculate secure digest
++ * \param[in] data_set     Cluster working set
++ *
++ * \return Pointer to node's digest cache entry
++ */
+ static op_digest_cache_t *
+-rsc_action_digest(resource_t * rsc, const char *task, const char *key,
+-                  node_t * node, xmlNode * xml_op, pe_working_set_t * data_set) 
++rsc_action_digest(pe_resource_t *rsc, const char *task, const char *key,
++                  pe_node_t *node, xmlNode *xml_op, bool calc_secure,
++                  pe_working_set_t *data_set)
+ {
+     op_digest_cache_t *data = NULL;
+ 
+@@ -2018,7 +2033,7 @@ rsc_action_digest(resource_t * rsc, const char *task, const char *key,
+ 
+         data->digest_all_calc = calculate_operation_digest(data->params_all, op_version);
+ 
+-        if (is_set(data_set->flags, pe_flag_sanitized)) {
++        if (calc_secure) {
+             data->params_secure = copy_xml(data->params_all);
+             if(secure_list) {
+                 filter_parameters(data->params_secure, secure_list, FALSE);
+@@ -2064,7 +2079,9 @@ rsc_action_digest_cmp(resource_t * rsc, xmlNode * xml_op, node_t * node,
+ 
+     interval = crm_parse_int(interval_s, "0");
+     key = generate_op_key(rsc->id, task, interval);
+-    data = rsc_action_digest(rsc, task, key, node, xml_op, data_set);
++    data = rsc_action_digest(rsc, task, key, node, xml_op,
++                             is_set(data_set->flags, pe_flag_sanitized),
++                             data_set);
+ 
+     data->rc = RSC_DIGEST_MATCH;
+     if (digest_restart && data->digest_restart_calc && strcmp(data->digest_restart_calc, digest_restart) != 0) {
+@@ -2178,7 +2195,7 @@ fencing_action_digest_cmp(pe_resource_t *rsc, const char *agent,
+     // Calculate device's current parameter digests
+     char *key = generate_op_key(rsc->id, STONITH_DIGEST_TASK, 0);
+     op_digest_cache_t *data = rsc_action_digest(rsc, STONITH_DIGEST_TASK, key,
+-                                                node, NULL, data_set);
++                                                node, NULL, TRUE, data_set);
+ 
+     free(key);
+ 
+-- 
+1.8.3.1
+
+
+From 8819c2f96f74ab4b4979df5ed04c16dd6bdad5f1 Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Sat, 8 Jun 2019 16:25:04 -0500
+Subject: [PATCH 3/6] Refactor: libpe_status: add function for checking
+ shutdown attribute
+
+... to reduce code duplication and allow further reuse
+---
+ include/crm/pengine/internal.h |  2 ++
+ lib/pengine/unpack.c           |  8 ++------
+ lib/pengine/utils.c            | 20 ++++++++++++++++++++
+ 3 files changed, 24 insertions(+), 6 deletions(-)
+
+diff --git a/include/crm/pengine/internal.h b/include/crm/pengine/internal.h
+index c40b075..c3f9f70 100644
+--- a/include/crm/pengine/internal.h
++++ b/include/crm/pengine/internal.h
+@@ -362,4 +362,6 @@ void pe__foreach_param_check(pe_working_set_t *data_set,
+                                         enum pe_check_parameters,
+                                         pe_working_set_t*));
+ void pe__free_param_checks(pe_working_set_t *data_set);
++
++bool pe__shutdown_requested(pe_node_t *node);
+ #endif
+diff --git a/lib/pengine/unpack.c b/lib/pengine/unpack.c
+index 619ccbf..cf725a1 100644
+--- a/lib/pengine/unpack.c
++++ b/lib/pengine/unpack.c
+@@ -1013,7 +1013,6 @@ unpack_handle_remote_attrs(node_t *this_node, xmlNode *state, pe_working_set_t *
+     const char *resource_discovery_enabled = NULL;
+     xmlNode *attrs = NULL;
+     resource_t *rsc = NULL;
+-    const char *shutdown = NULL;
+ 
+     if (crm_str_eq((const char *)state->name, XML_CIB_TAG_STATE, TRUE) == FALSE) {
+         return;
+@@ -1035,8 +1034,7 @@ unpack_handle_remote_attrs(node_t *this_node, xmlNode *state, pe_working_set_t *
+     attrs = find_xml_node(state, XML_TAG_TRANSIENT_NODEATTRS, FALSE);
+     add_node_attrs(attrs, this_node, TRUE, data_set);
+ 
+-    shutdown = pe_node_attribute_raw(this_node, XML_CIB_ATTR_SHUTDOWN);
+-    if (shutdown != NULL && safe_str_neq("0", shutdown)) {
++    if (pe__shutdown_requested(this_node)) {
+         crm_info("Node %s is shutting down", this_node->details->uname);
+         this_node->details->shutdown = TRUE;
+         if (rsc) {
+@@ -1512,7 +1510,6 @@ gboolean
+ determine_online_status(xmlNode * node_state, node_t * this_node, pe_working_set_t * data_set)
+ {
+     gboolean online = FALSE;
+-    const char *shutdown = NULL;
+     const char *exp_state = crm_element_value(node_state, XML_NODE_EXPECTED);
+ 
+     if (this_node == NULL) {
+@@ -1522,9 +1519,8 @@ determine_online_status(xmlNode * node_state, node_t * this_node, pe_working_set
+ 
+     this_node->details->shutdown = FALSE;
+     this_node->details->expected_up = FALSE;
+-    shutdown = pe_node_attribute_raw(this_node, XML_CIB_ATTR_SHUTDOWN);
+ 
+-    if (shutdown != NULL && safe_str_neq("0", shutdown)) {
++    if (pe__shutdown_requested(this_node)) {
+         this_node->details->shutdown = TRUE;
+ 
+     } else if (safe_str_eq(exp_state, CRMD_JOINSTATE_MEMBER)) {
+diff --git a/lib/pengine/utils.c b/lib/pengine/utils.c
+index f52f1c7..8eac2ce 100644
+--- a/lib/pengine/utils.c
++++ b/lib/pengine/utils.c
+@@ -2522,3 +2522,23 @@ void pe_action_set_reason(pe_action_t *action, const char *reason, bool overwrit
+         }
+     }
+ }
++
++/*!
++ * \internal
++ * \brief Check whether shutdown has been requested for a node
++ *
++ * \param[in] node  Node to check
++ *
++ * \return TRUE if node has shutdown attribute set and nonzero, FALSE otherwise
++ * \note This differs from simply using node->details->shutdown in that it can
++ *       be used before that has been determined (and in fact to determine it),
++ *       and it can also be used to distinguish requested shutdown from implicit
++ *       shutdown of remote nodes by virtue of their connection stopping.
++ */
++bool
++pe__shutdown_requested(pe_node_t *node)
++{
++    const char *shutdown = pe_node_attribute_raw(node, XML_CIB_ATTR_SHUTDOWN);
++
++    return shutdown && strcmp(shutdown, "0");
++}
+-- 
+1.8.3.1
+
+
+From 938e99f29ed5faaeb4015247e363ddc7e77208a3 Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Wed, 5 Jun 2019 16:37:26 -0500
+Subject: [PATCH 4/6] Fix: scheduler: remote state is failed if node is
+ shutting down with connection failure
+
+When determining remote state, if the connection resource is failed and not
+being started again, we consider the state to be unknown if the connection has
+a reconnect interval, because we won't know whether the connection can be
+recovered until the interval expires and we re-attempt connection.
+
+However, if the node is shutting down at the time, we won't re-attempt
+connection, so consider the state failed in that case. (Note that we check the
+actual shutdown node attribute, rather than node->details->shutdown, since that
+is set for remote nodes whenever the connection is stopping.)
+
+This avoids a situation where actions that cannot succeed can be scheduled on a
+remote node that's shutting down.
+---
+ pengine/allocate.c | 3 ++-
+ 1 file changed, 2 insertions(+), 1 deletion(-)
+
+diff --git a/pengine/allocate.c b/pengine/allocate.c
+index 578db2f..c9877a4 100644
+--- a/pengine/allocate.c
++++ b/pengine/allocate.c
+@@ -1998,7 +1998,8 @@ get_remote_node_state(pe_node_t *node)
+ 
+         if ((remote_rsc->next_role == RSC_ROLE_STOPPED)
+             && remote_rsc->remote_reconnect_interval
+-            && node->details->remote_was_fenced) {
++            && node->details->remote_was_fenced
++            && !pe__shutdown_requested(node)) {
+ 
+             /* We won't know whether the connection is recoverable until the
+              * reconnect interval expires and we reattempt connection.
+-- 
+1.8.3.1
+
+
+From c20f8920634f47bbdf699d80dafd50c6a72eac8b Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Wed, 5 Jun 2019 16:43:19 -0500
+Subject: [PATCH 5/6] Fix: libpe_status: don't order implied stops relative to
+ a remote connection
+
+Actions behind a remote connection are ordered relative to any start or stop of
+the remote connection. However, if the action is a stop implied due to fencing,
+it does not require the remote connection, and the ordering should not be done.
+
+This avoids a delay in the remote connection recovery if it is failed, e.g.
+previously the ordering would look like:
+
+   fence remote node -> implied stop of resource on remote -> stop connection
+
+Now, the connection stop can proceed simultaneously with the remote node
+fencing.
+---
+ pengine/allocate.c | 11 +++++------
+ 1 file changed, 5 insertions(+), 6 deletions(-)
+
+diff --git a/pengine/allocate.c b/pengine/allocate.c
+index c9877a4..c7c68f8 100644
+--- a/pengine/allocate.c
++++ b/pengine/allocate.c
+@@ -2091,14 +2091,13 @@ apply_remote_ordering(action_t *action, pe_working_set_t *data_set)
+                                        pe_order_implies_first, data_set);
+ 
+             } else if(state == remote_state_failed) {
+-                /* We would only be here if the resource is
+-                 * running on the remote node.  Since we have no
+-                 * way to stop it, it is necessary to fence the
+-                 * node.
++                /* The resource is active on the node, but since we don't have a
++                 * valid connection, the only way to stop the resource is by
++                 * fencing the node. There is no need to order the stop relative
++                 * to the remote connection, since the stop will become implied
++                 * by the fencing.
+                  */
+                 pe_fence_node(data_set, action->node, "resources are active and the connection is unrecoverable");
+-                order_action_then_stop(action, remote_rsc,
+-                                       pe_order_implies_first, data_set);
+ 
+             } else if(remote_rsc->next_role == RSC_ROLE_STOPPED) {
+                 /* State must be remote_state_unknown or remote_state_stopped.
+-- 
+1.8.3.1
+
+
+From 26a28ee80b7fc110125eedac377dfa4c0a8e8294 Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Fri, 14 Jun 2019 14:08:47 -0500
+Subject: [PATCH 6/6] Test: pengine: update regression tests for remote
+ connection ordering change
+
+---
+ pengine/test10/remote-connection-unrecoverable.dot     | 2 --
+ pengine/test10/remote-connection-unrecoverable.exp     | 6 ------
+ pengine/test10/remote-connection-unrecoverable.summary | 2 +-
+ pengine/test10/remote-fence-before-reconnect.dot       | 1 -
+ pengine/test10/remote-fence-before-reconnect.exp       | 6 +-----
+ pengine/test10/remote-fence-before-reconnect.summary   | 2 +-
+ pengine/test10/remote-recover-all.dot                  | 2 --
+ pengine/test10/remote-recover-all.exp                  | 6 ------
+ pengine/test10/remote-recover-all.summary              | 4 ++--
+ pengine/test10/remote-recover-no-resources.dot         | 1 -
+ pengine/test10/remote-recover-no-resources.exp         | 3 ---
+ pengine/test10/remote-recover-no-resources.summary     | 2 +-
+ pengine/test10/remote-recover-unknown.dot              | 1 -
+ pengine/test10/remote-recover-unknown.exp              | 3 ---
+ pengine/test10/remote-recover-unknown.summary          | 2 +-
+ 15 files changed, 7 insertions(+), 36 deletions(-)
+
+diff --git a/pengine/test10/remote-connection-unrecoverable.dot b/pengine/test10/remote-connection-unrecoverable.dot
+index 0360cd0..b5caca6 100644
+--- a/pengine/test10/remote-connection-unrecoverable.dot
++++ b/pengine/test10/remote-connection-unrecoverable.dot
+@@ -7,14 +7,12 @@ digraph "g" {
+ "remote1_stop_0 node1" [ style=bold color="green" fontcolor="orange"]
+ "rsc1_delete_0 remote1" -> "rsc1_start_0 node2" [ style = dashed]
+ "rsc1_delete_0 remote1" [ style=dashed color="red" fontcolor="black"]
+-"rsc1_monitor_0 node2" -> "remote1_stop_0 node1" [ style = bold]
+ "rsc1_monitor_0 node2" -> "rsc1_start_0 node2" [ style = bold]
+ "rsc1_monitor_0 node2" -> "rsc2-master_demote_0" [ style = bold]
+ "rsc1_monitor_0 node2" [ style=bold color="green" fontcolor="black"]
+ "rsc1_monitor_10000 node2" [ style=bold color="green" fontcolor="black"]
+ "rsc1_start_0 node2" -> "rsc1_monitor_10000 node2" [ style = bold]
+ "rsc1_start_0 node2" [ style=bold color="green" fontcolor="black"]
+-"rsc1_stop_0 remote1" -> "remote1_stop_0 node1" [ style = bold]
+ "rsc1_stop_0 remote1" -> "rsc1_delete_0 remote1" [ style = dashed]
+ "rsc1_stop_0 remote1" -> "rsc1_start_0 node2" [ style = bold]
+ "rsc1_stop_0 remote1" -> "rsc2-master_demote_0" [ style = bold]
+diff --git a/pengine/test10/remote-connection-unrecoverable.exp b/pengine/test10/remote-connection-unrecoverable.exp
+index 73fa7a1..339ad56 100644
+--- a/pengine/test10/remote-connection-unrecoverable.exp
++++ b/pengine/test10/remote-connection-unrecoverable.exp
+@@ -9,12 +9,6 @@
+       <trigger>
+         <crm_event id="1" operation="stonith" operation_key="stonith-node1-reboot" on_node="node1" on_node_uuid="1"/>
+       </trigger>
+-      <trigger>
+-        <pseudo_event id="6" operation="stop" operation_key="rsc1_stop_0"/>
+-      </trigger>
+-      <trigger>
+-        <rsc_op id="8" operation="monitor" operation_key="rsc1_monitor_0" on_node="node2" on_node_uuid="2"/>
+-      </trigger>
+     </inputs>
+   </synapse>
+   <synapse id="1">
+diff --git a/pengine/test10/remote-connection-unrecoverable.summary b/pengine/test10/remote-connection-unrecoverable.summary
+index efeb765..18f7dc7 100644
+--- a/pengine/test10/remote-connection-unrecoverable.summary
++++ b/pengine/test10/remote-connection-unrecoverable.summary
+@@ -24,12 +24,12 @@ Executing cluster transition:
+  * Resource action: killer          stop on node2
+  * Resource action: rsc1            monitor on node2
+  * Fencing node1 (reboot)
++ * Pseudo action:   remote1_stop_0
+  * Fencing remote1 (reboot)
+  * Resource action: killer          start on node2
+  * Resource action: killer          monitor=60000 on node2
+  * Pseudo action:   rsc1_stop_0
+  * Pseudo action:   rsc2-master_demote_0
+- * Pseudo action:   remote1_stop_0
+  * Resource action: rsc1            start on node2
+  * Pseudo action:   rsc2_demote_0
+  * Pseudo action:   rsc2-master_demoted_0
+diff --git a/pengine/test10/remote-fence-before-reconnect.dot b/pengine/test10/remote-fence-before-reconnect.dot
+index 4ced43e..5812b7f 100644
+--- a/pengine/test10/remote-fence-before-reconnect.dot
++++ b/pengine/test10/remote-fence-before-reconnect.dot
+@@ -3,7 +3,6 @@
+ "fake2_monitor_10000 c7auto1" [ style=bold color="green" fontcolor="black"]
+ "fake2_start_0 c7auto1" -> "fake2_monitor_10000 c7auto1" [ style = bold]
+ "fake2_start_0 c7auto1" [ style=bold color="green" fontcolor="black"]
+-"fake2_stop_0 c7auto4" -> "c7auto4_stop_0 c7auto1" [ style = bold]
+ "fake2_stop_0 c7auto4" -> "fake2_start_0 c7auto1" [ style = bold]
+ "fake2_stop_0 c7auto4" [ style=bold color="green" fontcolor="orange"]
+ "stonith 'reboot' c7auto4" -> "fake2_start_0 c7auto1" [ style = bold]
+diff --git a/pengine/test10/remote-fence-before-reconnect.exp b/pengine/test10/remote-fence-before-reconnect.exp
+index f99d9ef..f506f85 100644
+--- a/pengine/test10/remote-fence-before-reconnect.exp
++++ b/pengine/test10/remote-fence-before-reconnect.exp
+@@ -9,11 +9,7 @@
+         </downed>
+       </rsc_op>
+     </action_set>
+-    <inputs>
+-      <trigger>
+-        <pseudo_event id="13" operation="stop" operation_key="fake2_stop_0"/>
+-      </trigger>
+-    </inputs>
++    <inputs/>
+   </synapse>
+   <synapse id="1">
+     <action_set>
+diff --git a/pengine/test10/remote-fence-before-reconnect.summary b/pengine/test10/remote-fence-before-reconnect.summary
+index f61e18b..03eac20 100644
+--- a/pengine/test10/remote-fence-before-reconnect.summary
++++ b/pengine/test10/remote-fence-before-reconnect.summary
+@@ -17,9 +17,9 @@ Transition Summary:
+  * Move       fake2       ( c7auto4 -> c7auto1 )  
+ 
+ Executing cluster transition:
++ * Resource action: c7auto4         stop on c7auto1
+  * Fencing c7auto4 (reboot)
+  * Pseudo action:   fake2_stop_0
+- * Resource action: c7auto4         stop on c7auto1
+  * Resource action: fake2           start on c7auto1
+  * Resource action: fake2           monitor=10000 on c7auto1
+ 
+diff --git a/pengine/test10/remote-recover-all.dot b/pengine/test10/remote-recover-all.dot
+index 1f967c5..b48b04e 100644
+--- a/pengine/test10/remote-recover-all.dot
++++ b/pengine/test10/remote-recover-all.dot
+@@ -19,7 +19,6 @@ digraph "g" {
+ "galera_demote_0 galera-2" -> "galera_stop_0 galera-2" [ style = bold]
+ "galera_demote_0 galera-2" [ style=bold color="green" fontcolor="orange"]
+ "galera_monitor_10000 galera-0" [ style=bold color="green" fontcolor="black"]
+-"galera_stop_0 galera-2" -> "galera-2_stop_0 controller-1" [ style = bold]
+ "galera_stop_0 galera-2" -> "galera-master_stopped_0" [ style = bold]
+ "galera_stop_0 galera-2" [ style=bold color="green" fontcolor="orange"]
+ "haproxy-clone_stop_0" -> "haproxy-clone_stopped_0" [ style = bold]
+@@ -60,7 +59,6 @@ digraph "g" {
+ "rabbitmq_post_notify_stonith_0" -> "rabbitmq_post_notify_stonith_0 messaging-0" [ style = bold]
+ "rabbitmq_post_notify_stonith_0" -> "rabbitmq_post_notify_stonith_0 messaging-2" [ style = bold]
+ "rabbitmq_post_notify_stonith_0" [ style=bold color="green" fontcolor="orange"]
+-"rabbitmq_stop_0 messaging-1" -> "messaging-1_stop_0 controller-1" [ style = bold]
+ "rabbitmq_stop_0 messaging-1" -> "rabbitmq-clone_stopped_0" [ style = bold]
+ "rabbitmq_stop_0 messaging-1" [ style=bold color="green" fontcolor="orange"]
+ "redis-master_confirmed-post_notify_stopped_0" [ style=bold color="green" fontcolor="orange"]
+diff --git a/pengine/test10/remote-recover-all.exp b/pengine/test10/remote-recover-all.exp
+index 900781c..e61ad6a 100644
+--- a/pengine/test10/remote-recover-all.exp
++++ b/pengine/test10/remote-recover-all.exp
+@@ -9,9 +9,6 @@
+       <trigger>
+         <crm_event id="1" operation="stonith" operation_key="stonith-controller-1-reboot" on_node="controller-1" on_node_uuid="2"/>
+       </trigger>
+-      <trigger>
+-        <pseudo_event id="39" operation="stop" operation_key="rabbitmq_stop_0" internal_operation_key="rabbitmq:2_stop_0"/>
+-      </trigger>
+     </inputs>
+   </synapse>
+   <synapse id="1">
+@@ -64,9 +61,6 @@
+       <trigger>
+         <crm_event id="1" operation="stonith" operation_key="stonith-controller-1-reboot" on_node="controller-1" on_node_uuid="2"/>
+       </trigger>
+-      <trigger>
+-        <pseudo_event id="49" operation="stop" operation_key="galera_stop_0" internal_operation_key="galera:1_stop_0"/>
+-      </trigger>
+     </inputs>
+   </synapse>
+   <synapse id="5" priority="1000000">
+diff --git a/pengine/test10/remote-recover-all.summary b/pengine/test10/remote-recover-all.summary
+index 865f39a..cfeac3a 100644
+--- a/pengine/test10/remote-recover-all.summary
++++ b/pengine/test10/remote-recover-all.summary
+@@ -63,6 +63,8 @@ Executing cluster transition:
+  * Resource action: stonith-fence_ipmilan-525400b4f6bd stop on controller-0
+  * Pseudo action:   stonith-fence_ipmilan-5254005bdbb5_stop_0
+  * Fencing controller-1 (reboot)
++ * Pseudo action:   messaging-1_stop_0
++ * Pseudo action:   galera-2_stop_0
+  * Pseudo action:   redis_post_notify_stop_0
+  * Resource action: redis           notify on controller-0
+  * Resource action: redis           notify on controller-2
+@@ -94,7 +96,6 @@ Executing cluster transition:
+  * Resource action: stonith-fence_ipmilan-525400b4f6bd monitor=60000 on controller-0
+  * Resource action: stonith-fence_ipmilan-5254005bdbb5 start on controller-2
+  * Resource action: galera-0        monitor=20000 on controller-2
+- * Pseudo action:   galera-2_stop_0
+  * Resource action: rabbitmq        notify on messaging-2
+  * Resource action: rabbitmq        notify on messaging-0
+  * Pseudo action:   rabbitmq_notified_0
+@@ -107,7 +108,6 @@ Executing cluster transition:
+  * Resource action: ip-172.17.1.17  start on controller-2
+  * Resource action: ip-172.17.4.11  start on controller-2
+  * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
+- * Pseudo action:   messaging-1_stop_0
+  * Pseudo action:   redis_notified_0
+  * Resource action: ip-172.17.1.14  monitor=10000 on controller-2
+  * Resource action: ip-172.17.1.17  monitor=10000 on controller-2
+diff --git a/pengine/test10/remote-recover-no-resources.dot b/pengine/test10/remote-recover-no-resources.dot
+index a46c305..a0b1ecc 100644
+--- a/pengine/test10/remote-recover-no-resources.dot
++++ b/pengine/test10/remote-recover-no-resources.dot
+@@ -45,7 +45,6 @@ digraph "g" {
+ "rabbitmq_post_notify_stonith_0" -> "rabbitmq_post_notify_stonith_0 messaging-0" [ style = bold]
+ "rabbitmq_post_notify_stonith_0" -> "rabbitmq_post_notify_stonith_0 messaging-2" [ style = bold]
+ "rabbitmq_post_notify_stonith_0" [ style=bold color="green" fontcolor="orange"]
+-"rabbitmq_stop_0 messaging-1" -> "messaging-1_stop_0 controller-1" [ style = bold]
+ "rabbitmq_stop_0 messaging-1" -> "rabbitmq-clone_stopped_0" [ style = bold]
+ "rabbitmq_stop_0 messaging-1" [ style=bold color="green" fontcolor="orange"]
+ "redis-master_confirmed-post_notify_stopped_0" [ style=bold color="green" fontcolor="orange"]
+diff --git a/pengine/test10/remote-recover-no-resources.exp b/pengine/test10/remote-recover-no-resources.exp
+index 4d82aa4..27f18b5 100644
+--- a/pengine/test10/remote-recover-no-resources.exp
++++ b/pengine/test10/remote-recover-no-resources.exp
+@@ -9,9 +9,6 @@
+       <trigger>
+         <crm_event id="1" operation="stonith" operation_key="stonith-controller-1-reboot" on_node="controller-1" on_node_uuid="2"/>
+       </trigger>
+-      <trigger>
+-        <pseudo_event id="38" operation="stop" operation_key="rabbitmq_stop_0" internal_operation_key="rabbitmq:2_stop_0"/>
+-      </trigger>
+     </inputs>
+   </synapse>
+   <synapse id="1">
+diff --git a/pengine/test10/remote-recover-no-resources.summary b/pengine/test10/remote-recover-no-resources.summary
+index 9527161..c01eb87 100644
+--- a/pengine/test10/remote-recover-no-resources.summary
++++ b/pengine/test10/remote-recover-no-resources.summary
+@@ -60,6 +60,7 @@ Executing cluster transition:
+  * Resource action: stonith-fence_ipmilan-525400b4f6bd stop on controller-0
+  * Pseudo action:   stonith-fence_ipmilan-5254005bdbb5_stop_0
+  * Fencing controller-1 (reboot)
++ * Pseudo action:   messaging-1_stop_0
+  * Pseudo action:   galera-2_stop_0
+  * Pseudo action:   redis_post_notify_stop_0
+  * Resource action: redis           notify on controller-0
+@@ -92,7 +93,6 @@ Executing cluster transition:
+  * Pseudo action:   ip-172.17.1.17_stop_0
+  * Pseudo action:   ip-172.17.4.11_stop_0
+  * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
+- * Pseudo action:   messaging-1_stop_0
+  * Resource action: redis           notify on controller-0
+  * Resource action: redis           notify on controller-2
+  * Pseudo action:   redis-master_confirmed-post_notify_stopped_0
+diff --git a/pengine/test10/remote-recover-unknown.dot b/pengine/test10/remote-recover-unknown.dot
+index a883eb4..1d13e50 100644
+--- a/pengine/test10/remote-recover-unknown.dot
++++ b/pengine/test10/remote-recover-unknown.dot
+@@ -46,7 +46,6 @@ digraph "g" {
+ "rabbitmq_post_notify_stonith_0" -> "rabbitmq_post_notify_stonith_0 messaging-0" [ style = bold]
+ "rabbitmq_post_notify_stonith_0" -> "rabbitmq_post_notify_stonith_0 messaging-2" [ style = bold]
+ "rabbitmq_post_notify_stonith_0" [ style=bold color="green" fontcolor="orange"]
+-"rabbitmq_stop_0 messaging-1" -> "messaging-1_stop_0 controller-1" [ style = bold]
+ "rabbitmq_stop_0 messaging-1" -> "rabbitmq-clone_stopped_0" [ style = bold]
+ "rabbitmq_stop_0 messaging-1" [ style=bold color="green" fontcolor="orange"]
+ "redis-master_confirmed-post_notify_stopped_0" [ style=bold color="green" fontcolor="orange"]
+diff --git a/pengine/test10/remote-recover-unknown.exp b/pengine/test10/remote-recover-unknown.exp
+index 65677b4..13bd295 100644
+--- a/pengine/test10/remote-recover-unknown.exp
++++ b/pengine/test10/remote-recover-unknown.exp
+@@ -9,9 +9,6 @@
+       <trigger>
+         <crm_event id="1" operation="stonith" operation_key="stonith-controller-1-reboot" on_node="controller-1" on_node_uuid="2"/>
+       </trigger>
+-      <trigger>
+-        <pseudo_event id="39" operation="stop" operation_key="rabbitmq_stop_0" internal_operation_key="rabbitmq:2_stop_0"/>
+-      </trigger>
+     </inputs>
+   </synapse>
+   <synapse id="1">
+diff --git a/pengine/test10/remote-recover-unknown.summary b/pengine/test10/remote-recover-unknown.summary
+index 78a60d0..64f37cb 100644
+--- a/pengine/test10/remote-recover-unknown.summary
++++ b/pengine/test10/remote-recover-unknown.summary
+@@ -61,6 +61,7 @@ Executing cluster transition:
+  * Resource action: stonith-fence_ipmilan-525400b4f6bd stop on controller-0
+  * Pseudo action:   stonith-fence_ipmilan-5254005bdbb5_stop_0
+  * Fencing controller-1 (reboot)
++ * Pseudo action:   messaging-1_stop_0
+  * Pseudo action:   galera-2_stop_0
+  * Pseudo action:   redis_post_notify_stop_0
+  * Resource action: redis           notify on controller-0
+@@ -94,7 +95,6 @@ Executing cluster transition:
+  * Pseudo action:   ip-172.17.1.17_stop_0
+  * Pseudo action:   ip-172.17.4.11_stop_0
+  * Resource action: stonith-fence_ipmilan-5254005bdbb5 monitor=60000 on controller-2
+- * Pseudo action:   messaging-1_stop_0
+  * Resource action: redis           notify on controller-0
+  * Resource action: redis           notify on controller-2
+  * Pseudo action:   redis-master_confirmed-post_notify_stopped_0
+-- 
+1.8.3.1
+
diff --git a/SOURCES/012-tls-priorities.patch b/SOURCES/012-tls-priorities.patch
new file mode 100644
index 0000000..34396a2
--- /dev/null
+++ b/SOURCES/012-tls-priorities.patch
@@ -0,0 +1,189 @@
+From 7c3bc762a9cede20a0193f64ca1a36f507aeeeb3 Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Fri, 20 Apr 2018 13:23:10 -0500
+Subject: [PATCH 1/2] Build: libcrmcommon: configure option to specify GnuTLS
+ cipher priorities
+
+Default to current behavior, i.e. "NORMAL". Spec file overrides with "@SYSTEM"
+on distros that have it.
+
+Pacemaker does not use option value as-is; it adds "+ANON-DH" for CIB remote
+commands and "+DHE-PSK:+PSK" for Pacemaker Remote connections. In the longer
+term, we could consider moving to certificate-based connections in both cases,
+but that has backward compatibility issues as well as additional administrative
+burden.
+---
+ configure.ac        | 9 +++++++++
+ lib/common/remote.c | 4 ++--
+ pacemaker.spec.in   | 4 ++++
+ 3 files changed, 15 insertions(+), 2 deletions(-)
+
+diff --git a/configure.ac b/configure.ac
+index ce02777..a7084e2 100644
+--- a/configure.ac
++++ b/configure.ac
+@@ -290,6 +290,12 @@ AC_ARG_WITH(cibsecrets,
+     [ SUPPORT_CIBSECRETS=no ],
+ )
+ 
++AC_ARG_WITH(gnutls-priorities,
++    [  --with-gnutls-priorities  GnuTLS cipher priorities @<:@NORMAL@:>@ ],
++    [ PCMK_GNUTLS_PRIORITIES="$withval" ],
++    [ PCMK_GNUTLS_PRIORITIES="NORMAL" ],
++)
++
+ CSPREFIX=""
+ AC_ARG_WITH(ais-prefix,
+     [  --with-ais-prefix=DIR  Prefix used when Corosync was installed [$prefix]],
+@@ -453,6 +459,9 @@ if test x"${BUG_URL}" = x""; then
+ fi
+ AC_SUBST(BUG_URL)
+ 
++AC_DEFINE_UNQUOTED([PCMK_GNUTLS_PRIORITIES], ["$PCMK_GNUTLS_PRIORITIES"],
++		   [GnuTLS cipher priorities])
++
+ for j in prefix exec_prefix bindir sbindir libexecdir datadir sysconfdir \
+     sharedstatedir localstatedir libdir includedir oldincludedir infodir \
+     mandir INITDIR docdir CONFIGDIR
+diff --git a/lib/common/remote.c b/lib/common/remote.c
+index 12d25fa..1e4f8d8 100644
+--- a/lib/common/remote.c
++++ b/lib/common/remote.c
+@@ -244,9 +244,9 @@ pcmk__new_tls_session(int csock, unsigned int conn_type,
+ #  ifdef HAVE_GNUTLS_PRIORITY_SET_DIRECT
+     if (cred_type == GNUTLS_CRD_ANON) {
+         // http://www.manpagez.com/info/gnutls/gnutls-2.10.4/gnutls_81.php#Echo-Server-with-anonymous-authentication
+-        prio = "NORMAL:+ANON-DH";
++        prio = PCMK_GNUTLS_PRIORITIES ":+ANON-DH";
+     } else {
+-        prio = "NORMAL:+DHE-PSK:+PSK";
++        prio = PCMK_GNUTLS_PRIORITIES ":+DHE-PSK:+PSK";
+     }
+ #  endif
+ 
+diff --git a/pacemaker.spec.in b/pacemaker.spec.in
+index 3a26572..fd0e3c8 100644
+--- a/pacemaker.spec.in
++++ b/pacemaker.spec.in
+@@ -80,6 +80,9 @@
+   } || %{?__transaction_systemd_inhibit:1}%{!?__transaction_systemd_inhibit:0}%{nil \
+   } || %(test -f /usr/lib/os-release; test $? -ne 0; echo $?))
+ 
++%if 0%{?fedora} > 20 || 0%{?rhel} > 7
++%global gnutls_priorities @SYSTEM
++%endif
+ 
+ # Definitions for backward compatibility with older RPM versions
+ 
+@@ -403,6 +406,7 @@ export LDFLAGS_HARDENED_LIB="%{?_hardening_ldflags}"
+         --without-heartbeat                        \
+         %{!?with_doc:        --with-brand=}        \
+         %{!?with_hardening:  --disable-hardening}  \
++        %{?gnutls_priorities: --with-gnutls-priorities="%{gnutls_priorities}"} \
+         --with-initdir=%{_initrddir}               \
+         --localstatedir=%{_var}                    \
+         --with-version=%{version}-%{release}
+-- 
+1.8.3.1
+
+
+From 99a83b172544102ec32585514e5808585f2ce31c Mon Sep 17 00:00:00 2001
+From: Ken Gaillot <kgaillot@redhat.com>
+Date: Mon, 8 Jul 2019 17:39:12 -0500
+Subject: [PATCH 2/2] Feature: remote: allow run-time configurable TLS
+ priorities
+
+This also restores compilability with GnuTLS <2.1.7 (not that anyone is still
+using that ...), unintentionally broken in 5bded36 (1.1.20).
+---
+ lib/common/remote.c     | 34 +++++++++++++++++++++++++++-------
+ mcp/pacemaker.sysconfig |  9 +++++++++
+ 2 files changed, 36 insertions(+), 7 deletions(-)
+
+diff --git a/lib/common/remote.c b/lib/common/remote.c
+index 1e4f8d8..ccd0342 100644
+--- a/lib/common/remote.c
++++ b/lib/common/remote.c
+@@ -237,17 +237,25 @@ pcmk__new_tls_session(int csock, unsigned int conn_type,
+ {
+     int rc = GNUTLS_E_SUCCESS;
+ #  ifdef HAVE_GNUTLS_PRIORITY_SET_DIRECT
+-    const char *prio = NULL;
++    const char *prio_base = NULL;
++    char *prio = NULL;
+ #  endif
+     gnutls_session_t *session = NULL;
+ 
+ #  ifdef HAVE_GNUTLS_PRIORITY_SET_DIRECT
+-    if (cred_type == GNUTLS_CRD_ANON) {
+-        // http://www.manpagez.com/info/gnutls/gnutls-2.10.4/gnutls_81.php#Echo-Server-with-anonymous-authentication
+-        prio = PCMK_GNUTLS_PRIORITIES ":+ANON-DH";
+-    } else {
+-        prio = PCMK_GNUTLS_PRIORITIES ":+DHE-PSK:+PSK";
++    /* Determine list of acceptable ciphers, etc. Pacemaker always adds the
++     * values required for its functionality.
++     *
++     * For an example of anonymous authentication, see:
++     * http://www.manpagez.com/info/gnutls/gnutls-2.10.4/gnutls_81.php#Echo-Server-with-anonymous-authentication
++     */
++
++    prio_base = getenv("PCMK_tls_priorities");
++    if (prio_base == NULL) {
++        prio_base = PCMK_GNUTLS_PRIORITIES;
+     }
++    prio = crm_strdup_printf("%s:%s", prio_base,
++                             (cred_type == GNUTLS_CRD_ANON)? "+ANON-DH" : "+DHE-PSK:+PSK");
+ #  endif
+ 
+     session = gnutls_malloc(sizeof(gnutls_session_t));
+@@ -285,6 +293,9 @@ pcmk__new_tls_session(int csock, unsigned int conn_type,
+     if (rc != GNUTLS_E_SUCCESS) {
+         goto error;
+     }
++#  ifdef HAVE_GNUTLS_PRIORITY_SET_DIRECT
++    free(prio);
++#  endif
+     return session;
+ 
+ error:
+@@ -292,7 +303,16 @@ error:
+             CRM_XS " rc=%d priority='%s'",
+             (cred_type == GNUTLS_CRD_ANON)? "anonymous" : "PSK",
+             (conn_type == GNUTLS_SERVER)? "server" : "client",
+-            gnutls_strerror(rc), rc, prio);
++            gnutls_strerror(rc), rc,
++#  ifdef HAVE_GNUTLS_PRIORITY_SET_DIRECT
++            prio
++#  else
++            "default"
++#  endif
++            );
++#  ifdef HAVE_GNUTLS_PRIORITY_SET_DIRECT
++    free(prio);
++#  endif
+     if (session != NULL) {
+         gnutls_free(session);
+     }
+diff --git a/mcp/pacemaker.sysconfig b/mcp/pacemaker.sysconfig
+index a983011..0da401e 100644
+--- a/mcp/pacemaker.sysconfig
++++ b/mcp/pacemaker.sysconfig
+@@ -101,6 +101,15 @@
+ # value must be the same on all nodes. The default is "3121".
+ # PCMK_remote_port=3121
+ 
++# Use these GnuTLS cipher priorities for TLS connections. See:
++#
++#   https://gnutls.org/manual/html_node/Priority-Strings.html
++#
++# Pacemaker will append ":+ANON-DH" for remote CIB access (when enabled) and
++# ":+DHE-PSK:+PSK" for Pacemaker Remote connections, as they are required for
++# the respective functionality.
++# PCMK_tls_priorities="NORMAL"
++
+ # Set bounds on the bit length of the prime number generated for Diffie-Hellman
+ # parameters needed by TLS connections. The default is not to set any bounds.
+ #
+-- 
+1.8.3.1
+
diff --git a/SPECS/pacemaker.spec b/SPECS/pacemaker.spec
index 76c252d..3315f2c 100644
--- a/SPECS/pacemaker.spec
+++ b/SPECS/pacemaker.spec
@@ -80,6 +80,14 @@
   } || %{?__transaction_systemd_inhibit:1}%{!?__transaction_systemd_inhibit:0}%{nil \
   } || %(test -f /usr/lib/os-release; test $? -ne 0; echo $?))
 
+%if 0%{?fedora} > 20 || 0%{?rhel} > 7
+%global gnutls_priorities @SYSTEM
+%else
+%if 0%{?rhel} > 6
+%global gnutls_priorities NORMAL:-VERS-SSL3.0:-VERS-TLS1.0:-VERS-TLS1.1:-MD5:-3DES-CBC:-ARCFOUR-128:-ARCFOUR-40
+%endif
+%endif
+ 
 ## Upstream commit to use for nagios-agents-metadata package
 %global nagios_hash 105ab8a
 
@@ -160,7 +168,7 @@
 Name:          pacemaker
 Summary:       Scalable High-Availability cluster resource manager
 Version:       %{pcmkversion}
-Release:       %{pcmk_release}%{?dist}
+Release:       %{pcmk_release}%{?dist}.1
 %if %{defined _unitdir}
 License:       GPLv2+ and LGPLv2+
 %else
@@ -186,6 +194,8 @@ Patch7:        007-security.patch
 Patch8:        008-security-log.patch
 Patch9:        009-use-after-free.patch
 Patch10:       010-fork-callback.patch
+Patch11:       011-remote.patch
+Patch12:       012-tls-priorities.patch
 
 # patches that aren't from upstream
 Patch100:      lrmd-protocol-version.patch
@@ -204,7 +214,7 @@ Provides:      pcmk-cluster-manager
 
 %{?systemd_requires}
 
-ExclusiveArch: aarch64 i686 ppc64le s390x x86_64 %{arm}
+ExclusiveArch: aarch64 i686 ppc64le s390x x86_64
 
 # Pacemaker targets compatibility with python 2.6+ and 3.2+
 Requires:      python >= 2.6
@@ -460,6 +470,7 @@ export LDFLAGS_HARDENED_LIB="%{?_hardening_ldflags}"
         --without-heartbeat                        \
         %{!?with_doc:        --with-brand=}        \
         %{!?with_hardening:  --disable-hardening}  \
+        %{?gnutls_priorities: --with-gnutls-priorities="%{gnutls_priorities}"} \
         --with-initdir=%{_initrddir}               \
         --localstatedir=%{_var}                    \
         --with-bug-url=https://bugzilla.redhat.com/ \
@@ -872,6 +883,12 @@ exit 0
 %attr(0644,root,root) %{_datadir}/pacemaker/nagios/plugins-metadata/*
 
 %changelog
+* Tue Jul 23 2019 Ken Gaillot <kgaillot@redhat.com> - 1.1.20-5.1
+- Handle losing remote node while it is shutting down
+- Allow configurable GnuTLS cipher priorities and use stricter default
+- Resolves: rhbz#1732335
+- Resolves: rhbz#1733187
+
 * Fri May 24 2019 Ken Gaillot <kgaillot@redhat.com> - 1.1.20-5
 - Correct memory issue in fence agent output fix
 - Resolves: rhbz#1549366