diff options
| author | Atin Mukherjee <amukherj@redhat.com> | 2016-07-26 16:01:56 +0530 |
|---|---|---|
| committer | Atin Mukherjee <amukherj@redhat.com> | 2016-08-22 06:10:44 -0700 |
| commit | ca18f4bccd090e98ee5342ca05d3c0f9f94e9e2c (patch) | |
| tree | 4ea51c5424f1b0d927292c8b51177da99d1cd213 | |
| parent | fcb5b70b1099d0379b40c81f35750df8bb9545a5 (diff) | |
glusterd: Add async events
As the eventing framework is already in the code, this patch targets to capture
all the async glusterd events which are important to be notified to the higher
layers which consume the eventing framework.
I plan to break this work into two different patches where this patch set covers
the first set of events.
Change-Id: Ie1bd4f6fa84117b26ccb4c75bc4dc68e6ef19134
BUG: 1360809
Signed-off-by: Atin Mukherjee <amukherj@redhat.com>
Reviewed-on: http://review.gluster.org/15015
Smoke: Gluster Build System <jenkins@build.gluster.org>
NetBSD-regression: NetBSD Build System <jenkins@build.gluster.org>
CentOS-regression: Gluster Build System <jenkins@build.gluster.org>
Reviewed-by: Rohan Kanade <rkanade@redhat.com>
Reviewed-by: Samikshan Bairagya <samikshan@gmail.com>
| -rw-r--r-- | events/eventskeygen.py | 20 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-bitd-svc.c | 3 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-handler.c | 59 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-handshake.c | 3 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-nfs-svc.c | 3 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-op-sm.c | 11 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-pmap.c | 2 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-quotad-svc.c | 3 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-rebalance.c | 5 | ||||
| -rw-r--r-- | xlators/mgmt/glusterd/src/glusterd-svc-helper.c | 4 |
10 files changed, 98 insertions, 15 deletions
diff --git a/events/eventskeygen.py b/events/eventskeygen.py index 086f8a6307..282d0fe670 100644 --- a/events/eventskeygen.py +++ b/events/eventskeygen.py @@ -46,6 +46,26 @@ keys = ( "EVENT_BITROT_SCRUB_FREQ", "EVENT_BITROT_SCRUB_OPTION", "EVENT_BITROT_BAD_FILE", + + "EVENT_SVC_MANAGER_FAILED", + "EVENT_SVC_RECONFIGURE_FAILED", + + "EVENT_DEFRAG_STATUS_UPDATE_FAILED", + + "EVENT_PEER_STORE_FAILURE", + "EVENT_PEER_RPC_CREATE_FAILED", + "EVENT_PEER_REJECT", + "EVENT_PEER_CONNECT", + "EVENT_PEER_DISCONNECT", + "EVENT_PEER_NOT_FOUND", + "EVENT_UNKNOWN_PEER", + + "EVENT_BRICK_START_FAILED", + "EVENT_BRICK_STOP_FAILED", + "EVENT_BRICK_DISCONNECTED", + "EVENT_BRICK_CONNECTED", + + "EVENT_NOTIFY_UNKNOWN_OP", ) LAST_EVENT = "EVENT_LAST" diff --git a/xlators/mgmt/glusterd/src/glusterd-bitd-svc.c b/xlators/mgmt/glusterd/src/glusterd-bitd-svc.c index ee96ccbff8..f929401923 100644 --- a/xlators/mgmt/glusterd/src/glusterd-bitd-svc.c +++ b/xlators/mgmt/glusterd/src/glusterd-bitd-svc.c @@ -106,6 +106,9 @@ glusterd_bitdsvc_manager (glusterd_svc_t *svc, void *data, int flags) } out: + if (ret) + gf_event (EVENT_SVC_MANAGER_FAILED, "svc_name=%s", svc->name); + gf_msg_debug (THIS->name, 0, "Returning %d", ret); return ret; diff --git a/xlators/mgmt/glusterd/src/glusterd-handler.c b/xlators/mgmt/glusterd/src/glusterd-handler.c index 91ae6237c5..0085dce5d4 100644 --- a/xlators/mgmt/glusterd/src/glusterd-handler.c +++ b/xlators/mgmt/glusterd/src/glusterd-handler.c @@ -2771,6 +2771,8 @@ __glusterd_handle_friend_update (rpcsvc_request_t *req) GD_MSG_REQ_FROM_UNKNOWN_PEER, "Received friend update request " "from unknown peer %s", uuid_utoa (friend_req.uuid)); + gf_event (EVENT_UNKNOWN_PEER, "peer=%s", + uuid_utoa (friend_req.uuid)); goto out; } @@ -2867,10 +2869,13 @@ __glusterd_handle_friend_update (rpcsvc_request_t *req) goto unlock; } ret = glusterd_store_peerinfo (peerinfo); - if (ret) + if (ret) { gf_msg (this->name, GF_LOG_ERROR, 0, GD_MSG_PEERINFO_CREATE_FAIL, "Failed to store peerinfo"); + gf_event (EVENT_PEER_STORE_FAILURE, "peer=%s", + peerinfo->hostname); + } } unlock: rcu_read_unlock (); @@ -3525,6 +3530,8 @@ glusterd_friend_rpc_create (xlator_t *this, glusterd_peerinfo_t *peerinfo, GD_MSG_RPC_CREATE_FAIL, "failed to create rpc for" " peer %s", peerinfo->hostname); + gf_event (EVENT_PEER_RPC_CREATE_FAILED, "peer=%s", + peerinfo->hostname); goto out; } peerctx = NULL; @@ -3580,6 +3587,8 @@ glusterd_friend_add (const char *hoststr, int port, gf_msg (this->name, GF_LOG_ERROR, 0, GD_MSG_PEERINFO_CREATE_FAIL, "Failed to store peerinfo"); + gf_event (EVENT_PEER_STORE_FAILURE, "peer=%s", + (*friend)->hostname); } } @@ -3635,6 +3644,8 @@ glusterd_friend_add_from_peerinfo (glusterd_peerinfo_t *friend, gf_msg (this->name, GF_LOG_ERROR, 0, GD_MSG_PEERINFO_CREATE_FAIL, "Failed to store peerinfo"); + gf_event (EVENT_PEER_STORE_FAILURE, "peer=%s", + friend->hostname); } } @@ -4999,6 +5010,14 @@ __glusterd_brick_rpc_notify (struct rpc_clnt *rpc, void *mydata, switch (event) { case RPC_CLNT_CONNECT: + ret = get_volinfo_from_brickid (brickid, &volinfo); + if (ret) { + gf_msg (this->name, GF_LOG_ERROR, 0, + GD_MSG_VOLINFO_GET_FAIL, + "Failed to get volinfo from " + "brickid(%s)", brickid); + goto out; + } /* If a node on coming back up, already starts a brick * before the handshake, and the notification comes after * the handshake is done, then we need to check if this @@ -5012,15 +5031,6 @@ __glusterd_brick_rpc_notify (struct rpc_clnt *rpc, void *mydata, "Hence not starting the brick", brickinfo->hostname, brickinfo->path); - ret = get_volinfo_from_brickid (brickid, &volinfo); - if (ret) { - gf_msg (this->name, GF_LOG_ERROR, 0, - GD_MSG_VOLINFO_GET_FAIL, - "Failed to get volinfo from " - "brickid(%s)", brickid); - goto out; - } - ret = glusterd_brick_stop (volinfo, brickinfo, _gf_false); if (ret) { @@ -5037,17 +5047,34 @@ __glusterd_brick_rpc_notify (struct rpc_clnt *rpc, void *mydata, gf_msg_debug (this->name, 0, "Connected to %s:%s", brickinfo->hostname, brickinfo->path); glusterd_set_brick_status (brickinfo, GF_BRICK_STARTED); + gf_event (EVENT_BRICK_CONNECTED, "peer=%s;volume=%s;brick=%s", + brickinfo->hostname, volinfo->volname, + brickinfo->path); + ret = default_notify (this, GF_EVENT_CHILD_UP, NULL); break; case RPC_CLNT_DISCONNECT: rpc_clnt_unset_connected (&rpc->conn); - if (glusterd_is_brick_started (brickinfo)) + if (glusterd_is_brick_started (brickinfo)) { gf_msg (this->name, GF_LOG_INFO, 0, GD_MSG_BRICK_DISCONNECTED, "Brick %s:%s has disconnected from glusterd.", brickinfo->hostname, brickinfo->path); + ret = get_volinfo_from_brickid (brickid, &volinfo); + if (ret) { + gf_msg (this->name, GF_LOG_ERROR, 0, + GD_MSG_VOLINFO_GET_FAIL, + "Failed to get volinfo from " + "brickid(%s)", brickid); + goto out; + } + gf_event (EVENT_BRICK_DISCONNECTED, + "peer=%s;volume=%s;brick=%s", + brickinfo->hostname, volinfo->volname, + brickinfo->path); + } glusterd_set_brick_status (brickinfo, GF_BRICK_STOPPED); break; @@ -5174,6 +5201,11 @@ __glusterd_peer_rpc_notify (struct rpc_clnt *rpc, void *mydata, "%s(%s)", peerctx->peername, uuid_utoa (peerctx->peerid)); + if (RPC_CLNT_CONNECT == event) { + gf_event (EVENT_PEER_NOT_FOUND, "peer=%s;uuid=%s", + peerctx->peername, + uuid_utoa (peerctx->peerid)); + } ret = -1; goto out; } @@ -5188,6 +5220,8 @@ __glusterd_peer_rpc_notify (struct rpc_clnt *rpc, void *mydata, peerinfo->generation = uatomic_add_return (&conf->generation, 1); peerctx->peerinfo_gen = peerinfo->generation; + gf_event (EVENT_PEER_CONNECT, "host=%s;uuid=%s", + peerinfo->hostname, uuid_utoa (peerinfo->uuid)); ret = glusterd_peer_dump_version (this, rpc, peerctx); if (ret) @@ -5212,6 +5246,9 @@ __glusterd_peer_rpc_notify (struct rpc_clnt *rpc, void *mydata, "from glusterd.", peerinfo->hostname, uuid_utoa (peerinfo->uuid), glusterd_friend_sm_state_name_get (peerinfo->state.state)); + gf_event (EVENT_PEER_DISCONNECT, "peer=%s;uuid=%s;state=%s", + peerinfo->hostname, uuid_utoa (peerinfo->uuid), + glusterd_friend_sm_state_name_get (peerinfo->state.state)); if (peerinfo->connected) { if (conf->op_version < GD_OP_VERSION_3_6_0) { diff --git a/xlators/mgmt/glusterd/src/glusterd-handshake.c b/xlators/mgmt/glusterd/src/glusterd-handshake.c index 0ea66a027b..9f162d8afc 100644 --- a/xlators/mgmt/glusterd/src/glusterd-handshake.c +++ b/xlators/mgmt/glusterd/src/glusterd-handshake.c @@ -984,6 +984,7 @@ __server_event_notify (rpcsvc_request_t *req) gf_msg ("glusterd", GF_LOG_ERROR, EINVAL, GD_MSG_OP_UNSUPPORTED, "Unknown op received in event " "notify"); + gf_event (EVENT_NOTIFY_UNKNOWN_OP, "op=%d", args.op); ret = -1; break; } @@ -1118,6 +1119,8 @@ gd_validate_mgmt_hndsk_req (rpcsvc_request_t *req, dict_t *dict) GD_MSG_HANDSHAKE_REQ_REJECTED, "Rejecting management " "handshake request from unknown peer %s", req->trans->peerinfo.identifier); + gf_event (EVENT_PEER_REJECT, "peer=%s", + req->trans->peerinfo.identifier); return _gf_false; } diff --git a/xlators/mgmt/glusterd/src/glusterd-nfs-svc.c b/xlators/mgmt/glusterd/src/glusterd-nfs-svc.c index 60b792ffac..c6ab0c5d52 100644 --- a/xlators/mgmt/glusterd/src/glusterd-nfs-svc.c +++ b/xlators/mgmt/glusterd/src/glusterd-nfs-svc.c @@ -95,6 +95,9 @@ glusterd_nfssvc_manager (glusterd_svc_t *svc, void *data, int flags) goto out; } out: + if (ret) + gf_event (EVENT_SVC_MANAGER_FAILED, "svc_name=%s", svc->name); + gf_msg_debug (THIS->name, 0, "Returning %d", ret); return ret; diff --git a/xlators/mgmt/glusterd/src/glusterd-op-sm.c b/xlators/mgmt/glusterd/src/glusterd-op-sm.c index b06a597854..a0904fb963 100644 --- a/xlators/mgmt/glusterd/src/glusterd-op-sm.c +++ b/xlators/mgmt/glusterd/src/glusterd-op-sm.c @@ -2194,8 +2194,13 @@ glusterd_stop_bricks (glusterd_volinfo_t *volinfo) cds_list_for_each_entry (brickinfo, &volinfo->bricks, brick_list) { /*TODO: Need to change @del_brick in brick_stop to _gf_true * once we enable synctask in peer rpc prog */ - if (glusterd_brick_stop (volinfo, brickinfo, _gf_false)) + if (glusterd_brick_stop (volinfo, brickinfo, _gf_false)) { + gf_event (EVENT_BRICK_STOP_FAILED, + "peer=%s;volume=%s;brick=%s", + brickinfo->hostname, volinfo->volname, + brickinfo->path); return -1; + } } return 0; @@ -2217,6 +2222,10 @@ glusterd_start_bricks (glusterd_volinfo_t *volinfo) "Failed to start %s:%s for %s", brickinfo->hostname, brickinfo->path, volinfo->volname); + gf_event (EVENT_BRICK_START_FAILED, + "peer=%s;volume=%s;brick=%s", + brickinfo->hostname, volinfo->volname, + brickinfo->path); goto out; } } diff --git a/xlators/mgmt/glusterd/src/glusterd-pmap.c b/xlators/mgmt/glusterd/src/glusterd-pmap.c index 6c65da7939..2c27473f19 100644 --- a/xlators/mgmt/glusterd/src/glusterd-pmap.c +++ b/xlators/mgmt/glusterd/src/glusterd-pmap.c @@ -217,7 +217,7 @@ pmap_assign_port (xlator_t *this, int old_port, const char *path) GF_PMAP_PORT_BRICKSERVER, NULL); if (ret) { gf_msg (this->name, GF_LOG_WARNING, - GD_MSG_PMAP_REGISTRY_REMOVE_FAIL, 0, "Failed toi" + GD_MSG_PMAP_REGISTRY_REMOVE_FAIL, 0, "Failed to" "remove pmap registry for older signin for path" " %s", path); } diff --git a/xlators/mgmt/glusterd/src/glusterd-quotad-svc.c b/xlators/mgmt/glusterd/src/glusterd-quotad-svc.c index f3475a3f0e..3c457be96e 100644 --- a/xlators/mgmt/glusterd/src/glusterd-quotad-svc.c +++ b/xlators/mgmt/glusterd/src/glusterd-quotad-svc.c @@ -108,6 +108,9 @@ glusterd_quotadsvc_manager (glusterd_svc_t *svc, void *data, int flags) } } out: + if (ret) + gf_event (EVENT_SVC_MANAGER_FAILED, "svc_name=%s", svc->name); + gf_msg_debug (THIS->name, 0, "Returning %d", ret); return ret; diff --git a/xlators/mgmt/glusterd/src/glusterd-rebalance.c b/xlators/mgmt/glusterd/src/glusterd-rebalance.c index 35fa4627d0..86e21f396b 100644 --- a/xlators/mgmt/glusterd/src/glusterd-rebalance.c +++ b/xlators/mgmt/glusterd/src/glusterd-rebalance.c @@ -1124,10 +1124,13 @@ glusterd_defrag_event_notify_handle (dict_t *dict) ret = glusterd_defrag_volume_status_update (volinfo, dict); - if (ret) + if (ret) { gf_msg (this->name, GF_LOG_ERROR, 0, GD_MSG_DEFRAG_STATUS_UPDATE_FAIL, "Failed to update status"); + gf_event (EVENT_DEFRAG_STATUS_UPDATE_FAILED, "volume=%s", + volinfo->volname); + } out: return ret; diff --git a/xlators/mgmt/glusterd/src/glusterd-svc-helper.c b/xlators/mgmt/glusterd/src/glusterd-svc-helper.c index 44ee6d08d6..72f0092266 100644 --- a/xlators/mgmt/glusterd/src/glusterd-svc-helper.c +++ b/xlators/mgmt/glusterd/src/glusterd-svc-helper.c @@ -52,11 +52,13 @@ glusterd_svcs_reconfigure () ret = glusterd_bitdsvc_reconfigure (); if (ret) goto out; - ret = glusterd_scrubsvc_reconfigure (); if (ret) goto out; out: + if (ret) + gf_event (EVENT_SVC_RECONFIGURE_FAILED, ""); + return ret; } |
