Blob Blame History Raw
/*
   Copyright (c) 2010-2012 Red Hat, Inc. <http://www.redhat.com>
   This file is part of GlusterFS.

   This file is licensed to you under your choice of the GNU Lesser
   General Public License, version 3 or any later version (LGPLv3 or
   later), or the GNU General Public License, version 2 (GPLv2), in all
   cases as published by the Free Software Foundation.
*/

/* Widths of various columns in top read/write-perf output
 * Total width of top read/write-perf should be 80 chars
 * including one space between column
 */
#define VOL_TOP_PERF_FILENAME_DEF_WIDTH 47
#define VOL_TOP_PERF_FILENAME_ALT_WIDTH 44
#define VOL_TOP_PERF_SPEED_WIDTH 4
#define VOL_TOP_PERF_TIME_WIDTH 26

#define INDENT_MAIN_HEAD "%-25s %s "

/* Do not show estimates if greater than this number */
#define REBAL_ESTIMATE_SEC_UPPER_LIMIT (60 * 24 * 3600)
#define REBAL_ESTIMATE_START_TIME 600

#include "cli.h"
#include <glusterfs/compat-errno.h>
#include "cli-cmd.h"
#include <sys/uio.h>
#include <stdlib.h>
#include <sys/mount.h>
#include "cli1-xdr.h"
#include "xdr-generic.h"
#include "protocol-common.h"
#include "cli-mem-types.h"
#include <glusterfs/compat.h>
#include <glusterfs/upcall-utils.h>

#include <glusterfs/syscall.h>
#include "glusterfs3.h"
#include "portmap-xdr.h"
#include <glusterfs/byte-order.h>

#include "cli-quotad-client.h"
#include <glusterfs/run.h>
#include <glusterfs/quota-common-utils.h>
#include <glusterfs/events.h>

enum gf_task_types { GF_TASK_TYPE_REBALANCE, GF_TASK_TYPE_REMOVE_BRICK };

extern struct rpc_clnt *global_quotad_rpc;
extern rpc_clnt_prog_t cli_quotad_clnt;
extern rpc_clnt_prog_t *cli_rpc_prog;
extern int cli_op_ret;
extern int connected;

int32_t
gf_cli_remove_brick(call_frame_t *frame, xlator_t *this, void *data);

char *cli_vol_status_str[] = {
    "Created",
    "Started",
    "Stopped",
};

char *cli_vol_task_status_str[] = {"not started",
                                   "in progress",
                                   "stopped",
                                   "completed",
                                   "failed",
                                   "fix-layout in progress",
                                   "fix-layout stopped",
                                   "fix-layout completed",
                                   "fix-layout failed",
                                   "unknown"};

int32_t
gf_cli_snapshot(call_frame_t *frame, xlator_t *this, void *data);

int32_t
gf_cli_get_volume(call_frame_t *frame, xlator_t *this, void *data);

int
cli_to_glusterd(gf_cli_req *req, call_frame_t *frame, fop_cbk_fn_t cbkfn,
                xdrproc_t xdrproc, dict_t *dict, int procnum, xlator_t *this,
                rpc_clnt_prog_t *prog, struct iobref *iobref);

int
add_cli_cmd_timeout_to_dict(dict_t *dict);

rpc_clnt_prog_t cli_handshake_prog = {
    .progname = "cli handshake",
    .prognum = GLUSTER_HNDSK_PROGRAM,
    .progver = GLUSTER_HNDSK_VERSION,
};

rpc_clnt_prog_t cli_pmap_prog = {
    .progname = "cli portmap",
    .prognum = GLUSTER_PMAP_PROGRAM,
    .progver = GLUSTER_PMAP_VERSION,
};

static void
gf_free_xdr_cli_rsp(gf_cli_rsp rsp)
{
    if (rsp.dict.dict_val) {
        free(rsp.dict.dict_val);
    }
    if (rsp.op_errstr) {
        free(rsp.op_errstr);
    }
}

static void
gf_free_xdr_getspec_rsp(gf_getspec_rsp rsp)
{
    if (rsp.spec) {
        free(rsp.spec);
    }
    if (rsp.xdata.xdata_val) {
        free(rsp.xdata.xdata_val);
    }
}

static void
gf_free_xdr_fsm_log_rsp(gf1_cli_fsm_log_rsp rsp)
{
    if (rsp.op_errstr) {
        free(rsp.op_errstr);
    }
    if (rsp.fsm_log.fsm_log_val) {
        free(rsp.fsm_log.fsm_log_val);
    }
}

int
gf_cli_probe_cbk(struct rpc_req *req, struct iovec *iov, int count,
                 void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        // rsp.op_ret   = -1;
        // rsp.op_errno = EINVAL;
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to probe");

    if (rsp.op_errstr && (strlen(rsp.op_errstr) > 0)) {
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
        if (rsp.op_ret)
            gf_log("cli", GF_LOG_ERROR, "%s", msg);
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str(NULL, (rsp.op_ret) ? NULL : msg, rsp.op_ret,
                                 rsp.op_errno, (rsp.op_ret) ? msg : NULL);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (!rsp.op_ret)
        cli_out("peer probe: success. %s", msg);
    else
        cli_err("peer probe: failed: %s", msg);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_deprobe_cbk(struct rpc_req *req, struct iovec *iov, int count,
                   void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        // rsp.op_ret   = -1;
        // rsp.op_errno = EINVAL;
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to deprobe");

    if (rsp.op_ret) {
        if (strlen(rsp.op_errstr) > 0) {
            snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
            gf_log("cli", GF_LOG_ERROR, "%s", rsp.op_errstr);
        }
    } else {
        snprintf(msg, sizeof(msg), "success");
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str(NULL, (rsp.op_ret) ? NULL : msg, rsp.op_ret,
                                 rsp.op_errno, (rsp.op_ret) ? msg : NULL);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (!rsp.op_ret)
        cli_out("peer detach: %s", msg);
    else
        cli_err("peer detach: failed: %s", msg);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_output_peer_hostnames(dict_t *dict, int count, char *prefix)
{
    int ret = -1;
    char key[256] = {
        0,
    };
    int i = 0;
    char *hostname = NULL;

    cli_out("Other names:");
    /* Starting from friend.hostname1, as friend.hostname0 will be the same
     * as friend.hostname
     */
    for (i = 1; i < count; i++) {
        snprintf(key, sizeof(key), "%s.hostname%d", prefix, i);
        ret = dict_get_str(dict, key, &hostname);
        if (ret)
            break;
        cli_out("%s", hostname);
        hostname = NULL;
    }

    return ret;
}

int
gf_cli_output_peer_status(dict_t *dict, int count)
{
    int ret = -1;
    char *uuid_buf = NULL;
    char *hostname_buf = NULL;
    int32_t i = 1;
    char key[256] = {
        0,
    };
    char *state = NULL;
    int32_t connected = 0;
    char *connected_str = NULL;
    int hostname_count = 0;

    cli_out("Number of Peers: %d", count);
    i = 1;
    while (i <= count) {
        snprintf(key, 256, "friend%d.uuid", i);
        ret = dict_get_str(dict, key, &uuid_buf);
        if (ret)
            goto out;

        snprintf(key, 256, "friend%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname_buf);
        if (ret)
            goto out;

        snprintf(key, 256, "friend%d.connected", i);
        ret = dict_get_int32(dict, key, &connected);
        if (ret)
            goto out;
        if (connected)
            connected_str = "Connected";
        else
            connected_str = "Disconnected";

        snprintf(key, 256, "friend%d.state", i);
        ret = dict_get_str(dict, key, &state);
        if (ret)
            goto out;

        cli_out("\nHostname: %s\nUuid: %s\nState: %s (%s)", hostname_buf,
                uuid_buf, state, connected_str);

        snprintf(key, sizeof(key), "friend%d.hostname_count", i);
        ret = dict_get_int32(dict, key, &hostname_count);
        /* Print other addresses only if there are more than 1.
         */
        if ((ret == 0) && (hostname_count > 1)) {
            snprintf(key, sizeof(key), "friend%d", i);
            ret = gf_cli_output_peer_hostnames(dict, hostname_count, key);
            if (ret) {
                gf_log("cli", GF_LOG_WARNING,
                       "error outputting peer other names");
                goto out;
            }
        }
        i++;
    }

    ret = 0;
out:
    return ret;
}

int
gf_cli_output_pool_list(dict_t *dict, int count)
{
    int ret = -1;
    char *uuid_buf = NULL;
    char *hostname_buf = NULL;
    int32_t hostname_len = 8; /*min len 8 chars*/
    int32_t i = 1;
    char key[256] = {
        0,
    };
    int32_t connected = 0;
    char *connected_str = NULL;

    if (count <= 0)
        goto out;

    while (i <= count) {
        snprintf(key, 256, "friend%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname_buf);
        if (ret)
            goto out;

        ret = strlen(hostname_buf);
        if (ret > hostname_len)
            hostname_len = ret;

        i++;
    }

    cli_out("UUID\t\t\t\t\t%-*s\tState", hostname_len, "Hostname");

    i = 1;
    while (i <= count) {
        snprintf(key, 256, "friend%d.uuid", i);
        ret = dict_get_str(dict, key, &uuid_buf);
        if (ret)
            goto out;

        snprintf(key, 256, "friend%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname_buf);
        if (ret)
            goto out;

        snprintf(key, 256, "friend%d.connected", i);
        ret = dict_get_int32(dict, key, &connected);
        if (ret)
            goto out;
        if (connected)
            connected_str = "Connected";
        else
            connected_str = "Disconnected";

        cli_out("%s\t%-*s\t%s ", uuid_buf, hostname_len, hostname_buf,
                connected_str);
        i++;
    }

    ret = 0;
out:
    return ret;
}

/* function pointer for gf_cli_output_{pool_list,peer_status} */
typedef int (*cli_friend_output_fn)(dict_t *, int);

int
gf_cli_list_friends_cbk(struct rpc_req *req, struct iovec *iov, int count,
                        void *myframe)
{
    gf1_cli_peer_list_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char msg[1024] = {
        0,
    };
    char *cmd = NULL;
    cli_friend_output_fn friend_output_fn;
    call_frame_t *frame = NULL;
    unsigned long flags = 0;

    GF_ASSERT(myframe);

    frame = myframe;

    flags = (long)frame->local;

    if (flags == GF_CLI_LIST_POOL_NODES) {
        cmd = "pool list";
        friend_output_fn = &gf_cli_output_pool_list;
    } else {
        cmd = "peer status";
        friend_output_fn = &gf_cli_output_peer_status;
    }

    /* 'free' the flags set by gf_cli_list_friends */
    frame->local = NULL;

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_peer_list_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        // rsp.op_ret   = -1;
        // rsp.op_errno = EINVAL;
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to list: %d", rsp.op_ret);

    if (!rsp.op_ret) {
        if (!rsp.friends.friends_len) {
            snprintf(msg, sizeof(msg), "%s: No peers present", cmd);
            if (global_state->mode & GLUSTER_MODE_XML) {
                ret = cli_xml_output_peer_status(dict, rsp.op_ret, rsp.op_errno,
                                                 msg);
                if (ret)
                    gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
                goto out;
            }
            cli_err("%s", msg);
            ret = 0;
            goto out;
        }

        dict = dict_new();

        if (!dict) {
            ret = -1;
            goto out;
        }

        ret = dict_unserialize(rsp.friends.friends_val, rsp.friends.friends_len,
                               &dict);

        if (ret) {
            gf_log("", GF_LOG_ERROR, "Unable to allocate memory");
            goto out;
        }

        if (global_state->mode & GLUSTER_MODE_XML) {
            ret = cli_xml_output_peer_status(dict, rsp.op_ret, rsp.op_errno,
                                             msg);
            if (ret)
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
            goto out;
        }

        ret = dict_get_int32(dict, "count", &count);
        if (ret) {
            goto out;
        }

        ret = friend_output_fn(dict, count);
        if (ret) {
            goto out;
        }
    } else {
        if (global_state->mode & GLUSTER_MODE_XML) {
            ret = cli_xml_output_peer_status(dict, rsp.op_ret, rsp.op_errno,
                                             NULL);
            if (ret)
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        } else {
            ret = -1;
        }
        goto out;
    }

    ret = 0;

out:
    if (ret)
        cli_err("%s: failed", cmd);

    cli_cmd_broadcast_response(ret);

    if (dict)
        dict_unref(dict);

    if (rsp.friends.friends_val) {
        free(rsp.friends.friends_val);
    }
    return ret;
}

int
gf_cli_get_state_cbk(struct rpc_req *req, struct iovec *iov, int count,
                     void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char *daemon_name = NULL;
    char *ofilepath = NULL;

    GF_VALIDATE_OR_GOTO("cli", myframe, out);

    if (-1 == req->rpc_status) {
        goto out;
    }
    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret)
        goto out;

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            cli_err("Failed to get daemon state: %s", rsp.op_errstr);
        else
            cli_err(
                "Failed to get daemon state. Check glusterd"
                " log file for more details");
    } else {
        ret = dict_get_str(dict, "daemon", &daemon_name);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Couldn't get daemon name");

        ret = dict_get_str(dict, "ofilepath", &ofilepath);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Couldn't get filepath");

        if (daemon_name && ofilepath)
            cli_out("%s state dumped to %s", daemon_name, ofilepath);
    }

    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);

    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    return ret;
}

void
cli_out_options(char *substr, char *optstr, char *valstr)
{
    char *ptr1 = NULL;
    char *ptr2 = NULL;

    ptr1 = substr;
    ptr2 = optstr;

    while (ptr1) {
        /* Avoiding segmentation fault. */
        if (!ptr2)
            return;
        if (*ptr1 != *ptr2)
            break;
        ptr1++;
        ptr2++;
    }

    if (*ptr2 == '\0')
        return;
    cli_out("%s: %s", ptr2, valstr);
}

static int
_gf_cli_output_volinfo_opts(dict_t *d, char *k, data_t *v, void *tmp)
{
    int ret = 0;
    char *key = NULL;
    char *ptr = NULL;
    data_t *value = NULL;

    key = tmp;

    ptr = strstr(k, "option.");
    if (ptr) {
        value = v;
        if (!value) {
            ret = -1;
            goto out;
        }
        cli_out_options(key, k, v->data);
    }
out:
    return ret;
}

static int
print_brick_details(dict_t *dict, int volcount, int start_index, int end_index,
                    int replica_count)
{
    char key[1024] = {
        0,
    };
    int index = start_index;
    int isArbiter = 0;
    int ret = -1;
    char *brick = NULL;
#ifdef HAVE_BD_XLATOR
    char *caps = NULL;
#endif

    while (index <= end_index) {
        snprintf(key, 1024, "volume%d.brick%d", volcount, index);
        ret = dict_get_str(dict, key, &brick);
        if (ret)
            goto out;
        snprintf(key, sizeof(key), "volume%d.brick%d.isArbiter", volcount,
                 index);
        if (dict_get(dict, key))
            isArbiter = 1;
        else
            isArbiter = 0;

        if (isArbiter)
            cli_out("Brick%d: %s (arbiter)", index, brick);
        else
            cli_out("Brick%d: %s", index, brick);
#ifdef HAVE_BD_XLATOR
        snprintf(key, 1024, "volume%d.vg%d", volcount, index);
        ret = dict_get_str(dict, key, &caps);
        if (!ret)
            cli_out("Brick%d VG: %s", index, caps);
#endif
        index++;
    }
    ret = 0;
out:
    return ret;
}
void
gf_cli_print_number_of_bricks(int type, int brick_count, int dist_count,
                              int stripe_count, int replica_count,
                              int disperse_count, int redundancy_count,
                              int arbiter_count)
{
    if (type == GF_CLUSTER_TYPE_NONE || type == GF_CLUSTER_TYPE_TIER) {
        cli_out("Number of Bricks: %d", brick_count);
    } else if (type == GF_CLUSTER_TYPE_DISPERSE) {
        cli_out("Number of Bricks: %d x (%d + %d) = %d",
                (brick_count / dist_count), disperse_count - redundancy_count,
                redundancy_count, brick_count);
    } else {
        /* For both replicate and stripe, dist_count is
           good enough */
        if (arbiter_count == 0) {
            cli_out("Number of Bricks: %d x %d = %d",
                    (brick_count / dist_count), dist_count, brick_count);
        } else {
            cli_out("Number of Bricks: %d x (%d + %d) = %d",
                    (brick_count / dist_count), dist_count - arbiter_count,
                    arbiter_count, brick_count);
        }
    }
}

int
gf_cli_print_tier_info(dict_t *dict, int i, int brick_count)
{
    int hot_brick_count = -1;
    int cold_type = 0;
    int cold_brick_count = 0;
    int cold_replica_count = 0;
    int cold_arbiter_count = 0;
    int cold_disperse_count = 0;
    int cold_redundancy_count = 0;
    int cold_dist_count = 0;
    int hot_type = 0;
    int hot_replica_count = 0;
    int hot_dist_count = 0;
    int ret = -1;
    int vol_type = -1;
    char key[256] = {
        0,
    };

    GF_ASSERT(dict);

    snprintf(key, sizeof(key), "volume%d.cold_brick_count", i);
    ret = dict_get_int32(dict, key, &cold_brick_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.cold_type", i);
    ret = dict_get_int32(dict, key, &cold_type);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.cold_dist_count", i);
    ret = dict_get_int32(dict, key, &cold_dist_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.cold_replica_count", i);
    ret = dict_get_int32(dict, key, &cold_replica_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.cold_arbiter_count", i);
    ret = dict_get_int32(dict, key, &cold_arbiter_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.cold_disperse_count", i);
    ret = dict_get_int32(dict, key, &cold_disperse_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.cold_redundancy_count", i);
    ret = dict_get_int32(dict, key, &cold_redundancy_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.hot_brick_count", i);
    ret = dict_get_int32(dict, key, &hot_brick_count);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "volume%d.hot_type", i);
    ret = dict_get_int32(dict, key, &hot_type);
    if (ret)
        goto out;
    snprintf(key, sizeof(key), "volume%d.hot_replica_count", i);
    ret = dict_get_int32(dict, key, &hot_replica_count);
    if (ret)
        goto out;

    cli_out("Hot Tier :");
    hot_dist_count = (hot_replica_count ? hot_replica_count : 1);

    vol_type = get_vol_type(hot_type, hot_dist_count, hot_brick_count);
    cli_out("Hot Tier Type : %s", vol_type_str[vol_type]);

    gf_cli_print_number_of_bricks(hot_type, hot_brick_count, hot_dist_count, 0,
                                  hot_replica_count, 0, 0, 0);

    ret = print_brick_details(dict, i, 1, hot_brick_count, hot_replica_count);
    if (ret)
        goto out;

    cli_out("Cold Tier:");

    vol_type = get_vol_type(cold_type, cold_dist_count, cold_brick_count);
    cli_out("Cold Tier Type : %s", vol_type_str[vol_type]);

    gf_cli_print_number_of_bricks(cold_type, cold_brick_count, cold_dist_count,
                                  0, cold_replica_count, cold_disperse_count,
                                  cold_redundancy_count, cold_arbiter_count);

    ret = print_brick_details(dict, i, hot_brick_count + 1, brick_count,
                              cold_replica_count);
    if (ret)
        goto out;
out:
    return ret;
}

int
gf_cli_get_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                      void *myframe)
{
    int ret = -1;
    int opt_count = 0;
    int32_t i = 0;
    int32_t j = 1;
    int32_t status = 0;
    int32_t type = 0;
    int32_t brick_count = 0;
    int32_t dist_count = 0;
    int32_t stripe_count = 0;
    int32_t replica_count = 0;
    int32_t disperse_count = 0;
    int32_t redundancy_count = 0;
    int32_t arbiter_count = 0;
    int32_t snap_count = 0;
    int32_t vol_type = 0;
    int32_t transport = 0;
    char *volume_id_str = NULL;
    char *volname = NULL;
    dict_t *dict = NULL;
    cli_local_t *local = NULL;
    char key[1024] = {0};
    char err_str[2048] = {0};
    gf_cli_rsp rsp = {0};
    char *caps __attribute__((unused)) = NULL;
    int k __attribute__((unused)) = 0;
    call_frame_t *frame = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to get vol: %d", rsp.op_ret);

    if (!rsp.dict.dict_len) {
        if (global_state->mode & GLUSTER_MODE_XML)
            goto xml_output;
        cli_err("No volumes present");
        ret = 0;
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to allocate memory");
        goto out;
    }

    ret = dict_get_int32(dict, "count", &count);
    if (ret)
        goto out;

    if (!count) {
        switch (local->get_vol.flags) {
            case GF_CLI_GET_NEXT_VOLUME:
                GF_FREE(local->get_vol.volname);
                local->get_vol.volname = NULL;
                ret = 0;
                goto out;

            case GF_CLI_GET_VOLUME:
                snprintf(err_str, sizeof(err_str), "Volume %s does not exist",
                         local->get_vol.volname);
                ret = -1;
                if (!(global_state->mode & GLUSTER_MODE_XML))
                    goto out;
        }
    }

    if (rsp.op_ret) {
        if (global_state->mode & GLUSTER_MODE_XML)
            goto xml_output;
        ret = -1;
        goto out;
    }

xml_output:
    if (global_state->mode & GLUSTER_MODE_XML) {
        /* For GET_NEXT_VOLUME output is already begun in
         * and will also end in gf_cli_get_next_volume()
         */
        if (local->get_vol.flags == GF_CLI_GET_VOLUME) {
            ret = cli_xml_output_vol_info_begin(local, rsp.op_ret, rsp.op_errno,
                                                rsp.op_errstr);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
                goto out;
            }
        }

        if (dict) {
            ret = cli_xml_output_vol_info(local, dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
                goto out;
            }
        }

        if (local->get_vol.flags == GF_CLI_GET_VOLUME) {
            ret = cli_xml_output_vol_info_end(local);
            if (ret)
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        }
        goto out;
    }

    while (i < count) {
        cli_out(" ");
        snprintf(key, 256, "volume%d.name", i);
        ret = dict_get_str(dict, key, &volname);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.type", i);
        ret = dict_get_int32(dict, key, &type);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.status", i);
        ret = dict_get_int32(dict, key, &status);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.brick_count", i);
        ret = dict_get_int32(dict, key, &brick_count);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.dist_count", i);
        ret = dict_get_int32(dict, key, &dist_count);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.stripe_count", i);
        ret = dict_get_int32(dict, key, &stripe_count);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.replica_count", i);
        ret = dict_get_int32(dict, key, &replica_count);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.disperse_count", i);
        ret = dict_get_int32(dict, key, &disperse_count);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.redundancy_count", i);
        ret = dict_get_int32(dict, key, &redundancy_count);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "volume%d.arbiter_count", i);
        ret = dict_get_int32(dict, key, &arbiter_count);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.transport", i);
        ret = dict_get_int32(dict, key, &transport);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.volume_id", i);
        ret = dict_get_str(dict, key, &volume_id_str);
        if (ret)
            goto out;

        snprintf(key, 256, "volume%d.snap_count", i);
        ret = dict_get_int32(dict, key, &snap_count);
        if (ret)
            goto out;

        // Distributed (stripe/replicate/stripe-replica) setups
        vol_type = get_vol_type(type, dist_count, brick_count);

        cli_out("Volume Name: %s", volname);
        cli_out("Type: %s", vol_type_str[vol_type]);
        cli_out("Volume ID: %s", volume_id_str);
        cli_out("Status: %s", cli_vol_status_str[status]);
        cli_out("Snapshot Count: %d", snap_count);

#ifdef HAVE_BD_XLATOR
        k = 0;
        snprintf(key, sizeof(key), "volume%d.xlator%d", i, k);
        ret = dict_get_str(dict, key, &caps);
        if (ret)
            goto next;
        do {
            j = 0;
            cli_out("Xlator %d: %s", k + 1, caps);
            do {
                snprintf(key, sizeof(key), "volume%d.xlator%d.caps%d", i, k,
                         j++);
                ret = dict_get_str(dict, key, &caps);
                if (ret)
                    break;
                cli_out("Capability %d: %s", j, caps);
            } while (1);

            snprintf(key, sizeof(key), "volume%d.xlator%d", i, ++k);
            ret = dict_get_str(dict, key, &caps);
            if (ret)
                break;
        } while (1);

    next:
#endif
        gf_cli_print_number_of_bricks(
            type, brick_count, dist_count, stripe_count, replica_count,
            disperse_count, redundancy_count, arbiter_count);

        cli_out("Transport-type: %s",
                ((transport == 0) ? "tcp"
                                  : (transport == 1) ? "rdma" : "tcp,rdma"));
        j = 1;

        GF_FREE(local->get_vol.volname);
        local->get_vol.volname = gf_strdup(volname);

        if (type == GF_CLUSTER_TYPE_TIER) {
            ret = gf_cli_print_tier_info(dict, i, brick_count);
            if (ret)
                goto out;

        } else {
            cli_out("Bricks:");
            ret = print_brick_details(dict, i, j, brick_count, replica_count);
            if (ret)
                goto out;
        }

        snprintf(key, 256, "volume%d.opt_count", i);
        ret = dict_get_int32(dict, key, &opt_count);
        if (ret)
            goto out;

        if (!opt_count)
            goto out;

        cli_out("Options Reconfigured:");

        snprintf(key, 256, "volume%d.option.", i);

        ret = dict_foreach(dict, _gf_cli_output_volinfo_opts, key);
        if (ret)
            goto out;

        i++;
    }

    ret = 0;
out:
    if (ret)
        cli_err("%s", err_str);

    cli_cmd_broadcast_response(ret);

    if (dict)
        dict_unref(dict);

    gf_free_xdr_cli_rsp(rsp);

    gf_log("cli", GF_LOG_DEBUG, "Returning: %d", ret);
    return ret;
}

int
gf_cli_create_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                         void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    char *volname = NULL;
    dict_t *rsp_dict = NULL;
    call_frame_t *frame = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to create volume");

    ret = dict_get_str(local->dict, "volname", &volname);
    if (ret)
        goto out;

    if (global_state->mode & GLUSTER_MODE_XML) {
        if (rsp.op_ret == 0) {
            rsp_dict = dict_new();
            ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len,
                                   &rsp_dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization");
                goto out;
            }
        }

        ret = cli_xml_output_vol_create(rsp_dict, rsp.op_ret, rsp.op_errno,
                                        rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        cli_err("volume create: %s: failed: %s", volname, rsp.op_errstr);
    else if (rsp.op_ret)
        cli_err("volume create: %s: failed", volname);
    else
        cli_out(
            "volume create: %s: success: "
            "please start the volume to access data",
            volname);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (rsp_dict)
        dict_unref(rsp_dict);
    return ret;
}

int
gf_cli_delete_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                         void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    char *volname = NULL;
    call_frame_t *frame = NULL;
    dict_t *rsp_dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_str(local->dict, "volname", &volname);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "dict get failed");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to delete volume");

    if (global_state->mode & GLUSTER_MODE_XML) {
        if (rsp.op_ret == 0) {
            rsp_dict = dict_new();
            ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len,
                                   &rsp_dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization");
                goto out;
            }
        }

        ret = cli_xml_output_generic_volume("volDelete", rsp_dict, rsp.op_ret,
                                            rsp.op_errno, rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        cli_err("volume delete: %s: failed: %s", volname, rsp.op_errstr);
    else if (rsp.op_ret)
        cli_err("volume delete: %s: failed", volname);
    else
        cli_out("volume delete: %s: success", volname);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (rsp_dict)
        dict_unref(rsp_dict);
    gf_log("", GF_LOG_DEBUG, "Returning with %d", ret);
    return ret;
}

int
gf_cli3_1_uuid_get_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    char *uuid_str = NULL;
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    dict_t *dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    frame->local = NULL;

    gf_log("cli", GF_LOG_INFO, "Received resp to uuid get");

    dict = dict_new();
    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to unserialize "
               "response for uuid get");
        goto out;
    }

    ret = dict_get_str(dict, "uuid", &uuid_str);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to get uuid "
               "from dictionary");
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_dict("uuidGenerate", dict, rsp.op_ret,
                                  rsp.op_errno, rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, "") == 0)
            cli_err("Get uuid was unsuccessful");
        else
            cli_err("%s", rsp.op_errstr);

    } else {
        cli_out("UUID: %s", uuid_str);
    }
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    cli_local_wipe(local);
    gf_free_xdr_cli_rsp(rsp);

    if (dict)
        dict_unref(dict);

    gf_log("", GF_LOG_DEBUG, "Returning with %d", ret);
    return ret;
}

int
gf_cli3_1_uuid_reset_cbk(struct rpc_req *req, struct iovec *iov, int count,
                         void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    frame->local = NULL;

    gf_log("cli", GF_LOG_INFO, "Received resp to uuid reset");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_dict("uuidReset", NULL, rsp.op_ret, rsp.op_errno,
                                  rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        cli_err("%s", rsp.op_errstr);
    else
        cli_out("resetting the peer uuid has been %s",
                (rsp.op_ret) ? "unsuccessful" : "successful");
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    cli_local_wipe(local);
    gf_free_xdr_cli_rsp(rsp);

    gf_log("", GF_LOG_DEBUG, "Returning with %d", ret);
    return ret;
}

int
gf_cli_start_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                        void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    char *volname = NULL;
    call_frame_t *frame = NULL;
    dict_t *rsp_dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_str(local->dict, "volname", &volname);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "dict get failed");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to start volume");

    if (global_state->mode & GLUSTER_MODE_XML) {
        if (rsp.op_ret == 0) {
            rsp_dict = dict_new();
            ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len,
                                   &rsp_dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization");
                goto out;
            }
        }

        ret = cli_xml_output_generic_volume("volStart", rsp_dict, rsp.op_ret,
                                            rsp.op_errno, rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        cli_err("volume start: %s: failed: %s", volname, rsp.op_errstr);
    else if (rsp.op_ret)
        cli_err("volume start: %s: failed", volname);
    else
        cli_out("volume start: %s: success", volname);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (rsp_dict)
        dict_unref(rsp_dict);
    return ret;
}

int
gf_cli_stop_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    char *volname = NULL;
    call_frame_t *frame = NULL;
    dict_t *rsp_dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_str(local->dict, "volname", &volname);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Unable to get volname from dict");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to stop volume");

    if (global_state->mode & GLUSTER_MODE_XML) {
        if (rsp.op_ret == 0) {
            rsp_dict = dict_new();
            ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len,
                                   &rsp_dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization");
                goto out;
            }
        }

        ret = cli_xml_output_generic_volume("volStop", rsp_dict, rsp.op_ret,
                                            rsp.op_errno, rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        cli_err("volume stop: %s: failed: %s", volname, rsp.op_errstr);
    else if (rsp.op_ret)
        cli_err("volume stop: %s: failed", volname);
    else
        cli_out("volume stop: %s: success", volname);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (rsp_dict)
        dict_unref(rsp_dict);

    return ret;
}

int
gf_cli_print_rebalance_status(dict_t *dict, enum gf_task_types task_type,
                              gf_boolean_t is_tier)
{
    int ret = -1;
    int count = 0;
    int i = 1;
    char key[256] = {
        0,
    };
    gf_defrag_status_t status_rcd = GF_DEFRAG_STATUS_NOT_STARTED;
    uint64_t files = 0;
    uint64_t size = 0;
    uint64_t lookup = 0;
    char *node_name = NULL;
    uint64_t failures = 0;
    uint64_t skipped = 0;
    double elapsed = 0;
    char *status_str = NULL;
    char *size_str = NULL;
    int32_t hrs = 0;
    uint32_t min = 0;
    uint32_t sec = 0;
    gf_boolean_t down = _gf_false;
    gf_boolean_t fix_layout = _gf_false;
    uint64_t max_time = 0;
    uint64_t max_elapsed = 0;
    uint64_t time_left = 0;
    gf_boolean_t show_estimates = _gf_false;

    ret = dict_get_int32(dict, "count", &count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "count not set");
        goto out;
    }

    for (i = 1; i <= count; i++) {
        snprintf(key, sizeof(key), "status-%d", i);
        ret = dict_get_int32(dict, key, (int32_t *)&status_rcd);
        /* If information from a node is missing we should skip
         * the node and try to fetch information of other nodes.
         * If information is not found for all nodes, we should
         * error out.
         */
        if (!ret)
            break;
        if (ret && i == count) {
            gf_log("cli", GF_LOG_TRACE, "failed to get status");
            goto out;
        }
    }

    /* Fix layout will be sent to all nodes for the volume
       so every status should be of type
       GF_DEFRAG_STATUS_LAYOUT_FIX*
    */

    if ((task_type == GF_TASK_TYPE_REBALANCE) &&
        (status_rcd >= GF_DEFRAG_STATUS_LAYOUT_FIX_STARTED)) {
        fix_layout = _gf_true;
    }

    if (fix_layout) {
        cli_out("%35s %41s %27s", "Node", "status", "run time in h:m:s");
        cli_out("%35s %41s %27s", "---------", "-----------", "------------");
    } else {
        cli_out("%40s %16s %13s %13s %13s %13s %20s %18s", "Node",
                "Rebalanced-files", "size", "scanned", "failures", "skipped",
                "status",
                "run time in"
                " h:m:s");
        cli_out("%40s %16s %13s %13s %13s %13s %20s %18s", "---------",
                "-----------", "-----------", "-----------", "-----------",
                "-----------", "------------", "--------------");
    }

    for (i = 1; i <= count; i++) {
        /* Reset the variables to prevent carryover of values */
        node_name = NULL;
        files = 0;
        size = 0;
        lookup = 0;
        skipped = 0;
        status_str = NULL;
        elapsed = 0;
        time_left = 0;

        /* Check if status is NOT_STARTED, and continue early */
        snprintf(key, sizeof(key), "status-%d", i);

        ret = dict_get_int32(dict, key, (int32_t *)&status_rcd);
        if (ret == -ENOENT) {
            gf_log("cli", GF_LOG_TRACE, "count %d %d", count, i);
            gf_log("cli", GF_LOG_TRACE, "failed to get status");
            gf_log("cli", GF_LOG_ERROR,
                   "node down and has failed"
                   " to set dict");
            down = _gf_true;
            continue;
            /* skip this node if value not available*/
        } else if (ret) {
            gf_log("cli", GF_LOG_TRACE, "count %d %d", count, i);
            gf_log("cli", GF_LOG_TRACE, "failed to get status");
            continue;
            /* skip this node if value not available*/
        }

        if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd)
            continue;

        if (GF_DEFRAG_STATUS_STARTED == status_rcd)
            show_estimates = _gf_true;

        snprintf(key, 256, "node-name-%d", i);
        ret = dict_get_str(dict, key, &node_name);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get node-name");

        snprintf(key, sizeof(key), "files-%d", i);
        ret = dict_get_uint64(dict, key, &files);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get file count");

        snprintf(key, sizeof(key), "size-%d", i);
        ret = dict_get_uint64(dict, key, &size);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get size of xfer");

        snprintf(key, sizeof(key), "lookups-%d", i);
        ret = dict_get_uint64(dict, key, &lookup);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get lookedup file count");

        snprintf(key, sizeof(key), "failures-%d", i);
        ret = dict_get_uint64(dict, key, &failures);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get failures count");

        snprintf(key, sizeof(key), "skipped-%d", i);
        ret = dict_get_uint64(dict, key, &skipped);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get skipped count");

        /* For remove-brick include skipped count into failure count*/
        if (task_type != GF_TASK_TYPE_REBALANCE) {
            failures += skipped;
            skipped = 0;
        }

        snprintf(key, sizeof(key), "run-time-%d", i);
        ret = dict_get_double(dict, key, &elapsed);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get run-time");

        snprintf(key, sizeof(key), "time-left-%d", i);
        ret = dict_get_uint64(dict, key, &time_left);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get time left");

        if (elapsed > max_elapsed)
            max_elapsed = elapsed;

        if (time_left > max_time)
            max_time = time_left;

        /* Check for array bound */
        if (status_rcd >= GF_DEFRAG_STATUS_MAX)
            status_rcd = GF_DEFRAG_STATUS_MAX;

        status_str = cli_vol_task_status_str[status_rcd];
        size_str = gf_uint64_2human_readable(size);
        hrs = elapsed / 3600;
        min = ((uint64_t)elapsed % 3600) / 60;
        sec = ((uint64_t)elapsed % 3600) % 60;

        if (fix_layout) {
            cli_out("%35s %50s %8d:%d:%d", node_name, status_str, hrs, min,
                    sec);
        } else {
            if (size_str) {
                cli_out("%40s %16" PRIu64
                        " %13s"
                        " %13" PRIu64 " %13" PRIu64 " %13" PRIu64
                        " %20s "
                        "%8d:%02d:%02d",
                        node_name, files, size_str, lookup, failures, skipped,
                        status_str, hrs, min, sec);
            } else {
                cli_out("%40s %16" PRIu64 " %13" PRIu64 " %13" PRIu64
                        " %13" PRIu64 " %13" PRIu64
                        " %20s"
                        " %8d:%02d:%02d",
                        node_name, files, size, lookup, failures, skipped,
                        status_str, hrs, min, sec);
            }
        }
        GF_FREE(size_str);
    }
    if (is_tier && down)
        cli_out(
            "WARNING: glusterd might be down on one or more nodes."
            " Please check the nodes that are down using \'gluster"
            " peer status\' and start the glusterd on those nodes,"
            " else tier detach commit might fail!");

    /* Max time will be non-zero if rebalance is still running */
    if (max_time) {
        hrs = max_time / 3600;
        min = (max_time % 3600) / 60;
        sec = (max_time % 3600) % 60;

        if (hrs < REBAL_ESTIMATE_SEC_UPPER_LIMIT) {
            cli_out(
                "Estimated time left for rebalance to "
                "complete : %8d:%02d:%02d",
                hrs, min, sec);
        } else {
            cli_out(
                "Estimated time left for rebalance to "
                "complete : > 2 months. Please try again "
                "later.");
        }
    } else {
        /* Rebalance will return 0 if it could not calculate the
         * estimates or if it is complete.
         */
        if (!show_estimates) {
            goto out;
        }
        if (max_elapsed <= REBAL_ESTIMATE_START_TIME) {
            cli_out(
                "The estimated time for rebalance to complete "
                "will be unavailable for the first 10 "
                "minutes.");
        } else {
            cli_out(
                "Rebalance estimated time unavailable. Please "
                "try again later.");
        }
    }
out:
    return ret;
}

int
gf_cli_print_tier_status(dict_t *dict, enum gf_task_types task_type)
{
    int ret = -1;
    int count = 0;
    int i = 1;
    uint64_t promoted = 0;
    uint64_t demoted = 0;
    char key[256] = {
        0,
    };
    char *node_name = NULL;
    gf_defrag_status_t status_rcd = GF_DEFRAG_STATUS_NOT_STARTED;
    char *status_str = NULL;
    gf_boolean_t down = _gf_false;
    double elapsed = 0;
    int hrs = 0;
    int min = 0;
    int sec = 0;

    ret = dict_get_int32(dict, "count", &count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "count not set");
        goto out;
    }

    cli_out("%-20s %-20s %-20s %-20s %-20s", "Node", "Promoted files",
            "Demoted files", "Status", "run time in h:m:s");
    cli_out("%-20s %-20s %-20s %-20s %-20s", "---------", "---------",
            "---------", "---------", "---------");

    for (i = 1; i <= count; i++) {
        /* Reset the variables to prevent carryover of values */
        node_name = NULL;
        promoted = 0;
        demoted = 0;

        /* Check if status is NOT_STARTED, and continue early */
        snprintf(key, sizeof(key), "status-%d", i);

        ret = dict_get_int32(dict, key, (int32_t *)&status_rcd);
        if (ret == -ENOENT) {
            gf_log("cli", GF_LOG_TRACE,
                   "count: %d, %d,"
                   "failed to get status",
                   count, i);
            gf_log("cli", GF_LOG_ERROR,
                   "node down and has failed"
                   " to set dict");
            down = _gf_true;
            continue;
            /*skipping this node as value unavailable*/
        } else if (ret) {
            gf_log("cli", GF_LOG_TRACE,
                   "count: %d, %d,"
                   "failed to get status",
                   count, i);
            continue;
        }

        if (GF_DEFRAG_STATUS_NOT_STARTED == status_rcd)
            continue;

        snprintf(key, sizeof(key), "node-name-%d", i);
        ret = dict_get_str(dict, key, &node_name);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get node-name");

        snprintf(key, sizeof(key), "promoted-%d", i);
        ret = dict_get_uint64(dict, key, &promoted);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get promoted count");

        snprintf(key, sizeof(key), "demoted-%d", i);
        ret = dict_get_uint64(dict, key, &demoted);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get demoted count");

        snprintf(key, sizeof(key), "run-time-%d", i);
        ret = dict_get_double(dict, key, &elapsed);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get run-time");

        /* Check for array bound */
        if (status_rcd >= GF_DEFRAG_STATUS_MAX)
            status_rcd = GF_DEFRAG_STATUS_MAX;

        hrs = elapsed / 3600;
        min = ((int)elapsed % 3600) / 60;
        sec = ((int)elapsed % 3600) % 60;

        status_str = cli_vol_task_status_str[status_rcd];
        cli_out("%-20s %-20" PRIu64 " %-20" PRIu64
                " %-20s"
                " %d:%d:%d",
                node_name, promoted, demoted, status_str, hrs, min, sec);
    }
    if (down)
        cli_out(
            "WARNING: glusterd might be down on one or more nodes."
            " Please check the nodes that are down using \'gluster"
            " peer status\' and start the glusterd on those nodes.");
out:
    return ret;
}

int
gf_cli_defrag_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                         void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    cli_local_t *local = NULL;
    char *volname = NULL;
    call_frame_t *frame = NULL;
    int cmd = 0;
    int ret = -1;
    dict_t *dict = NULL;
    char msg[1024] = {
        0,
    };
    char *task_id_str = NULL;

    if (-1 == req->rpc_status) {
        goto out;
    }

    GF_ASSERT(myframe);

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_str(local->dict, "volname", &volname);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "Failed to get volname");
        goto out;
    }

    ret = dict_get_int32(local->dict, "rebalance-command", (int32_t *)&cmd);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get command");
        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
        if (ret < 0) {
            gf_log("glusterd", GF_LOG_ERROR,
                   "failed to "
                   "unserialize req-buffer to dictionary");
            goto out;
        }
    }

    if (!((cmd == GF_DEFRAG_CMD_STOP) || (cmd == GF_DEFRAG_CMD_STATUS) ||
          (cmd == GF_DEFRAG_CMD_STATUS_TIER)) &&
        !(global_state->mode & GLUSTER_MODE_XML)) {
        ret = dict_get_str(dict, GF_REBALANCE_TID_KEY, &task_id_str);
        if (ret) {
            gf_log("cli", GF_LOG_WARNING, "failed to get %s from dict",
                   GF_REBALANCE_TID_KEY);
        }
        if (rsp.op_ret && strcmp(rsp.op_errstr, "")) {
            snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
        } else {
            if (!rsp.op_ret) {
                /* append errstr in the cli msg for successful
                 * case since unlock failures can be highlighted
                 * event though rebalance command was successful
                 */
                if (cmd == GF_DEFRAG_CMD_START_TIER) {
                    snprintf(msg, sizeof(msg),
                             "Tier "
                             "start is successful on %s.",
                             volname);
                } else if (cmd == GF_DEFRAG_CMD_STOP_TIER) {
                    snprintf(msg, sizeof(msg),
                             "Tier "
                             "daemon stopped "
                             "on %s.",
                             volname);
                } else {
                    snprintf(msg, sizeof(msg),
                             "Rebalance on %s has been "
                             "started successfully. Use "
                             "rebalance status command to"
                             " check status of the "
                             "rebalance process.\nID: %s",
                             volname, task_id_str);
                }
            } else {
                snprintf(msg, sizeof(msg),
                         "Starting rebalance on volume %s has "
                         "been unsuccessful.",
                         volname);
            }
        }
        goto done;
    }

    if (cmd == GF_DEFRAG_CMD_STOP) {
        if (rsp.op_ret == -1) {
            if (strcmp(rsp.op_errstr, ""))
                snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
            else
                snprintf(msg, sizeof(msg), "rebalance volume %s stop failed",
                         volname);
            goto done;
        } else {
            /* append errstr in the cli msg for successful case
             * since unlock failures can be highlighted event though
             * rebalance command was successful */
            snprintf(msg, sizeof(msg),
                     "rebalance process may be in the middle of a "
                     "file migration.\nThe process will be fully "
                     "stopped once the migration of the file is "
                     "complete.\nPlease check rebalance process "
                     "for completion before doing any further "
                     "brick related tasks on the volume.\n%s",
                     rsp.op_errstr);
        }
    }
    if (cmd == GF_DEFRAG_CMD_STATUS || cmd == GF_DEFRAG_CMD_STATUS_TIER) {
        if (rsp.op_ret == -1) {
            if (strcmp(rsp.op_errstr, ""))
                snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
            else
                snprintf(msg, sizeof(msg),
                         "Failed to get the status of "
                         "rebalance process");
            goto done;
        } else {
            snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
        }
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_rebalance(cmd, dict, rsp.op_ret, rsp.op_errno,
                                           rsp.op_errstr);
        goto out;
    }

    if (cmd == GF_DEFRAG_CMD_STATUS_TIER)
        ret = gf_cli_print_tier_status(dict, GF_TASK_TYPE_REBALANCE);
    else if (cmd == GF_DEFRAG_CMD_DETACH_STATUS)
        ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REBALANCE,
                                            _gf_true);
    else
        ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REBALANCE,
                                            _gf_false);

    if (ret)
        gf_log("cli", GF_LOG_ERROR, "Failed to print rebalance status");

done:
    if (global_state->mode & GLUSTER_MODE_XML)
        cli_xml_output_str("volRebalance", msg, rsp.op_ret, rsp.op_errno,
                           rsp.op_errstr);
    else {
        if (rsp.op_ret)

            if (cmd == GF_DEFRAG_CMD_START_TIER ||
                cmd == GF_DEFRAG_CMD_STATUS_TIER) {
                cli_err(
                    "Tiering Migration Functionality: %s:"
                    " failed%s%s",
                    volname, strlen(msg) ? ": " : "", msg);
            } else
                cli_err("volume rebalance: %s: failed%s%s", volname,
                        strlen(msg) ? ": " : "", msg);
        else if (cmd == GF_DEFRAG_CMD_START_TIER ||
                 cmd == GF_DEFRAG_CMD_STATUS_TIER) {
            cli_out(
                "Tiering Migration Functionality: %s:"
                " success%s%s",
                volname, strlen(msg) ? ": " : "", msg);
        } else
            cli_out("volume rebalance: %s: success%s%s", volname,
                    strlen(msg) ? ": " : "", msg);
    }
    ret = rsp.op_ret;

out:
    gf_free_xdr_cli_rsp(rsp);
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    return ret;
}

int
gf_cli_rename_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                         void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to probe");
    snprintf(msg, sizeof(msg), "Rename volume %s",
             (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volRename", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume rename: failed");
    else
        cli_out("volume rename: success");

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_reset_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                        void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to reset");

    if (strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "reset volume %s",
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volReset", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume reset: failed: %s", msg);
    else
        cli_out("volume reset: success: %s", msg);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_ganesha_cbk(struct rpc_req *req, struct iovec *iov, int count,
                   void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to ganesha");

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret)
        goto out;

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            cli_err("nfs-ganesha: failed: %s", rsp.op_errstr);
        else
            cli_err("nfs-ganesha: failed");
    }

    else {
        cli_out("nfs-ganesha : success ");
    }

    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    return ret;
}

char *
is_server_debug_xlator(void *myframe)
{
    call_frame_t *frame = NULL;
    cli_local_t *local = NULL;
    char **words = NULL;
    char *key = NULL;
    char *value = NULL;
    char *debug_xlator = NULL;

    frame = myframe;
    local = frame->local;
    words = (char **)local->words;

    while (*words != NULL) {
        if (strstr(*words, "trace") == NULL &&
            strstr(*words, "error-gen") == NULL) {
            words++;
            continue;
        }

        key = *words;
        words++;
        value = *words;
        if (value == NULL)
            break;
        if (strstr(value, "client")) {
            words++;
            continue;
        } else {
            if (!(strstr(value, "posix") || strstr(value, "acl") ||
                  strstr(value, "locks") || strstr(value, "io-threads") ||
                  strstr(value, "marker") || strstr(value, "index"))) {
                words++;
                continue;
            } else {
                debug_xlator = gf_strdup(key);
                break;
            }
        }
    }

    return debug_xlator;
}

int
gf_cli_set_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                      void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char *help_str = NULL;
    char msg[1024] = {
        0,
    };
    char *debug_xlator = NULL;
    char tmp_str[512] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to set");

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "failed to unserialize volume set respone dict");
        goto out;
    }

    /* For brick processes graph change does not happen on the fly.
     * The process has to be restarted. So this is a check from the
     * volume set option such that if debug xlators such as trace/errorgen
     * are provided in the set command, warn the user.
     */
    debug_xlator = is_server_debug_xlator(myframe);

    if (dict_get_str(dict, "help-str", &help_str) && !msg[0])
        snprintf(msg, sizeof(msg), "Set volume %s",
                 (rsp.op_ret) ? "unsuccessful" : "successful");
    if (rsp.op_ret == 0 && debug_xlator) {
        snprintf(tmp_str, sizeof(tmp_str),
                 "\n%s translator has been "
                 "added to the server volume file. Please restart the"
                 " volume for enabling the translator",
                 debug_xlator);
    }

    if ((global_state->mode & GLUSTER_MODE_XML) && (help_str == NULL)) {
        ret = cli_xml_output_str("volSet", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            cli_err("volume set: failed: %s", rsp.op_errstr);
        else
            cli_err("volume set: failed");
    } else {
        if (help_str == NULL) {
            if (debug_xlator == NULL)
                cli_out("volume set: success");
            else
                cli_out("volume set: success%s", tmp_str);
        } else {
            cli_out("%s", help_str);
        }
    }

    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);
    GF_FREE(debug_xlator);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_add_tier_brick_cbk(struct rpc_req *req, struct iovec *iov, int count,
                          void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_VALIDATE_OR_GOTO("cli", myframe, out);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to attach tier");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "Attach tier %s",
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volAttachTier", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume attach-tier: failed: %s", msg);
    else
        cli_out("volume attach-tier: success");
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_attach_tier_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_VALIDATE_OR_GOTO("cli", myframe, out);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to attach tier");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "Attach tier %s",
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volAttachTier", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume attach-tier: failed: %s", msg);
    else
        cli_out("volume attach-tier: success");
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_remove_tier_brick_cbk(struct rpc_req *req, struct iovec *iov, int count,
                             void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };
    char *cmd_str = "unknown";
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    char *task_id_str = NULL;
    dict_t *rsp_dict = NULL;
    int32_t command = 0;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_int32(local->dict, "command", &command);
    if (ret) {
        gf_log("", GF_LOG_ERROR, "failed to get command");
        goto out;
    }

    if (rsp.dict.dict_len) {
        rsp_dict = dict_new();
        if (!rsp_dict) {
            ret = -1;
            goto out;
        }

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Failed to unserialize rsp_dict");
            goto out;
        }
    }

    switch (command) {
        case GF_DEFRAG_CMD_DETACH_START:
            cmd_str = "start";

            ret = dict_get_str(rsp_dict, GF_REMOVE_BRICK_TID_KEY, &task_id_str);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "remove-brick-id is not present in dict");
            }
            break;
        case GF_DEFRAG_CMD_DETACH_COMMIT:
            cmd_str = "commit";
            break;
        case GF_DEFRAG_CMD_DETACH_COMMIT_FORCE:
            cmd_str = "commit force";
            break;
        case GF_DEFRAG_CMD_DETACH_STOP:
            cmd_str = "stop";
            break;
        case GF_DEFRAG_CMD_DETACH_STATUS:
            cmd_str = "status";
            break;

        default:
            cmd_str = "unknown";
            break;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to detach tier");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "Detach tier %s %s", cmd_str,
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    ret = rsp.op_ret;
    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_remove_brick_detach_tier(
            _gf_true, rsp_dict, rsp.op_ret, rsp.op_errno, msg, "volDetachTier");

        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    } else {
        if (rsp.op_ret) {
            if (strcmp(rsp.op_errstr, ""))
                snprintf(msg, sizeof(msg),
                         "volume tier "
                         "detach %s: failed: %s",
                         cmd_str, rsp.op_errstr);
            else
                snprintf(msg, sizeof(msg),
                         "volume tier "
                         "detach %s: failed",
                         cmd_str);

            cli_err("%s", msg);
            goto out;

        } else {
            cli_out("volume detach tier %s: success", cmd_str);
            if (GF_DEFRAG_CMD_DETACH_START == command && task_id_str != NULL)
                cli_out("ID: %s", task_id_str);
            if (GF_DEFRAG_CMD_DETACH_COMMIT == command)
                cli_out(
                    "Check the detached bricks to ensure "
                    "all files are migrated.\nIf files "
                    "with data are found on the brick "
                    "path, copy them via a gluster mount "
                    "point before re-purposing the "
                    "removed brick. ");
        }
    }
    if (command == GF_DEFRAG_CMD_DETACH_STOP ||
        command == GF_DEFRAG_CMD_DETACH_STATUS)
        ret = gf_cli_print_rebalance_status(rsp_dict, GF_TASK_TYPE_REMOVE_BRICK,
                                            _gf_true);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to print remove-brick "
               "rebalance status");
        goto out;
    }

    if ((command == GF_DEFRAG_CMD_DETACH_STOP) && (rsp.op_ret == 0)) {
        cli_out(
            "'detach tier' process may be in the middle of a "
            "file migration.\nThe process will be fully stopped "
            "once the migration of the file is complete.\nPlease "
            "check detach tier process for completion before "
            "doing any further brick related tasks on the "
            "volume.");
    }
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (rsp_dict)
        dict_unref(rsp_dict);
    return ret;
}

int
gf_cli_detach_tier_status_cbk(struct rpc_req *req, struct iovec *iov, int count,
                              void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char msg[1024] = {
        0,
    };
    int32_t command = 0;
    gf1_op_commands cmd = GF_OP_CMD_NONE;
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    char *cmd_str = "unknown";

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_int32(local->dict, "command", &command);
    if (ret)
        goto out;

    cmd = command;

    switch (cmd) {
        case GF_OP_CMD_STOP_DETACH_TIER:
            cmd_str = "stop";
            break;
        case GF_OP_CMD_STATUS:
            cmd_str = "status";
            break;
        default:
            break;
    }

    ret = rsp.op_ret;
    if (rsp.op_ret == -1) {
        if (strcmp(rsp.op_errstr, ""))
            snprintf(msg, sizeof(msg),
                     "volume tier detach %s: "
                     "failed: %s",
                     cmd_str, rsp.op_errstr);
        else
            snprintf(msg, sizeof(msg),
                     "volume tier detach %s: "
                     "failed",
                     cmd_str);

        if (global_state->mode & GLUSTER_MODE_XML)
            goto xml_output;

        cli_err("%s", msg);
        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
        if (ret < 0) {
            strncpy(msg,
                    "failed to unserialize req-buffer to "
                    "dictionary",
                    sizeof(msg));

            if (global_state->mode & GLUSTER_MODE_XML) {
                rsp.op_ret = -1;
                goto xml_output;
            }

            gf_log("cli", GF_LOG_ERROR, "%s", msg);
            goto out;
        }
    }
xml_output:
    if (global_state->mode & GLUSTER_MODE_XML) {
        if (strcmp(rsp.op_errstr, "")) {
            ret = cli_xml_output_vol_remove_brick_detach_tier(
                _gf_true, dict, rsp.op_ret, rsp.op_errno, rsp.op_errstr,
                "volDetachTier");
        } else {
            ret = cli_xml_output_vol_remove_brick_detach_tier(
                _gf_true, dict, rsp.op_ret, rsp.op_errno, msg, "volDetachTier");
        }
        goto out;
    }

    ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REMOVE_BRICK,
                                        _gf_true);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to print remove-brick "
               "rebalance status");
        goto out;
    }

    if ((cmd == GF_OP_CMD_STOP_DETACH_TIER) && (rsp.op_ret == 0)) {
        cli_out(
            "'detach tier' process may be in the middle of a "
            "file migration.\nThe process will be fully stopped "
            "once the migration of the file is complete.\nPlease "
            "check detach tier process for completion before "
            "doing any further brick related tasks on the "
            "volume.");
    }

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_add_brick_cbk(struct rpc_req *req, struct iovec *iov, int count,
                     void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to add brick");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "Add Brick %s",
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volAddBrick", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume add-brick: failed: %s", rsp.op_errstr);
    else
        cli_out("volume add-brick: success");
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli3_remove_brick_status_cbk(struct rpc_req *req, struct iovec *iov,
                                int count, void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char msg[1024] = {
        0,
    };
    int32_t command = 0;
    gf1_op_commands cmd = GF_OP_CMD_NONE;
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    char *cmd_str = "unknown";

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_int32(local->dict, "command", &command);
    if (ret)
        goto out;

    cmd = command;

    switch (cmd) {
        case GF_OP_CMD_STOP:
            cmd_str = "stop";
            break;
        case GF_OP_CMD_STATUS:
            cmd_str = "status";
            break;
        default:
            break;
    }

    ret = rsp.op_ret;
    if (rsp.op_ret == -1) {
        if (strcmp(rsp.op_errstr, ""))
            snprintf(msg, sizeof(msg),
                     "volume remove-brick %s: "
                     "failed: %s",
                     cmd_str, rsp.op_errstr);
        else
            snprintf(msg, sizeof(msg),
                     "volume remove-brick %s: "
                     "failed",
                     cmd_str);

        if (global_state->mode & GLUSTER_MODE_XML)
            goto xml_output;

        cli_err("%s", msg);
        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
        if (ret < 0) {
            strncpy(msg,
                    "failed to unserialize req-buffer to "
                    "dictionary",
                    sizeof(msg));

            if (global_state->mode & GLUSTER_MODE_XML) {
                rsp.op_ret = -1;
                goto xml_output;
            }

            gf_log("cli", GF_LOG_ERROR, "%s", msg);
            goto out;
        }
    }

xml_output:
    if (global_state->mode & GLUSTER_MODE_XML) {
        if (strcmp(rsp.op_errstr, "")) {
            ret = cli_xml_output_vol_remove_brick_detach_tier(
                _gf_true, dict, rsp.op_ret, rsp.op_errno, rsp.op_errstr,
                "volRemoveBrick");
        } else {
            ret = cli_xml_output_vol_remove_brick_detach_tier(
                _gf_true, dict, rsp.op_ret, rsp.op_errno, msg,
                "volRemoveBrick");
        }
        goto out;
    }

    ret = gf_cli_print_rebalance_status(dict, GF_TASK_TYPE_REMOVE_BRICK,
                                        _gf_false);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to print remove-brick "
               "rebalance status");
        goto out;
    }

    if ((cmd == GF_OP_CMD_STOP) && (rsp.op_ret == 0)) {
        cli_out(
            "'remove-brick' process may be in the middle of a "
            "file migration.\nThe process will be fully stopped "
            "once the migration of the file is complete.\nPlease "
            "check remove-brick process for completion before "
            "doing any further brick related tasks on the "
            "volume.");
    }

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_remove_brick_cbk(struct rpc_req *req, struct iovec *iov, int count,
                        void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };
    gf1_op_commands cmd = GF_OP_CMD_NONE;
    char *cmd_str = "unknown";
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    char *task_id_str = NULL;
    dict_t *rsp_dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_int32(local->dict, "command", (int32_t *)&cmd);
    if (ret) {
        gf_log("", GF_LOG_ERROR, "failed to get command");
        goto out;
    }

    if (rsp.dict.dict_len) {
        rsp_dict = dict_new();
        if (!rsp_dict) {
            ret = -1;
            goto out;
        }

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Failed to unserialize rsp_dict");
            goto out;
        }
    }

    switch (cmd) {
        case GF_OP_CMD_DETACH_START:
        case GF_OP_CMD_START:
            cmd_str = "start";

            ret = dict_get_str(rsp_dict, GF_REMOVE_BRICK_TID_KEY, &task_id_str);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "remove-brick-id is not present in dict");
            }
            break;
        case GF_OP_CMD_COMMIT:
            cmd_str = "commit";
            break;
        case GF_OP_CMD_COMMIT_FORCE:
            cmd_str = "commit force";
            break;
        default:
            cmd_str = "unknown";
            break;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to remove brick");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "Remove Brick %s %s", cmd_str,
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_remove_brick_detach_tier(
            _gf_false, rsp_dict, rsp.op_ret, rsp.op_errno, msg,
            "volRemoveBrick");
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret) {
        cli_err("volume remove-brick %s: failed: %s", cmd_str, msg);
    } else {
        cli_out("volume remove-brick %s: success", cmd_str);
        if (GF_OP_CMD_START == cmd && task_id_str != NULL)
            cli_out("ID: %s", task_id_str);
        if (GF_OP_CMD_COMMIT == cmd)
            cli_out(
                "Check the removed bricks to ensure all files "
                "are migrated.\nIf files with data are "
                "found on the brick path, copy them via a "
                "gluster mount point before re-purposing the "
                "removed brick. ");
    }

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (rsp_dict)
        dict_unref(rsp_dict);

    return ret;
}

int
gf_cli_reset_brick_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    char *rb_operation_str = NULL;
    dict_t *rsp_dict = NULL;
    char msg[1024] = {
        0,
    };
    char *reset_op = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_str(local->dict, "operation", &reset_op);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "dict_get on operation failed");
        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        rsp_dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict);
        if (ret < 0) {
            gf_log(frame->this->name, GF_LOG_ERROR,
                   "failed to "
                   "unserialize rsp buffer to dictionary");
            goto out;
        }
    }

    if (strcmp(reset_op, "GF_RESET_OP_START") &&
        strcmp(reset_op, "GF_RESET_OP_COMMIT") &&
        strcmp(reset_op, "GF_RESET_OP_COMMIT_FORCE")) {
        rb_operation_str = gf_strdup("Unknown operation");
        ret = -1;
        goto out;
    }

    if (rsp.op_ret && (strcmp(rsp.op_errstr, ""))) {
        rb_operation_str = gf_strdup(rsp.op_errstr);
    } else {
        if (!strcmp(reset_op, "GF_RESET_OP_START")) {
            if (rsp.op_ret)
                rb_operation_str = gf_strdup(
                    "reset-brick "
                    "start "
                    "operation "
                    "failed");
            else
                rb_operation_str = gf_strdup(
                    "reset-brick "
                    "start "
                    "operation "
                    "successful");
        } else if (!strcmp(reset_op, "GF_RESET_OP_COMMIT")) {
            if (rsp.op_ret)
                rb_operation_str = gf_strdup(
                    "reset-brick "
                    "commit "
                    "operation "
                    "failed");
            else
                rb_operation_str = gf_strdup(
                    "reset-brick "
                    "commit "
                    "operation "
                    "successful");
        } else if (!strcmp(reset_op, "GF_RESET_OP_COMMIT_FORCE")) {
            if (rsp.op_ret)
                rb_operation_str = gf_strdup(
                    "reset-brick "
                    "commit "
                    "force operation "
                    "failed");
            else
                rb_operation_str = gf_strdup(
                    "reset-brick "
                    "commit "
                    "force operation "
                    "successful");
        }
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to reset brick");
    snprintf(msg, sizeof(msg), "%s",
             rb_operation_str ? rb_operation_str : "Unknown operation");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_replace_brick(rsp_dict, rsp.op_ret,
                                               rsp.op_errno, msg);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume reset-brick: failed: %s", msg);
    else
        cli_out("volume reset-brick: success: %s", msg);
    ret = rsp.op_ret;

out:
    if (frame)
        frame->local = NULL;

    if (local)
        cli_local_wipe(local);

    if (rb_operation_str)
        GF_FREE(rb_operation_str);

    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    if (rsp_dict)
        dict_unref(rsp_dict);

    return ret;
}
int
gf_cli_replace_brick_cbk(struct rpc_req *req, struct iovec *iov, int count,
                         void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    call_frame_t *frame = NULL;
    char *rb_operation_str = NULL;
    dict_t *rsp_dict = NULL;
    char msg[1024] = {
        0,
    };
    char *replace_op = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_str(local->dict, "operation", &replace_op);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "dict_get on operation failed");
        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        rsp_dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &rsp_dict);
        if (ret < 0) {
            gf_log(frame->this->name, GF_LOG_ERROR,
                   "failed to "
                   "unserialize rsp buffer to dictionary");
            goto out;
        }
    }

    if (!strcmp(replace_op, "GF_REPLACE_OP_COMMIT_FORCE")) {
        if (rsp.op_ret || ret)
            rb_operation_str = gf_strdup(
                "replace-brick commit "
                "force operation failed");
        else
            rb_operation_str = gf_strdup(
                "replace-brick commit "
                "force operation "
                "successful");
    } else {
        gf_log(frame->this->name, GF_LOG_DEBUG, "Unknown operation");
    }

    if (rsp.op_ret && (strcmp(rsp.op_errstr, ""))) {
        rb_operation_str = gf_strdup(rsp.op_errstr);
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to replace brick");
    snprintf(msg, sizeof(msg), "%s",
             rb_operation_str ? rb_operation_str : "Unknown operation");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_replace_brick(rsp_dict, rsp.op_ret,
                                               rsp.op_errno, msg);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume replace-brick: failed: %s", msg);
    else
        cli_out("volume replace-brick: success: %s", msg);
    ret = rsp.op_ret;

out:
    if (frame)
        frame->local = NULL;

    if (local)
        cli_local_wipe(local);

    if (rb_operation_str)
        GF_FREE(rb_operation_str);

    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    if (rsp_dict)
        dict_unref(rsp_dict);

    return ret;
}

static int
gf_cli_log_rotate_cbk(struct rpc_req *req, struct iovec *iov, int count,
                      void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to log rotate");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "log rotate %s",
                 (rsp.op_ret) ? "unsuccessful" : "successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volLogRotate", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume log-rotate: failed: %s", msg);
    else
        cli_out("volume log-rotate: success");
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    return ret;
}

static int
gf_cli_sync_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to sync");

    if (rsp.op_ret && strcmp(rsp.op_errstr, ""))
        snprintf(msg, sizeof(msg), "volume sync: failed: %s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "volume sync: %s",
                 (rsp.op_ret) ? "failed" : "success");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volSync", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("%s", msg);
    else
        cli_out("%s", msg);
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

static int
print_quota_list_usage_output(cli_local_t *local, char *path, int64_t avail,
                              char *sl_str, quota_limits_t *limits,
                              quota_meta_t *used_space, gf_boolean_t sl,
                              gf_boolean_t hl, double sl_num,
                              gf_boolean_t limit_set)
{
    int32_t ret = -1;
    char *used_str = NULL;
    char *avail_str = NULL;
    char *hl_str = NULL;
    char *sl_val = NULL;

    used_str = gf_uint64_2human_readable(used_space->size);

    if (limit_set) {
        hl_str = gf_uint64_2human_readable(limits->hl);
        avail_str = gf_uint64_2human_readable(avail);

        sl_val = gf_uint64_2human_readable(sl_num);
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_quota_xml_output(local, path, limits->hl, sl_str, sl_num,
                                   used_space->size, avail, sl ? "Yes" : "No",
                                   hl ? "Yes" : "No", limit_set);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to "
                   "output in xml format for quota "
                   "list command");
        }
        goto out;
    }

    if (limit_set) {
        if (!used_str) {
            cli_out("%-40s %7s %7s(%s) %8" PRIu64 "%9" PRIu64
                    ""
                    "%15s %18s",
                    path, hl_str, sl_str, sl_val, used_space->size, avail,
                    sl ? "Yes" : "No", hl ? "Yes" : "No");
        } else {
            cli_out("%-40s %7s %7s(%s) %8s %7s %15s %20s", path, hl_str, sl_str,
                    sl_val, used_str, avail_str, sl ? "Yes" : "No",
                    hl ? "Yes" : "No");
        }
    } else {
        cli_out("%-36s %10s %10s %14s %9s %15s %18s", path, "N/A", "N/A",
                used_str, "N/A", "N/A", "N/A");
    }

    ret = 0;
out:
    GF_FREE(hl_str);
    GF_FREE(used_str);
    GF_FREE(avail_str);
    GF_FREE(sl_val);

    return ret;
}

static int
print_quota_list_object_output(cli_local_t *local, char *path, int64_t avail,
                               char *sl_str, quota_limits_t *limits,
                               quota_meta_t *used_space, gf_boolean_t sl,
                               gf_boolean_t hl, double sl_num,
                               gf_boolean_t limit_set)
{
    int32_t ret = -1;
    int64_t sl_val = sl_num;

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_quota_object_xml_output(local, path, sl_str, sl_val, limits,
                                          used_space, avail, sl ? "Yes" : "No",
                                          hl ? "Yes" : "No", limit_set);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to "
                   "output in xml format for quota "
                   "list command");
        }
        goto out;
    }

    if (limit_set) {
        cli_out("%-40s %9" PRIu64 " %9s(%" PRId64 ") %10" PRIu64
                ""
                "%10" PRIu64 " %11" PRIu64 " %15s %20s",
                path, limits->hl, sl_str, sl_val, used_space->file_count,
                used_space->dir_count, avail, sl ? "Yes" : "No",
                hl ? "Yes" : "No");
    } else {
        cli_out("%-40s %9s %9s %10" PRIu64 " %10" PRIu64 " %11s %15s %20s",
                path, "N/A", "N/A", used_space->file_count,
                used_space->dir_count, "N/A", "N/A", "N/A");
    }
    ret = 0;

out:

    return ret;
}

static int
print_quota_list_output(cli_local_t *local, char *path, char *default_sl,
                        quota_limits_t *limits, quota_meta_t *used_space,
                        int type, gf_boolean_t limit_set)
{
    int64_t avail = 0;
    char percent_str[20] = {0};
    char *sl_final = NULL;
    int ret = -1;
    double sl_num = 0;
    gf_boolean_t sl = _gf_false;
    gf_boolean_t hl = _gf_false;
    int64_t used_size = 0;

    GF_ASSERT(local);
    GF_ASSERT(path);

    if (limit_set) {
        if (limits->sl < 0) {
            ret = gf_string2percent(default_sl, &sl_num);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "could not convert default soft limit"
                       " to percent");
                goto out;
            }
            sl_num = (sl_num * limits->hl) / 100;
            sl_final = default_sl;
        } else {
            sl_num = (limits->sl * limits->hl) / 100;
            snprintf(percent_str, sizeof(percent_str), "%" PRIu64 "%%",
                     limits->sl);
            sl_final = percent_str;
        }
        if (type == GF_QUOTA_OPTION_TYPE_LIST)
            used_size = used_space->size;
        else
            used_size = used_space->file_count + used_space->dir_count;

        if (limits->hl > used_size) {
            avail = limits->hl - used_size;
            hl = _gf_false;
            if (used_size > sl_num)
                sl = _gf_true;
            else
                sl = _gf_false;
        } else {
            avail = 0;
            hl = sl = _gf_true;
        }
    }

    if (type == GF_QUOTA_OPTION_TYPE_LIST)
        ret = print_quota_list_usage_output(local, path, avail, sl_final,
                                            limits, used_space, sl, hl, sl_num,
                                            limit_set);
    else
        ret = print_quota_list_object_output(local, path, avail, sl_final,
                                             limits, used_space, sl, hl, sl_num,
                                             limit_set);
out:
    return ret;
}

static int
print_quota_list_from_mountdir(cli_local_t *local, char *mountdir,
                               char *default_sl, char *path, int type)
{
    int ret = -1;
    ssize_t xattr_size = 0;
    quota_limits_t limits = {
        0,
    };
    quota_meta_t used_space = {
        0,
    };
    char *key = NULL;
    gf_boolean_t limit_set = _gf_true;

    GF_ASSERT(local);
    GF_ASSERT(mountdir);
    GF_ASSERT(path);

    if (type == GF_QUOTA_OPTION_TYPE_LIST)
        key = QUOTA_LIMIT_KEY;
    else
        key = QUOTA_LIMIT_OBJECTS_KEY;

    ret = sys_lgetxattr(mountdir, key, (void *)&limits, sizeof(limits));
    if (ret < 0) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to get the xattr %s "
               "on %s. Reason : %s",
               key, mountdir, strerror(errno));

        switch (errno) {
#if defined(ENODATA)
            case ENODATA:
#endif
#if defined(ENOATTR) && (ENOATTR != ENODATA)
            case ENOATTR:
#endif
                /* If it's an ENOATTR, quota/inode-quota is
                 * configured(limit is set at least for one directory).
                 * The user is trying to issue 'list/list-objects'
                 * command for a directory on which quota limit is
                 * not set and we are showing the used-space in case
                 * of list-usage and showing (dir_count, file_count)
                 * in case of list-objects. Other labels are
                 * shown "N/A".
                 */

                limit_set = _gf_false;
                goto enoattr;
                break;

            default:
                cli_err("%-40s %s", path, strerror(errno));
                break;
        }

        goto out;
    }

    limits.hl = ntoh64(limits.hl);
    limits.sl = ntoh64(limits.sl);

enoattr:
    xattr_size = sys_lgetxattr(mountdir, QUOTA_SIZE_KEY, NULL, 0);
    if (xattr_size < (sizeof(int64_t) * 2) &&
        type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS) {
        ret = -1;

        /* This can happen when glusterfs is upgraded from 3.6 to 3.7
         * and the xattr healing is not completed.
         */
    } else if (xattr_size > (sizeof(int64_t) * 2)) {
        ret = sys_lgetxattr(mountdir, QUOTA_SIZE_KEY, &used_space,
                            sizeof(used_space));
    } else if (xattr_size > 0) {
        /* This is for compatibility.
         * Older version had only file usage
         */
        ret = sys_lgetxattr(mountdir, QUOTA_SIZE_KEY, &(used_space.size),
                            sizeof(used_space.size));
        used_space.file_count = 0;
        used_space.dir_count = 0;
    } else {
        ret = -1;
    }

    if (ret < 0) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to get quota size "
               "on path %s: %s",
               mountdir, strerror(errno));
        print_quota_list_empty(path, type);
        goto out;
    }

    used_space.size = ntoh64(used_space.size);
    used_space.file_count = ntoh64(used_space.file_count);
    used_space.dir_count = ntoh64(used_space.dir_count);

    ret = print_quota_list_output(local, path, default_sl, &limits, &used_space,
                                  type, limit_set);
out:
    return ret;
}

int
gluster_remove_auxiliary_mount(char *volname)
{
    int ret = -1;
    char mountdir[PATH_MAX] = {
        0,
    };
    xlator_t *this = NULL;

    this = THIS;
    GF_ASSERT(this);

    GLUSTERD_GET_QUOTA_LIST_MOUNT_PATH(mountdir, volname, "/");
    ret = gf_umount_lazy(this->name, mountdir, 1);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "umount on %s failed, "
               "reason : %s",
               mountdir, strerror(errno));
    }

    return ret;
}

int
gf_cli_print_limit_list_from_dict(cli_local_t *local, char *volname,
                                  dict_t *dict, char *default_sl, int count,
                                  int op_ret, int op_errno, char *op_errstr)
{
    int ret = -1;
    int i = 0;
    char key[1024] = {
        0,
    };
    char mountdir[PATH_MAX] = {
        0,
    };
    char *path = NULL;
    int type = -1;

    if (!dict || count <= 0)
        goto out;

    ret = dict_get_int32(dict, "type", &type);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get quota type");
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_quota_limit_list_begin(local, op_ret, op_errno,
                                                        op_errstr);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Error outputting xml begin");
            goto out;
        }
    } else {
        print_quota_list_header(type);
    }

    while (count--) {
        snprintf(key, sizeof(key), "path%d", i++);

        ret = dict_get_str(dict, key, &path);
        if (ret < 0) {
            gf_log("cli", GF_LOG_DEBUG,
                   "Path not present in limit"
                   " list");
            continue;
        }

        ret = gf_canonicalize_path(path);
        if (ret)
            goto out;
        GLUSTERD_GET_QUOTA_LIST_MOUNT_PATH(mountdir, volname, path);
        ret = print_quota_list_from_mountdir(local, mountdir, default_sl, path,
                                             type);
    }

out:
    return ret;
}

int
print_quota_list_from_quotad(call_frame_t *frame, dict_t *rsp_dict)
{
    char *path = NULL;
    char *default_sl = NULL;
    int ret = -1;
    cli_local_t *local = NULL;
    dict_t *gd_rsp_dict = NULL;
    quota_meta_t used_space = {
        0,
    };
    quota_limits_t limits = {
        0,
    };
    quota_limits_t *size_limits = NULL;
    int32_t type = 0;
    int32_t success_count = 0;

    GF_ASSERT(frame);

    local = frame->local;
    gd_rsp_dict = local->dict;

    ret = dict_get_int32(rsp_dict, "type", &type);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get type");
        goto out;
    }

    ret = dict_get_str(rsp_dict, GET_ANCESTRY_PATH_KEY, &path);
    if (ret) {
        gf_log("cli", GF_LOG_WARNING,
               "path key is not present "
               "in dict");
        goto out;
    }

    ret = dict_get_str(gd_rsp_dict, "default-soft-limit", &default_sl);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "failed to "
               "get default soft limit");
        goto out;
    }

    if (type == GF_QUOTA_OPTION_TYPE_LIST) {
        ret = dict_get_bin(rsp_dict, QUOTA_LIMIT_KEY, (void **)&size_limits);
        if (ret) {
            gf_log("cli", GF_LOG_WARNING, "limit key not present in dict on %s",
                   path);
            goto out;
        }
    } else {
        ret = dict_get_bin(rsp_dict, QUOTA_LIMIT_OBJECTS_KEY,
                           (void **)&size_limits);
        if (ret) {
            gf_log("cli", GF_LOG_WARNING,
                   "object limit key not present in dict on %s", path);
            goto out;
        }
    }

    limits.hl = ntoh64(size_limits->hl);
    limits.sl = ntoh64(size_limits->sl);

    if (type == GF_QUOTA_OPTION_TYPE_LIST)
        ret = quota_dict_get_meta(rsp_dict, QUOTA_SIZE_KEY,
                                  SLEN(QUOTA_SIZE_KEY), &used_space);
    else
        ret = quota_dict_get_inode_meta(rsp_dict, QUOTA_SIZE_KEY,
                                        SLEN(QUOTA_SIZE_KEY), &used_space);

    if (ret < 0) {
        gf_log("cli", GF_LOG_WARNING, "size key not present in dict");
        print_quota_list_empty(path, type);
        goto out;
    }

    LOCK(&local->lock);
    {
        ret = dict_get_int32(gd_rsp_dict, "quota-list-success-count",
                             &success_count);
        if (ret)
            success_count = 0;

        ret = dict_set_int32(gd_rsp_dict, "quota-list-success-count",
                             success_count + 1);
    }
    UNLOCK(&local->lock);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to set "
               "quota-list-success-count in dict");
        goto out;
    }

    if (success_count == 0) {
        if (!(global_state->mode & GLUSTER_MODE_XML)) {
            print_quota_list_header(type);
        } else {
            ret = cli_xml_output_vol_quota_limit_list_begin(local, 0, 0, NULL);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Error in "
                       "printing xml output");
                goto out;
            }
        }
    }

    ret = print_quota_list_output(local, path, default_sl, &limits, &used_space,
                                  type, _gf_true);
out:
    return ret;
}

void *
cli_cmd_broadcast_response_detached(void *opaque)
{
    int32_t ret = 0;

    ret = (intptr_t)opaque;
    cli_cmd_broadcast_response(ret);

    return NULL;
}

int32_t
cli_quota_compare_path(struct list_head *list1, struct list_head *list2)
{
    struct list_node *node1 = NULL;
    struct list_node *node2 = NULL;
    dict_t *dict1 = NULL;
    dict_t *dict2 = NULL;
    char *path1 = NULL;
    char *path2 = NULL;
    int ret = 0;

    node1 = list_entry(list1, struct list_node, list);
    node2 = list_entry(list2, struct list_node, list);

    dict1 = node1->ptr;
    dict2 = node2->ptr;

    ret = dict_get_str(dict1, GET_ANCESTRY_PATH_KEY, &path1);
    if (ret < 0)
        return 0;

    ret = dict_get_str(dict2, GET_ANCESTRY_PATH_KEY, &path2);
    if (ret < 0)
        return 0;

    return strcmp(path1, path2);
}

int
cli_quotad_getlimit_cbk(struct rpc_req *req, struct iovec *iov, int count,
                        void *myframe)
{
    /*TODO: we need to gather the path, hard-limit, soft-limit and used space*/
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    struct list_node *node = NULL;
    struct list_node *tmpnode = NULL;
    call_frame_t *frame = NULL;
    cli_local_t *local = NULL;
    int32_t list_count = 0;
    pthread_t th_id = {
        0,
    };
    int32_t max_count = 0;

    GF_ASSERT(myframe);

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    LOCK(&local->lock);
    {
        ret = dict_get_int32(local->dict, "quota-list-count", &list_count);
        if (ret)
            list_count = 0;

        list_count++;
        ret = dict_set_int32(local->dict, "quota-list-count", list_count);
    }
    UNLOCK(&local->lock);

    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to set "
               "quota-list-count in dict");
        goto out;
    }

    if (-1 == req->rpc_status) {
        if (list_count == 0)
            cli_err(
                "Connection failed. Please check if quota "
                "daemon is operational.");
        ret = -1;
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret) {
        ret = -1;
        if (strcmp(rsp.op_errstr, ""))
            cli_err("quota command failed : %s", rsp.op_errstr);
        else
            cli_err("quota command : failed");
        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
        if (ret < 0) {
            gf_log("cli", GF_LOG_ERROR,
                   "failed to "
                   "unserialize req-buffer to dictionary");
            goto out;
        }

        node = list_node_add_order(dict, &local->dict_list,
                                   cli_quota_compare_path);
        if (node == NULL) {
            gf_log("cli", GF_LOG_ERROR, "failed to add node to the list");
            dict_unref(dict);
            ret = -1;
            goto out;
        }
    }

    ret = dict_get_int32(local->dict, "max_count", &max_count);
    if (ret < 0) {
        gf_log("cli", GF_LOG_ERROR, "failed to get max_count");
        goto out;
    }

    if (list_count == max_count) {
        list_for_each_entry_safe(node, tmpnode, &local->dict_list, list)
        {
            dict = node->ptr;
            print_quota_list_from_quotad(frame, dict);
            list_node_del(node);
            dict_unref(dict);
        }
    }

out:
    /* Bad Fix: CLI holds the lock to process a command.
     * When processing quota list command, below sequence of steps executed
     * in the same thread and causing deadlock
     *
     * 1) CLI holds the lock
     * 2) Send rpc_clnt_submit request to quotad for quota usage
     * 3) If quotad is down, rpc_clnt_submit invokes cbk function with error
     * 4) cbk function cli_quotad_getlimit_cbk invokes
     *    cli_cmd_broadcast_response which tries to hold lock to broadcast
     *    the results and hangs, because same thread has already holding
     *    the lock
     *
     * Broadcasting response in a separate thread which is not a
     * good fix. This needs to be re-visted with better solution
     */
    if (ret == -1) {
        ret = pthread_create(&th_id, NULL, cli_cmd_broadcast_response_detached,
                             (void *)-1);
        if (ret)
            gf_log("cli", GF_LOG_ERROR,
                   "pthread_create failed: "
                   "%s",
                   strerror(errno));
    } else {
        cli_cmd_broadcast_response(ret);
    }
    gf_free_xdr_cli_rsp(rsp);

    return ret;
}

int
cli_quotad_getlimit(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;
    ret = add_cli_cmd_timeout_to_dict(dict);

    ret = dict_allocate_and_serialize(dict, &req.dict.dict_val,
                                      &req.dict.dict_len);
    if (ret < 0) {
        gf_log(this->name, GF_LOG_ERROR, "failed to serialize the data");

        goto out;
    }

    ret = cli_cmd_submit(global_quotad_rpc, &req, frame, &cli_quotad_clnt,
                         GF_AGGREGATOR_GETLIMIT, NULL, this,
                         cli_quotad_getlimit_cbk, (xdrproc_t)xdr_gf_cli_req);

out:
    GF_FREE(req.dict.dict_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

void
gf_cli_quota_list(cli_local_t *local, char *volname, dict_t *dict,
                  char *default_sl, int count, int op_ret, int op_errno,
                  char *op_errstr)
{
    GF_VALIDATE_OR_GOTO("cli", volname, out);

    if (!connected)
        goto out;

    if (count > 0)
        gf_cli_print_limit_list_from_dict(local, volname, dict, default_sl,
                                          count, op_ret, op_errno, op_errstr);
out:
    return;
}

int
gf_cli_quota_cbk(struct rpc_req *req, struct iovec *iov, int count,
                 void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char *volname = NULL;
    int32_t type = 0;
    call_frame_t *frame = NULL;
    char *default_sl = NULL;
    cli_local_t *local = NULL;
    char *default_sl_dup = NULL;
    int32_t entry_count = 0;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret) {
        ret = -1;
        if (global_state->mode & GLUSTER_MODE_XML)
            goto xml_output;

        if (strcmp(rsp.op_errstr, "")) {
            cli_err("quota command failed : %s", rsp.op_errstr);
            if (rsp.op_ret == -ENOENT)
                cli_err(
                    "please enter the path relative to "
                    "the volume");
        } else {
            cli_err("quota command : failed");
        }

        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        dict = dict_new();

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
        if (ret < 0) {
            gf_log("cli", GF_LOG_ERROR,
                   "failed to "
                   "unserialize req-buffer to dictionary");
            goto out;
        }
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to quota command");

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        gf_log(frame->this->name, GF_LOG_ERROR, "failed to get volname");

    ret = dict_get_str(dict, "default-soft-limit", &default_sl);
    if (ret)
        gf_log(frame->this->name, GF_LOG_TRACE,
               "failed to get "
               "default soft limit");

    // default-soft-limit is part of rsp_dict only iff we sent
    // GLUSTER_CLI_QUOTA with type being GF_QUOTA_OPTION_TYPE_LIST
    if (default_sl) {
        default_sl_dup = gf_strdup(default_sl);
        if (!default_sl_dup) {
            ret = -1;
            goto out;
        }
        ret = dict_set_dynstr(local->dict, "default-soft-limit",
                              default_sl_dup);
        if (ret) {
            gf_log(frame->this->name, GF_LOG_TRACE,
                   "failed to set default soft limit");
            GF_FREE(default_sl_dup);
        }
    }

    ret = dict_get_int32(dict, "type", &type);
    if (ret)
        gf_log(frame->this->name, GF_LOG_TRACE, "failed to get type");

    ret = dict_get_int32(dict, "count", &entry_count);
    if (ret)
        gf_log(frame->this->name, GF_LOG_TRACE, "failed to get count");

    if ((type == GF_QUOTA_OPTION_TYPE_LIST) ||
        (type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)) {
        gf_cli_quota_list(local, volname, dict, default_sl, entry_count,
                          rsp.op_ret, rsp.op_errno, rsp.op_errstr);

        if (global_state->mode & GLUSTER_MODE_XML) {
            ret = cli_xml_output_vol_quota_limit_list_end(local);
            if (ret < 0) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR,
                       "Error in printing"
                       " xml output");
            }
            goto out;
        }
    }

xml_output:

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volQuota", NULL, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (!rsp.op_ret && type != GF_QUOTA_OPTION_TYPE_LIST &&
        type != GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)
        cli_out("volume quota : success");

    ret = rsp.op_ret;
out:

    if ((type == GF_QUOTA_OPTION_TYPE_LIST) ||
        (type == GF_QUOTA_OPTION_TYPE_LIST_OBJECTS)) {
        gluster_remove_auxiliary_mount(volname);
    }

    cli_cmd_broadcast_response(ret);
    if (dict)
        dict_unref(dict);

    gf_free_xdr_cli_rsp(rsp);

    return ret;
}

int
gf_cli_getspec_cbk(struct rpc_req *req, struct iovec *iov, int count,
                   void *myframe)
{
    gf_getspec_rsp rsp = {
        0,
    };
    int ret = -1;
    char *spec = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_getspec_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret == -1) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "getspec failed");
        ret = -1;
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to getspec");

    spec = GF_MALLOC(rsp.op_ret + 1, cli_mt_char);
    if (!spec) {
        gf_log("", GF_LOG_ERROR, "out of memory");
        ret = -1;
        goto out;
    }
    memcpy(spec, rsp.spec, rsp.op_ret);
    spec[rsp.op_ret] = '\0';
    cli_out("%s", spec);
    GF_FREE(spec);

    ret = 0;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_getspec_rsp(rsp);
    return ret;
}

int
gf_cli_pmap_b2p_cbk(struct rpc_req *req, struct iovec *iov, int count,
                    void *myframe)
{
    pmap_port_by_brick_rsp rsp = {
        0,
    };
    int ret = -1;
    char *spec = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_pmap_port_by_brick_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret == -1) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "pump_b2p failed");
        ret = -1;
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to pmap b2p");

    cli_out("%d", rsp.port);
    GF_FREE(spec);

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    return ret;
}

int32_t
gf_cli_probe(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {
        {
            0,
        },
    };
    int ret = 0;
    dict_t *dict = NULL;
    int port = 0;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_int32(dict, "port", &port);
    if (ret) {
        ret = dict_set_int32(dict, "port", CLI_GLUSTERD_PORT);
        if (ret)
            goto out;
    }

    ret = cli_to_glusterd(&req, frame, gf_cli_probe_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_PROBE,
                          this, cli_rpc_prog, NULL);

out:
    GF_FREE(req.dict.dict_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int32_t
gf_cli_deprobe(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {
        {
            0,
        },
    };
    int ret = 0;
    dict_t *dict = NULL;
    int port = 0;
    int flags = 0;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;
    ret = dict_get_int32(dict, "port", &port);
    if (ret) {
        ret = dict_set_int32(dict, "port", CLI_GLUSTERD_PORT);
        if (ret)
            goto out;
    }

    ret = dict_get_int32(dict, "flags", &flags);
    if (ret) {
        ret = dict_set_int32(dict, "flags", 0);
        if (ret)
            goto out;
    }

    ret = cli_to_glusterd(&req, frame, gf_cli_deprobe_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_DEPROBE,
                          this, cli_rpc_prog, NULL);

out:
    GF_FREE(req.dict.dict_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int32_t
gf_cli_list_friends(call_frame_t *frame, xlator_t *this, void *data)
{
    gf1_cli_peer_list_req req = {
        0,
    };
    int ret = 0;
    unsigned long flags = 0;

    if (!frame || !this) {
        ret = -1;
        goto out;
    }

    GF_ASSERT(frame->local == NULL);

    flags = (long)data;
    req.flags = flags;
    frame->local = (void *)flags;
    ret = cli_cmd_submit(
        NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_LIST_FRIENDS, NULL, this,
        gf_cli_list_friends_cbk, (xdrproc_t)xdr_gf1_cli_peer_list_req);

out:
    if (ret && frame) {
        /*
         * If everything goes fine, gf_cli_list_friends_cbk()
         * [invoked through cli_cmd_submit()]resets the
         * frame->local to NULL. In case cli_cmd_submit()
         * fails in between, RESET frame->local here.
         */
        frame->local = NULL;
    }
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

int32_t
gf_cli_get_state(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {
        {
            0,
        },
    };
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_get_state_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_GET_STATE, this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_get_next_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = 0;
    cli_cmd_volume_get_ctx_t *ctx = NULL;
    cli_local_t *local = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    ctx = data;
    local = frame->local;

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_info_begin(local, 0, 0, "");
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
            goto out;
        }
    }

    ret = gf_cli_get_volume(frame, this, data);

    if (!local || !local->get_vol.volname) {
        if ((global_state->mode & GLUSTER_MODE_XML))
            goto end_xml;

        cli_err("No volumes present");
        goto out;
    }

    ctx->volname = local->get_vol.volname;

    while (ctx->volname) {
        ret = gf_cli_get_volume(frame, this, ctx);
        if (ret)
            goto out;
        ctx->volname = local->get_vol.volname;
    }

end_xml:
    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_info_end(local);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
    }

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

int32_t
gf_cli_get_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    cli_cmd_volume_get_ctx_t *ctx = NULL;
    dict_t *dict = NULL;
    int32_t flags = 0;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    ctx = data;

    dict = dict_new();
    if (!dict) {
        gf_log(THIS->name, GF_LOG_ERROR, "Failed to create the dict");
        ret = -1;
        goto out;
    }

    if (ctx->volname) {
        ret = dict_set_str(dict, "volname", ctx->volname);
        if (ret)
            goto out;
    }

    flags = ctx->flags;
    ret = dict_set_int32(dict, "flags", flags);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "failed to set flags");
        goto out;
    }

    ret = dict_allocate_and_serialize(dict, &req.dict.dict_val,
                                      &req.dict.dict_len);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "failed to serialize dict");
        goto out;
    }

    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog,
                         GLUSTER_CLI_GET_VOLUME, NULL, this,
                         gf_cli_get_volume_cbk, (xdrproc_t)xdr_gf_cli_req);

out:
    if (dict)
        dict_unref(dict);

    GF_FREE(req.dict.dict_val);

    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

int32_t
gf_cli3_1_uuid_get(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;
    ret = cli_to_glusterd(&req, frame, gf_cli3_1_uuid_get_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_UUID_GET,
                          this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli3_1_uuid_reset(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;
    ret = cli_to_glusterd(&req, frame, gf_cli3_1_uuid_reset_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_UUID_RESET, this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_create_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_create_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_CREATE_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_delete_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_delete_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_DELETE_VOLUME, this, cli_rpc_prog, NULL);

out:
    GF_FREE(req.dict.dict_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int32_t
gf_cli_start_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_start_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_START_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_stop_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = data;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_stop_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_STOP_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_defrag_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_defrag_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_DEFRAG_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_rename_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_allocate_and_serialize(dict, &req.dict.dict_val,
                                      &req.dict.dict_len);
    if (ret < 0) {
        gf_log(this->name, GF_LOG_ERROR, "failed to serialize the data");

        goto out;
    }

    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog,
                         GLUSTER_CLI_RENAME_VOLUME, NULL, this,
                         gf_cli_rename_volume_cbk, (xdrproc_t)xdr_gf_cli_req);

out:
    GF_FREE(req.dict.dict_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int32_t
gf_cli_reset_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_reset_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_RESET_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_ganesha(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_ganesha_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_GANESHA,
                          this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int32_t
gf_cli_set_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_set_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_SET_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_add_brick(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;
    char *volname = NULL;
    int32_t count = 0;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_str(dict, "volname", &volname);

    if (ret)
        goto out;

    ret = dict_get_int32(dict, "count", &count);
    if (ret)
        goto out;

    ret = cli_to_glusterd(&req, frame, gf_cli_add_brick_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_ADD_BRICK, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_tier(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = 0;
    int32_t command = 0;
    gf_cli_req req = {{
        0,
    }};
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }
    dict = data;

    ret = dict_get_int32(dict, "rebalance-command", &command);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get rebalance-command");
        goto out;
    }

    ret = cli_to_glusterd(&req, frame, gf_cli_defrag_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_TIER,
                          this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_add_tier_brick(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_add_tier_brick_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_ADD_TIER_BRICK, this, cli_rpc_prog, NULL);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to send request to "
               "glusterd");
        goto out;
    }

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_attach_tier(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;
    dict_t *newdict = NULL;
    char *tierwords[] = {"volume", "tier", "", "start", NULL};
    const char **words = (const char **)tierwords;
    char *volname = NULL;
    cli_local_t *local = NULL;
    cli_local_t *oldlocal = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_attach_tier_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_ATTACH_TIER, this, cli_rpc_prog, NULL);
    if (ret)
        goto out;
    ret = dict_get_str(dict, "volname", &volname);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get volume name");
        goto notify_cli;
    }

    words[2] = volname;
    ret = cli_cmd_volume_old_tier_parse((const char **)words, 4, &newdict);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to parse tier start "
               "command");
        goto notify_cli;
    }

    gf_log("cli", GF_LOG_DEBUG, "Sending tier start");

    oldlocal = frame->local;
    CLI_LOCAL_INIT(local, words, frame, newdict);
    ret = gf_cli_tier(frame, this, newdict);
    frame->local = oldlocal;
    cli_local_wipe(local);

notify_cli:
    if (ret) {
        cli_out(
            "Failed to run tier start. Please execute tier start "
            "command explicitly");
        cli_out("Usage : gluster volume tier <volname> start");
    }

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_remove_tier_brick(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req status_req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;
    int32_t command = 0;
    char *volname = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "command", &command);
    if (ret)
        goto out;

    ret = dict_set_int32(dict, "rebalance-command", (int32_t)command);
    if (ret) {
        gf_log(this->name, GF_LOG_ERROR, "Failed to set dict");
        goto out;
    }

    ret = cli_to_glusterd(&status_req, frame, gf_cli_remove_tier_brick_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_TIER,
                          this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(status_req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_remove_brick(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    ;
    gf_cli_req status_req = {{
        0,
    }};
    ;
    int ret = 0;
    dict_t *dict = NULL;
    int32_t command = 0;
    char *volname = NULL;
    int32_t cmd = 0;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "command", &command);
    if (ret)
        goto out;

    if ((command != GF_OP_CMD_STATUS) && (command != GF_OP_CMD_STOP)) {
        ret = cli_to_glusterd(
            &req, frame, gf_cli_remove_brick_cbk, (xdrproc_t)xdr_gf_cli_req,
            dict, GLUSTER_CLI_REMOVE_BRICK, this, cli_rpc_prog, NULL);
    } else {
        /* Need rebalance status to be sent :-) */
        if (command == GF_OP_CMD_STATUS)
            cmd |= GF_DEFRAG_CMD_STATUS;
        else
            cmd |= GF_DEFRAG_CMD_STOP;

        ret = dict_set_int32(dict, "rebalance-command", (int32_t)cmd);
        if (ret) {
            gf_log(this->name, GF_LOG_ERROR, "Failed to set dict");
            goto out;
        }

        ret = cli_to_glusterd(
            &status_req, frame, gf_cli3_remove_brick_status_cbk,
            (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_DEFRAG_VOLUME, this,
            cli_rpc_prog, NULL);
    }

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    GF_FREE(status_req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_reset_brick(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;
    char *dst_brick = NULL;
    char *src_brick = NULL;
    char *volname = NULL;
    char *op = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_str(dict, "operation", &op);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on operation failed");
        goto out;
    }

    ret = dict_get_str(dict, "volname", &volname);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on volname failed");
        goto out;
    }

    ret = dict_get_str(dict, "src-brick", &src_brick);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on src-brick failed");
        goto out;
    }

    if (!strcmp(op, "GF_RESET_OP_COMMIT") ||
        !strcmp(op, "GF_RESET_OP_COMMIT_FORCE")) {
        ret = dict_get_str(dict, "dst-brick", &dst_brick);
        if (ret) {
            gf_log(this->name, GF_LOG_DEBUG, "dict_get on dst-brick failed");
            goto out;
        }
    }

    gf_log(this->name, GF_LOG_DEBUG, "Received command reset-brick %s on %s.",
           op, src_brick);

    ret = cli_to_glusterd(&req, frame, gf_cli_reset_brick_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_RESET_BRICK, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_replace_brick(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;
    char *src_brick = NULL;
    char *dst_brick = NULL;
    char *volname = NULL;
    int32_t op = 0;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_int32(dict, "operation", &op);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on operation failed");
        goto out;
    }
    ret = dict_get_str(dict, "volname", &volname);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on volname failed");
        goto out;
    }

    ret = dict_get_str(dict, "src-brick", &src_brick);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on src-brick failed");
        goto out;
    }

    ret = dict_get_str(dict, "dst-brick", &dst_brick);
    if (ret) {
        gf_log(this->name, GF_LOG_DEBUG, "dict_get on dst-brick failed");
        goto out;
    }

    gf_log(this->name, GF_LOG_DEBUG,
           "Received command replace-brick %s with "
           "%s with operation=%d",
           src_brick, dst_brick, op);

    ret = cli_to_glusterd(&req, frame, gf_cli_replace_brick_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_REPLACE_BRICK, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_log_rotate(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_log_rotate_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_LOG_ROTATE, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_sync_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = 0;
    gf_cli_req req = {{
        0,
    }};
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_sync_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_SYNC_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_getspec(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_getspec_req req = {
        0,
    };
    int ret = 0;
    dict_t *dict = NULL;
    dict_t *op_dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_str(dict, "volid", &req.key);
    if (ret)
        goto out;

    op_dict = dict_new();
    if (!op_dict) {
        ret = -1;
        goto out;
    }

    // Set the supported min and max op-versions, so glusterd can make a
    // decision
    ret = dict_set_int32(op_dict, "min-op-version", GD_OP_VERSION_MIN);
    if (ret) {
        gf_log(THIS->name, GF_LOG_ERROR,
               "Failed to set min-op-version"
               " in request dict");
        goto out;
    }

    ret = dict_set_int32(op_dict, "max-op-version", GD_OP_VERSION_MAX);
    if (ret) {
        gf_log(THIS->name, GF_LOG_ERROR,
               "Failed to set max-op-version"
               " in request dict");
        goto out;
    }

    ret = dict_allocate_and_serialize(op_dict, &req.xdata.xdata_val,
                                      &req.xdata.xdata_len);
    if (ret < 0) {
        gf_log(THIS->name, GF_LOG_ERROR, "Failed to serialize dictionary");
        goto out;
    }

    ret = cli_cmd_submit(NULL, &req, frame, &cli_handshake_prog,
                         GF_HNDSK_GETSPEC, NULL, this, gf_cli_getspec_cbk,
                         (xdrproc_t)xdr_gf_getspec_req);

out:
    if (op_dict) {
        dict_unref(op_dict);
    }
    GF_FREE(req.xdata.xdata_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int32_t
gf_cli_quota(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_quota_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_QUOTA,
                          this, cli_rpc_prog, NULL);

out:
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_pmap_b2p(call_frame_t *frame, xlator_t *this, void *data)
{
    pmap_port_by_brick_req req = {
        0,
    };
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = dict_get_str(dict, "brick", &req.brick);
    if (ret)
        goto out;

    ret = cli_cmd_submit(NULL, &req, frame, &cli_pmap_prog, GF_PMAP_PORTBYBRICK,
                         NULL, this, gf_cli_pmap_b2p_cbk,
                         (xdrproc_t)xdr_pmap_port_by_brick_req);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

static int
gf_cli_fsm_log_cbk(struct rpc_req *req, struct iovec *iov, int count,
                   void *myframe)
{
    gf1_cli_fsm_log_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    int tr_count = 0;
    char key[256] = {0};
    int i = 0;
    char *old_state = NULL;
    char *new_state = NULL;
    char *event = NULL;
    char *time = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_fsm_log_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            cli_err("%s", rsp.op_errstr);
        cli_err("fsm log unsuccessful");
        ret = rsp.op_ret;
        goto out;
    }

    dict = dict_new();
    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.fsm_log.fsm_log_val, rsp.fsm_log.fsm_log_len,
                           &dict);

    if (ret) {
        cli_err("bad response");
        goto out;
    }

    ret = dict_get_int32(dict, "count", &tr_count);
    if (tr_count)
        cli_out("number of transitions: %d", tr_count);
    else
        cli_err("No transitions");
    for (i = 0; i < tr_count; i++) {
        snprintf(key, sizeof(key), "log%d-old-state", i);
        ret = dict_get_str(dict, key, &old_state);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "log%d-event", i);
        ret = dict_get_str(dict, key, &event);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "log%d-new-state", i);
        ret = dict_get_str(dict, key, &new_state);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "log%d-time", i);
        ret = dict_get_str(dict, key, &time);
        if (ret)
            goto out;
        cli_out(
            "Old State: [%s]\n"
            "New State: [%s]\n"
            "Event    : [%s]\n"
            "timestamp: [%s]\n",
            old_state, new_state, event, time);
    }

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    if (dict) {
        dict_unref(dict);
    }
    gf_free_xdr_fsm_log_rsp(rsp);

    return ret;
}

int32_t
gf_cli_fsm_log(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = -1;
    gf1_cli_fsm_log_req req = {
        0,
    };

    GF_ASSERT(frame);
    GF_ASSERT(this);
    GF_ASSERT(data);

    if (!frame || !this || !data)
        goto out;
    req.name = data;
    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_FSM_LOG,
                         NULL, this, gf_cli_fsm_log_cbk,
                         (xdrproc_t)xdr_gf1_cli_fsm_log_req);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

int
gf_cli_gsync_config_command(dict_t *dict)
{
    runner_t runner = {
        0,
    };
    char *subop = NULL;
    char *gwd = NULL;
    char *slave = NULL;
    char *confpath = NULL;
    char *master = NULL;
    char *op_name = NULL;
    int ret = -1;
    char conf_path[PATH_MAX] = "";

    if (dict_get_str(dict, "subop", &subop) != 0)
        return -1;

    if (strcmp(subop, "get") != 0 && strcmp(subop, "get-all") != 0) {
        cli_out(GEOREP " config updated successfully");
        return 0;
    }

    if (dict_get_str(dict, "glusterd_workdir", &gwd) != 0 ||
        dict_get_str(dict, "slave", &slave) != 0)
        return -1;

    if (dict_get_str(dict, "master", &master) != 0)
        master = NULL;
    if (dict_get_str(dict, "op_name", &op_name) != 0)
        op_name = NULL;

    ret = dict_get_str(dict, "conf_path", &confpath);
    if (ret || !confpath) {
        ret = snprintf(conf_path, sizeof(conf_path) - 1,
                       "%s/" GEOREP "/gsyncd_template.conf", gwd);
        conf_path[ret] = '\0';
        confpath = conf_path;
    }

    runinit(&runner);
    runner_add_args(&runner, GSYNCD_PREFIX "/gsyncd", "-c", NULL);
    runner_argprintf(&runner, "%s", confpath);
    runner_argprintf(&runner, "--iprefix=%s", DATADIR);
    if (master)
        runner_argprintf(&runner, ":%s", master);
    runner_add_arg(&runner, slave);
    runner_argprintf(&runner, "--config-%s", subop);
    if (op_name)
        runner_add_arg(&runner, op_name);

    return runner_run(&runner);
}

int
gf_cli_print_status(char **title_values, gf_gsync_status_t **sts_vals,
                    int *spacing, int gsync_count, int number_of_fields,
                    int is_detail)
{
    int i = 0;
    int j = 0;
    int ret = 0;
    int status_fields = 8; /* Indexed at 0 */
    int total_spacing = 0;
    char **output_values = NULL;
    char *tmp = NULL;
    char *hyphens = NULL;

    /* calculating spacing for hyphens */
    for (i = 0; i < number_of_fields; i++) {
        /* Suppressing detail output for status */
        if ((!is_detail) && (i > status_fields)) {
            /* Suppressing detailed output for
             * status */
            continue;
        }
        spacing[i] += 3; /* Adding extra space to
                            distinguish between fields */
        total_spacing += spacing[i];
    }
    total_spacing += 4; /* For the spacing between the fields */

    /* char pointers for each field */
    output_values = GF_CALLOC(number_of_fields, sizeof(char *),
                              gf_common_mt_char);
    if (!output_values) {
        ret = -1;
        goto out;
    }
    for (i = 0; i < number_of_fields; i++) {
        output_values[i] = GF_CALLOC(spacing[i] + 1, sizeof(char),
                                     gf_common_mt_char);
        if (!output_values[i]) {
            ret = -1;
            goto out;
        }
    }

    hyphens = GF_CALLOC(total_spacing + 1, sizeof(char), gf_common_mt_char);
    if (!hyphens) {
        ret = -1;
        goto out;
    }

    cli_out(" ");

    /* setting the title "NODE", "MASTER", etc. from title_values[]
       and printing the same */
    for (j = 0; j < number_of_fields; j++) {
        if ((!is_detail) && (j > status_fields)) {
            /* Suppressing detailed output for
             * status */
            output_values[j][0] = '\0';
            continue;
        }
        memset(output_values[j], ' ', spacing[j]);
        memcpy(output_values[j], title_values[j], strlen(title_values[j]));
        output_values[j][spacing[j]] = '\0';
    }
    cli_out("%s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s", output_values[0],
            output_values[1], output_values[2], output_values[3],
            output_values[4], output_values[5], output_values[6],
            output_values[7], output_values[8], output_values[9],
            output_values[10], output_values[11], output_values[12],
            output_values[13], output_values[14], output_values[15]);

    /* setting and printing the hyphens */
    memset(hyphens, '-', total_spacing);
    hyphens[total_spacing] = '\0';
    cli_out("%s", hyphens);

    for (i = 0; i < gsync_count; i++) {
        for (j = 0; j < number_of_fields; j++) {
            if ((!is_detail) && (j > status_fields)) {
                /* Suppressing detailed output for
                 * status */
                output_values[j][0] = '\0';
                continue;
            }
            tmp = get_struct_variable(j, sts_vals[i]);
            if (!tmp) {
                gf_log("", GF_LOG_ERROR, "struct member empty.");
                ret = -1;
                goto out;
            }
            memset(output_values[j], ' ', spacing[j]);
            memcpy(output_values[j], tmp, strlen(tmp));
            output_values[j][spacing[j]] = '\0';
        }

        cli_out("%s %s %s %s %s %s %s %s %s %s %s %s %s %s %s %s",
                output_values[0], output_values[1], output_values[2],
                output_values[3], output_values[4], output_values[5],
                output_values[6], output_values[7], output_values[8],
                output_values[9], output_values[10], output_values[11],
                output_values[12], output_values[13], output_values[14],
                output_values[15]);
    }

out:
    if (output_values) {
        for (i = 0; i < number_of_fields; i++) {
            if (output_values[i])
                GF_FREE(output_values[i]);
        }
        GF_FREE(output_values);
    }

    if (hyphens)
        GF_FREE(hyphens);

    return ret;
}

int
gf_gsync_status_t_comparator(const void *p, const void *q)
{
    char *slavekey1 = NULL;
    char *slavekey2 = NULL;

    slavekey1 = get_struct_variable(20, (*(gf_gsync_status_t **)p));
    slavekey2 = get_struct_variable(20, (*(gf_gsync_status_t **)q));
    if (!slavekey1 || !slavekey2) {
        gf_log("cli", GF_LOG_ERROR, "struct member empty.");
        return 0;
    }

    return strcmp(slavekey1, slavekey2);
}

int
gf_cli_read_status_data(dict_t *dict, gf_gsync_status_t **sts_vals,
                        int *spacing, int gsync_count, int number_of_fields)
{
    char *tmp = NULL;
    char sts_val_name[PATH_MAX] = "";
    int ret = 0;
    int i = 0;
    int j = 0;

    /* Storing per node status info in each object */
    for (i = 0; i < gsync_count; i++) {
        snprintf(sts_val_name, sizeof(sts_val_name), "status_value%d", i);

        /* Fetching the values from dict, and calculating
           the max length for each field */
        ret = dict_get_bin(dict, sts_val_name, (void **)&(sts_vals[i]));
        if (ret)
            goto out;

        for (j = 0; j < number_of_fields; j++) {
            tmp = get_struct_variable(j, sts_vals[i]);
            if (!tmp) {
                gf_log("", GF_LOG_ERROR, "struct member empty.");
                ret = -1;
                goto out;
            }
            if (strlen(tmp) > spacing[j])
                spacing[j] = strlen(tmp);
        }
    }

    /* Sort based on Session Slave */
    qsort(sts_vals, gsync_count, sizeof(gf_gsync_status_t *),
          gf_gsync_status_t_comparator);

out:
    return ret;
}

int
gf_cli_gsync_status_output(dict_t *dict, gf_boolean_t is_detail)
{
    int gsync_count = 0;
    int i = 0;
    int ret = 0;
    int spacing[16] = {0};
    int num_of_fields = 16;
    char errmsg[1024] = "";
    char *master = NULL;
    char *slave = NULL;
    char *title_values[] = {"MASTER NODE",
                            "MASTER VOL",
                            "MASTER BRICK",
                            "SLAVE USER",
                            "SLAVE",
                            "SLAVE NODE",
                            "STATUS",
                            "CRAWL STATUS",
                            "LAST_SYNCED",
                            "ENTRY",
                            "DATA",
                            "META",
                            "FAILURES",
                            "CHECKPOINT TIME",
                            "CHECKPOINT COMPLETED",
                            "CHECKPOINT COMPLETION TIME"};
    gf_gsync_status_t **sts_vals = NULL;

    /* Checks if any session is active or not */
    ret = dict_get_int32(dict, "gsync-count", &gsync_count);
    if (ret) {
        ret = dict_get_str(dict, "master", &master);

        ret = dict_get_str(dict, "slave", &slave);

        if (master) {
            if (slave)
                snprintf(errmsg, sizeof(errmsg),
                         "No active "
                         "geo-replication sessions between %s"
                         " and %s",
                         master, slave);
            else
                snprintf(errmsg, sizeof(errmsg),
                         "No active "
                         "geo-replication sessions for %s",
                         master);
        } else
            snprintf(errmsg, sizeof(errmsg),
                     "No active "
                     "geo-replication sessions");

        gf_log("cli", GF_LOG_INFO, "%s", errmsg);
        cli_out("%s", errmsg);
        ret = -1;
        goto out;
    }

    for (i = 0; i < num_of_fields; i++)
        spacing[i] = strlen(title_values[i]);

    /* gsync_count = number of nodes reporting output.
       each sts_val object will store output of each
       node */
    sts_vals = GF_MALLOC(gsync_count * sizeof(gf_gsync_status_t *),
                         gf_common_mt_char);
    if (!sts_vals) {
        ret = -1;
        goto out;
    }
    for (i = 0; i < gsync_count; i++) {
        sts_vals[i] = GF_CALLOC(1, sizeof(gf_gsync_status_t),
                                gf_common_mt_char);
        if (!sts_vals[i]) {
            ret = -1;
            goto out;
        }
    }

    ret = gf_cli_read_status_data(dict, sts_vals, spacing, gsync_count,
                                  num_of_fields);
    if (ret) {
        gf_log("", GF_LOG_ERROR, "Unable to read status data");
        goto out;
    }

    ret = gf_cli_print_status(title_values, sts_vals, spacing, gsync_count,
                              num_of_fields, is_detail);
    if (ret) {
        gf_log("", GF_LOG_ERROR, "Unable to print status output");
        goto out;
    }

out:
    if (sts_vals)
        GF_FREE(sts_vals);

    return ret;
}

static int32_t
write_contents_to_common_pem_file(dict_t *dict, int output_count)
{
    char *workdir = NULL;
    char common_pem_file[PATH_MAX] = "";
    char *output = NULL;
    char output_name[PATH_MAX] = "";
    int bytes_written = 0;
    int fd = -1;
    int ret = -1;
    int i = -1;

    ret = dict_get_str(dict, "glusterd_workdir", &workdir);
    if (ret || !workdir) {
        gf_log("", GF_LOG_ERROR, "Unable to fetch workdir");
        ret = -1;
        goto out;
    }

    snprintf(common_pem_file, sizeof(common_pem_file),
             "%s/geo-replication/common_secret.pem.pub", workdir);

    sys_unlink(common_pem_file);

    fd = open(common_pem_file, O_WRONLY | O_CREAT, 0600);
    if (fd == -1) {
        gf_log("", GF_LOG_ERROR,
               "Failed to open %s"
               " Error : %s",
               common_pem_file, strerror(errno));
        ret = -1;
        goto out;
    }

    for (i = 1; i <= output_count; i++) {
        snprintf(output_name, sizeof(output_name), "output_%d", i);
        ret = dict_get_str(dict, output_name, &output);
        if (ret) {
            gf_log("", GF_LOG_ERROR, "Failed to get %s.", output_name);
            cli_out("Unable to fetch output.");
        }
        if (output) {
            bytes_written = sys_write(fd, output, strlen(output));
            if (bytes_written != strlen(output)) {
                gf_log("", GF_LOG_ERROR,
                       "Failed to write "
                       "to %s",
                       common_pem_file);
                ret = -1;
                goto out;
            }
            /* Adding the new line character */
            bytes_written = sys_write(fd, "\n", 1);
            if (bytes_written != 1) {
                gf_log("", GF_LOG_ERROR, "Failed to add new line char");
                ret = -1;
                goto out;
            }
            output = NULL;
        }
    }

    cli_out("Common secret pub file present at %s", common_pem_file);
    ret = 0;
out:
    if (fd >= 0)
        sys_close(fd);

    gf_log("", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

int
gf_cli_sys_exec_cbk(struct rpc_req *req, struct iovec *iov, int count,
                    void *myframe)
{
    int ret = -1;
    int output_count = -1;
    int i = -1;
    char *output = NULL;
    char *command = NULL;
    char output_name[PATH_MAX] = "";
    gf_cli_rsp rsp = {
        0,
    };
    dict_t *dict = NULL;

    GF_ASSERT(myframe);

    if (req->rpc_status == -1) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret)
        goto out;

    if (rsp.op_ret) {
        cli_err("%s", rsp.op_errstr ? rsp.op_errstr : "Command failed.");
        ret = rsp.op_ret;
        goto out;
    }

    ret = dict_get_int32(dict, "output_count", &output_count);
    if (ret) {
        cli_out("Command executed successfully.");
        ret = 0;
        goto out;
    }

    ret = dict_get_str(dict, "command", &command);
    if (ret) {
        gf_log("", GF_LOG_ERROR, "Unable to get command from dict");
        goto out;
    }

    if (!strcmp(command, "gsec_create")) {
        ret = write_contents_to_common_pem_file(dict, output_count);
        if (!ret)
            goto out;
    }

    for (i = 1; i <= output_count; i++) {
        snprintf(output_name, sizeof(output_name), "output_%d", i);
        ret = dict_get_str(dict, output_name, &output);
        if (ret) {
            gf_log("", GF_LOG_ERROR, "Failed to get %s.", output_name);
            cli_out("Unable to fetch output.");
        }
        if (output) {
            cli_out("%s", output);
            output = NULL;
        }
    }

    ret = 0;
out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_copy_file_cbk(struct rpc_req *req, struct iovec *iov, int count,
                     void *myframe)
{
    int ret = -1;
    gf_cli_rsp rsp = {
        0,
    };
    dict_t *dict = NULL;

    GF_ASSERT(myframe);

    if (req->rpc_status == -1) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret)
        goto out;

    if (rsp.op_ret) {
        cli_err("%s", rsp.op_errstr ? rsp.op_errstr : "Copy unsuccessful");
        ret = rsp.op_ret;
        goto out;
    }

    cli_out("Successfully copied file.");

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_gsync_set_cbk(struct rpc_req *req, struct iovec *iov, int count,
                     void *myframe)
{
    int ret = -1;
    gf_cli_rsp rsp = {
        0,
    };
    dict_t *dict = NULL;
    char *gsync_status = NULL;
    char *master = NULL;
    char *slave = NULL;
    int32_t type = 0;
    gf_boolean_t status_detail = _gf_false;

    GF_ASSERT(myframe);

    if (req->rpc_status == -1) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret)
        goto out;

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_gsync(dict, rsp.op_ret, rsp.op_errno,
                                       rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret) {
        cli_err("%s",
                rsp.op_errstr ? rsp.op_errstr : GEOREP " command unsuccessful");
        ret = rsp.op_ret;
        goto out;
    }

    ret = dict_get_str(dict, "gsync-status", &gsync_status);
    if (!ret)
        cli_out("%s", gsync_status);

    ret = dict_get_int32(dict, "type", &type);
    if (ret) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "failed to get type");
        goto out;
    }

    switch (type) {
        case GF_GSYNC_OPTION_TYPE_START:
        case GF_GSYNC_OPTION_TYPE_STOP:
            if (dict_get_str(dict, "master", &master) != 0)
                master = "???";
            if (dict_get_str(dict, "slave", &slave) != 0)
                slave = "???";

            cli_out(
                "%s " GEOREP
                " session between %s & %s"
                " has been successful",
                type == GF_GSYNC_OPTION_TYPE_START ? "Starting" : "Stopping",
                master, slave);
            break;

        case GF_GSYNC_OPTION_TYPE_PAUSE:
        case GF_GSYNC_OPTION_TYPE_RESUME:
            if (dict_get_str(dict, "master", &master) != 0)
                master = "???";
            if (dict_get_str(dict, "slave", &slave) != 0)
                slave = "???";

            cli_out("%s " GEOREP
                    " session between %s & %s"
                    " has been successful",
                    type == GF_GSYNC_OPTION_TYPE_PAUSE ? "Pausing" : "Resuming",
                    master, slave);
            break;

        case GF_GSYNC_OPTION_TYPE_CONFIG:
            ret = gf_cli_gsync_config_command(dict);
            break;

        case GF_GSYNC_OPTION_TYPE_STATUS:
            status_detail = dict_get_str_boolean(dict, "status-detail",
                                                 _gf_false);
            ret = gf_cli_gsync_status_output(dict, status_detail);
            break;

        case GF_GSYNC_OPTION_TYPE_DELETE:
            if (dict_get_str(dict, "master", &master) != 0)
                master = "???";
            if (dict_get_str(dict, "slave", &slave) != 0)
                slave = "???";
            cli_out("Deleting " GEOREP
                    " session between %s & %s"
                    " has been successful",
                    master, slave);
            break;

        case GF_GSYNC_OPTION_TYPE_CREATE:
            if (dict_get_str(dict, "master", &master) != 0)
                master = "???";
            if (dict_get_str(dict, "slave", &slave) != 0)
                slave = "???";
            cli_out("Creating " GEOREP
                    " session between %s & %s"
                    " has been successful",
                    master, slave);
            break;

        default:
            cli_out(GEOREP " command executed successfully");
    }

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int32_t
gf_cli_sys_exec(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = 0;
    dict_t *dict = NULL;
    gf_cli_req req = {{
        0,
    }};

    if (!frame || !this || !data) {
        ret = -1;
        gf_log("cli", GF_LOG_ERROR, "Invalid data");
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_sys_exec_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict, GLUSTER_CLI_SYS_EXEC,
                          this, cli_rpc_prog, NULL);
out:
    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_copy_file(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = 0;
    dict_t *dict = NULL;
    gf_cli_req req = {{
        0,
    }};

    if (!frame || !this || !data) {
        ret = -1;
        gf_log("cli", GF_LOG_ERROR, "Invalid data");
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_copy_file_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_COPY_FILE, this, cli_rpc_prog, NULL);
out:
    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_gsync_set(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = 0;
    dict_t *dict = NULL;
    gf_cli_req req = {{
        0,
    }};

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_gsync_set_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_GSYNC_SET, this, cli_rpc_prog, NULL);

out:
    GF_FREE(req.dict.dict_val);

    return ret;
}

int
cli_profile_info_percentage_cmp(void *a, void *b)
{
    cli_profile_info_t *ia = NULL;
    cli_profile_info_t *ib = NULL;
    int ret = 0;

    ia = a;
    ib = b;
    if (ia->percentage_avg_latency < ib->percentage_avg_latency)
        ret = -1;
    else if (ia->percentage_avg_latency > ib->percentage_avg_latency)
        ret = 1;
    else
        ret = 0;
    return ret;
}

void
cmd_profile_volume_brick_out(dict_t *dict, int count, int interval)
{
    char key[256] = {0};
    int i = 0;
    uint64_t sec = 0;
    uint64_t r_count = 0;
    uint64_t w_count = 0;
    uint64_t rb_counts[32] = {0};
    uint64_t wb_counts[32] = {0};
    cli_profile_info_t profile_info[GF_FOP_MAXVALUE] = {{0}};
    cli_profile_info_t upcall_info[GF_UPCALL_FLAGS_MAXVALUE] = {
        {0},
    };
    char output[128] = {0};
    int per_line = 0;
    char read_blocks[128] = {0};
    char write_blocks[128] = {0};
    int index = 0;
    int is_header_printed = 0;
    int ret = 0;
    double total_percentage_latency = 0;

    for (i = 0; i < 32; i++) {
        snprintf(key, sizeof(key), "%d-%d-read-%d", count, interval, (1 << i));
        ret = dict_get_uint64(dict, key, &rb_counts[i]);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }
    }

    for (i = 0; i < 32; i++) {
        snprintf(key, sizeof(key), "%d-%d-write-%d", count, interval, (1 << i));
        ret = dict_get_uint64(dict, key, &wb_counts[i]);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }
    }

    for (i = 0; i < GF_UPCALL_FLAGS_MAXVALUE; i++) {
        snprintf(key, sizeof(key), "%d-%d-%d-upcall-hits", count, interval, i);
        ret = dict_get_uint64(dict, key, &upcall_info[i].fop_hits);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }
        upcall_info[i].fop_name = (char *)gf_upcall_list[i];
    }

    for (i = 0; i < GF_FOP_MAXVALUE; i++) {
        snprintf(key, sizeof(key), "%d-%d-%d-hits", count, interval, i);
        ret = dict_get_uint64(dict, key, &profile_info[i].fop_hits);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }

        snprintf(key, sizeof(key), "%d-%d-%d-avglatency", count, interval, i);
        ret = dict_get_double(dict, key, &profile_info[i].avg_latency);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }

        snprintf(key, sizeof(key), "%d-%d-%d-minlatency", count, interval, i);
        ret = dict_get_double(dict, key, &profile_info[i].min_latency);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }

        snprintf(key, sizeof(key), "%d-%d-%d-maxlatency", count, interval, i);
        ret = dict_get_double(dict, key, &profile_info[i].max_latency);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
        }
        profile_info[i].fop_name = (char *)gf_fop_list[i];

        total_percentage_latency += (profile_info[i].fop_hits *
                                     profile_info[i].avg_latency);
    }
    if (total_percentage_latency) {
        for (i = 0; i < GF_FOP_MAXVALUE; i++) {
            profile_info[i]
                .percentage_avg_latency = 100 * ((profile_info[i].avg_latency *
                                                  profile_info[i].fop_hits) /
                                                 total_percentage_latency);
        }
        gf_array_insertionsort(profile_info, 1, GF_FOP_MAXVALUE - 1,
                               sizeof(cli_profile_info_t),
                               cli_profile_info_percentage_cmp);
    }
    snprintf(key, sizeof(key), "%d-%d-duration", count, interval);
    ret = dict_get_uint64(dict, key, &sec);
    if (ret) {
        gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
    }

    snprintf(key, sizeof(key), "%d-%d-total-read", count, interval);
    ret = dict_get_uint64(dict, key, &r_count);
    if (ret) {
        gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
    }

    snprintf(key, sizeof(key), "%d-%d-total-write", count, interval);
    ret = dict_get_uint64(dict, key, &w_count);
    if (ret) {
        gf_log("cli", GF_LOG_DEBUG, "failed to get %s from dict", key);
    }

    if (interval == -1)
        cli_out("Cumulative Stats:");
    else
        cli_out("Interval %d Stats:", interval);
    snprintf(output, sizeof(output), "%14s", "Block Size:");
    snprintf(read_blocks, sizeof(read_blocks), "%14s", "No. of Reads:");
    snprintf(write_blocks, sizeof(write_blocks), "%14s", "No. of Writes:");
    index = 14;
    for (i = 0; i < 32; i++) {
        if ((rb_counts[i] == 0) && (wb_counts[i] == 0))
            continue;
        per_line++;
        snprintf(output + index, sizeof(output) - index, "%19db+ ", (1 << i));
        if (rb_counts[i]) {
            snprintf(read_blocks + index, sizeof(read_blocks) - index,
                     "%21" PRId64 " ", rb_counts[i]);
        } else {
            snprintf(read_blocks + index, sizeof(read_blocks) - index, "%21s ",
                     "0");
        }
        if (wb_counts[i]) {
            snprintf(write_blocks + index, sizeof(write_blocks) - index,
                     "%21" PRId64 " ", wb_counts[i]);
        } else {
            snprintf(write_blocks + index, sizeof(write_blocks) - index,
                     "%21s ", "0");
        }
        index += 22;
        if (per_line == 3) {
            cli_out("%s", output);
            cli_out("%s", read_blocks);
            cli_out("%s", write_blocks);
            cli_out(" ");
            per_line = 0;
            snprintf(output, sizeof(output), "%14s", "Block Size:");
            snprintf(read_blocks, sizeof(read_blocks), "%14s", "No. of Reads:");
            snprintf(write_blocks, sizeof(write_blocks), "%14s",
                     "No. of Writes:");
            index = 14;
        }
    }

    if (per_line != 0) {
        cli_out("%s", output);
        cli_out("%s", read_blocks);
        cli_out("%s", write_blocks);
    }
    for (i = 0; i < GF_FOP_MAXVALUE; i++) {
        if (profile_info[i].fop_hits == 0)
            continue;
        if (is_header_printed == 0) {
            cli_out("%10s %13s %13s %13s %14s %11s", "%-latency", "Avg-latency",
                    "Min-Latency", "Max-Latency", "No. of calls", "Fop");
            cli_out("%10s %13s %13s %13s %14s %11s", "---------", "-----------",
                    "-----------", "-----------", "------------", "----");
            is_header_printed = 1;
        }
        if (profile_info[i].fop_hits) {
            cli_out(
                "%10.2lf %10.2lf us %10.2lf us %10.2lf us"
                " %14" PRId64 " %11s",
                profile_info[i].percentage_avg_latency,
                profile_info[i].avg_latency, profile_info[i].min_latency,
                profile_info[i].max_latency, profile_info[i].fop_hits,
                profile_info[i].fop_name);
        }
    }

    for (i = 0; i < GF_UPCALL_FLAGS_MAXVALUE; i++) {
        if (upcall_info[i].fop_hits == 0)
            continue;
        if (upcall_info[i].fop_hits) {
            cli_out(
                "%10.2lf %10.2lf us %10.2lf us %10.2lf us"
                " %14" PRId64 " %11s",
                upcall_info[i].percentage_avg_latency,
                upcall_info[i].avg_latency, upcall_info[i].min_latency,
                upcall_info[i].max_latency, upcall_info[i].fop_hits,
                upcall_info[i].fop_name);
        }
    }

    cli_out(" ");
    cli_out("%12s: %" PRId64 " seconds", "Duration", sec);
    cli_out("%12s: %" PRId64 " bytes", "Data Read", r_count);
    cli_out("%12s: %" PRId64 " bytes", "Data Written", w_count);
    cli_out(" ");
}

int32_t
gf_cli_profile_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                          void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    gf1_cli_stats_op op = GF_CLI_STATS_NONE;
    char key[256] = {0};
    int interval = 0;
    int i = 1;
    int32_t brick_count = 0;
    char *volname = NULL;
    char *brick = NULL;
    char str[1024] = {
        0,
    };
    int stats_cleared = 0;
    gf1_cli_info_op info_op = GF_CLI_INFO_NONE;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to profile");
    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret) {
        gf_log("", GF_LOG_ERROR, "Unable to allocate memory");
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_profile(dict, rsp.op_ret, rsp.op_errno,
                                         rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "op", (int32_t *)&op);
    if (ret)
        goto out;

    if (rsp.op_ret && strcmp(rsp.op_errstr, "")) {
        cli_err("%s", rsp.op_errstr);
    } else {
        switch (op) {
            case GF_CLI_STATS_START:
                cli_out("Starting volume profile on %s has been %s ", volname,
                        (rsp.op_ret) ? "unsuccessful" : "successful");
                break;
            case GF_CLI_STATS_STOP:
                cli_out("Stopping volume profile on %s has been %s ", volname,
                        (rsp.op_ret) ? "unsuccessful" : "successful");
                break;
            case GF_CLI_STATS_INFO:
                break;
            default:
                cli_out("volume profile on %s has been %s ", volname,
                        (rsp.op_ret) ? "unsuccessful" : "successful");
                break;
        }
    }

    if (rsp.op_ret) {
        ret = rsp.op_ret;
        goto out;
    }

    if (GF_CLI_STATS_INFO != op) {
        ret = 0;
        goto out;
    }

    ret = dict_get_int32(dict, "info-op", (int32_t *)&info_op);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "count", &brick_count);
    if (ret)
        goto out;

    if (!brick_count) {
        cli_err("All bricks of volume %s are down.", volname);
        goto out;
    }

    while (i <= brick_count) {
        snprintf(key, sizeof(key), "%d-brick", i);
        ret = dict_get_str(dict, key, &brick);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Couldn't get brick name");
            goto out;
        }

        ret = dict_get_str_boolean(dict, "nfs", _gf_false);

        if (ret)
            snprintf(str, sizeof(str), "NFS Server : %s", brick);
        else
            snprintf(str, sizeof(str), "Brick: %s", brick);
        cli_out("%s", str);
        memset(str, '-', strlen(str));
        cli_out("%s", str);

        if (GF_CLI_INFO_CLEAR == info_op) {
            snprintf(key, sizeof(key), "%d-stats-cleared", i);
            ret = dict_get_int32(dict, key, &stats_cleared);
            if (ret)
                goto out;
            cli_out(stats_cleared ? "Cleared stats."
                                  : "Failed to clear stats.");
        } else {
            snprintf(key, sizeof(key), "%d-cumulative", i);
            ret = dict_get_int32(dict, key, &interval);
            if (ret == 0)
                cmd_profile_volume_brick_out(dict, i, interval);

            snprintf(key, sizeof(key), "%d-interval", i);
            ret = dict_get_int32(dict, key, &interval);
            if (ret == 0)
                cmd_profile_volume_brick_out(dict, i, interval);
        }
        i++;
    }
    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int32_t
gf_cli_profile_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = -1;
    gf_cli_req req = {{
        0,
    }};
    dict_t *dict = NULL;

    GF_ASSERT(frame);
    GF_ASSERT(this);
    GF_ASSERT(data);

    if (!frame || !this || !data)
        goto out;
    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_profile_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_PROFILE_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_top_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                      void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    gf1_cli_stats_op op = GF_CLI_STATS_NONE;
    char key[256] = {0};
    int i = 0;
    int32_t brick_count = 0;
    char brick[1024];
    int32_t members = 0;
    char *filename;
    char *bricks;
    uint64_t value = 0;
    int32_t j = 0;
    gf1_cli_top_op top_op = GF_CLI_TOP_NONE;
    uint64_t nr_open = 0;
    uint64_t max_nr_open = 0;
    double throughput = 0;
    double time = 0;
    int32_t time_sec = 0;
    long int time_usec = 0;
    char timestr[256] = {
        0,
    };
    char *openfd_str = NULL;
    gf_boolean_t nfs = _gf_false;
    gf_boolean_t clear_stats = _gf_false;
    int stats_cleared = 0;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to top");
    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            cli_err("%s", rsp.op_errstr);
        cli_err("volume top unsuccessful");
        ret = rsp.op_ret;
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret) {
        gf_log("", GF_LOG_ERROR, "Unable to allocate memory");
        goto out;
    }

    ret = dict_get_int32(dict, "op", (int32_t *)&op);

    if (op != GF_CLI_STATS_TOP) {
        ret = 0;
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_top(dict, rsp.op_ret, rsp.op_errno,
                                     rsp.op_errstr);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        }
        goto out;
    }

    ret = dict_get_int32(dict, "count", &brick_count);
    if (ret)
        goto out;
    snprintf(key, sizeof(key), "%d-top-op", 1);
    ret = dict_get_int32(dict, key, (int32_t *)&top_op);
    if (ret)
        goto out;

    clear_stats = dict_get_str_boolean(dict, "clear-stats", _gf_false);

    while (i < brick_count) {
        i++;
        snprintf(brick, sizeof(brick), "%d-brick", i);
        ret = dict_get_str(dict, brick, &bricks);
        if (ret)
            goto out;

        nfs = dict_get_str_boolean(dict, "nfs", _gf_false);

        if (clear_stats) {
            snprintf(key, sizeof(key), "%d-stats-cleared", i);
            ret = dict_get_int32(dict, key, &stats_cleared);
            if (ret)
                goto out;
            cli_out(stats_cleared ? "Cleared stats for %s %s"
                                  : "Failed to clear stats for %s %s",
                    nfs ? "NFS server on" : "brick", bricks);
            continue;
        }

        if (nfs)
            cli_out("NFS Server : %s", bricks);
        else
            cli_out("Brick: %s", bricks);

        snprintf(key, sizeof(key), "%d-members", i);
        ret = dict_get_int32(dict, key, &members);

        switch (top_op) {
            case GF_CLI_TOP_OPEN:
                snprintf(key, sizeof(key), "%d-current-open", i);
                ret = dict_get_uint64(dict, key, &nr_open);
                if (ret)
                    break;
                snprintf(key, sizeof(key), "%d-max-open", i);
                ret = dict_get_uint64(dict, key, &max_nr_open);
                if (ret)
                    goto out;
                snprintf(key, sizeof(key), "%d-max-openfd-time", i);
                ret = dict_get_str(dict, key, &openfd_str);
                if (ret)
                    goto out;
                cli_out("Current open fds: %" PRIu64
                        ", Max open"
                        " fds: %" PRIu64 ", Max openfd time: %s",
                        nr_open, max_nr_open, openfd_str);
            case GF_CLI_TOP_READ:
            case GF_CLI_TOP_WRITE:
            case GF_CLI_TOP_OPENDIR:
            case GF_CLI_TOP_READDIR:
                if (!members) {
                    continue;
                }
                cli_out("Count\t\tfilename\n=======================");
                break;
            case GF_CLI_TOP_READ_PERF:
            case GF_CLI_TOP_WRITE_PERF:
                snprintf(key, sizeof(key), "%d-throughput", i);
                ret = dict_get_double(dict, key, &throughput);
                if (!ret) {
                    snprintf(key, sizeof(key), "%d-time", i);
                    ret = dict_get_double(dict, key, &time);
                }
                if (!ret)
                    cli_out("Throughput %.2f MBps time %.4f secs", throughput,
                            time / 1e6);

                if (!members) {
                    continue;
                }
                cli_out("%*s %-*s %-*s", VOL_TOP_PERF_SPEED_WIDTH, "MBps",
                        VOL_TOP_PERF_FILENAME_DEF_WIDTH, "Filename",
                        VOL_TOP_PERF_TIME_WIDTH, "Time");
                cli_out("%*s %-*s %-*s", VOL_TOP_PERF_SPEED_WIDTH,
                        "====", VOL_TOP_PERF_FILENAME_DEF_WIDTH,
                        "========", VOL_TOP_PERF_TIME_WIDTH, "====");
                break;
            default:
                goto out;
        }

        for (j = 1; j <= members; j++) {
            snprintf(key, sizeof(key), "%d-filename-%d", i, j);
            ret = dict_get_str(dict, key, &filename);
            if (ret)
                break;
            snprintf(key, sizeof(key), "%d-value-%d", i, j);
            ret = dict_get_uint64(dict, key, &value);
            if (ret)
                goto out;
            if (top_op == GF_CLI_TOP_READ_PERF ||
                top_op == GF_CLI_TOP_WRITE_PERF) {
                snprintf(key, sizeof(key), "%d-time-sec-%d", i, j);
                ret = dict_get_int32(dict, key, (int32_t *)&time_sec);
                if (ret)
                    goto out;
                snprintf(key, sizeof(key), "%d-time-usec-%d", i, j);
                ret = dict_get_int32(dict, key, (int32_t *)&time_usec);
                if (ret)
                    goto out;
                gf_time_fmt(timestr, sizeof timestr, time_sec, gf_timefmt_FT);
                snprintf(timestr + strlen(timestr),
                         sizeof timestr - strlen(timestr), ".%ld", time_usec);
                if (strlen(filename) < VOL_TOP_PERF_FILENAME_DEF_WIDTH)
                    cli_out("%*" PRIu64 " %-*s %-*s", VOL_TOP_PERF_SPEED_WIDTH,
                            value, VOL_TOP_PERF_FILENAME_DEF_WIDTH, filename,
                            VOL_TOP_PERF_TIME_WIDTH, timestr);
                else
                    cli_out("%*" PRIu64 " ...%-*s %-*s",
                            VOL_TOP_PERF_SPEED_WIDTH, value,
                            VOL_TOP_PERF_FILENAME_ALT_WIDTH,
                            filename + strlen(filename) -
                                VOL_TOP_PERF_FILENAME_ALT_WIDTH,
                            VOL_TOP_PERF_TIME_WIDTH, timestr);
            } else {
                cli_out("%" PRIu64 "\t\t%s", value, filename);
            }
        }
    }
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);

    if (dict)
        dict_unref(dict);

    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int32_t
gf_cli_top_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = -1;
    gf_cli_req req = {{
        0,
    }};
    dict_t *dict = NULL;

    GF_ASSERT(frame);
    GF_ASSERT(this);
    GF_ASSERT(data);

    if (!frame || !this || !data)
        goto out;
    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_top_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_PROFILE_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    GF_FREE(req.dict.dict_val);
    return ret;
}

int
gf_cli_getwd_cbk(struct rpc_req *req, struct iovec *iov, int count,
                 void *myframe)
{
    gf1_cli_getwd_rsp rsp = {
        0,
    };
    int ret = -1;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_getwd_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret == -1) {
        cli_err("getwd failed");
        ret = rsp.op_ret;
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to getwd");

    cli_out("%s", rsp.wd);

    ret = 0;

out:
    cli_cmd_broadcast_response(ret);
    if (rsp.wd) {
        free(rsp.wd);
    }

    return ret;
}

int32_t
gf_cli_getwd(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = -1;
    gf1_cli_getwd_req req = {
        0,
    };

    GF_ASSERT(frame);
    GF_ASSERT(this);

    if (!frame || !this)
        goto out;

    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_GETWD,
                         NULL, this, gf_cli_getwd_cbk,
                         (xdrproc_t)xdr_gf1_cli_getwd_req);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    return ret;
}

void
cli_print_volume_status_mempool(dict_t *dict, char *prefix)
{
    int ret = -1;
    int32_t mempool_count = 0;
    char *name = NULL;
    int32_t hotcount = 0;
    int32_t coldcount = 0;
    uint64_t paddedsizeof = 0;
    uint64_t alloccount = 0;
    int32_t maxalloc = 0;
    uint64_t pool_misses = 0;
    int32_t maxstdalloc = 0;
    char key[1024] = {
        0,
    };
    int i = 0;

    GF_ASSERT(dict);
    GF_ASSERT(prefix);

    snprintf(key, sizeof(key), "%s.mempool-count", prefix);
    ret = dict_get_int32(dict, key, &mempool_count);
    if (ret)
        goto out;

    cli_out("Mempool Stats\n-------------");
    cli_out("%-30s %9s %9s %12s %10s %8s %8s %12s", "Name", "HotCount",
            "ColdCount", "PaddedSizeof", "AllocCount", "MaxAlloc", "Misses",
            "Max-StdAlloc");
    cli_out("%-30s %9s %9s %12s %10s %8s %8s %12s", "----", "--------",
            "---------", "------------", "----------", "--------", "--------",
            "------------");

    for (i = 0; i < mempool_count; i++) {
        snprintf(key, sizeof(key), "%s.pool%d.name", prefix, i);
        ret = dict_get_str(dict, key, &name);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.hotcount", prefix, i);
        ret = dict_get_int32(dict, key, &hotcount);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.coldcount", prefix, i);
        ret = dict_get_int32(dict, key, &coldcount);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.paddedsizeof", prefix, i);
        ret = dict_get_uint64(dict, key, &paddedsizeof);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.alloccount", prefix, i);
        ret = dict_get_uint64(dict, key, &alloccount);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.max_alloc", prefix, i);
        ret = dict_get_int32(dict, key, &maxalloc);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.max-stdalloc", prefix, i);
        ret = dict_get_int32(dict, key, &maxstdalloc);
        if (ret)
            goto out;

        snprintf(key, sizeof(key), "%s.pool%d.pool-misses", prefix, i);
        ret = dict_get_uint64(dict, key, &pool_misses);
        if (ret)
            goto out;

        cli_out("%-30s %9d %9d %12" PRIu64 " %10" PRIu64 " %8d %8" PRIu64
                " %12d",
                name, hotcount, coldcount, paddedsizeof, alloccount, maxalloc,
                pool_misses, maxstdalloc);
    }

out:
    return;
}

void
cli_print_volume_status_mem(dict_t *dict, gf_boolean_t notbrick)
{
    int ret = -1;
    char *volname = NULL;
    char *hostname = NULL;
    char *path = NULL;
    int online = -1;
    char key[1024] = {
        0,
    };
    int brick_index_max = -1;
    int other_count = 0;
    int index_max = 0;
    int val = 0;
    int i = 0;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;
    cli_out("Memory status for volume : %s", volname);

    ret = dict_get_int32(dict, "brick-index-max", &brick_index_max);
    if (ret)
        goto out;
    ret = dict_get_int32(dict, "other-count", &other_count);
    if (ret)
        goto out;

    index_max = brick_index_max + other_count;

    for (i = 0; i <= index_max; i++) {
        cli_out("----------------------------------------------");

        snprintf(key, sizeof(key), "brick%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname);
        if (ret)
            continue;
        snprintf(key, sizeof(key), "brick%d.path", i);
        ret = dict_get_str(dict, key, &path);
        if (ret)
            continue;
        if (notbrick)
            cli_out("%s : %s", hostname, path);
        else
            cli_out("Brick : %s:%s", hostname, path);

        snprintf(key, sizeof(key), "brick%d.status", i);
        ret = dict_get_int32(dict, key, &online);
        if (ret)
            goto out;
        if (!online) {
            if (notbrick)
                cli_out("%s is offline", hostname);
            else
                cli_out("Brick is offline");
            continue;
        }

        cli_out("Mallinfo\n--------");

        snprintf(key, sizeof(key), "brick%d.mallinfo.arena", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Arena", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.ordblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Ordblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.smblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Smblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.hblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Hblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.hblkhd", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Hblkhd", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.usmblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Usmblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.fsmblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Fsmblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.uordblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Uordblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.fordblks", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Fordblks", val);

        snprintf(key, sizeof(key), "brick%d.mallinfo.keepcost", i);
        ret = dict_get_int32(dict, key, &val);
        if (ret)
            goto out;
        cli_out("%-8s : %d", "Keepcost", val);

        cli_out(" ");
        snprintf(key, sizeof(key), "brick%d", i);
        cli_print_volume_status_mempool(dict, key);
    }
out:
    cli_out("----------------------------------------------\n");
    return;
}

void
cli_print_volume_status_client_list(dict_t *dict, gf_boolean_t notbrick)
{
    int ret = -1;
    char *volname = NULL;
    int client_count = 0;
    int current_count = 0;
    char key[1024] = {
        0,
    };
    int i = 0;
    int total = 0;
    char *name = NULL;
    gf_boolean_t is_fuse_done = _gf_false;
    gf_boolean_t is_gfapi_done = _gf_false;
    gf_boolean_t is_tierd_done = _gf_false;
    gf_boolean_t is_rebalance_done = _gf_false;
    gf_boolean_t is_glustershd_done = _gf_false;
    gf_boolean_t is_quotad_done = _gf_false;
    gf_boolean_t is_snapd_done = _gf_false;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;
    cli_out("Client connections for volume %s", volname);

    ret = dict_get_int32(dict, "client-count", &client_count);
    if (ret)
        goto out;

    cli_out("%-48s %15s", "Name", "count");
    cli_out("%-48s %15s", "-----", "------");
    for (i = 0; i < client_count; i++) {
        name = NULL;
        snprintf(key, sizeof(key), "client%d.name", i);
        ret = dict_get_str(dict, key, &name);

        if (!strncmp(name, "fuse", 4)) {
            if (!is_fuse_done) {
                is_fuse_done = _gf_true;
                ret = dict_get_int32(dict, "fuse-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        } else if (!strncmp(name, "gfapi", 5)) {
            if (!is_gfapi_done) {
                is_gfapi_done = _gf_true;
                ret = dict_get_int32(dict, "gfapi-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        } else if (!strcmp(name, "tierd")) {
            if (!is_tierd_done) {
                is_tierd_done = _gf_true;
                ret = dict_get_int32(dict, "tierd-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        } else if (!strcmp(name, "rebalance")) {
            if (!is_rebalance_done) {
                is_rebalance_done = _gf_true;
                ret = dict_get_int32(dict, "rebalance-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        } else if (!strcmp(name, "glustershd")) {
            if (!is_glustershd_done) {
                is_glustershd_done = _gf_true;
                ret = dict_get_int32(dict, "glustershd-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        } else if (!strcmp(name, "quotad")) {
            if (!is_quotad_done) {
                is_quotad_done = _gf_true;
                ret = dict_get_int32(dict, "quotad-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        } else if (!strcmp(name, "snapd")) {
            if (!is_snapd_done) {
                is_snapd_done = _gf_true;
                ret = dict_get_int32(dict, "snapd-count", &current_count);
                if (ret)
                    goto out;
                total = total + current_count;
                goto print;
            }
            continue;
        }

    print:
        cli_out("%-48s %15d", name, current_count);
    }
out:
    cli_out("\ntotal clients for volume %s : %d ", volname, total);
    cli_out(
        "-----------------------------------------------------------------\n");
    return;
}

void
cli_print_volume_status_clients(dict_t *dict, gf_boolean_t notbrick)
{
    int ret = -1;
    char *volname = NULL;
    int brick_index_max = -1;
    int other_count = 0;
    int index_max = 0;
    char *hostname = NULL;
    char *path = NULL;
    int online = -1;
    int client_count = 0;
    char *clientname = NULL;
    uint64_t bytesread = 0;
    uint64_t byteswrite = 0;
    uint32_t opversion = 0;
    char key[1024] = {
        0,
    };
    int i = 0;
    int j = 0;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;
    cli_out("Client connections for volume %s", volname);

    ret = dict_get_int32(dict, "brick-index-max", &brick_index_max);
    if (ret)
        goto out;
    ret = dict_get_int32(dict, "other-count", &other_count);
    if (ret)
        goto out;

    index_max = brick_index_max + other_count;

    for (i = 0; i <= index_max; i++) {
        hostname = NULL;
        path = NULL;
        online = -1;
        client_count = 0;
        clientname = NULL;
        bytesread = 0;
        byteswrite = 0;

        cli_out("----------------------------------------------");

        snprintf(key, sizeof(key), "brick%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname);

        snprintf(key, sizeof(key), "brick%d.path", i);
        ret = dict_get_str(dict, key, &path);

        if (hostname && path) {
            if (notbrick)
                cli_out("%s : %s", hostname, path);
            else
                cli_out("Brick : %s:%s", hostname, path);
        }
        snprintf(key, sizeof(key), "brick%d.status", i);
        ret = dict_get_int32(dict, key, &online);
        if (!online) {
            if (notbrick)
                cli_out("%s is offline", hostname);
            else
                cli_out("Brick is offline");
            continue;
        }

        snprintf(key, sizeof(key), "brick%d.clientcount", i);
        ret = dict_get_int32(dict, key, &client_count);

        if (hostname && path)
            cli_out("Clients connected : %d", client_count);
        if (client_count == 0)
            continue;

        cli_out("%-48s %15s %15s %15s", "Hostname", "BytesRead", "BytesWritten",
                "OpVersion");
        cli_out("%-48s %15s %15s %15s", "--------", "---------", "------------",
                "---------");
        for (j = 0; j < client_count; j++) {
            snprintf(key, sizeof(key), "brick%d.client%d.hostname", i, j);
            ret = dict_get_str(dict, key, &clientname);

            snprintf(key, sizeof(key), "brick%d.client%d.bytesread", i, j);
            ret = dict_get_uint64(dict, key, &bytesread);

            snprintf(key, sizeof(key), "brick%d.client%d.byteswrite", i, j);
            ret = dict_get_uint64(dict, key, &byteswrite);

            snprintf(key, sizeof(key), "brick%d.client%d.opversion", i, j);
            ret = dict_get_uint32(dict, key, &opversion);

            cli_out("%-48s %15" PRIu64 " %15" PRIu64 " %15" PRIu32, clientname,
                    bytesread, byteswrite, opversion);
        }
    }
out:
    cli_out("----------------------------------------------\n");
    return;
}

void
cli_print_volume_status_inode_entry(dict_t *dict, char *prefix)
{
    int ret = -1;
    char key[1024] = {
        0,
    };
    char *gfid = NULL;
    uint64_t nlookup = 0;
    uint32_t ref = 0;
    int ia_type = 0;
    char inode_type;

    GF_ASSERT(dict);
    GF_ASSERT(prefix);

    snprintf(key, sizeof(key), "%s.gfid", prefix);
    ret = dict_get_str(dict, key, &gfid);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "%s.nlookup", prefix);
    ret = dict_get_uint64(dict, key, &nlookup);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "%s.ref", prefix);
    ret = dict_get_uint32(dict, key, &ref);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "%s.ia_type", prefix);
    ret = dict_get_int32(dict, key, &ia_type);
    if (ret)
        goto out;

    switch (ia_type) {
        case IA_IFREG:
            inode_type = 'R';
            break;
        case IA_IFDIR:
            inode_type = 'D';
            break;
        case IA_IFLNK:
            inode_type = 'L';
            break;
        case IA_IFBLK:
            inode_type = 'B';
            break;
        case IA_IFCHR:
            inode_type = 'C';
            break;
        case IA_IFIFO:
            inode_type = 'F';
            break;
        case IA_IFSOCK:
            inode_type = 'S';
            break;
        default:
            inode_type = 'I';
            break;
    }

    cli_out("%-40s %14" PRIu64 " %14" PRIu32 " %9c", gfid, nlookup, ref,
            inode_type);

out:
    return;
}

void
cli_print_volume_status_itables(dict_t *dict, char *prefix)
{
    int ret = -1;
    char key[1024] = {
        0,
    };
    uint32_t active_size = 0;
    uint32_t lru_size = 0;
    uint32_t purge_size = 0;
    uint32_t lru_limit = 0;
    int i = 0;

    GF_ASSERT(dict);
    GF_ASSERT(prefix);

    snprintf(key, sizeof(key), "%s.lru_limit", prefix);
    ret = dict_get_uint32(dict, key, &lru_limit);
    if (ret)
        goto out;
    cli_out("LRU limit     : %u", lru_limit);

    snprintf(key, sizeof(key), "%s.active_size", prefix);
    ret = dict_get_uint32(dict, key, &active_size);
    if (ret)
        goto out;

#ifdef DEBUG
    if (active_size != 0) {
        cli_out("Active inodes:");
        cli_out("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", "IA type");
        cli_out("%-40s %14s %14s %9s", "----", "-------", "---", "-------");
    }
    for (i = 0; i < active_size; i++) {
        snprintf(key, sizeof(key), "%s.active%d", prefix, i);
        cli_print_volume_status_inode_entry(dict, key);
    }
    cli_out(" ");

#else
    cli_out("Active Inodes : %u", active_size);

#endif
    snprintf(key, sizeof(key), "%s.lru_size", prefix);
    ret = dict_get_uint32(dict, key, &lru_size);
    if (ret)
        goto out;

#ifdef DEBUG
    if (lru_size != 0) {
        cli_out("LRU inodes:");
        cli_out("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", "IA type");
        cli_out("%-40s %14s %14s %9s", "----", "-------", "---", "-------");
    }
    for (i = 0; i < lru_size; i++) {
        snprintf(key, sizeof(key), "%s.lru%d", prefix, i);
        cli_print_volume_status_inode_entry(dict, key);
    }
    cli_out(" ");
#else
    cli_out("LRU Inodes    : %u", lru_size);
#endif

    snprintf(key, sizeof(key), "%s.purge_size", prefix);
    ret = dict_get_uint32(dict, key, &purge_size);
    if (ret)
        goto out;
#ifdef DEBUG
    if (purge_size != 0) {
        cli_out("Purged inodes:");
        cli_out("%-40s %14s %14s %9s", "GFID", "Lookups", "Ref", "IA type");
        cli_out("%-40s %14s %14s %9s", "----", "-------", "---", "-------");
    }
    for (i = 0; i < purge_size; i++) {
        snprintf(key, sizeof(key), "%s.purge%d", prefix, i);
        cli_print_volume_status_inode_entry(dict, key);
    }
#else
    cli_out("Purge Inodes  : %u", purge_size);
#endif
out:
    return;
}

void
cli_print_volume_status_inode(dict_t *dict, gf_boolean_t notbrick)
{
    int ret = -1;
    char *volname = NULL;
    int brick_index_max = -1;
    int other_count = 0;
    int index_max = 0;
    char *hostname = NULL;
    char *path = NULL;
    int online = -1;
    int conn_count = 0;
    char key[1024] = {
        0,
    };
    int i = 0;
    int j = 0;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;
    cli_out("Inode tables for volume %s", volname);

    ret = dict_get_int32(dict, "brick-index-max", &brick_index_max);
    if (ret)
        goto out;
    ret = dict_get_int32(dict, "other-count", &other_count);
    if (ret)
        goto out;

    index_max = brick_index_max + other_count;

    for (i = 0; i <= index_max; i++) {
        cli_out("----------------------------------------------");

        snprintf(key, sizeof(key), "brick%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname);
        if (ret)
            goto out;
        snprintf(key, sizeof(key), "brick%d.path", i);
        ret = dict_get_str(dict, key, &path);
        if (ret)
            goto out;
        if (notbrick)
            cli_out("%s : %s", hostname, path);
        else
            cli_out("Brick : %s:%s", hostname, path);

        snprintf(key, sizeof(key), "brick%d.status", i);
        ret = dict_get_int32(dict, key, &online);
        if (ret)
            goto out;
        if (!online) {
            if (notbrick)
                cli_out("%s is offline", hostname);
            else
                cli_out("Brick is offline");
            continue;
        }

        snprintf(key, sizeof(key), "brick%d.conncount", i);
        ret = dict_get_int32(dict, key, &conn_count);
        if (ret)
            goto out;

        for (j = 0; j < conn_count; j++) {
            if (conn_count > 1)
                cli_out("Connection %d:", j + 1);

            snprintf(key, sizeof(key), "brick%d.conn%d.itable", i, j);
            cli_print_volume_status_itables(dict, key);
            cli_out(" ");
        }
    }
out:
    cli_out("----------------------------------------------");
    return;
}

void
cli_print_volume_status_fdtable(dict_t *dict, char *prefix)
{
    int ret = -1;
    char key[1024] = {
        0,
    };
    int refcount = 0;
    uint32_t maxfds = 0;
    int firstfree = 0;
    int openfds = 0;
    int fd_pid = 0;
    int fd_refcount = 0;
    int fd_flags = 0;
    int i = 0;

    GF_ASSERT(dict);
    GF_ASSERT(prefix);

    snprintf(key, sizeof(key), "%s.refcount", prefix);
    ret = dict_get_int32(dict, key, &refcount);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "%s.maxfds", prefix);
    ret = dict_get_uint32(dict, key, &maxfds);
    if (ret)
        goto out;

    snprintf(key, sizeof(key), "%s.firstfree", prefix);
    ret = dict_get_int32(dict, key, &firstfree);
    if (ret)
        goto out;

    cli_out("RefCount = %d  MaxFDs = %d  FirstFree = %d", refcount, maxfds,
            firstfree);

    snprintf(key, sizeof(key), "%s.openfds", prefix);
    ret = dict_get_int32(dict, key, &openfds);
    if (ret)
        goto out;
    if (0 == openfds) {
        cli_err("No open fds");
        goto out;
    }

    cli_out("%-19s %-19s %-19s %-19s", "FD Entry", "PID", "RefCount", "Flags");
    cli_out("%-19s %-19s %-19s %-19s", "--------", "---", "--------", "-----");

    for (i = 0; i < maxfds; i++) {
        snprintf(key, sizeof(key), "%s.fdentry%d.pid", prefix, i);
        ret = dict_get_int32(dict, key, &fd_pid);
        if (ret)
            continue;

        snprintf(key, sizeof(key), "%s.fdentry%d.refcount", prefix, i);
        ret = dict_get_int32(dict, key, &fd_refcount);
        if (ret)
            continue;

        snprintf(key, sizeof(key), "%s.fdentry%d.flags", prefix, i);
        ret = dict_get_int32(dict, key, &fd_flags);
        if (ret)
            continue;

        cli_out("%-19d %-19d %-19d %-19d", i, fd_pid, fd_refcount, fd_flags);
    }

out:
    return;
}

void
cli_print_volume_status_fd(dict_t *dict, gf_boolean_t notbrick)
{
    int ret = -1;
    char *volname = NULL;
    int brick_index_max = -1;
    int other_count = 0;
    int index_max = 0;
    char *hostname = NULL;
    char *path = NULL;
    int online = -1;
    int conn_count = 0;
    char key[1024] = {
        0,
    };
    int i = 0;
    int j = 0;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;
    cli_out("FD tables for volume %s", volname);

    ret = dict_get_int32(dict, "brick-index-max", &brick_index_max);
    if (ret)
        goto out;
    ret = dict_get_int32(dict, "other-count", &other_count);
    if (ret)
        goto out;

    index_max = brick_index_max + other_count;

    for (i = 0; i <= index_max; i++) {
        cli_out("----------------------------------------------");

        snprintf(key, sizeof(key), "brick%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname);
        if (ret)
            goto out;
        snprintf(key, sizeof(key), "brick%d.path", i);
        ret = dict_get_str(dict, key, &path);
        if (ret)
            goto out;

        if (notbrick)
            cli_out("%s : %s", hostname, path);
        else
            cli_out("Brick : %s:%s", hostname, path);

        snprintf(key, sizeof(key), "brick%d.status", i);
        ret = dict_get_int32(dict, key, &online);
        if (ret)
            goto out;
        if (!online) {
            if (notbrick)
                cli_out("%s is offline", hostname);
            else
                cli_out("Brick is offline");
            continue;
        }

        snprintf(key, sizeof(key), "brick%d.conncount", i);
        ret = dict_get_int32(dict, key, &conn_count);
        if (ret)
            goto out;

        for (j = 0; j < conn_count; j++) {
            cli_out("Connection %d:", j + 1);

            snprintf(key, sizeof(key), "brick%d.conn%d.fdtable", i, j);
            cli_print_volume_status_fdtable(dict, key);
            cli_out(" ");
        }
    }
out:
    cli_out("----------------------------------------------");
    return;
}

void
cli_print_volume_status_call_frame(dict_t *dict, char *prefix)
{
    int ret = -1;
    char key[1024] = {
        0,
    };
    int ref_count = 0;
    char *translator = 0;
    int complete = 0;
    char *parent = NULL;
    char *wind_from = NULL;
    char *wind_to = NULL;
    char *unwind_from = NULL;
    char *unwind_to = NULL;

    if (!dict || !prefix)
        return;

    snprintf(key, sizeof(key), "%s.refcount", prefix);
    ret = dict_get_int32(dict, key, &ref_count);
    if (ret)
        return;

    snprintf(key, sizeof(key), "%s.translator", prefix);
    ret = dict_get_str(dict, key, &translator);
    if (ret)
        return;

    snprintf(key, sizeof(key), "%s.complete", prefix);
    ret = dict_get_int32(dict, key, &complete);
    if (ret)
        return;

    cli_out("  Ref Count   = %d", ref_count);
    cli_out("  Translator  = %s", translator);
    cli_out("  Completed   = %s", (complete ? "Yes" : "No"));

    snprintf(key, sizeof(key), "%s.parent", prefix);
    ret = dict_get_str(dict, key, &parent);
    if (!ret)
        cli_out("  Parent      = %s", parent);

    snprintf(key, sizeof(key), "%s.windfrom", prefix);
    ret = dict_get_str(dict, key, &wind_from);
    if (!ret)
        cli_out("  Wind From   = %s", wind_from);

    snprintf(key, sizeof(key), "%s.windto", prefix);
    ret = dict_get_str(dict, key, &wind_to);
    if (!ret)
        cli_out("  Wind To     = %s", wind_to);

    snprintf(key, sizeof(key), "%s.unwindfrom", prefix);
    ret = dict_get_str(dict, key, &unwind_from);
    if (!ret)
        cli_out("  Unwind From = %s", unwind_from);

    snprintf(key, sizeof(key), "%s.unwindto", prefix);
    ret = dict_get_str(dict, key, &unwind_to);
    if (!ret)
        cli_out("  Unwind To   = %s", unwind_to);
}

void
cli_print_volume_status_call_stack(dict_t *dict, char *prefix)
{
    int ret = -1;
    char key[1024] = {
        0,
    };
    int uid = 0;
    int gid = 0;
    int pid = 0;
    uint64_t unique = 0;
    // char            *op = NULL;
    int count = 0;
    int i = 0;

    if (!dict || !prefix)
        return;

    snprintf(key, sizeof(key), "%s.uid", prefix);
    ret = dict_get_int32(dict, key, &uid);
    if (ret)
        return;

    snprintf(key, sizeof(key), "%s.gid", prefix);
    ret = dict_get_int32(dict, key, &gid);
    if (ret)
        return;

    snprintf(key, sizeof(key), "%s.pid", prefix);
    ret = dict_get_int32(dict, key, &pid);
    if (ret)
        return;

    snprintf(key, sizeof(key), "%s.unique", prefix);
    ret = dict_get_uint64(dict, key, &unique);
    if (ret)
        return;

    /*
    snprintf (key, sizeof (key), "%s.op", prefix);
    ret = dict_get_str (dict, key, &op);
    if (ret)
            return;
    */

    snprintf(key, sizeof(key), "%s.count", prefix);
    ret = dict_get_int32(dict, key, &count);
    if (ret)
        return;

    cli_out(" UID    : %d", uid);
    cli_out(" GID    : %d", gid);
    cli_out(" PID    : %d", pid);
    cli_out(" Unique : %" PRIu64, unique);
    // cli_out ("\tOp     : %s", op);
    cli_out(" Frames : %d", count);

    for (i = 0; i < count; i++) {
        cli_out(" Frame %d", i + 1);

        snprintf(key, sizeof(key), "%s.frame%d", prefix, i);
        cli_print_volume_status_call_frame(dict, key);
    }

    cli_out(" ");
}

void
cli_print_volume_status_callpool(dict_t *dict, gf_boolean_t notbrick)
{
    int ret = -1;
    char *volname = NULL;
    int brick_index_max = -1;
    int other_count = 0;
    int index_max = 0;
    char *hostname = NULL;
    char *path = NULL;
    int online = -1;
    int call_count = 0;
    char key[1024] = {
        0,
    };
    int i = 0;
    int j = 0;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;
    cli_out("Pending calls for volume %s", volname);

    ret = dict_get_int32(dict, "brick-index-max", &brick_index_max);
    if (ret)
        goto out;
    ret = dict_get_int32(dict, "other-count", &other_count);
    if (ret)
        goto out;

    index_max = brick_index_max + other_count;

    for (i = 0; i <= index_max; i++) {
        cli_out("----------------------------------------------");

        snprintf(key, sizeof(key), "brick%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname);
        if (ret)
            goto out;
        snprintf(key, sizeof(key), "brick%d.path", i);
        ret = dict_get_str(dict, key, &path);
        if (ret)
            goto out;

        if (notbrick)
            cli_out("%s : %s", hostname, path);
        else
            cli_out("Brick : %s:%s", hostname, path);

        snprintf(key, sizeof(key), "brick%d.status", i);
        ret = dict_get_int32(dict, key, &online);
        if (ret)
            goto out;
        if (!online) {
            if (notbrick)
                cli_out("%s is offline", hostname);
            else
                cli_out("Brick is offline");
            continue;
        }

        snprintf(key, sizeof(key), "brick%d.callpool.count", i);
        ret = dict_get_int32(dict, key, &call_count);
        if (ret)
            goto out;
        cli_out("Pending calls: %d", call_count);

        if (0 == call_count)
            continue;

        for (j = 0; j < call_count; j++) {
            cli_out("Call Stack%d", j + 1);

            snprintf(key, sizeof(key), "brick%d.callpool.stack%d", i, j);
            cli_print_volume_status_call_stack(dict, key);
        }
    }

out:
    cli_out("----------------------------------------------");
    return;
}

static void
cli_print_volume_status_tasks(dict_t *dict)
{
    int ret = -1;
    int i = 0;
    int j = 0;
    int count = 0;
    int task_count = 0;
    int status = 0;
    char *op = NULL;
    char *task_id_str = NULL;
    char *volname = NULL;
    char key[64] = {
        0,
    };
    char task[32] = {
        0,
    };
    char *brick = NULL;

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "tasks", &task_count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get tasks count");
        return;
    }

    cli_out("Task Status of Volume %s", volname);
    cli_print_line(CLI_BRICK_STATUS_LINE_LEN);

    if (task_count == 0) {
        cli_out("There are no active volume tasks");
        cli_out(" ");
        return;
    }

    for (i = 0; i < task_count; i++) {
        snprintf(key, sizeof(key), "task%d.type", i);
        ret = dict_get_str(dict, key, &op);
        if (ret)
            return;
        cli_out("%-20s : %-20s", "Task", op);

        snprintf(key, sizeof(key), "task%d.id", i);
        ret = dict_get_str(dict, key, &task_id_str);
        if (ret)
            return;
        cli_out("%-20s : %-20s", "ID", task_id_str);

        snprintf(key, sizeof(key), "task%d.status", i);
        ret = dict_get_int32(dict, key, &status);
        if (ret)
            return;

        snprintf(task, sizeof(task), "task%d", i);

        if (!strcmp(op, "Remove brick")) {
            snprintf(key, sizeof(key), "%s.count", task);
            ret = dict_get_int32(dict, key, &count);
            if (ret)
                goto out;

            cli_out("%-20s", "Removed bricks:");

            for (j = 1; j <= count; j++) {
                snprintf(key, sizeof(key), "%s.brick%d", task, j);
                ret = dict_get_str(dict, key, &brick);
                if (ret)
                    goto out;

                cli_out("%-20s", brick);
            }
        }
        cli_out("%-20s : %-20s", "Status", cli_vol_task_status_str[status]);
        cli_out(" ");
    }

out:
    return;
}

static int
gf_cli_status_cbk(struct rpc_req *req, struct iovec *iov, int count,
                  void *myframe)
{
    int ret = -1;
    int brick_index_max = -1;
    int other_count = 0;
    int index_max = 0;
    int i = 0;
    int type = -1;
    int hot_brick_count = -1;
    int pid = -1;
    uint32_t cmd = 0;
    gf_boolean_t notbrick = _gf_false;
    char key[1024] = {
        0,
    };
    char *hostname = NULL;
    char *path = NULL;
    char *volname = NULL;
    dict_t *dict = NULL;
    gf_cli_rsp rsp = {
        0,
    };
    cli_volume_status_t status = {0};
    cli_local_t *local = NULL;
    gf_boolean_t wipe_local = _gf_false;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (req->rpc_status == -1)
        goto out;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_DEBUG, "Received response to status cmd");

    local = ((call_frame_t *)myframe)->local;
    if (!local) {
        local = cli_local_get();
        if (!local) {
            ret = -1;
            gf_log("cli", GF_LOG_ERROR, "Failed to get local");
            goto out;
        }
        wipe_local = _gf_true;
    }

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
        else
            snprintf(msg, sizeof(msg),
                     "Unable to obtain volume "
                     "status information.");

        if (global_state->mode & GLUSTER_MODE_XML) {
            if (!local->all)
                cli_xml_output_str("volStatus", msg, rsp.op_ret, rsp.op_errno,
                                   rsp.op_errstr);
            ret = 0;
            goto out;
        }

        cli_err("%s", msg);
        if (local && local->all) {
            ret = 0;
            cli_out(" ");
        } else
            ret = -1;

        goto out;
    }

    dict = dict_new();
    if (!dict) {
        gf_log(THIS->name, GF_LOG_ERROR, "Failed to create the dict");
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret)
        goto out;

    ret = dict_get_uint32(dict, "cmd", &cmd);
    if (ret)
        goto out;

    if ((cmd & GF_CLI_STATUS_ALL)) {
        if (local && local->dict) {
            dict_ref(dict);
            ret = dict_set_static_ptr(local->dict, "rsp-dict", dict);
            ret = 0;
        } else {
            gf_log("cli", GF_LOG_ERROR, "local not found");
            ret = -1;
        }
        goto out;
    }

    if ((cmd & GF_CLI_STATUS_NFS) || (cmd & GF_CLI_STATUS_SHD) ||
        (cmd & GF_CLI_STATUS_QUOTAD) || (cmd & GF_CLI_STATUS_SNAPD) ||
        (cmd & GF_CLI_STATUS_BITD) || (cmd & GF_CLI_STATUS_SCRUB) ||
        (cmd & GF_CLI_STATUS_TIERD))
        notbrick = _gf_true;

    if (global_state->mode & GLUSTER_MODE_XML) {
        if (!local->all) {
            ret = cli_xml_output_vol_status_begin(local, rsp.op_ret,
                                                  rsp.op_errno, rsp.op_errstr);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
                goto xml_end;
            }
        }
        if (cmd & GF_CLI_STATUS_TASKS) {
            ret = cli_xml_output_vol_status_tasks_detail(local, dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Error outputting "
                       "to xml");
                goto xml_end;
            }
        } else {
            ret = cli_xml_output_vol_status(local, dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
                goto xml_end;
            }
        }

    xml_end:
        if (!local->all) {
            ret = cli_xml_output_vol_status_end(local);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
            }
        }
        goto out;
    }

    status.brick = GF_MALLOC(PATH_MAX + 256, gf_common_mt_strdup);
    if (!status.brick) {
        errno = ENOMEM;
        ret = -1;
        goto out;
    }
    switch (cmd & GF_CLI_STATUS_MASK) {
        case GF_CLI_STATUS_MEM:
            cli_print_volume_status_mem(dict, notbrick);
            goto cont;
            break;
        case GF_CLI_STATUS_CLIENTS:
            cli_print_volume_status_clients(dict, notbrick);
            goto cont;
            break;
        case GF_CLI_STATUS_CLIENT_LIST:
            cli_print_volume_status_client_list(dict, notbrick);
            goto cont;
            break;
        case GF_CLI_STATUS_INODE:
            cli_print_volume_status_inode(dict, notbrick);
            goto cont;
            break;
        case GF_CLI_STATUS_FD:
            cli_print_volume_status_fd(dict, notbrick);
            goto cont;
            break;
        case GF_CLI_STATUS_CALLPOOL:
            cli_print_volume_status_callpool(dict, notbrick);
            goto cont;
            break;
        case GF_CLI_STATUS_TASKS:
            cli_print_volume_status_tasks(dict);
            goto cont;
            break;
        default:
            break;
    }

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "brick-index-max", &brick_index_max);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "other-count", &other_count);
    if (ret)
        goto out;

    index_max = brick_index_max + other_count;

    ret = dict_get_int32(dict, "type", &type);
    if (ret)
        goto out;

    ret = dict_get_int32(dict, "hot_brick_count", &hot_brick_count);
    if (ret)
        goto out;

    cli_out("Status of volume: %s", volname);

    if ((cmd & GF_CLI_STATUS_DETAIL) == 0) {
        cli_out("%-*s %s  %s  %s  %s", CLI_VOL_STATUS_BRICK_LEN,
                "Gluster process", "TCP Port", "RDMA Port", "Online", "Pid");
        cli_print_line(CLI_BRICK_STATUS_LINE_LEN);
    }
    if (type == GF_CLUSTER_TYPE_TIER) {
        cli_out("Hot Bricks:");
    }
    for (i = 0; i <= index_max; i++) {
        if (type == GF_CLUSTER_TYPE_TIER && i == hot_brick_count) {
            cli_out("Cold Bricks:");
        }
        status.rdma_port = 0;

        snprintf(key, sizeof(key), "brick%d.hostname", i);
        ret = dict_get_str(dict, key, &hostname);
        if (ret)
            continue;

        snprintf(key, sizeof(key), "brick%d.path", i);
        ret = dict_get_str(dict, key, &path);
        if (ret)
            continue;

        /* Brick/not-brick is handled separately here as all
         * types of nodes are contained in the default output
         */
        memset(status.brick, 0, PATH_MAX + 255);
        if (!strcmp(hostname, "NFS Server") ||
            !strcmp(hostname, "Self-heal Daemon") ||
            !strcmp(hostname, "Quota Daemon") ||
            !strcmp(hostname, "Snapshot Daemon") ||
            !strcmp(hostname, "Scrubber Daemon") ||
            !strcmp(hostname, "Bitrot Daemon") ||
            !strcmp(hostname, "Tier Daemon"))
            snprintf(status.brick, PATH_MAX + 255, "%s on %s", hostname, path);
        else {
            snprintf(key, sizeof(key), "brick%d.rdma_port", i);
            ret = dict_get_int32(dict, key, &(status.rdma_port));
            if (ret)
                continue;
            snprintf(status.brick, PATH_MAX + 255, "Brick %s:%s", hostname,
                     path);
        }

        snprintf(key, sizeof(key), "brick%d.port", i);
        ret = dict_get_int32(dict, key, &(status.port));
        if (ret)
            continue;

        snprintf(key, sizeof(key), "brick%d.status", i);
        ret = dict_get_int32(dict, key, &(status.online));
        if (ret)
            continue;

        snprintf(key, sizeof(key), "brick%d.pid", i);
        ret = dict_get_int32(dict, key, &pid);
        if (ret)
            continue;
        if (pid == -1)
            ret = gf_asprintf(&(status.pid_str), "%s", "N/A");
        else
            ret = gf_asprintf(&(status.pid_str), "%d", pid);

        if (ret == -1)
            goto out;

        if ((cmd & GF_CLI_STATUS_DETAIL)) {
            ret = cli_get_detail_status(dict, i, &status);
            if (ret)
                goto out;
            cli_print_line(CLI_BRICK_STATUS_LINE_LEN);
            cli_print_detailed_status(&status);
        } else {
            cli_print_brick_status(&status);
        }

        /* Allocatated memory using gf_asprintf*/
        GF_FREE(status.pid_str);
    }
    cli_out(" ");

    if ((cmd & GF_CLI_STATUS_MASK) == GF_CLI_STATUS_NONE)
        cli_print_volume_status_tasks(dict);
cont:
    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);
    GF_FREE(status.brick);
    if (local && wipe_local) {
        cli_local_wipe(local);
    }

    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int32_t
gf_cli_status_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = -1;
    dict_t *dict = NULL;

    if (!frame || !this || !data)
        goto out;

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_status_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_STATUS_VOLUME, this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning: %d", ret);
    GF_FREE(req.dict.dict_val);
    return ret;
}

int
gf_cli_status_volume_all(call_frame_t *frame, xlator_t *this, void *data)
{
    int i = 0;
    int ret = -1;
    int vol_count = -1;
    uint32_t cmd = 0;
    char key[1024] = {0};
    char *volname = NULL;
    void *vol_dict = NULL;
    dict_t *dict = NULL;
    cli_local_t *local = NULL;

    if (!frame)
        goto out;

    if (!frame->local)
        goto out;

    local = frame->local;

    ret = dict_get_uint32(local->dict, "cmd", &cmd);
    if (ret)
        goto out;

    local->all = _gf_true;

    ret = gf_cli_status_volume(frame, this, data);

    if (ret)
        goto out;

    ret = dict_get_ptr(local->dict, "rsp-dict", &vol_dict);
    if (ret)
        goto out;

    ret = dict_get_int32((dict_t *)vol_dict, "vol_count", &vol_count);
    if (ret) {
        cli_err("Failed to get names of volumes");
        goto out;
    }

    /* remove the "all" flag in cmd */
    cmd &= ~GF_CLI_STATUS_ALL;
    cmd |= GF_CLI_STATUS_VOL;

    if (global_state->mode & GLUSTER_MODE_XML) {
        // TODO: Pass proper op_* values
        ret = cli_xml_output_vol_status_begin(local, 0, 0, NULL);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
            goto xml_end;
        }
    }

    if (vol_count == 0 && !(global_state->mode & GLUSTER_MODE_XML)) {
        cli_err("No volumes present");
        ret = 0;
        goto out;
    }

    for (i = 0; i < vol_count; i++) {
        dict = dict_new();
        if (!dict)
            goto out;

        snprintf(key, sizeof(key), "vol%d", i);
        ret = dict_get_str(vol_dict, key, &volname);
        if (ret)
            goto out;

        ret = dict_set_str(dict, "volname", volname);
        if (ret)
            goto out;

        ret = dict_set_uint32(dict, "cmd", cmd);
        if (ret)
            goto out;

        ret = gf_cli_status_volume(frame, this, dict);
        if (ret)
            goto out;

        dict_unref(dict);
    }

xml_end:
    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_status_end(local);
    }

out:
    if (ret)
        gf_log("cli", GF_LOG_ERROR, "status all failed");

    if (vol_dict)
        dict_unref(vol_dict);

    if (ret && dict)
        dict_unref(dict);

    if (local)
        cli_local_wipe(local);

    if (frame)
        frame->local = NULL;

    return ret;
}

static int
gf_cli_mount_cbk(struct rpc_req *req, struct iovec *iov, int count,
                 void *myframe)
{
    gf1_cli_mount_rsp rsp = {
        0,
    };
    int ret = -1;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_mount_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to mount");

    if (rsp.op_ret == 0) {
        ret = 0;
        cli_out("%s", rsp.path);
    } else {
        /* weird sounding but easy to parse... */
        cli_err("%d : failed with this errno (%s)", rsp.op_errno,
                strerror(rsp.op_errno));
        ret = -1;
    }

out:
    cli_cmd_broadcast_response(ret);
    if (rsp.path) {
        free(rsp.path);
    }
    return ret;
}

int32_t
gf_cli_mount(call_frame_t *frame, xlator_t *this, void *data)
{
    gf1_cli_mount_req req = {
        0,
    };
    int ret = -1;
    void **dataa = data;
    char *label = NULL;
    dict_t *dict = NULL;

    if (!frame || !this || !data)
        goto out;

    label = dataa[0];
    dict = dataa[1];

    req.label = label;
    ret = dict_allocate_and_serialize(dict, &req.dict.dict_val,
                                      &req.dict.dict_len);
    if (ret) {
        ret = -1;
        goto out;
    }

    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_MOUNT,
                         NULL, this, gf_cli_mount_cbk,
                         (xdrproc_t)xdr_gf1_cli_mount_req);

out:
    GF_FREE(req.dict.dict_val);
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

static int
gf_cli_umount_cbk(struct rpc_req *req, struct iovec *iov, int count,
                  void *myframe)
{
    gf1_cli_umount_rsp rsp = {
        0,
    };
    int ret = -1;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf1_cli_umount_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to mount");

    if (rsp.op_ret == 0)
        ret = 0;
    else {
        cli_err("umount failed");
        ret = -1;
    }

out:
    cli_cmd_broadcast_response(ret);
    return ret;
}

int32_t
gf_cli_umount(call_frame_t *frame, xlator_t *this, void *data)
{
    gf1_cli_umount_req req = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;

    if (!frame || !this || !data)
        goto out;

    dict = data;

    ret = dict_get_str(dict, "path", &req.path);
    if (ret == 0)
        ret = dict_get_int32(dict, "lazy", &req.lazy);

    if (ret) {
        ret = -1;
        goto out;
    }

    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog, GLUSTER_CLI_UMOUNT,
                         NULL, this, gf_cli_umount_cbk,
                         (xdrproc_t)xdr_gf1_cli_umount_req);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

void
cmd_heal_volume_statistics_out(dict_t *dict, int brick)
{
    uint64_t num_entries = 0;
    int ret = 0;
    char key[256] = {0};
    char *hostname = NULL;
    uint64_t i = 0;
    uint64_t healed_count = 0;
    uint64_t split_brain_count = 0;
    uint64_t heal_failed_count = 0;
    char *start_time_str = NULL;
    char *end_time_str = NULL;
    char *crawl_type = NULL;
    int progress = -1;

    snprintf(key, sizeof key, "%d-hostname", brick);
    ret = dict_get_str(dict, key, &hostname);
    if (ret)
        goto out;
    cli_out("------------------------------------------------");
    cli_out("\nCrawl statistics for brick no %d", brick);
    cli_out("Hostname of brick %s", hostname);

    snprintf(key, sizeof key, "statistics-%d-count", brick);
    ret = dict_get_uint64(dict, key, &num_entries);
    if (ret)
        goto out;

    for (i = 0; i < num_entries; i++) {
        snprintf(key, sizeof key, "statistics_crawl_type-%d-%" PRIu64, brick,
                 i);
        ret = dict_get_str(dict, key, &crawl_type);
        if (ret)
            goto out;

        snprintf(key, sizeof key, "statistics_healed_cnt-%d-%" PRIu64, brick,
                 i);
        ret = dict_get_uint64(dict, key, &healed_count);
        if (ret)
            goto out;

        snprintf(key, sizeof key, "statistics_sb_cnt-%d-%" PRIu64, brick, i);
        ret = dict_get_uint64(dict, key, &split_brain_count);
        if (ret)
            goto out;
        snprintf(key, sizeof key, "statistics_heal_failed_cnt-%d-%" PRIu64,
                 brick, i);
        ret = dict_get_uint64(dict, key, &heal_failed_count);
        if (ret)
            goto out;
        snprintf(key, sizeof key, "statistics_strt_time-%d-%" PRIu64, brick, i);
        ret = dict_get_str(dict, key, &start_time_str);
        if (ret)
            goto out;
        snprintf(key, sizeof key, "statistics_end_time-%d-%" PRIu64, brick, i);
        ret = dict_get_str(dict, key, &end_time_str);
        if (ret)
            goto out;
        snprintf(key, sizeof key, "statistics_inprogress-%d-%" PRIu64, brick,
                 i);
        ret = dict_get_int32(dict, key, &progress);
        if (ret)
            goto out;

        cli_out("\nStarting time of crawl: %s", start_time_str);
        if (progress == 1)
            cli_out("Crawl is in progress");
        else
            cli_out("Ending time of crawl: %s", end_time_str);

        cli_out("Type of crawl: %s", crawl_type);
        cli_out("No. of entries healed: %" PRIu64, healed_count);
        cli_out("No. of entries in split-brain: %" PRIu64, split_brain_count);
        cli_out("No. of heal failed entries: %" PRIu64, heal_failed_count);
    }

out:
    return;
}

void
cmd_heal_volume_brick_out(dict_t *dict, int brick)
{
    uint64_t num_entries = 0;
    int ret = 0;
    char key[256] = {0};
    char *hostname = NULL;
    char *path = NULL;
    char *status = NULL;
    uint64_t i = 0;
    uint32_t time = 0;
    char timestr[32] = {0};
    char *shd_status = NULL;

    snprintf(key, sizeof key, "%d-hostname", brick);
    ret = dict_get_str(dict, key, &hostname);
    if (ret)
        goto out;
    snprintf(key, sizeof key, "%d-path", brick);
    ret = dict_get_str(dict, key, &path);
    if (ret)
        goto out;
    cli_out("\nBrick %s:%s", hostname, path);

    snprintf(key, sizeof key, "%d-status", brick);
    ret = dict_get_str(dict, key, &status);
    if (status && strlen(status))
        cli_out("Status: %s", status);

    snprintf(key, sizeof key, "%d-shd-status", brick);
    ret = dict_get_str(dict, key, &shd_status);

    if (!shd_status) {
        snprintf(key, sizeof key, "%d-count", brick);
        ret = dict_get_uint64(dict, key, &num_entries);
        cli_out("Number of entries: %" PRIu64, num_entries);

        for (i = 0; i < num_entries; i++) {
            snprintf(key, sizeof key, "%d-%" PRIu64, brick, i);
            ret = dict_get_str(dict, key, &path);
            if (ret)
                continue;
            time = 0;
            snprintf(key, sizeof key, "%d-%" PRIu64 "-time", brick, i);
            ret = dict_get_uint32(dict, key, &time);
            if (ret || !time) {
                cli_out("%s", path);
            } else {
                gf_time_fmt(timestr, sizeof timestr, time, gf_timefmt_FT);
                if (i == 0) {
                    cli_out("at                    path on brick");
                    cli_out("-----------------------------------");
                }
                cli_out("%s %s", timestr, path);
            }
        }
    }

out:
    return;
}

void
cmd_heal_volume_statistics_heal_count_out(dict_t *dict, int brick)
{
    uint64_t num_entries = 0;
    int ret = 0;
    char key[256] = {0};
    char *hostname = NULL;
    char *path = NULL;
    char *status = NULL;
    char *shd_status = NULL;

    snprintf(key, sizeof key, "%d-hostname", brick);
    ret = dict_get_str(dict, key, &hostname);
    if (ret)
        goto out;
    snprintf(key, sizeof key, "%d-path", brick);
    ret = dict_get_str(dict, key, &path);
    if (ret)
        goto out;
    cli_out("\nBrick %s:%s", hostname, path);

    snprintf(key, sizeof key, "%d-status", brick);
    ret = dict_get_str(dict, key, &status);
    if (status && strlen(status))
        cli_out("Status: %s", status);

    snprintf(key, sizeof key, "%d-shd-status", brick);
    ret = dict_get_str(dict, key, &shd_status);

    if (!shd_status) {
        snprintf(key, sizeof key, "%d-hardlinks", brick);
        ret = dict_get_uint64(dict, key, &num_entries);
        if (ret)
            cli_out("No gathered input for this brick");
        else
            cli_out("Number of entries: %" PRIu64, num_entries);
    }

out:
    return;
}

int
gf_is_cli_heal_get_command(gf_xl_afr_op_t heal_op)
{
    /* If the command is get command value is 1 otherwise 0, for
       invalid commands -1 */
    int get_cmds[GF_SHD_OP_HEAL_DISABLE + 1] = {
        [GF_SHD_OP_INVALID] = -1,
        [GF_SHD_OP_HEAL_INDEX] = 0,
        [GF_SHD_OP_HEAL_FULL] = 0,
        [GF_SHD_OP_INDEX_SUMMARY] = 1,
        [GF_SHD_OP_HEALED_FILES] = 1,
        [GF_SHD_OP_HEAL_FAILED_FILES] = 1,
        [GF_SHD_OP_SPLIT_BRAIN_FILES] = 1,
        [GF_SHD_OP_STATISTICS] = 1,
        [GF_SHD_OP_STATISTICS_HEAL_COUNT] = 1,
        [GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA] = 1,
        [GF_SHD_OP_HEAL_ENABLE] = 0,
        [GF_SHD_OP_HEAL_DISABLE] = 0,
    };

    if (heal_op > GF_SHD_OP_INVALID && heal_op <= GF_SHD_OP_HEAL_DISABLE)
        return get_cmds[heal_op] == 1;
    return _gf_false;
}

int
gf_cli_heal_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    cli_local_t *local = NULL;
    char *volname = NULL;
    call_frame_t *frame = NULL;
    dict_t *dict = NULL;
    int brick_count = 0;
    int i = 0;
    gf_xl_afr_op_t heal_op = GF_SHD_OP_INVALID;
    char *operation = NULL;
    char *substr = NULL;
    char *heal_op_str = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status) {
        goto out;
    }

    frame = myframe;

    GF_ASSERT(frame->local);

    local = frame->local;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    ret = dict_get_int32(local->dict, "heal-op", (int32_t *)&heal_op);
    // TODO: Proper XML output
    //#if (HAVE_LIB_XML)
    //        if (global_state->mode & GLUSTER_MODE_XML) {
    //                ret = cli_xml_output_dict ("volHeal", dict, rsp.op_ret,
    //                                           rsp.op_errno, rsp.op_errstr);
    //                if (ret)
    //                        gf_log ("cli", GF_LOG_ERROR,
    //                                "Error outputting to xml");
    //                goto out;
    //        }
    //#endif

    ret = dict_get_str(local->dict, "volname", &volname);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "failed to get volname");
        goto out;
    }

    gf_log("cli", GF_LOG_INFO, "Received resp to heal volume");

    operation = "Gathering ";
    substr = "";
    switch (heal_op) {
        case GF_SHD_OP_HEAL_INDEX:
            operation = "Launching heal operation ";
            heal_op_str = "to perform index self heal";
            substr =
                "\nUse heal info commands to check"
                " status.";
            break;
        case GF_SHD_OP_HEAL_FULL:
            operation = "Launching heal operation ";
            heal_op_str = "to perform full self heal";
            substr =
                "\nUse heal info commands to check"
                " status.";
            break;
        case GF_SHD_OP_INDEX_SUMMARY:
            heal_op_str = "list of entries to be healed";
            break;
        case GF_SHD_OP_HEALED_FILES:
            heal_op_str = "list of healed entries";
            break;
        case GF_SHD_OP_HEAL_FAILED_FILES:
            heal_op_str = "list of heal failed entries";
            break;
        case GF_SHD_OP_SPLIT_BRAIN_FILES:
            heal_op_str = "list of split brain entries";
            break;
        case GF_SHD_OP_STATISTICS:
            heal_op_str = "crawl statistics";
            break;
        case GF_SHD_OP_STATISTICS_HEAL_COUNT:
            heal_op_str = "count of entries to be healed";
            break;
        case GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA:
            heal_op_str = "count of entries to be healed per replica";
            break;
        /* The below 4 cases are never hit; they're coded only to make
         * compiler warnings go away.*/
        case GF_SHD_OP_SBRAIN_HEAL_FROM_BIGGER_FILE:
        case GF_SHD_OP_SBRAIN_HEAL_FROM_LATEST_MTIME:
        case GF_SHD_OP_SBRAIN_HEAL_FROM_BRICK:
        case GF_SHD_OP_HEAL_SUMMARY:
            break;

        case GF_SHD_OP_INVALID:
            heal_op_str = "invalid heal op";
            break;
        case GF_SHD_OP_HEAL_ENABLE:
            operation = "";
            heal_op_str = "Enable heal";
            break;
        case GF_SHD_OP_HEAL_DISABLE:
            operation = "";
            heal_op_str = "Disable heal";
            break;
        case GF_SHD_OP_GRANULAR_ENTRY_HEAL_ENABLE:
            operation = "";
            heal_op_str = "Enable granular entry heal";
            break;
        case GF_SHD_OP_GRANULAR_ENTRY_HEAL_DISABLE:
            operation = "";
            heal_op_str = "Disable granular entry heal";
            break;
    }

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, "")) {
            cli_err("%s%s on volume %s has been unsuccessful:", operation,
                    heal_op_str, volname);
            cli_err("%s", rsp.op_errstr);
        }
        ret = rsp.op_ret;
        goto out;
    } else {
        cli_out("%s%s on volume %s has been successful %s", operation,
                heal_op_str, volname, substr);
    }

    ret = rsp.op_ret;
    if (!gf_is_cli_heal_get_command(heal_op))
        goto out;

    dict = dict_new();
    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret) {
        gf_log("", GF_LOG_ERROR, "Unable to allocate memory");
        goto out;
    }

    ret = dict_get_int32(dict, "count", &brick_count);
    if (ret)
        goto out;

    if (!brick_count) {
        cli_err("All bricks of volume %s are down.", volname);
        ret = -1;
        goto out;
    }

    switch (heal_op) {
        case GF_SHD_OP_STATISTICS:
            for (i = 0; i < brick_count; i++)
                cmd_heal_volume_statistics_out(dict, i);
            break;
        case GF_SHD_OP_STATISTICS_HEAL_COUNT:
        case GF_SHD_OP_STATISTICS_HEAL_COUNT_PER_REPLICA:
            for (i = 0; i < brick_count; i++)
                cmd_heal_volume_statistics_heal_count_out(dict, i);
            break;
        case GF_SHD_OP_INDEX_SUMMARY:
        case GF_SHD_OP_HEALED_FILES:
        case GF_SHD_OP_HEAL_FAILED_FILES:
        case GF_SHD_OP_SPLIT_BRAIN_FILES:
            for (i = 0; i < brick_count; i++)
                cmd_heal_volume_brick_out(dict, i);
            break;
        default:
            break;
    }

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    if (dict)
        dict_unref(dict);
    return ret;
}

int32_t
gf_cli_heal_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = 0;
    dict_t *dict = NULL;

    if (!frame || !this || !data) {
        ret = -1;
        goto out;
    }

    dict = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_heal_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, dict,
                          GLUSTER_CLI_HEAL_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_statedump_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                            void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char msg[1024] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }
    gf_log("cli", GF_LOG_DEBUG, "Received response to statedump");
    if (rsp.op_ret)
        snprintf(msg, sizeof(msg), "%s", rsp.op_errstr);
    else
        snprintf(msg, sizeof(msg), "Volume statedump successful");

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_str("volStatedump", msg, rsp.op_ret, rsp.op_errno,
                                 rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("volume statedump: failed: %s", msg);
    else
        cli_out("volume statedump: success");
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int32_t
gf_cli_statedump_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    dict_t *options = NULL;
    int ret = -1;

    if (!frame || !this || !data)
        goto out;

    options = data;

    ret = cli_to_glusterd(
        &req, frame, gf_cli_statedump_volume_cbk, (xdrproc_t)xdr_gf_cli_req,
        options, GLUSTER_CLI_STATEDUMP_VOLUME, this, cli_rpc_prog, NULL);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_list_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    int ret = -1;
    gf_cli_rsp rsp = {
        0,
    };
    dict_t *dict = NULL;
    int vol_count = 0;
    ;
    char *volname = NULL;
    char key[1024] = {
        0,
    };
    int i = 0;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();
    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to allocate memory");
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_list(dict, rsp.op_ret, rsp.op_errno,
                                      rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (rsp.op_ret)
        cli_err("%s", rsp.op_errstr);
    else {
        ret = dict_get_int32(dict, "count", &vol_count);
        if (ret)
            goto out;

        if (vol_count == 0) {
            cli_err("No volumes present in cluster");
            goto out;
        }
        for (i = 0; i < vol_count; i++) {
            snprintf(key, sizeof(key), "volume%d", i);
            ret = dict_get_str(dict, key, &volname);
            if (ret)
                goto out;
            cli_out("%s", volname);
        }
    }

    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);

    if (dict)
        dict_unref(dict);
    return ret;
}

int32_t
gf_cli_list_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    int ret = -1;
    gf_cli_req req = {{
        0,
    }};

    if (!frame || !this)
        goto out;

    ret = cli_cmd_submit(NULL, &req, frame, cli_rpc_prog,
                         GLUSTER_CLI_LIST_VOLUME, NULL, this,
                         gf_cli_list_volume_cbk, (xdrproc_t)xdr_gf_cli_req);

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);
    return ret;
}

int32_t
gf_cli_clearlocks_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                             void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    char *lk_summary = NULL;
    char *volname = NULL;
    dict_t *dict = NULL;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }
    gf_log("cli", GF_LOG_DEBUG, "Received response to clear-locks");

    if (rsp.op_ret) {
        cli_err("Volume clear-locks unsuccessful");
        cli_err("%s", rsp.op_errstr);

    } else {
        if (!rsp.dict.dict_len) {
            cli_err("Possibly no locks cleared");
            ret = 0;
            goto out;
        }

        dict = dict_new();

        if (!dict) {
            ret = -1;
            goto out;
        }

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Unable to serialize response dictionary");
            goto out;
        }

        ret = dict_get_str(dict, "volname", &volname);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Unable to get volname "
                   "from dictionary");
            goto out;
        }

        ret = dict_get_str(dict, "lk-summary", &lk_summary);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Unable to get lock "
                   "summary from dictionary");
            goto out;
        }
        cli_out("Volume clear-locks successful");
        cli_out("%s", lk_summary);
    }

    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int32_t
gf_cli_clearlocks_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    dict_t *options = NULL;
    int ret = -1;

    if (!frame || !this || !data)
        goto out;

    options = data;

    ret = cli_to_glusterd(
        &req, frame, gf_cli_clearlocks_volume_cbk, (xdrproc_t)xdr_gf_cli_req,
        options, GLUSTER_CLI_CLRLOCKS_VOLUME, this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
cli_snapshot_remove_reply(gf_cli_rsp *rsp, dict_t *dict, call_frame_t *frame)
{
    int32_t ret = -1;
    char *snap_name = NULL;
    int32_t delete_cmd = -1;
    cli_local_t *local = NULL;

    GF_ASSERT(frame);
    GF_ASSERT(rsp);
    GF_ASSERT(dict);

    local = frame->local;

    ret = dict_get_int32(dict, "sub-cmd", &delete_cmd);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not get sub-cmd");
        goto end;
    }

    if ((global_state->mode & GLUSTER_MODE_XML) &&
        (delete_cmd == GF_SNAP_DELETE_TYPE_SNAP)) {
        ret = cli_xml_output_snap_delete_begin(local, rsp->op_ret,
                                               rsp->op_errno, rsp->op_errstr);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to create "
                   "xml output for delete");
            goto end;
        }
    }

    if (rsp->op_ret && !(global_state->mode & GLUSTER_MODE_XML)) {
        cli_err("snapshot delete: failed: %s",
                rsp->op_errstr ? rsp->op_errstr
                               : "Please check log file for details");
        ret = rsp->op_ret;
        goto out;
    }

    if (delete_cmd == GF_SNAP_DELETE_TYPE_ALL ||
        delete_cmd == GF_SNAP_DELETE_TYPE_VOL) {
        local = ((call_frame_t *)frame)->local;
        if (!local) {
            ret = -1;
            gf_log("cli", GF_LOG_ERROR, "frame->local is NULL");
            goto out;
        }

        /* During first call back of snapshot delete of type
         * ALL and VOL, We will get the snapcount and snapnames.
         * Hence to make the subsequent rpc calls for individual
         * snapshot delete, We need to save it in local dictionary.
         */
        dict_copy(dict, local->dict);
        ret = 0;
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_snapshot_delete(local, dict, rsp);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to create "
                   "xml output for snapshot delete command");
            goto out;
        }
        /* Error out in case of the op already failed */
        if (rsp->op_ret) {
            ret = rsp->op_ret;
            goto out;
        }
    } else {
        ret = dict_get_str(dict, "snapname", &snap_name);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Failed to get snapname");
            goto out;
        }

        cli_out("snapshot delete: %s: snap removed successfully", snap_name);
    }
    ret = 0;

out:
    if ((global_state->mode & GLUSTER_MODE_XML) &&
        (delete_cmd == GF_SNAP_DELETE_TYPE_SNAP)) {
        ret = cli_xml_output_snap_delete_end(local);
    }
end:
    return ret;
}

int
cli_snapshot_config_display(dict_t *dict, gf_cli_rsp *rsp)
{
    char buf[PATH_MAX] = "";
    char *volname = NULL;
    int ret = -1;
    int config_command = 0;
    uint64_t value = 0;
    uint64_t hard_limit = 0;
    uint64_t soft_limit = 0;
    uint64_t i = 0;
    uint64_t voldisplaycount = 0;
    char *auto_delete = NULL;
    char *snap_activate = NULL;

    GF_ASSERT(dict);
    GF_ASSERT(rsp);

    if (rsp->op_ret) {
        cli_err("Snapshot Config : failed: %s",
                rsp->op_errstr ? rsp->op_errstr
                               : "Please check log file for details");
        ret = rsp->op_ret;
        goto out;
    }

    ret = dict_get_int32(dict, "config-command", &config_command);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not fetch config type");
        goto out;
    }

    ret = dict_get_str(dict, "volname", &volname);
    /* Ignore the error, as volname is optional */

    if (!volname) {
        volname = "System";
    }

    ret = dict_get_uint64(dict, "snap-max-hard-limit", &hard_limit);
    /* Ignore the error, as the key specified is optional */
    ret = dict_get_uint64(dict, "snap-max-soft-limit", &soft_limit);

    ret = dict_get_str(dict, "auto-delete", &auto_delete);

    ret = dict_get_str(dict, "snap-activate-on-create", &snap_activate);

    if (!hard_limit && !soft_limit &&
        config_command != GF_SNAP_CONFIG_DISPLAY && !auto_delete &&
        !snap_activate) {
        ret = -1;
        gf_log(THIS->name, GF_LOG_ERROR, "Could not fetch config-key");
        goto out;
    }

    switch (config_command) {
        case GF_SNAP_CONFIG_TYPE_SET:
            if (hard_limit && soft_limit) {
                cli_out(
                    "snapshot config: snap-max-hard-limit "
                    "& snap-max-soft-limit for system set "
                    "successfully");
            } else if (hard_limit) {
                cli_out(
                    "snapshot config: snap-max-hard-limit "
                    "for %s set successfully",
                    volname);
            } else if (soft_limit) {
                cli_out(
                    "snapshot config: snap-max-soft-limit "
                    "for %s set successfully",
                    volname);
            } else if (auto_delete) {
                cli_out(
                    "snapshot config: auto-delete "
                    "successfully set");
            } else if (snap_activate) {
                cli_out(
                    "snapshot config: activate-on-create "
                    "successfully set");
            }
            break;

        case GF_SNAP_CONFIG_DISPLAY:
            cli_out("\nSnapshot System Configuration:");
            ret = dict_get_uint64(dict, "snap-max-hard-limit", &value);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Could not fetch "
                       "snap_max_hard_limit for %s",
                       volname);
                ret = -1;
                goto out;
            }
            cli_out("snap-max-hard-limit : %" PRIu64, value);

            ret = dict_get_uint64(dict, "snap-max-soft-limit", &soft_limit);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Could not fetch "
                       "snap-max-soft-limit for %s",
                       volname);
                ret = -1;
                goto out;
            }
            cli_out("snap-max-soft-limit : %" PRIu64 "%%", soft_limit);

            cli_out("auto-delete : %s", auto_delete);

            cli_out("activate-on-create : %s\n", snap_activate);

            cli_out("Snapshot Volume Configuration:");

            ret = dict_get_uint64(dict, "voldisplaycount", &voldisplaycount);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Could not fetch voldisplaycount");
                ret = -1;
                goto out;
            }

            for (i = 0; i < voldisplaycount; i++) {
                snprintf(buf, sizeof(buf), "volume%" PRIu64 "-volname", i);
                ret = dict_get_str(dict, buf, &volname);
                if (ret) {
                    gf_log("cli", GF_LOG_ERROR,
                           "Could not fetch "
                           " %s",
                           buf);
                    ret = -1;
                    goto out;
                }
                cli_out("\nVolume : %s", volname);

                snprintf(buf, sizeof(buf),
                         "volume%" PRIu64 "-snap-max-hard-limit", i);
                ret = dict_get_uint64(dict, buf, &value);
                if (ret) {
                    gf_log("cli", GF_LOG_ERROR,
                           "Could not fetch "
                           " %s",
                           buf);
                    ret = -1;
                    goto out;
                }
                cli_out("snap-max-hard-limit : %" PRIu64, value);

                snprintf(buf, sizeof(buf),
                         "volume%" PRIu64 "-active-hard-limit", i);
                ret = dict_get_uint64(dict, buf, &value);
                if (ret) {
                    gf_log("cli", GF_LOG_ERROR,
                           "Could not fetch"
                           " effective snap_max_hard_limit for "
                           "%s",
                           volname);
                    ret = -1;
                    goto out;
                }
                cli_out("Effective snap-max-hard-limit : %" PRIu64, value);

                snprintf(buf, sizeof(buf),
                         "volume%" PRIu64 "-snap-max-soft-limit", i);
                ret = dict_get_uint64(dict, buf, &value);
                if (ret) {
                    gf_log("cli", GF_LOG_ERROR,
                           "Could not fetch "
                           " %s",
                           buf);
                    ret = -1;
                    goto out;
                }
                cli_out("Effective snap-max-soft-limit : %" PRIu64
                        " "
                        "(%" PRIu64 "%%)",
                        value, soft_limit);
            }
            break;
        default:
            break;
    }

    ret = 0;
out:
    return ret;
}

/* This function is used to print the volume related information
 * of a snap.
 *
 * arg - 0, dict       : Response Dictionary.
 * arg - 1, prefix str : snaplist.snap{0..}.vol{0..}.*
 */
int
cli_get_each_volinfo_in_snap(dict_t *dict, char *keyprefix,
                             gf_boolean_t snap_driven)
{
    char key[PATH_MAX] = "";
    char *get_buffer = NULL;
    int value = 0;
    int ret = -1;
    char indent[5] = "\t";
    char *volname = NULL;

    GF_ASSERT(dict);
    GF_ASSERT(keyprefix);

    if (snap_driven) {
        ret = snprintf(key, sizeof(key), "%s.volname", keyprefix);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &get_buffer);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key);
            goto out;
        }
        cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Snap Volume Name", ":",
                get_buffer);

        ret = snprintf(key, sizeof(key), "%s.origin-volname", keyprefix);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &volname);
        if (ret) {
            gf_log("cli", GF_LOG_WARNING, "Failed to get %s", key);
            cli_out("%-12s", "Origin:");
        }
        cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Origin Volume name", ":",
                volname);

        ret = snprintf(key, sizeof(key), "%s.snapcount", keyprefix);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_int32(dict, key, &value);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key);
            goto out;
        }
        cli_out("%s%s %s      %s %d", indent, "Snaps taken for", volname, ":",
                value);

        ret = snprintf(key, sizeof(key), "%s.snaps-available", keyprefix);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_int32(dict, key, &value);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key);
            goto out;
        }
        cli_out("%s%s %s  %s %d", indent, "Snaps available for", volname, ":",
                value);
    }

    ret = snprintf(key, sizeof(key), "%s.vol-status", keyprefix);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key, &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get %s", key);
        goto out;
    }
    cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Status", ":", get_buffer);
out:
    return ret;
}

/* This function is used to print snap related information
 * arg - 0, dict       : Response dictionary.
 * arg - 1, prefix_str : snaplist.snap{0..}.*
 */
int
cli_get_volinfo_in_snap(dict_t *dict, char *keyprefix)
{
    char key[PATH_MAX] = "";
    int i = 0;
    int volcount = 0;
    int ret = -1;

    GF_ASSERT(dict);
    GF_ASSERT(keyprefix);

    ret = snprintf(key, sizeof(key), "%s.vol-count", keyprefix);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_int32(dict, key, &volcount);
    for (i = 1; i <= volcount; i++) {
        ret = snprintf(key, sizeof(key), "%s.vol%d", keyprefix, i);
        if (ret < 0) {
            goto out;
        }
        ret = cli_get_each_volinfo_in_snap(dict, key, _gf_true);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Could not list "
                   "details of volume in a snap");
            goto out;
        }
        cli_out(" ");
    }

out:
    return ret;
}

int
cli_get_each_snap_info(dict_t *dict, char *prefix_str, gf_boolean_t snap_driven)
{
    char key_buffer[PATH_MAX] = "";
    char *get_buffer = NULL;
    int ret = -1;
    char indent[5] = "";

    GF_ASSERT(dict);
    GF_ASSERT(prefix_str);

    if (!snap_driven)
        strcat(indent, "\t");

    ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snapname", prefix_str);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key_buffer, &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to fetch snapname %s ", key_buffer);
        goto out;
    }
    cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Snapshot", ":", get_buffer);

    ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snap-id", prefix_str);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key_buffer, &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to fetch snap-id %s ", key_buffer);
        goto out;
    }
    cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Snap UUID", ":", get_buffer);

    ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snap-desc", prefix_str);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key_buffer, &get_buffer);
    if (!ret) {
        /* Ignore error for description */
        cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Description", ":",
                get_buffer);
    }

    ret = snprintf(key_buffer, sizeof(key_buffer), "%s.snap-time", prefix_str);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key_buffer, &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to fetch snap-time %s ",
               prefix_str);
        goto out;
    }
    cli_out("%s" INDENT_MAIN_HEAD "%s", indent, "Created", ":", get_buffer);

    if (snap_driven) {
        cli_out("%-12s", "Snap Volumes:\n");
        ret = cli_get_volinfo_in_snap(dict, prefix_str);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Unable to list details "
                   "of the snaps");
            goto out;
        }
    }
out:
    return ret;
}

/* This is a generic function to print snap related information.
 * arg - 0, dict : Response Dictionary
 */
int
cli_call_snapshot_info(dict_t *dict, gf_boolean_t bool_snap_driven)
{
    int snap_count = 0;
    char key[PATH_MAX] = "";
    int ret = -1;
    int i = 0;

    GF_ASSERT(dict);

    ret = dict_get_int32(dict, "snapcount", &snap_count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to get snapcount");
        goto out;
    }

    if (snap_count == 0) {
        cli_out("No snapshots present");
    }

    for (i = 1; i <= snap_count; i++) {
        ret = snprintf(key, sizeof(key), "snap%d", i);
        if (ret < 0) {
            goto out;
        }
        ret = cli_get_each_snap_info(dict, key, bool_snap_driven);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Unable to print snap details");
            goto out;
        }
    }
out:
    return ret;
}

int
cli_get_snaps_in_volume(dict_t *dict)
{
    int ret = -1;
    int i = 0;
    int count = 0;
    int avail = 0;
    char key[PATH_MAX] = "";
    char *get_buffer = NULL;

    GF_ASSERT(dict);

    ret = dict_get_str(dict, "origin-volname", &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not fetch origin-volname");
        goto out;
    }
    cli_out(INDENT_MAIN_HEAD "%s", "Volume Name", ":", get_buffer);

    ret = dict_get_int32(dict, "snapcount", &avail);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not fetch snapcount");
        goto out;
    }
    cli_out(INDENT_MAIN_HEAD "%d", "Snaps Taken", ":", avail);

    ret = dict_get_int32(dict, "snaps-available", &count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not fetch snaps-available");
        goto out;
    }
    cli_out(INDENT_MAIN_HEAD "%d", "Snaps Available", ":", count);

    for (i = 1; i <= avail; i++) {
        snprintf(key, sizeof(key), "snap%d", i);
        ret = cli_get_each_snap_info(dict, key, _gf_false);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Unable to print snap details");
            goto out;
        }

        ret = snprintf(key, sizeof(key), "snap%d.vol1", i);
        if (ret < 0) {
            goto out;
        }
        ret = cli_get_each_volinfo_in_snap(dict, key, _gf_false);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Could not get volume "
                   "related information");
            goto out;
        }

        cli_out(" ");
    }
out:
    return ret;
}

int
cli_snapshot_list(dict_t *dict)
{
    int snapcount = 0;
    char key[PATH_MAX] = "";
    int ret = -1;
    int i = 0;
    char *get_buffer = NULL;

    GF_ASSERT(dict);

    ret = dict_get_int32(dict, "snapcount", &snapcount);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not fetch snap count");
        goto out;
    }

    if (snapcount == 0) {
        cli_out("No snapshots present");
    }

    for (i = 1; i <= snapcount; i++) {
        ret = snprintf(key, sizeof(key), "snapname%d", i);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &get_buffer);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Could not get %s ", key);
            goto out;
        } else {
            cli_out("%s", get_buffer);
        }
    }
out:
    return ret;
}

int
cli_get_snap_volume_status(dict_t *dict, char *key_prefix)
{
    int ret = -1;
    char key[PATH_MAX] = "";
    char *buffer = NULL;
    int brickcount = 0;
    int i = 0;
    int pid = 0;

    GF_ASSERT(dict);
    GF_ASSERT(key_prefix);

    ret = snprintf(key, sizeof(key), "%s.brickcount", key_prefix);
    if (ret < 0) {
        goto out;
    }
    ret = dict_get_int32(dict, key, &brickcount);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to fetch brickcount");
        goto out;
    }

    for (i = 0; i < brickcount; i++) {
        ret = snprintf(key, sizeof(key), "%s.brick%d.path", key_prefix, i);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &buffer);
        if (ret) {
            gf_log("cli", GF_LOG_INFO, "Unable to get Brick Path");
            continue;
        }
        cli_out("\n\t%-17s %s   %s", "Brick Path", ":", buffer);

        ret = snprintf(key, sizeof(key), "%s.brick%d.vgname", key_prefix, i);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &buffer);
        if (ret) {
            gf_log("cli", GF_LOG_INFO, "Unable to get Volume Group");
            cli_out("\t%-17s %s   %s", "Volume Group", ":", "N/A");
        } else
            cli_out("\t%-17s %s   %s", "Volume Group", ":", buffer);

        ret = snprintf(key, sizeof(key), "%s.brick%d.status", key_prefix, i);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &buffer);
        if (ret) {
            gf_log("cli", GF_LOG_INFO, "Unable to get Brick Running");
            cli_out("\t%-17s %s   %s", "Brick Running", ":", "N/A");
        } else
            cli_out("\t%-17s %s   %s", "Brick Running", ":", buffer);

        ret = snprintf(key, sizeof(key), "%s.brick%d.pid", key_prefix, i);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_int32(dict, key, &pid);
        if (ret) {
            gf_log("cli", GF_LOG_INFO, "Unable to get pid");
            cli_out("\t%-17s %s   %s", "Brick PID", ":", "N/A");
        } else
            cli_out("\t%-17s %s   %d", "Brick PID", ":", pid);

        ret = snprintf(key, sizeof(key), "%s.brick%d.data", key_prefix, i);
        if (ret < 0) {
            goto out;
        }

        ret = dict_get_str(dict, key, &buffer);
        if (ret) {
            gf_log("cli", GF_LOG_INFO, "Unable to get Data Percent");
            cli_out("\t%-17s %s   %s", "Data Percentage", ":", "N/A");
        } else
            cli_out("\t%-17s %s   %s", "Data Percentage", ":", buffer);

        ret = snprintf(key, sizeof(key), "%s.brick%d.lvsize", key_prefix, i);
        if (ret < 0) {
            goto out;
        }
        ret = dict_get_str(dict, key, &buffer);
        if (ret) {
            gf_log("cli", GF_LOG_INFO, "Unable to get LV Size");
            cli_out("\t%-17s %s   %s", "LV Size", ":", "N/A");
        } else
            cli_out("\t%-17s %s   %s", "LV Size", ":", buffer);
    }

    ret = 0;
out:
    return ret;
}

int
cli_get_single_snap_status(dict_t *dict, char *keyprefix)
{
    int ret = -1;
    char key[PATH_MAX] = "";
    int i = 0;
    int volcount = 0;
    char *get_buffer = NULL;

    GF_ASSERT(dict);
    GF_ASSERT(keyprefix);

    ret = snprintf(key, sizeof(key), "%s.snapname", keyprefix);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key, &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to get snapname");
        goto out;
    }
    cli_out("\nSnap Name : %s", get_buffer);

    ret = snprintf(key, sizeof(key), "%s.uuid", keyprefix);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key, &get_buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to get snap UUID");
        goto out;
    }
    cli_out("Snap UUID : %s", get_buffer);

    ret = snprintf(key, sizeof(key), "%s.volcount", keyprefix);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_int32(dict, key, &volcount);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Unable to get volume count");
        goto out;
    }

    for (i = 0; i < volcount; i++) {
        ret = snprintf(key, sizeof(key), "%s.vol%d", keyprefix, i);
        if (ret < 0) {
            goto out;
        }

        ret = cli_get_snap_volume_status(dict, key);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Could not get snap volume status");
            goto out;
        }
    }
out:
    return ret;
}

int32_t
cli_populate_req_dict_for_delete(dict_t *snap_dict, dict_t *dict, size_t index)
{
    int32_t ret = -1;
    char key[PATH_MAX] = "";
    char *buffer = NULL;

    GF_ASSERT(snap_dict);
    GF_ASSERT(dict);

    ret = dict_set_int32(snap_dict, "sub-cmd", GF_SNAP_DELETE_TYPE_ITER);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Could not save command "
               "type in snap dictionary");
        goto out;
    }

    ret = snprintf(key, sizeof(key), "snapname%zu", index);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key, &buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get snapname");
        goto out;
    }

    ret = dict_set_dynstr_with_alloc(snap_dict, "snapname", buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to save snapname");
        goto out;
    }

    ret = dict_set_int32(snap_dict, "type", GF_SNAP_OPTION_TYPE_DELETE);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to save command type");
        goto out;
    }

    ret = dict_set_dynstr_with_alloc(snap_dict, "cmd-str", "snapshot delete");
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not save command string as delete");
        goto out;
    }
out:
    return ret;
}

int
cli_populate_req_dict_for_status(dict_t *snap_dict, dict_t *dict, int index)
{
    int ret = -1;
    char key[PATH_MAX] = "";
    char *buffer = NULL;

    GF_ASSERT(snap_dict);
    GF_ASSERT(dict);

    ret = dict_set_uint32(snap_dict, "sub-cmd", GF_SNAP_STATUS_TYPE_ITER);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Could not save command "
               "type in snap dict");
        goto out;
    }

    ret = snprintf(key, sizeof(key), "status.snap%d.snapname", index);
    if (ret < 0) {
        goto out;
    }

    ret = dict_get_str(dict, key, &buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not get snapname");
        goto out;
    }

    ret = dict_set_str(snap_dict, "snapname", buffer);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Could not save snapname "
               "in snap dict");
        goto out;
    }

    ret = dict_set_int32(snap_dict, "type", GF_SNAP_OPTION_TYPE_STATUS);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not save command type");
        goto out;
    }

    ret = dict_set_dynstr_with_alloc(snap_dict, "cmd-str", "snapshot status");
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not save command string as status");
        goto out;
    }

    ret = dict_set_int32(snap_dict, "hold_vol_locks", _gf_false);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Setting volume lock flag failed");
        goto out;
    }

out:
    return ret;
}

int
cli_snapshot_status(dict_t *dict, gf_cli_rsp *rsp, call_frame_t *frame)
{
    char key[PATH_MAX] = "";
    int ret = -1;
    int status_cmd = -1;
    cli_local_t *local = NULL;

    GF_ASSERT(dict);
    GF_ASSERT(rsp);
    GF_ASSERT(frame);

    local = ((call_frame_t *)frame)->local;
    if (!local) {
        gf_log("cli", GF_LOG_ERROR, "frame->local is NULL");
        goto out;
    }

    if (rsp->op_ret) {
        if (rsp->op_errstr) {
            ret = dict_set_dynstr_with_alloc(local->dict, "op_err_str",
                                             rsp->op_errstr);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Failed to set "
                       "op_errstr in local dictionary");
                goto out;
            }
        }
        ret = rsp->op_ret;
        goto out;
    }

    ret = dict_get_int32(dict, "sub-cmd", &status_cmd);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not fetch status type");
        goto out;
    }

    if ((status_cmd != GF_SNAP_STATUS_TYPE_SNAP) &&
        (status_cmd != GF_SNAP_STATUS_TYPE_ITER)) {
        dict_copy(dict, local->dict);
        goto out;
    }

    ret = snprintf(key, sizeof(key), "status.snap0");
    if (ret < 0) {
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_snapshot_status_single_snap(local, dict, key);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to create "
                   "xml output for snapshot status");
            goto out;
        }
    } else {
        ret = cli_get_single_snap_status(dict, key);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Could not fetch "
                   "status of snap");
            goto out;
        }
    }

    ret = 0;
out:
    return ret;
}

int
gf_cli_generate_snapshot_event(gf_cli_rsp *rsp, dict_t *dict, int32_t type,
                               char *snap_name, char *volname, char *snap_uuid,
                               char *clone_name)
{
    int ret = -1;
    int config_command = 0;
    int32_t delete_cmd = -1;
    uint64_t hard_limit = 0;
    uint64_t soft_limit = 0;
    char *auto_delete = NULL;
    char *snap_activate = NULL;
    char msg[PATH_MAX] = {
        0,
    };
    char option[512] = {
        0,
    };

    GF_VALIDATE_OR_GOTO("cli", dict, out);
    GF_VALIDATE_OR_GOTO("cli", rsp, out);

    switch (type) {
        case GF_SNAP_OPTION_TYPE_CREATE:
            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            if (!volname) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get volume name");
                goto out;
            }

            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_CREATE_FAILED,
                         "snapshot_name=%s;volume_name=%s;error=%s", snap_name,
                         volname,
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            if (!snap_uuid) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid");
                goto out;
            }

            gf_event(EVENT_SNAPSHOT_CREATED,
                     "snapshot_name=%s;"
                     "volume_name=%s;snapshot_uuid=%s",
                     snap_name, volname, snap_uuid);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_ACTIVATE:
            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_ACTIVATE_FAILED,
                         "snapshot_name=%s;error=%s", snap_name,
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            if (!snap_uuid) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid");
                goto out;
            }

            gf_event(EVENT_SNAPSHOT_ACTIVATED,
                     "snapshot_name=%s;"
                     "snapshot_uuid=%s",
                     snap_name, snap_uuid);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_DEACTIVATE:
            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_DEACTIVATE_FAILED,
                         "snapshot_name=%s;error=%s", snap_name,
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            if (!snap_uuid) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid");
                goto out;
            }

            gf_event(EVENT_SNAPSHOT_DEACTIVATED,
                     "snapshot_name=%s;"
                     "snapshot_uuid=%s",
                     snap_name, snap_uuid);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_RESTORE:
            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_RESTORE_FAILED,
                         "snapshot_name=%s;error=%s", snap_name,
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            if (!snap_uuid) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid");
                goto out;
            }

            if (!volname) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR, "Failed to get volname");
                goto out;
            }

            gf_event(EVENT_SNAPSHOT_RESTORED,
                     "snapshot_name=%s;"
                     "snapshot_uuid=%s;volume_name=%s",
                     snap_name, snap_uuid, volname);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_DELETE:
            ret = dict_get_int32(dict, "sub-cmd", &delete_cmd);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Could not get sub-cmd");
                goto out;
            }

            /*
             * Need not generate any event (success or failure) for delete *
             * all, as it will trigger individual delete for all snapshots *
             */
            if (delete_cmd == GF_SNAP_DELETE_TYPE_ALL) {
                ret = 0;
                break;
            }

            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_DELETE_FAILED,
                         "snapshot_name=%s;error=%s", snap_name,
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            if (!snap_uuid) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid");
                goto out;
            }

            gf_event(EVENT_SNAPSHOT_DELETED,
                     "snapshot_name=%s;"
                     "snapshot_uuid=%s",
                     snap_name, snap_uuid);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_CLONE:
            if (!clone_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get clone name");
                goto out;
            }

            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snapname name");
                goto out;
            }

            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_CLONE_FAILED,
                         "snapshot_name=%s;clone_name=%s;"
                         "error=%s",
                         snap_name, clone_name,
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            if (!snap_uuid) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap uuid");
                goto out;
            }

            gf_event(EVENT_SNAPSHOT_CLONED,
                     "snapshot_name=%s;"
                     "clone_name=%s;clone_uuid=%s",
                     snap_name, clone_name, snap_uuid);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_CONFIG:
            if (rsp->op_ret != 0) {
                gf_event(EVENT_SNAPSHOT_CONFIG_UPDATE_FAILED, "error=%s",
                         rsp->op_errstr ? rsp->op_errstr
                                        : "Please check log file for details");
                ret = 0;
                break;
            }

            ret = dict_get_int32(dict, "config-command", &config_command);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Could not fetch config type");
                goto out;
            }

            if (config_command == GF_SNAP_CONFIG_DISPLAY) {
                ret = 0;
                break;
            }

            /* These are optional parameters therefore ignore the error */
            ret = dict_get_uint64(dict, "snap-max-hard-limit", &hard_limit);
            ret = dict_get_uint64(dict, "snap-max-soft-limit", &soft_limit);
            ret = dict_get_str(dict, "auto-delete", &auto_delete);
            ret = dict_get_str(dict, "snap-activate-on-create", &snap_activate);

            if (!hard_limit && !soft_limit && !auto_delete && !snap_activate) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR,
                       "At least one option from "
                       "snap-max-hard-limit, snap-max-soft-limit, "
                       "auto-delete and snap-activate-on-create "
                       "should be set");
                goto out;
            }

            volname = NULL;
            ret = dict_get_str(dict, "volname", &volname);

            if (hard_limit || soft_limit) {
                snprintf(option, sizeof(option), "%s=%" PRIu64,
                         hard_limit ? "hard_limit" : "soft_limit",
                         hard_limit ? hard_limit : soft_limit);
            } else if (auto_delete || snap_activate) {
                snprintf(option, sizeof(option), "%s=%s",
                         auto_delete ? "auto-delete" : "snap-activate",
                         auto_delete ? auto_delete : snap_activate);
            }

            snprintf(msg, sizeof(msg), "config_type=%s;%s",
                     volname ? "volume_config" : "system_config", option);

            gf_event(EVENT_SNAPSHOT_CONFIG_UPDATED, "%s", msg);

            ret = 0;
            break;

        default:
            gf_log("cli", GF_LOG_WARNING,
                   "Cannot generate event for unknown type.");
            ret = 0;
            goto out;
    }

out:
    return ret;
}

/*
 * Fetch necessary data from dict at one place instead of *
 * repeating the same code again and again.               *
 */
int
gf_cli_snapshot_get_data_from_dict(dict_t *dict, char **snap_name,
                                   char **volname, char **snap_uuid,
                                   int8_t *soft_limit_flag, char **clone_name)
{
    int ret = -1;

    GF_VALIDATE_OR_GOTO("cli", dict, out);

    if (snap_name) {
        ret = dict_get_str(dict, "snapname", snap_name);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get snapname from dict");
        }
    }

    if (volname) {
        ret = dict_get_str(dict, "volname1", volname);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get volname1 from dict");
        }
    }

    if (snap_uuid) {
        ret = dict_get_str(dict, "snapuuid", snap_uuid);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get snapuuid from dict");
        }
    }

    if (soft_limit_flag) {
        ret = dict_get_int8(dict, "soft-limit-reach", soft_limit_flag);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG,
                   "failed to get soft-limit-reach from dict");
        }
    }

    if (clone_name) {
        ret = dict_get_str(dict, "clonename", clone_name);
        if (ret) {
            gf_log("cli", GF_LOG_DEBUG, "failed to get clonename from dict");
        }
    }

    ret = 0;
out:
    return ret;
}

int
gf_cli_snapshot_cbk(struct rpc_req *req, struct iovec *iov, int count,
                    void *myframe)
{
    int ret = -1;
    gf_cli_rsp rsp = {
        0,
    };
    dict_t *dict = NULL;
    char *snap_name = NULL;
    char *clone_name = NULL;
    int32_t type = 0;
    call_frame_t *frame = NULL;
    gf_boolean_t snap_driven = _gf_false;
    int8_t soft_limit_flag = -1;
    char *volname = NULL;
    char *snap_uuid = NULL;

    GF_ASSERT(myframe);

    if (req->rpc_status == -1) {
        goto out;
    }

    frame = myframe;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(frame->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

    if (ret)
        goto out;

    ret = dict_get_int32(dict, "type", &type);
    if (ret) {
        gf_log(frame->this->name, GF_LOG_ERROR, "failed to get type");
        goto out;
    }

    ret = gf_cli_snapshot_get_data_from_dict(
        dict, &snap_name, &volname, &snap_uuid, &soft_limit_flag, &clone_name);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to fetch data from dict.");
        goto out;
    }

#if (USE_EVENTS)
    ret = gf_cli_generate_snapshot_event(&rsp, dict, type, snap_name, volname,
                                         snap_uuid, clone_name);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to generate snapshot event");
        goto out;
    }
#endif

    /* Snapshot status and delete command is handled separately */
    if (global_state->mode & GLUSTER_MODE_XML &&
        GF_SNAP_OPTION_TYPE_STATUS != type &&
        GF_SNAP_OPTION_TYPE_DELETE != type) {
        ret = cli_xml_output_snapshot(type, dict, rsp.op_ret, rsp.op_errno,
                                      rsp.op_errstr);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        }

        goto out;
    }

    switch (type) {
        case GF_SNAP_OPTION_TYPE_CREATE:
            if (rsp.op_ret) {
                cli_err("snapshot create: failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            if (!snap_name) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            if (!volname) {
                ret = -1;
                gf_log("cli", GF_LOG_ERROR, "Failed to get volume name");
                goto out;
            }

            cli_out(
                "snapshot create: success: Snap %s created "
                "successfully",
                snap_name);

            if (soft_limit_flag == 1) {
                cli_out(
                    "Warning: Soft-limit of volume (%s) is "
                    "reached. Snapshot creation is not possible "
                    "once hard-limit is reached.",
                    volname);
            }
            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_CLONE:
            if (rsp.op_ret) {
                cli_err("snapshot clone: failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            if (!clone_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get clone name");
                goto out;
            }

            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snapname name");
                goto out;
            }

            cli_out(
                "snapshot clone: success: Clone %s created "
                "successfully",
                clone_name);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_RESTORE:
            if (rsp.op_ret) {
                cli_err("snapshot restore: failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            cli_out(
                "Snapshot restore: %s: Snap restored "
                "successfully",
                snap_name);

            ret = 0;
            break;
        case GF_SNAP_OPTION_TYPE_ACTIVATE:
            if (rsp.op_ret) {
                cli_err("snapshot activate: failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            cli_out(
                "Snapshot activate: %s: Snap activated "
                "successfully",
                snap_name);

            ret = 0;
            break;

        case GF_SNAP_OPTION_TYPE_DEACTIVATE:
            if (rsp.op_ret) {
                cli_err("snapshot deactivate: failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            if (!snap_name) {
                gf_log("cli", GF_LOG_ERROR, "Failed to get snap name");
                goto out;
            }

            cli_out(
                "Snapshot deactivate: %s: Snap deactivated "
                "successfully",
                snap_name);

            ret = 0;
            break;
        case GF_SNAP_OPTION_TYPE_INFO:
            if (rsp.op_ret) {
                cli_err("Snapshot info : failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            snap_driven = dict_get_str_boolean(dict, "snap-driven", _gf_false);
            if (snap_driven == _gf_true) {
                ret = cli_call_snapshot_info(dict, snap_driven);
                if (ret) {
                    gf_log("cli", GF_LOG_ERROR, "Snapshot info failed");
                    goto out;
                }
            } else if (snap_driven == _gf_false) {
                ret = cli_get_snaps_in_volume(dict);
                if (ret) {
                    gf_log("cli", GF_LOG_ERROR, "Snapshot info failed");
                    goto out;
                }
            }
            break;

        case GF_SNAP_OPTION_TYPE_CONFIG:
            ret = cli_snapshot_config_display(dict, &rsp);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Failed to display "
                       "snapshot config output.");
                goto out;
            }
            break;

        case GF_SNAP_OPTION_TYPE_LIST:
            if (rsp.op_ret) {
                cli_err("Snapshot list : failed: %s",
                        rsp.op_errstr ? rsp.op_errstr
                                      : "Please check log file for details");
                ret = rsp.op_ret;
                goto out;
            }

            ret = cli_snapshot_list(dict);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Failed to display "
                       "snapshot list");
                goto out;
            }
            break;

        case GF_SNAP_OPTION_TYPE_DELETE:
            ret = cli_snapshot_remove_reply(&rsp, dict, frame);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Failed to delete snap");
                goto out;
            }
            break;

        case GF_SNAP_OPTION_TYPE_STATUS:
            ret = cli_snapshot_status(dict, &rsp, frame);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR,
                       "Failed to display "
                       "snapshot status output.");
                goto out;
            }
            break;

        default:
            cli_err("Unknown command executed");
            ret = -1;
            goto out;
    }
out:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);

    free(rsp.dict.dict_val);
    free(rsp.op_errstr);

    return ret;
}

int32_t
gf_cli_snapshot_for_delete(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int32_t ret = -1;
    int32_t cmd = -1;
    cli_local_t *local = NULL;
    dict_t *snap_dict = NULL;
    int32_t snapcount = 0;
    int i = 0;
    char question[PATH_MAX] = "";
    char *volname = NULL;
    gf_answer_t answer = GF_ANSWER_NO;

    GF_VALIDATE_OR_GOTO("cli", frame, out);
    GF_VALIDATE_OR_GOTO("cli", frame->local, out);
    GF_VALIDATE_OR_GOTO("cli", this, out);
    GF_VALIDATE_OR_GOTO("cli", data, out);

    local = frame->local;

    ret = dict_get_int32(local->dict, "sub-cmd", &cmd);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to get "
               "sub-cmd");
        goto out;
    }

    /* No need multiple RPCs for individual snapshot delete*/
    if (cmd == GF_SNAP_DELETE_TYPE_SNAP) {
        ret = 0;
        goto out;
    }

    ret = dict_get_int32(local->dict, "snapcount", &snapcount);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Could not get "
               "snapcount");
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
#ifdef HAVE_LIB_XML
        ret = xmlTextWriterWriteFormatElement(
            local->writer, (xmlChar *)"snapCount", "%d", snapcount);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to write "
                   "xml element \"snapCount\"");
            goto out;
        }
#endif /* HAVE_LIB_XML */
    } else if (snapcount == 0) {
        cli_out("No snapshots present");
        goto out;
    }

    if (cmd == GF_SNAP_DELETE_TYPE_ALL) {
        snprintf(question, sizeof(question),
                 "System contains %d "
                 "snapshot(s).\nDo you still "
                 "want to continue and delete them? ",
                 snapcount);
    } else {
        ret = dict_get_str(local->dict, "volname", &volname);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to fetch "
                   "volname from local dictionary");
            goto out;
        }

        snprintf(question, sizeof(question),
                 "Volume (%s) contains "
                 "%d snapshot(s).\nDo you still want to "
                 "continue and delete them? ",
                 volname, snapcount);
    }

    answer = cli_cmd_get_confirmation(global_state, question);
    if (GF_ANSWER_NO == answer) {
        ret = 0;
        gf_log("cli", GF_LOG_DEBUG,
               "User cancelled "
               "snapshot delete operation for snap delete");
        goto out;
    }

    for (i = 1; i <= snapcount; i++) {
        ret = -1;

        snap_dict = dict_new();
        if (!snap_dict)
            goto out;

        ret = cli_populate_req_dict_for_delete(snap_dict, local->dict, i);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Could not "
                   "populate snap request dictionary");
            goto out;
        }

        ret = cli_to_glusterd(&req, frame, gf_cli_snapshot_cbk,
                              (xdrproc_t)xdr_gf_cli_req, snap_dict,
                              GLUSTER_CLI_SNAP, this, cli_rpc_prog, NULL);
        if (ret) {
            /* Fail the operation if deleting one of the
             * snapshots is failed
             */
            gf_log("cli", GF_LOG_ERROR,
                   "cli_to_glusterd "
                   "for snapshot delete failed");
            goto out;
        }
        dict_unref(snap_dict);
        snap_dict = NULL;
    }

out:
    if (snap_dict)
        dict_unref(snap_dict);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_snapshot_for_status(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    int ret = -1;
    int32_t cmd = -1;
    cli_local_t *local = NULL;
    dict_t *snap_dict = NULL;
    int snapcount = 0;
    int i = 0;

    GF_VALIDATE_OR_GOTO("cli", frame, out);
    GF_VALIDATE_OR_GOTO("cli", frame->local, out);
    GF_VALIDATE_OR_GOTO("cli", this, out);
    GF_VALIDATE_OR_GOTO("cli", data, out);

    local = frame->local;

    ret = dict_get_int32(local->dict, "sub-cmd", &cmd);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get sub-cmd");
        goto out;
    }

    /* Snapshot status of single snap (i.e. GF_SNAP_STATUS_TYPE_SNAP)
     * is already handled. Therefore we can return from here.
     * If want to get status of all snaps in the system or volume then
     * we should get them one by one.*/
    if ((cmd == GF_SNAP_STATUS_TYPE_SNAP) ||
        (cmd == GF_SNAP_STATUS_TYPE_ITER)) {
        ret = 0;
        goto out;
    }

    ret = dict_get_int32(local->dict, "status.snapcount", &snapcount);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Could not get snapcount");
        goto out;
    }

    if (snapcount == 0 && !(global_state->mode & GLUSTER_MODE_XML)) {
        cli_out("No snapshots present");
    }

    for (i = 0; i < snapcount; i++) {
        ret = -1;

        snap_dict = dict_new();
        if (!snap_dict)
            goto out;

        ret = cli_populate_req_dict_for_status(snap_dict, local->dict, i);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Could not "
                   "populate snap request dictionary");
            goto out;
        }

        ret = cli_to_glusterd(&req, frame, gf_cli_snapshot_cbk,
                              (xdrproc_t)xdr_gf_cli_req, snap_dict,
                              GLUSTER_CLI_SNAP, this, cli_rpc_prog, NULL);

        /* Ignore the return value and error for snapshot
         * status of type "ALL" or "VOL"
         *
         * Scenario : There might be case where status command
         * and delete command might be issued at the same time.
         * In that case when status tried to fetch detail of
         * snap which has been deleted by concurrent command,
         * then it will show snapshot not present. Which will
         * not be appropriate.
         */
        if (ret && (cmd != GF_SNAP_STATUS_TYPE_ALL &&
                    cmd != GF_SNAP_STATUS_TYPE_VOL)) {
            gf_log("cli", GF_LOG_ERROR,
                   "cli_to_glusterd for snapshot status failed");
            goto out;
        }
        dict_unref(snap_dict);
        snap_dict = NULL;
    }

    ret = 0;
out:
    if (snap_dict)
        dict_unref(snap_dict);
    GF_FREE(req.dict.dict_val);

    return ret;
}

int32_t
gf_cli_snapshot(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    dict_t *options = NULL;
    int ret = -1;
    int tmp_ret = -1;
    cli_local_t *local = NULL;
    char *err_str = NULL;
    int type = -1;

    if (!frame || !this || !data)
        goto out;

    if (!frame->local)
        goto out;

    local = frame->local;

    options = data;

    ret = dict_get_int32(local->dict, "type", &type);

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_snapshot_begin_composite_op(local);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to begin "
                   "snapshot xml composite op");
            goto out;
        }
    }

    ret = cli_to_glusterd(&req, frame, gf_cli_snapshot_cbk,
                          (xdrproc_t)xdr_gf_cli_req, options, GLUSTER_CLI_SNAP,
                          this, cli_rpc_prog, NULL);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "cli_to_glusterd for "
               "snapshot failed");
        goto xmlend;
    }

    if (GF_SNAP_OPTION_TYPE_STATUS == type) {
        ret = gf_cli_snapshot_for_status(frame, this, data);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "cli to glusterd "
                   "for snapshot status command failed");
        }

        goto xmlend;
    }

    if (GF_SNAP_OPTION_TYPE_DELETE == type) {
        ret = gf_cli_snapshot_for_delete(frame, this, data);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "cli to glusterd "
                   "for snapshot delete command failed");
        }

        goto xmlend;
    }

    ret = 0;

xmlend:
    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_snapshot_end_composite_op(local);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to end "
                   "snapshot xml composite op");
            goto out;
        }
    }
out:
    if (ret && local && GF_SNAP_OPTION_TYPE_STATUS == type) {
        tmp_ret = dict_get_str(local->dict, "op_err_str", &err_str);
        if (tmp_ret || !err_str) {
            cli_err("Snapshot Status : failed: %s",
                    "Please "
                    "check log file for details");
        } else {
            cli_err("Snapshot Status : failed: %s", err_str);
            dict_del(local->dict, "op_err_str");
        }
    }

    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    if (global_state->mode & GLUSTER_MODE_XML) {
        /* XML mode handles its own error */
        ret = 0;
    }
    return ret;
}

int32_t
gf_cli_barrier_volume_cbk(struct rpc_req *req, struct iovec *iov, int count,
                          void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }
    gf_log("cli", GF_LOG_DEBUG, "Received response to barrier");

    if (rsp.op_ret) {
        if (rsp.op_errstr && (strlen(rsp.op_errstr) > 1)) {
            cli_err("volume barrier: command unsuccessful : %s", rsp.op_errstr);
        } else {
            cli_err("volume barrier: command unsuccessful");
        }
    } else {
        cli_out("volume barrier: command successful");
    }
    ret = rsp.op_ret;

out:
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_barrier_volume(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    dict_t *options = NULL;
    int ret = -1;

    if (!frame || !this || !data)
        goto out;

    options = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_barrier_volume_cbk,
                          (xdrproc_t)xdr_gf_cli_req, options,
                          GLUSTER_CLI_BARRIER_VOLUME, this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int32_t
gf_cli_get_vol_opt_cbk(struct rpc_req *req, struct iovec *iov, int count,
                       void *myframe)
{
    gf_cli_rsp rsp = {
        0,
    };
    int ret = -1;
    dict_t *dict = NULL;
    char *key = NULL;
    char *value = NULL;
    char msg[1024] = {
        0,
    };
    int i = 0;
    char dict_key[50] = {
        0,
    };

    GF_ASSERT(myframe);

    if (-1 == req->rpc_status)
        goto out;

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }
    gf_log("cli", GF_LOG_DEBUG, "Received response to get volume option");

    if (rsp.op_ret) {
        if (strcmp(rsp.op_errstr, ""))
            snprintf(msg, sizeof(msg),
                     "volume get option: "
                     "failed: %s",
                     rsp.op_errstr);
        else
            snprintf(msg, sizeof(msg),
                     "volume get option: "
                     "failed");

        if (global_state->mode & GLUSTER_MODE_XML) {
            ret = cli_xml_output_str("volGetopts", msg, rsp.op_ret,
                                     rsp.op_errno, rsp.op_errstr);
            if (ret) {
                gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
            }
        } else {
            cli_err("%s", msg);
        }
        ret = rsp.op_ret;
        goto out_nolog;
    }
    dict = dict_new();

    if (!dict) {
        ret = -1;
        goto out;
    }

    ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed rsp_dict unserialization");
        goto out;
    }

    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_getopts(dict, rsp.op_ret, rsp.op_errno,
                                         rsp.op_errstr);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "xml output generation "
                   "failed");
            ret = 0;
        }
        goto out;
    }

    ret = dict_get_str(dict, "warning", &value);
    if (!ret) {
        cli_out("%s", value);
    }

    ret = dict_get_int32(dict, "count", &count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "Failed to retrieve count "
               "from the dictionary");
        goto out;
    }

    if (count <= 0) {
        gf_log("cli", GF_LOG_ERROR,
               "Value of count :%d is "
               "invalid",
               count);
        ret = -1;
        goto out;
    }

    cli_out("%-40s%-40s", "Option", "Value");
    cli_out("%-40s%-40s", "------", "-----");
    for (i = 1; i <= count; i++) {
        snprintf(dict_key, sizeof dict_key, "key%d", i);
        ret = dict_get_str(dict, dict_key, &key);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to"
                   " retrieve %s from the "
                   "dictionary",
                   dict_key);
            goto out;
        }
        snprintf(dict_key, sizeof dict_key, "value%d", i);
        ret = dict_get_str(dict, dict_key, &value);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to "
                   "retrieve key value for %s from"
                   "the dictionary",
                   dict_key);
            goto out;
        }
        cli_out("%-40s%-40s", key, value);
    }

out:
    if (ret) {
        cli_out(
            "volume get option failed. Check the cli/glusterd log "
            "file for more details");
    }

out_nolog:
    if (dict)
        dict_unref(dict);
    cli_cmd_broadcast_response(ret);
    gf_free_xdr_cli_rsp(rsp);
    return ret;
}

int
gf_cli_get_vol_opt(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    dict_t *options = NULL;
    int ret = -1;

    if (!frame || !this || !data)
        goto out;

    options = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_get_vol_opt_cbk,
                          (xdrproc_t)xdr_gf_cli_req, options,
                          GLUSTER_CLI_GET_VOL_OPT, this, cli_rpc_prog, NULL);
out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);
    return ret;
}

int
add_cli_cmd_timeout_to_dict(dict_t *dict)
{
    int ret = 0;

    if (cli_default_conn_timeout > 120) {
        ret = dict_set_uint32(dict, "timeout", cli_default_conn_timeout);
        if (ret) {
            gf_log("cli", GF_LOG_INFO,
                   "Failed to save"
                   "timeout to dict");
        }
    }
    return ret;
}

int
cli_to_glusterd(gf_cli_req *req, call_frame_t *frame, fop_cbk_fn_t cbkfn,
                xdrproc_t xdrproc, dict_t *dict, int procnum, xlator_t *this,
                rpc_clnt_prog_t *prog, struct iobref *iobref)
{
    int ret = 0;
    size_t len = 0;
    char *cmd = NULL;
    int i = 0;
    const char **words = NULL;
    cli_local_t *local = NULL;

    if (!this || !frame || !dict) {
        ret = -1;
        goto out;
    }

    if (!frame->local) {
        ret = -1;
        goto out;
    }

    local = frame->local;

    if (!local->words) {
        ret = -1;
        goto out;
    }

    words = local->words;

    while (words[i])
        len += strlen(words[i++]) + 1;

    cmd = GF_CALLOC(1, len, gf_common_mt_char);

    if (!cmd) {
        ret = -1;
        goto out;
    }

    for (i = 0; words[i]; i++) {
        strncat(cmd, words[i], len - 1);
        if (words[i + 1] != NULL)
            strncat(cmd, " ", len - 1);
    }

    cmd[len - 1] = '\0';

    ret = dict_set_dynstr(dict, "cmd-str", cmd);
    if (ret)
        goto out;

    ret = add_cli_cmd_timeout_to_dict(dict);

    ret = dict_allocate_and_serialize(dict, &(req->dict).dict_val,
                                      &(req->dict).dict_len);

    if (ret < 0) {
        gf_log(this->name, GF_LOG_DEBUG,
               "failed to get serialized length of dict");
        goto out;
    }

    ret = cli_cmd_submit(NULL, req, frame, prog, procnum, iobref, this, cbkfn,
                         (xdrproc_t)xdrproc);
out:
    return ret;
}

int
gf_cli_print_bitrot_scrub_status(dict_t *dict)
{
    int i = 1;
    int j = 0;
    int ret = -1;
    int count = 0;
    char key[256] = {
        0,
    };
    char *volname = NULL;
    char *node_name = NULL;
    char *scrub_freq = NULL;
    char *state_scrub = NULL;
    char *scrub_impact = NULL;
    char *bad_file_str = NULL;
    char *scrub_log_file = NULL;
    char *bitrot_log_file = NULL;
    uint64_t scrub_files = 0;
    uint64_t unsigned_files = 0;
    uint64_t scrub_time = 0;
    uint64_t days = 0;
    uint64_t hours = 0;
    uint64_t minutes = 0;
    uint64_t seconds = 0;
    char *last_scrub = NULL;
    uint64_t error_count = 0;
    int8_t scrub_running = 0;
    char *scrub_state_op = NULL;

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        gf_log("cli", GF_LOG_TRACE, "failed to get volume name");

    ret = dict_get_str(dict, "features.scrub", &state_scrub);
    if (ret)
        gf_log("cli", GF_LOG_TRACE, "failed to get scrub state value");

    ret = dict_get_str(dict, "features.scrub-throttle", &scrub_impact);
    if (ret)
        gf_log("cli", GF_LOG_TRACE,
               "failed to get scrub impact "
               "value");

    ret = dict_get_str(dict, "features.scrub-freq", &scrub_freq);
    if (ret)
        gf_log("cli", GF_LOG_TRACE, "failed to get scrub -freq value");

    ret = dict_get_str(dict, "bitrot_log_file", &bitrot_log_file);
    if (ret)
        gf_log("cli", GF_LOG_TRACE,
               "failed to get bitrot log file "
               "location");

    ret = dict_get_str(dict, "scrub_log_file", &scrub_log_file);
    if (ret)
        gf_log("cli", GF_LOG_TRACE,
               "failed to get scrubber log file "
               "location");

    ret = dict_get_int32(dict, "count", &count);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "count not get count value from"
               " dictionary");
        goto out;
    }

    for (i = 1; i <= count; i++) {
        snprintf(key, 256, "scrub-running-%d", i);
        ret = dict_get_int8(dict, key, &scrub_running);
        if (ret)
            gf_log("cli", GF_LOG_TRACE,
                   "failed to get scrubbed "
                   "files");
        if (scrub_running)
            break;
    }

    if (scrub_running)
        gf_asprintf(&scrub_state_op, "%s (In Progress)", state_scrub);
    else
        gf_asprintf(&scrub_state_op, "%s (Idle)", state_scrub);

    cli_out("\n%s: %s\n", "Volume name ", volname);

    cli_out("%s: %s\n", "State of scrub", scrub_state_op);

    cli_out("%s: %s\n", "Scrub impact", scrub_impact);

    cli_out("%s: %s\n", "Scrub frequency", scrub_freq);

    cli_out("%s: %s\n", "Bitrot error log location", bitrot_log_file);

    cli_out("%s: %s\n", "Scrubber error log location", scrub_log_file);

    for (i = 1; i <= count; i++) {
        /* Reset the variables to prevent carryover of values */
        node_name = NULL;
        last_scrub = NULL;
        scrub_time = 0;
        days = 0;
        hours = 0;
        minutes = 0;
        seconds = 0;
        error_count = 0;
        scrub_files = 0;
        unsigned_files = 0;

        snprintf(key, 256, "node-name-%d", i);
        ret = dict_get_str(dict, key, &node_name);
        if (ret)
            gf_log("cli", GF_LOG_TRACE, "failed to get node-name");

        snprintf(key, 256, "scrubbed-files-%d", i);
        ret = dict_get_uint64(dict, key, &scrub_files);
        if (ret)
            gf_log("cli", GF_LOG_TRACE,
                   "failed to get scrubbed "
                   "files");

        snprintf(key, 256, "unsigned-files-%d", i);
        ret = dict_get_uint64(dict, key, &unsigned_files);
        if (ret)
            gf_log("cli", GF_LOG_TRACE,
                   "failed to get unsigned "
                   "files");

        snprintf(key, 256, "scrub-duration-%d", i);
        ret = dict_get_uint64(dict, key, &scrub_time);
        if (ret)
            gf_log("cli", GF_LOG_TRACE,
                   "failed to get last scrub "
                   "duration");

        snprintf(key, 256, "last-scrub-time-%d", i);
        ret = dict_get_str(dict, key, &last_scrub);
        if (ret)
            gf_log("cli", GF_LOG_TRACE,
                   "failed to get last scrub"
                   " time");
        snprintf(key, 256, "error-count-%d", i);
        ret = dict_get_uint64(dict, key, &error_count);
        if (ret)
            gf_log("cli", GF_LOG_TRACE,
                   "failed to get error "
                   "count");

        cli_out("\n%s\n",
                "=========================================="
                "===============");

        cli_out("%s: %s\n", "Node", node_name);

        cli_out("%s: %" PRIu64 "\n", "Number of Scrubbed files", scrub_files);

        cli_out("%s: %" PRIu64 "\n", "Number of Skipped files", unsigned_files);

        if ((!last_scrub) || !strcmp(last_scrub, ""))
            cli_out("%s: %s\n", "Last completed scrub time",
                    "Scrubber pending to complete.");
        else
            cli_out("%s: %s\n", "Last completed scrub time", last_scrub);

        /* Printing last scrub duration time in human readable form*/
        seconds = scrub_time % 60;
        minutes = (scrub_time / 60) % 60;
        hours = (scrub_time / 3600) % 24;
        days = scrub_time / 86400;
        cli_out("%s: %" PRIu64 ":%" PRIu64 ":%" PRIu64 ":%" PRIu64 "\n",
                "Duration of last scrub (D:M:H:M:S)", days, hours, minutes,
                seconds);

        cli_out("%s: %" PRIu64 "\n", "Error count", error_count);

        if (error_count) {
            cli_out("%s:\n", "Corrupted object's [GFID]");
            /* Printing list of bad file's (Corrupted object's)*/
            for (j = 0; j < error_count; j++) {
                snprintf(key, 256, "quarantine-%d-%d", j, i);
                ret = dict_get_str(dict, key, &bad_file_str);
                if (!ret) {
                    cli_out("%s\n", bad_file_str);
                }
            }
        }
    }
    cli_out("%s\n",
            "=========================================="
            "===============");

out:
    GF_FREE(scrub_state_op);
    return 0;
}

int
gf_cli_bitrot_cbk(struct rpc_req *req, struct iovec *iov, int count,
                  void *myframe)
{
    int ret = -1;
    int type = 0;
    gf_cli_rsp rsp = {
        0,
    };
    dict_t *dict = NULL;
    char *scrub_cmd = NULL;
    char *volname = NULL;
    char *cmd_str = NULL;
    char *cmd_op = NULL;

    GF_ASSERT(myframe);

    if (req->rpc_status == -1) {
        goto out;
    }

    ret = xdr_to_generic(*iov, &rsp, (xdrproc_t)xdr_gf_cli_rsp);
    if (ret < 0) {
        gf_log(((call_frame_t *)myframe)->this->name, GF_LOG_ERROR,
               "Failed to decode xdr response");
        goto out;
    }

    if (rsp.op_ret) {
        ret = -1;
        if (global_state->mode & GLUSTER_MODE_XML)
            goto xml_output;

        if (strcmp(rsp.op_errstr, ""))
            cli_err("Bitrot command failed : %s", rsp.op_errstr);
        else
            cli_err("Bitrot command : failed");

        goto out;
    }

    if (rsp.dict.dict_len) {
        /* Unserialize the dictionary */
        dict = dict_new();

        if (!dict) {
            ret = -1;
            goto out;
        }

        ret = dict_unserialize(rsp.dict.dict_val, rsp.dict.dict_len, &dict);

        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "failed to unserialize "
                   "req-buffer to dictionary");
            goto out;
        }
    }

    gf_log("cli", GF_LOG_DEBUG, "Received resp to bit rot command");

    ret = dict_get_int32(dict, "type", &type);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR, "Failed to get command type");
        goto out;
    }

    /* Ignoring the error, as using dict val for cli output only */
    ret = dict_get_str(dict, "scrub-value", &scrub_cmd);
    if (ret)
        gf_log("cli", GF_LOG_TRACE, "Failed to get scrub command");

    ret = dict_get_str(dict, "volname", &volname);
    if (ret)
        gf_log("cli", GF_LOG_TRACE, "failed to get volume name");

    ret = dict_get_str(dict, "cmd-str", &cmd_str);
    if (ret)
        gf_log("cli", GF_LOG_TRACE, "failed to get command string");

    if (cmd_str)
        cmd_op = strrchr(cmd_str, ' ') + 1;

    if ((type == GF_BITROT_CMD_SCRUB_STATUS) &&
        !(global_state->mode & GLUSTER_MODE_XML)) {
        ret = gf_cli_print_bitrot_scrub_status(dict);
        if (ret) {
            gf_log("cli", GF_LOG_ERROR,
                   "Failed to print bitrot "
                   "scrub status");
        }
        goto out;
    }

    switch (type) {
        case GF_BITROT_OPTION_TYPE_ENABLE:
            cli_out(
                "volume bitrot: success bitrot enabled "
                "for volume %s",
                volname);
            ret = 0;
            goto out;
        case GF_BITROT_OPTION_TYPE_DISABLE:
            cli_out(
                "volume bitrot: success bitrot disabled "
                "for volume %s",
                volname);
            ret = 0;
            goto out;
        case GF_BITROT_CMD_SCRUB_ONDEMAND:
            cli_out(
                "volume bitrot: scrubber started ondemand "
                "for volume %s",
                volname);
            ret = 0;
            goto out;
        case GF_BITROT_OPTION_TYPE_SCRUB:
            if (!strncmp("pause", scrub_cmd, sizeof("pause")))
                cli_out(
                    "volume bitrot: scrubber paused "
                    "for volume %s",
                    volname);
            if (!strncmp("resume", scrub_cmd, sizeof("resume")))
                cli_out(
                    "volume bitrot: scrubber resumed "
                    "for volume %s",
                    volname);
            ret = 0;
            goto out;
        case GF_BITROT_OPTION_TYPE_SCRUB_FREQ:
            cli_out(
                "volume bitrot: scrub-frequency is set to %s "
                "successfully for volume %s",
                cmd_op, volname);
            ret = 0;
            goto out;
        case GF_BITROT_OPTION_TYPE_SCRUB_THROTTLE:
            cli_out(
                "volume bitrot: scrub-throttle is set to %s "
                "successfully for volume %s",
                cmd_op, volname);
            ret = 0;
            goto out;
    }

xml_output:
    if (global_state->mode & GLUSTER_MODE_XML) {
        ret = cli_xml_output_vol_profile(dict, rsp.op_ret, rsp.op_errno,
                                         rsp.op_errstr);
        if (ret)
            gf_log("cli", GF_LOG_ERROR, "Error outputting to xml");
        goto out;
    }

    if (!rsp.op_ret)
        cli_out("volume bitrot: success");

    ret = rsp.op_ret;

out:
    if (dict)
        dict_unref(dict);

    gf_free_xdr_cli_rsp(rsp);
    cli_cmd_broadcast_response(ret);
    return ret;
}

int32_t
gf_cli_bitrot(call_frame_t *frame, xlator_t *this, void *data)
{
    gf_cli_req req = {{
        0,
    }};
    dict_t *options = NULL;
    int ret = -1;

    if (!frame || !this || !data)
        goto out;

    options = data;

    ret = cli_to_glusterd(&req, frame, gf_cli_bitrot_cbk,
                          (xdrproc_t)xdr_gf_cli_req, options,
                          GLUSTER_CLI_BITROT, this, cli_rpc_prog, NULL);
    if (ret) {
        gf_log("cli", GF_LOG_ERROR,
               "cli_to_glusterd for "
               "bitrot failed");
        goto out;
    }

out:
    gf_log("cli", GF_LOG_DEBUG, "Returning %d", ret);

    GF_FREE(req.dict.dict_val);

    return ret;
}

struct rpc_clnt_procedure gluster_cli_actors[GLUSTER_CLI_MAXVALUE] = {
    [GLUSTER_CLI_NULL] = {"NULL", NULL},
    [GLUSTER_CLI_PROBE] = {"PROBE_QUERY", gf_cli_probe},
    [GLUSTER_CLI_DEPROBE] = {"DEPROBE_QUERY", gf_cli_deprobe},
    [GLUSTER_CLI_LIST_FRIENDS] = {"LIST_FRIENDS", gf_cli_list_friends},
    [GLUSTER_CLI_UUID_RESET] = {"UUID_RESET", gf_cli3_1_uuid_reset},
    [GLUSTER_CLI_UUID_GET] = {"UUID_GET", gf_cli3_1_uuid_get},
    [GLUSTER_CLI_CREATE_VOLUME] = {"CREATE_VOLUME", gf_cli_create_volume},
    [GLUSTER_CLI_DELETE_VOLUME] = {"DELETE_VOLUME", gf_cli_delete_volume},
    [GLUSTER_CLI_START_VOLUME] = {"START_VOLUME", gf_cli_start_volume},
    [GLUSTER_CLI_STOP_VOLUME] = {"STOP_VOLUME", gf_cli_stop_volume},
    [GLUSTER_CLI_RENAME_VOLUME] = {"RENAME_VOLUME", gf_cli_rename_volume},
    [GLUSTER_CLI_DEFRAG_VOLUME] = {"DEFRAG_VOLUME", gf_cli_defrag_volume},
    [GLUSTER_CLI_GET_VOLUME] = {"GET_VOLUME", gf_cli_get_volume},
    [GLUSTER_CLI_GET_NEXT_VOLUME] = {"GET_NEXT_VOLUME", gf_cli_get_next_volume},
    [GLUSTER_CLI_SET_VOLUME] = {"SET_VOLUME", gf_cli_set_volume},
    [GLUSTER_CLI_ADD_BRICK] = {"ADD_BRICK", gf_cli_add_brick},
    [GLUSTER_CLI_REMOVE_BRICK] = {"REMOVE_BRICK", gf_cli_remove_brick},
    [GLUSTER_CLI_REPLACE_BRICK] = {"REPLACE_BRICK", gf_cli_replace_brick},
    [GLUSTER_CLI_LOG_ROTATE] = {"LOG ROTATE", gf_cli_log_rotate},
    [GLUSTER_CLI_GETSPEC] = {"GETSPEC", gf_cli_getspec},
    [GLUSTER_CLI_PMAP_PORTBYBRICK] = {"PMAP PORTBYBRICK", gf_cli_pmap_b2p},
    [GLUSTER_CLI_SYNC_VOLUME] = {"SYNC_VOLUME", gf_cli_sync_volume},
    [GLUSTER_CLI_RESET_VOLUME] = {"RESET_VOLUME", gf_cli_reset_volume},
    [GLUSTER_CLI_FSM_LOG] = {"FSM_LOG", gf_cli_fsm_log},
    [GLUSTER_CLI_GSYNC_SET] = {"GSYNC_SET", gf_cli_gsync_set},
    [GLUSTER_CLI_PROFILE_VOLUME] = {"PROFILE_VOLUME", gf_cli_profile_volume},
    [GLUSTER_CLI_QUOTA] = {"QUOTA", gf_cli_quota},
    [GLUSTER_CLI_TOP_VOLUME] = {"TOP_VOLUME", gf_cli_top_volume},
    [GLUSTER_CLI_GETWD] = {"GETWD", gf_cli_getwd},
    [GLUSTER_CLI_STATUS_VOLUME] = {"STATUS_VOLUME", gf_cli_status_volume},
    [GLUSTER_CLI_STATUS_ALL] = {"STATUS_ALL", gf_cli_status_volume_all},
    [GLUSTER_CLI_MOUNT] = {"MOUNT", gf_cli_mount},
    [GLUSTER_CLI_UMOUNT] = {"UMOUNT", gf_cli_umount},
    [GLUSTER_CLI_HEAL_VOLUME] = {"HEAL_VOLUME", gf_cli_heal_volume},
    [GLUSTER_CLI_STATEDUMP_VOLUME] = {"STATEDUMP_VOLUME",
                                      gf_cli_statedump_volume},
    [GLUSTER_CLI_LIST_VOLUME] = {"LIST_VOLUME", gf_cli_list_volume},
    [GLUSTER_CLI_CLRLOCKS_VOLUME] = {"CLEARLOCKS_VOLUME",
                                     gf_cli_clearlocks_volume},
    [GLUSTER_CLI_COPY_FILE] = {"COPY_FILE", gf_cli_copy_file},
    [GLUSTER_CLI_SYS_EXEC] = {"SYS_EXEC", gf_cli_sys_exec},
    [GLUSTER_CLI_SNAP] = {"SNAP", gf_cli_snapshot},
    [GLUSTER_CLI_BARRIER_VOLUME] = {"BARRIER VOLUME", gf_cli_barrier_volume},
    [GLUSTER_CLI_GANESHA] = {"GANESHA", gf_cli_ganesha},
    [GLUSTER_CLI_GET_VOL_OPT] = {"GET_VOL_OPT", gf_cli_get_vol_opt},
    [GLUSTER_CLI_BITROT] = {"BITROT", gf_cli_bitrot},
    [GLUSTER_CLI_ATTACH_TIER] = {"ATTACH_TIER", gf_cli_attach_tier},
    [GLUSTER_CLI_TIER] = {"TIER", gf_cli_tier},
    [GLUSTER_CLI_GET_STATE] = {"GET_STATE", gf_cli_get_state},
    [GLUSTER_CLI_RESET_BRICK] = {"RESET_BRICK", gf_cli_reset_brick},
    [GLUSTER_CLI_REMOVE_TIER_BRICK] = {"DETACH_TIER", gf_cli_remove_tier_brick},
    [GLUSTER_CLI_ADD_TIER_BRICK] = {"ADD_TIER_BRICK", gf_cli_add_tier_brick}};

struct rpc_clnt_program cli_prog = {
    .progname = "Gluster CLI",
    .prognum = GLUSTER_CLI_PROGRAM,
    .progver = GLUSTER_CLI_VERSION,
    .numproc = GLUSTER_CLI_MAXVALUE,
    .proctable = gluster_cli_actors,
};

struct rpc_clnt_procedure cli_quotad_procs[GF_AGGREGATOR_MAXVALUE] = {
    [GF_AGGREGATOR_NULL] = {"NULL", NULL},
    [GF_AGGREGATOR_LOOKUP] = {"LOOKUP", NULL},
    [GF_AGGREGATOR_GETLIMIT] = {"GETLIMIT", cli_quotad_getlimit},
};

struct rpc_clnt_program cli_quotad_clnt = {
    .progname = "CLI Quotad client",
    .prognum = GLUSTER_AGGREGATOR_PROGRAM,
    .progver = GLUSTER_AGGREGATOR_VERSION,
    .numproc = GF_AGGREGATOR_MAXVALUE,
    .proctable = cli_quotad_procs,
};