<feed xmlns='http://www.w3.org/2005/Atom'>
<title>glusterfs.git/cli, branch v3.5.6</title>
<subtitle>GlusterFS is a distributed file-system capable of scaling to several petabytes. It aggregates various storage bricks over Infiniband RDMA or TCP/IP interconnect into one large parallel network file system.</subtitle>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/'/>
<entry>
<title>features/quota : Fix XML output for quota list command</title>
<updated>2015-07-07T16:11:38+00:00</updated>
<author>
<name>vmallika</name>
<email>vmallika@redhat.com</email>
</author>
<published>2015-06-15T07:01:48+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=c9e92231e8fb31e6e4a9f061746daaedd77ad1b7'/>
<id>c9e92231e8fb31e6e4a9f061746daaedd77ad1b7</id>
<content type='text'>
This is a backport of http://review.gluster.org/#/c/9481/

&gt; Sample output:
&gt; ---------------
&gt;
&gt; Sample 1)
&gt; ----------
&gt; [root@snapshot-28 glusterfs]# gluster volume quota vol1 list /dir1 /dir4
&gt; /dir5 --xml
&gt; &lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&gt; &lt;cliOutput&gt;
&gt;   &lt;opRet&gt;0&lt;/opRet&gt;
&gt;   &lt;opErrno&gt;0&lt;/opErrno&gt;
&gt;   &lt;opErrstr/&gt;
&gt;   &lt;volQuota&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir1&lt;/path&gt;
&gt;       &lt;hard_limit&gt;10.0MB&lt;/hard_limit&gt;
&gt;       &lt;soft_limit&gt;80%&lt;/soft_limit&gt;
&gt;       &lt;used_space&gt;0Bytes&lt;/used_space&gt;
&gt;       &lt;avail_space&gt;10.0MB&lt;/avail_space&gt;
&gt;     &lt;/limit&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir4&lt;/path&gt;
&gt;       &lt;path&gt;No such file or directory&lt;/path&gt;
&gt;     &lt;/limit&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir5&lt;/path&gt;
&gt;       &lt;path&gt;No such file or directory&lt;/path&gt;
&gt;     &lt;/limit&gt;
&gt;   &lt;/volQuota&gt;
&gt; &lt;/cliOutput&gt;
&gt;
&gt; Sample 2)
&gt; ---------
&gt; gluster volume quota vol1 list --xml
&gt; &lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&gt; &lt;cliOutput&gt;
&gt;   &lt;opRet&gt;0&lt;/opRet&gt;
&gt;   &lt;opErrno&gt;0&lt;/opErrno&gt;
&gt;   &lt;opErrstr/&gt;
&gt;   &lt;volQuota/&gt;
&gt; &lt;/cliOutput&gt;
&gt; &lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&gt; &lt;cliOutput&gt;
&gt;   &lt;volQuota&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir&lt;/path&gt;
&gt;       &lt;hard_limit&gt;10.0MB&lt;/hard_limit&gt;
&gt;       &lt;soft_limit&gt;80%&lt;/soft_limit&gt;
&gt;       &lt;used_space&gt;0Bytes&lt;/used_space&gt;
&gt;       &lt;avail_space&gt;10.0MB&lt;/avail_space&gt;
&gt;     &lt;/limit&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir1&lt;/path&gt;
&gt;       &lt;hard_limit&gt;10.0MB&lt;/hard_limit&gt;
&gt;       &lt;soft_limit&gt;80%&lt;/soft_limit&gt;
&gt;       &lt;used_space&gt;0Bytes&lt;/used_space&gt;
&gt;       &lt;avail_space&gt;10.0MB&lt;/avail_space&gt;
&gt;     &lt;/limit&gt;
&gt;   &lt;/volQuota&gt;
&gt; &lt;/cliOutput&gt;
&gt;
&gt; Change-Id: I8a8d83cff88f778e5ee01fbca07d9f94c412317a
&gt; BUG: 1185259
&gt; Signed-off-by: Sachin Pandit &lt;spandit@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/9481
&gt; Reviewed-by: Vijaikumar Mallikarjuna &lt;vmallika@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;

Change-Id: Ibdf51db626a07e68b5ace98140877f6d21918c20
BUG: 1231641
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11220
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Sachin Pandit &lt;spandit@redhat.com&gt;
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is a backport of http://review.gluster.org/#/c/9481/

&gt; Sample output:
&gt; ---------------
&gt;
&gt; Sample 1)
&gt; ----------
&gt; [root@snapshot-28 glusterfs]# gluster volume quota vol1 list /dir1 /dir4
&gt; /dir5 --xml
&gt; &lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&gt; &lt;cliOutput&gt;
&gt;   &lt;opRet&gt;0&lt;/opRet&gt;
&gt;   &lt;opErrno&gt;0&lt;/opErrno&gt;
&gt;   &lt;opErrstr/&gt;
&gt;   &lt;volQuota&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir1&lt;/path&gt;
&gt;       &lt;hard_limit&gt;10.0MB&lt;/hard_limit&gt;
&gt;       &lt;soft_limit&gt;80%&lt;/soft_limit&gt;
&gt;       &lt;used_space&gt;0Bytes&lt;/used_space&gt;
&gt;       &lt;avail_space&gt;10.0MB&lt;/avail_space&gt;
&gt;     &lt;/limit&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir4&lt;/path&gt;
&gt;       &lt;path&gt;No such file or directory&lt;/path&gt;
&gt;     &lt;/limit&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir5&lt;/path&gt;
&gt;       &lt;path&gt;No such file or directory&lt;/path&gt;
&gt;     &lt;/limit&gt;
&gt;   &lt;/volQuota&gt;
&gt; &lt;/cliOutput&gt;
&gt;
&gt; Sample 2)
&gt; ---------
&gt; gluster volume quota vol1 list --xml
&gt; &lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&gt; &lt;cliOutput&gt;
&gt;   &lt;opRet&gt;0&lt;/opRet&gt;
&gt;   &lt;opErrno&gt;0&lt;/opErrno&gt;
&gt;   &lt;opErrstr/&gt;
&gt;   &lt;volQuota/&gt;
&gt; &lt;/cliOutput&gt;
&gt; &lt;?xml version="1.0" encoding="UTF-8" standalone="yes"?&gt;
&gt; &lt;cliOutput&gt;
&gt;   &lt;volQuota&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir&lt;/path&gt;
&gt;       &lt;hard_limit&gt;10.0MB&lt;/hard_limit&gt;
&gt;       &lt;soft_limit&gt;80%&lt;/soft_limit&gt;
&gt;       &lt;used_space&gt;0Bytes&lt;/used_space&gt;
&gt;       &lt;avail_space&gt;10.0MB&lt;/avail_space&gt;
&gt;     &lt;/limit&gt;
&gt;     &lt;limit&gt;
&gt;       &lt;path&gt;/dir1&lt;/path&gt;
&gt;       &lt;hard_limit&gt;10.0MB&lt;/hard_limit&gt;
&gt;       &lt;soft_limit&gt;80%&lt;/soft_limit&gt;
&gt;       &lt;used_space&gt;0Bytes&lt;/used_space&gt;
&gt;       &lt;avail_space&gt;10.0MB&lt;/avail_space&gt;
&gt;     &lt;/limit&gt;
&gt;   &lt;/volQuota&gt;
&gt; &lt;/cliOutput&gt;
&gt;
&gt; Change-Id: I8a8d83cff88f778e5ee01fbca07d9f94c412317a
&gt; BUG: 1185259
&gt; Signed-off-by: Sachin Pandit &lt;spandit@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/9481
&gt; Reviewed-by: Vijaikumar Mallikarjuna &lt;vmallika@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;

Change-Id: Ibdf51db626a07e68b5ace98140877f6d21918c20
BUG: 1231641
Signed-off-by: vmallika &lt;vmallika@redhat.com&gt;
Reviewed-on: http://review.gluster.org/11220
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Sachin Pandit &lt;spandit@redhat.com&gt;
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cli: Fix xml output for volume status</title>
<updated>2014-07-21T07:11:19+00:00</updated>
<author>
<name>Niels de Vos</name>
<email>ndevos@redhat.com</email>
</author>
<published>2014-07-18T17:05:26+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=d5f72dc49604aec2643d92a1b4e321c532ef8d05'/>
<id>d5f72dc49604aec2643d92a1b4e321c532ef8d05</id>
<content type='text'>
The XML output for volume status was malformed when one of the nodes is
down, leading to outputs like
-------
          &lt;node&gt;
             &lt;node&gt;
               &lt;hostname&gt;NFS Server&lt;/hostname&gt;
               &lt;path&gt;localhost&lt;/path&gt;
               &lt;peerid&gt;63ca3d2f-8c1f-4b84-b797-b4baddab81fb&lt;/peerid&gt;
               &lt;status&gt;1&lt;/status&gt;
               &lt;port&gt;2049&lt;/port&gt;
               &lt;pid&gt;2130&lt;/pid&gt;
             &lt;/node&gt;
-----

This was happening because we were starting the &lt;node&gt; element before
determining if node was present, and were not closing it or clearing it
when not finding the node in the dict.

To fix this, the &lt;node&gt; element is only started once a node has been
found in the dict.

Cherry picked from commit 2ba42d07eb967472227eb0a93e4ca2cac7a197b5:
&gt; Change-Id: I6b6205f14b27a69adb95d85db7b48999aa48d400
&gt; BUG: 1046020
&gt; Signed-off-by: Kaushal M &lt;kaushal@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/6571
&gt; Reviewed-by: Aravinda VK &lt;avishwan@redhat.com&gt;
&gt; Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;

Change-Id: I6b6205f14b27a69adb95d85db7b48999aa48d400
BUG: 1117241
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/8334
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
The XML output for volume status was malformed when one of the nodes is
down, leading to outputs like
-------
          &lt;node&gt;
             &lt;node&gt;
               &lt;hostname&gt;NFS Server&lt;/hostname&gt;
               &lt;path&gt;localhost&lt;/path&gt;
               &lt;peerid&gt;63ca3d2f-8c1f-4b84-b797-b4baddab81fb&lt;/peerid&gt;
               &lt;status&gt;1&lt;/status&gt;
               &lt;port&gt;2049&lt;/port&gt;
               &lt;pid&gt;2130&lt;/pid&gt;
             &lt;/node&gt;
-----

This was happening because we were starting the &lt;node&gt; element before
determining if node was present, and were not closing it or clearing it
when not finding the node in the dict.

To fix this, the &lt;node&gt; element is only started once a node has been
found in the dict.

Cherry picked from commit 2ba42d07eb967472227eb0a93e4ca2cac7a197b5:
&gt; Change-Id: I6b6205f14b27a69adb95d85db7b48999aa48d400
&gt; BUG: 1046020
&gt; Signed-off-by: Kaushal M &lt;kaushal@redhat.com&gt;
&gt; Reviewed-on: http://review.gluster.org/6571
&gt; Reviewed-by: Aravinda VK &lt;avishwan@redhat.com&gt;
&gt; Reviewed-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
&gt; Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
&gt; Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;

Change-Id: I6b6205f14b27a69adb95d85db7b48999aa48d400
BUG: 1117241
Signed-off-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Reviewed-on: http://review.gluster.org/8334
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>NetBSD build fixes for release-3.5</title>
<updated>2014-06-08T09:18:13+00:00</updated>
<author>
<name>Emmanuel Dreyfus</name>
<email>manu@netbsd.org</email>
</author>
<published>2014-05-29T12:42:24+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=587bd2b4b7b7076dd469cabf9b8ccf77fca87cfc'/>
<id>587bd2b4b7b7076dd469cabf9b8ccf77fca87cfc</id>
<content type='text'>
1) Make sure __THROW is definedThis is a backport of I6e7cb1eb59b84988e155e9a8b696e842b7ff8f7f

2) include &lt;rpc/xdr.h&gt; before &lt;rpc/auth.h&gt; so that XDR is defined
This was fixed in master within I20193d3f8904388e47344e523b3787dbeab044acbut weonly pull up

3) NetBSD's gettext is in libintl, hence search it at configure time
This is a backport of I651a74fe49c3f087fe135dab3453fd5b18b4268a

4) include &lt;sys/wait.h&gt; to have WEXITSTATUS defined
This problem does not exist in master as WEXITSTATUS is not used

5) Do not define popcountl() on NetBSD as it is in &lt;strings.h&gt;
This is a backport of I4428a88b1e0d7c5f6740022861ffe230dbbd84bd

BUG: 764655
Change-Id: Ieea5a2a627e2b7930525d6c525f1602073574a97
Signed-off-by: Emmanuel Dreyfus &lt;manu@netbsd.org&gt;
Reviewed-on: http://review.gluster.org/7925
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
1) Make sure __THROW is definedThis is a backport of I6e7cb1eb59b84988e155e9a8b696e842b7ff8f7f

2) include &lt;rpc/xdr.h&gt; before &lt;rpc/auth.h&gt; so that XDR is defined
This was fixed in master within I20193d3f8904388e47344e523b3787dbeab044acbut weonly pull up

3) NetBSD's gettext is in libintl, hence search it at configure time
This is a backport of I651a74fe49c3f087fe135dab3453fd5b18b4268a

4) include &lt;sys/wait.h&gt; to have WEXITSTATUS defined
This problem does not exist in master as WEXITSTATUS is not used

5) Do not define popcountl() on NetBSD as it is in &lt;strings.h&gt;
This is a backport of I4428a88b1e0d7c5f6740022861ffe230dbbd84bd

BUG: 764655
Change-Id: Ieea5a2a627e2b7930525d6c525f1602073574a97
Signed-off-by: Emmanuel Dreyfus &lt;manu@netbsd.org&gt;
Reviewed-on: http://review.gluster.org/7925
Reviewed-by: Kaleb KEITHLEY &lt;kkeithle@redhat.com&gt;
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>heal: glfs-heal implementation</title>
<updated>2014-04-28T16:39:39+00:00</updated>
<author>
<name>Pranith Kumar K</name>
<email>pkarampu@redhat.com</email>
</author>
<published>2013-12-15T02:35:04+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=75cb75fe8207eb4407253b42bb2d0d8f2c7341ee'/>
<id>75cb75fe8207eb4407253b42bb2d0d8f2c7341ee</id>
<content type='text'>
Change-Id: I634f24d90cd856ceab3cc0c6e9a91003f443403e
BUG: 1039544
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6511
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Change-Id: I634f24d90cd856ceab3cc0c6e9a91003f443403e
BUG: 1039544
Signed-off-by: Pranith Kumar K &lt;pkarampu@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6511
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Ravishankar N &lt;ravishankar@redhat.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cli: Deprecation message added for remove-brick op with default behaviour.</title>
<updated>2014-04-03T09:07:05+00:00</updated>
<author>
<name>Atin Mukherjee</name>
<email>amukherj@redhat.com</email>
</author>
<published>2014-03-20T05:16:25+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=92d3d8b8cb8fc5886d71fe184339fdcbeb5439db'/>
<id>92d3d8b8cb8fc5886d71fe184339fdcbeb5439db</id>
<content type='text'>
Background : From 3.6 version and onwards, remove-brick can be executed with
explicit options only.

Change-Id: Ibe376e371c5aa7a68621cf4ec2e74c6809614f9b
BUG: 1077682
Signed-off-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/7302
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Background : From 3.6 version and onwards, remove-brick can be executed with
explicit options only.

Change-Id: Ibe376e371c5aa7a68621cf4ec2e74c6809614f9b
BUG: 1077682
Signed-off-by: Atin Mukherjee &lt;amukherj@redhat.com&gt;
Reviewed-on: http://review.gluster.org/7302
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cli: xml: Rebalance status(xml) was empty when a glusterd down</title>
<updated>2014-02-06T09:40:33+00:00</updated>
<author>
<name>Aravinda VK</name>
<email>avishwan@redhat.com</email>
</author>
<published>2013-12-02T09:49:17+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=cfaab6bcce48fcd369d9c8d66be4413d1d0b8356'/>
<id>cfaab6bcce48fcd369d9c8d66be4413d1d0b8356</id>
<content type='text'>
When a glusterd is down in cluster rebalance/remove-brick status
--xml will fail to get status and returns null.

This patch skips collecting status if glusterd is down, and
collects status from all the other up nodes.

BUG: 1036564
Change-Id: Id8fbb63476e136296231d6652a8bd1a4547edbf5
Signed-off-by: Aravinda VK &lt;avishwan@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6391
Reviewed-on: http://review.gluster.org/6848
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
When a glusterd is down in cluster rebalance/remove-brick status
--xml will fail to get status and returns null.

This patch skips collecting status if glusterd is down, and
collects status from all the other up nodes.

BUG: 1036564
Change-Id: Id8fbb63476e136296231d6652a8bd1a4547edbf5
Signed-off-by: Aravinda VK &lt;avishwan@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6391
Reviewed-on: http://review.gluster.org/6848
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd/geo-rep: more glusterd and cli fixes for geo-rep.</title>
<updated>2014-01-27T17:44:33+00:00</updated>
<author>
<name>Ajeet Jha</name>
<email>ajha@redhat.com</email>
</author>
<published>2013-12-02T07:25:18+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=f33d09f23b089bd07437eb714f8dffa43460d6b5'/>
<id>f33d09f23b089bd07437eb714f8dffa43460d6b5</id>
<content type='text'>
    -&gt; handle option validation cases in reset case.
    -&gt; Creating valid conf path when glusterd restarts.
    -&gt; Reading the gsyncd worker thread status and displaying it.
    -&gt; Displaying status-detail per worker.
    -&gt; Fetch checkpoint info in geo-rep status.
    -&gt; use-tarssh value validation added.

misc: misc geo-rep fixes based on cluster, logrotate etc..
    -&gt; cluster/dht: fix 'stime' getxattr getting overwritten.
    -&gt; cluster/afr: return max of 'stime' values in subvol.
    -&gt; geo-rep-logrotate: Sending SIGHUP to geo-rep auxiliary.
    -&gt; cluster/dht: fix convoluted logic while aggregating.
    -&gt; cluster/*: fix 'stime' min/max fetch logic.

Change-Id: I811acea0bbd6194797a3e55d89295d1ea021ac85
BUG: 1036552
Signed-off-by: Ajeet Jha &lt;ajha@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6405
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amarts@gmail.com&gt;
Reviewed-by: Anand Avati &lt;avati@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6810
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
    -&gt; handle option validation cases in reset case.
    -&gt; Creating valid conf path when glusterd restarts.
    -&gt; Reading the gsyncd worker thread status and displaying it.
    -&gt; Displaying status-detail per worker.
    -&gt; Fetch checkpoint info in geo-rep status.
    -&gt; use-tarssh value validation added.

misc: misc geo-rep fixes based on cluster, logrotate etc..
    -&gt; cluster/dht: fix 'stime' getxattr getting overwritten.
    -&gt; cluster/afr: return max of 'stime' values in subvol.
    -&gt; geo-rep-logrotate: Sending SIGHUP to geo-rep auxiliary.
    -&gt; cluster/dht: fix convoluted logic while aggregating.
    -&gt; cluster/*: fix 'stime' min/max fetch logic.

Change-Id: I811acea0bbd6194797a3e55d89295d1ea021ac85
BUG: 1036552
Signed-off-by: Ajeet Jha &lt;ajha@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6405
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Amar Tumballi &lt;amarts@gmail.com&gt;
Reviewed-by: Anand Avati &lt;avati@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6810
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>Create /var/run/gluster at install time</title>
<updated>2014-01-23T04:43:29+00:00</updated>
<author>
<name>Emmanuel Dreyfus</name>
<email>manu@netbsd.org</email>
</author>
<published>2014-01-22T10:13:05+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=0f8fb23ec622ec8f96649780804019721e1bd364'/>
<id>0f8fb23ec622ec8f96649780804019721e1bd364</id>
<content type='text'>
This is a backport of I67c33984a797bc92c7035724695c5ef2bbdf5267

BUG: 764655
Change-Id: I9b202bd8ede20dbf2f55af2361492f5d242547b2
Signed-off-by: Emmanuel Dreyfus &lt;manu@netbsd.org&gt;
Reviewed-on: http://review.gluster.org/6760
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
This is a backport of I67c33984a797bc92c7035724695c5ef2bbdf5267

BUG: 764655
Change-Id: I9b202bd8ede20dbf2f55af2361492f5d242547b2
Signed-off-by: Emmanuel Dreyfus &lt;manu@netbsd.org&gt;
Reviewed-on: http://review.gluster.org/6760
Reviewed-by: Niels de Vos &lt;ndevos@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>cli: Addition of new child elements under brick in volume info xml.</title>
<updated>2014-01-17T10:12:55+00:00</updated>
<author>
<name>ndarshan</name>
<email>dnarayan@redhat.com</email>
</author>
<published>2013-12-27T08:11:19+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=16218e529a7b38434d3618b551de1496456ee580'/>
<id>16218e529a7b38434d3618b551de1496456ee580</id>
<content type='text'>
Added new child elements; name and hostUuid under brick in
volume info xml where name and host uuid of the bricks are stored,
This does not break backward compatibility as the old value under
brick is not removed.

Change-Id: I95b690d90fc5df40aa62bc76621afa18f7c6073b
Signed-off-by: ndarshan &lt;dnarayan@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6604
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-on: http://review.gluster.org/6721
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
Added new child elements; name and hostUuid under brick in
volume info xml where name and host uuid of the bricks are stored,
This does not break backward compatibility as the old value under
brick is not removed.

Change-Id: I95b690d90fc5df40aa62bc76621afa18f7c6073b
Signed-off-by: ndarshan &lt;dnarayan@redhat.com&gt;
Reviewed-on: http://review.gluster.org/6604
Reviewed-by: Kaushal M &lt;kaushal@redhat.com&gt;
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-on: http://review.gluster.org/6721
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
<entry>
<title>glusterd: Aggregate tasks status in 'volume status [tasks]'</title>
<updated>2013-12-23T14:56:34+00:00</updated>
<author>
<name>Krishnan Parthasarathi</name>
<email>kparthas@redhat.com</email>
</author>
<published>2013-12-23T08:37:45+00:00</published>
<link rel='alternate' type='text/html' href='https://fedorapeople.org/cgit/anoopcs/public_git/glusterfs.git/commit/?id=9d592246d6121aa38cd6fb6a875be4473d4979c8'/>
<id>9d592246d6121aa38cd6fb6a875be4473d4979c8</id>
<content type='text'>
        Backport of http://review.gluster.org/6230
Previously, glusterd used to just send back the local status of a task
in a 'volume status [tasks]' command. As the rebalance operation is
distributed and asynchronus, this meant that different peers could give
different status values for a rebalance or remove-brick task.

With this patch, all the peers will send back the tasks status as a part
of the 'volume status' commit op, and the origin peer will aggregate
these to arrive at a final status for the task.

The aggregation is only done for rebalance or remove-brick tasks. The
replace-brick task will have the same status on all the peers (see
comment in glusterd_volume_status_aggregate_tasks_status() for more
information) and need not be aggregated.

The rebalance process has 5 states,
 NOT_STARTED - rebalance process has not been started on this node
 STARTED - rebalance process has been started and is still running
 STOPPED - rebalance process was stopped by a 'rebalance/remove-brick
           stop' command
 COMPLETED - rebalance process completed successfully
 FAILED - rebalance process failed to complete successfully
The aggregation is done using the following precedence,
 STARTED &gt; FAILED &gt; STOPPED &gt; COMPLETED &gt; NOT_STARTED

The new changes make the 'volume status tasks' command a distributed
command as we need to get the task status from all peers.

The following tests were performed,
- Start a remove-brick task and do a status command on a peer which
  doesn't have the brick being removed. The remove-brick status was
  given correctly as 'in progress' and 'completed', instead of 'not
  started'
- Start a rebalance task, run the status command. The status moved to
  'completed' only after rebalance completed on all nodes.

Also, change the CLI xml output code for rebalance status to use the
same algorithm for status aggregation.

Change-Id: Ifd4aff705aa51609a612d5a9194acc73e10a82c0
BUG: 1027094
Signed-off-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
 http://review.gluster.org/6230
Reviewed-on: http://review.gluster.org/6562
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</content>
<content type='xhtml'>
<div xmlns='http://www.w3.org/1999/xhtml'>
<pre>
        Backport of http://review.gluster.org/6230
Previously, glusterd used to just send back the local status of a task
in a 'volume status [tasks]' command. As the rebalance operation is
distributed and asynchronus, this meant that different peers could give
different status values for a rebalance or remove-brick task.

With this patch, all the peers will send back the tasks status as a part
of the 'volume status' commit op, and the origin peer will aggregate
these to arrive at a final status for the task.

The aggregation is only done for rebalance or remove-brick tasks. The
replace-brick task will have the same status on all the peers (see
comment in glusterd_volume_status_aggregate_tasks_status() for more
information) and need not be aggregated.

The rebalance process has 5 states,
 NOT_STARTED - rebalance process has not been started on this node
 STARTED - rebalance process has been started and is still running
 STOPPED - rebalance process was stopped by a 'rebalance/remove-brick
           stop' command
 COMPLETED - rebalance process completed successfully
 FAILED - rebalance process failed to complete successfully
The aggregation is done using the following precedence,
 STARTED &gt; FAILED &gt; STOPPED &gt; COMPLETED &gt; NOT_STARTED

The new changes make the 'volume status tasks' command a distributed
command as we need to get the task status from all peers.

The following tests were performed,
- Start a remove-brick task and do a status command on a peer which
  doesn't have the brick being removed. The remove-brick status was
  given correctly as 'in progress' and 'completed', instead of 'not
  started'
- Start a rebalance task, run the status command. The status moved to
  'completed' only after rebalance completed on all nodes.

Also, change the CLI xml output code for rebalance status to use the
same algorithm for status aggregation.

Change-Id: Ifd4aff705aa51609a612d5a9194acc73e10a82c0
BUG: 1027094
Signed-off-by: Krishnan Parthasarathi &lt;kparthas@redhat.com&gt;
 http://review.gluster.org/6230
Reviewed-on: http://review.gluster.org/6562
Tested-by: Gluster Build System &lt;jenkins@build.gluster.com&gt;
Reviewed-by: Vijay Bellur &lt;vbellur@redhat.com&gt;
</pre>
</div>
</content>
</entry>
</feed>
