On 05/27/2014 08:47 PM, Ivano Talamo wrote:
Dear all,
we have a replicated volume (2 servers with 1 brick each) on
Scientific Linux 6.2 with gluster 3.4.
Everything was running fine until we shutdown of of the two and kept
it down for 2 months.
When it came up again the volume could be healed and we have the
following symptoms
(call #1 the always-up server, #2 the server that was kept down):
-doing I/O on the volume has very bad performances (impossible to keep
VM images on it)
A replica's bricks are not supposed to be intentionally kept down even
for hours, leave alone months :-( ; If you do; then when it does come
backup, there would be tons of stuff to heal, so a performance hit is
expected.
-on #1 there's 3997354 files on .glusterfs/indices/xattrop/ and the
number doesn't go down
When #2 was down, did the I/O involve directory renames? (see if there
are entries on .glusterfs/landfill on #2). If yes then this is a known
issue and a fix is in progress : http://review.gluster.org/#/c/7879/
-on #1 gluster volume heal vol1 info the first time takes a lot to end
and doesn't show nothing.
This is fixed in glusterfs 3.5 where heal info is much more responsive.
after that it prints "Another transaction is in progress. Please try
again after sometime."
Furthermore on #1 glusterhd.log is full of messages like this:
[2014-05-27 15:07:44.145326] W
[client-rpc-fops.c:1538:client3_3_inodelk_cbk] 0-vol1-client-0: remote
operation failed: No such file or directory
[2014-05-27 15:07:44.145880] W
[client-rpc-fops.c:1640:client3_3_entrylk_cbk] 0-vol1-client-0: remote
operation failed: No such file or directory
[2014-05-27 15:07:44.146070] E
[afr-self-heal-entry.c:2296:afr_sh_post_nonblocking_entry_cbk]
0-vol1-replicate-0: Non Blocking entrylks failed for
<gfid:bfbe65db-7426-4ca0-bf0b-7d1a28de2052>.
[2014-05-27 15:13:34.772856] E
[afr-self-heal-data.c:1270:afr_sh_data_open_cbk] 0-vol1-replicate-0:
open of <gfid:18a358e0-23d3-4f56-8d74-f5cc38a0d0ea> failed on child
vol1-client-0 (No such file or directory)
On #2 bricks I see some updates, ie. new filenames appearing and
.glusterfs/indices/xattrop/ is usually empy.
Do you know what's happening? How can we fix this?
You could try a `gluster volume heal vol1 full` to see if the bricks get
synced.
Regards,
Ravi
thank you,
Ivano
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users
_______________________________________________
Gluster-users mailing list
[email protected]
http://supercolony.gluster.org/mailman/listinfo/gluster-users