Hey Milos,
I see that gfid got healed for those directories from the getfattr output
and the glfsheal log also has messages corresponding to deleting the
entries on one brick as part of healing which then got recreated on the
brick with the correct gfid. Can you run the "guster volume heal "
& "gl
Can you give me the stat & getfattr output of all those 6 entries from both
the bricks and the glfsheal-.log file from the node where you run
this command?
Meanwhile can you also try running this with the source-brick option?
On Thu, Mar 21, 2019 at 5:22 PM Milos Cuculovic wrote:
> Thank you Kar
Thank you Karthik,
I have run this for all files (see example below) and it says the file is not
in split-brain:
sudo gluster volume heal storage2 split-brain latest-mtime
/dms/final_archive/41be9ff5ec05c4b1c989c6053e709e59
Healing /dms/final_archive/41be9ff5ec05c4b1c989c6053e709e59 failed: Fi
Hi Milos,
Thanks for the logs and the getfattr output.
>From the logs I can see that there are 6 entries under the
directory "/data/data-cluster/dms/final_archive" named
41be9ff5ec05c4b1c989c6053e709e59
5543982fab4b56060aa09f667a8ae617
a8b7f31775eebc8d1867e7f9de7b6eaf
c1d3f3c2d7ae90e891e671e2f20d5
Can you attach the "glustershd.log" file which will be present under
"/var/log/glusterfs/" from both the nodes and the "stat" & "getfattr -d -m
. -e hex " output of all the entries listed in the heal
info output from both the bricks?
On Thu, Mar 21, 2019 at 3:54 PM Milos Cuculovic wrote:
> Than
Thanks Karthik!
I was trying to find some resolution methods from [2] but unfortunately none
worked (I can explain what I tried if needed).
> I guess the volume you are talking about is of type replica-2 (1x2).
That’s correct, aware of the arbiter solution but still didn’t took time to
implemen
Hi,
Note: I guess the volume you are talking about is of type replica-2 (1x2).
Usually replica 2 volumes are prone to split-brain. If you can consider
converting them to arbiter or replica-3, they will handle most of the cases
which can lead to slit-brains. For more information see [1].
Resolving
Since 24h, after upgrading from 4.0 to 4.1.7 one of the servers, the heal shows
this:
sudo gluster volume heal storage2 info
Brick storage3:/data/data-cluster
/dms/final_archive - Possibly undergoing heal
Status: Connected
Number of entries: 3
Brick storage4:/data/data-cluster
/dms/final_
I was now able to catch the split brain log:
sudo gluster volume heal storage2 info
Brick storage3:/data/data-cluster
/dms/final_archive - Is in split-brain
Status: Connected
Number of entries: 3
Brick storage4:/data/data-cluster
/dms/final_archive - Is in split-brain
Status: Connected
Nu