I don't understand why there's such a complicated process to recover when I can just look at both files, decide which one I need and delete another one.
On Thu, Sep 11, 2014 at 7:56 AM, Pranith Kumar Karampuri < pkara...@redhat.com> wrote: > > On 09/11/2014 09:29 AM, Ilya Ivanov wrote: > > Right... I deleted it and now all appears to be fine. > > Still, could you please elaborate on "gfid split-brain"? > > Could you go through > https://github.com/gluster/glusterfs/blob/master/doc/debugging/split-brain.md > Let us know if you would like something to be more clearer and we can add > that and improve the document. > > Pranith > > > > On Thu, Sep 11, 2014 at 5:32 AM, Pranith Kumar Karampuri < > pkara...@redhat.com> wrote: > >> >> On 09/11/2014 12:16 AM, Ilya Ivanov wrote: >> >> Any insight? >> >> Was the other file's gfid d3def9e1-c6d0-4b7d-a322-b5019305182e? >> Could you check if this file exists in <brick>/.glusterfs/d3/de/ >> When a file is deleted this file also needs to be deleted if there are no >> more hardlinks to the file >> >> Pranith >> >> >> On Tue, Sep 9, 2014 at 8:35 AM, Ilya Ivanov <bearw...@gmail.com> wrote: >> >>> What's a gfid split-brain and how is it different from "normal" >>> split-brain? >>> >>> I accessed the file with stat, but "heal info" still shows "Number of >>> entries: 1" >>> >>> [root@gluster1 gluster]# getfattr -d -m. -e hex gv01/123 >>> # getfattr -d -m. -e hex gv01/123 >>> # file: gv01/123 >>> trusted.afr.gv01-client-0=0x000000000000000000000000 >>> trusted.afr.gv01-client-1=0x000000000000000000000000 >>> trusted.gfid=0x35f86f4561134ba0bd1b94ef70179d4d >>> >>> [root@gluster1 gluster]# getfattr -d -m. -e hex gv01 >>> # file: gv01 >>> trusted.afr.gv01-client-0=0x000000000000000000000000 >>> trusted.afr.gv01-client-1=0x000000000000000000000000 >>> trusted.gfid=0x00000000000000000000000000000001 >>> trusted.glusterfs.dht=0x000000010000000000000000ffffffff >>> trusted.glusterfs.volume-id=0x31a2c4c486ca4344b838d2c2e6c716c1 >>> >>> >>> >>> On Tue, Sep 9, 2014 at 8:19 AM, Pranith Kumar Karampuri < >>> pkara...@redhat.com> wrote: >>> >>>> >>>> On 09/09/2014 11:35 AM, Ilya Ivanov wrote: >>>> >>>> Ahh, thank you, now I get it. I deleted it on one node and it >>>> replicated to another one. Now I get the following output: >>>> >>>> [root@gluster1 var]# gluster volume heal gv01 info >>>> Brick gluster1:/home/gluster/gv01/ >>>> <gfid:d3def9e1-c6d0-4b7d-a322-b5019305182e> >>>> Number of entries: 1 >>>> >>>> Brick gluster2:/home/gluster/gv01/ >>>> Number of entries: 0 >>>> >>>> Is it normal? Why the number of entries isn't reset to 0? >>>> >>>> If you access the file using ls/stat etc, it will be fixed. But before >>>> that could you please post the output of 'getfattr -d -m. -e hex >>>> file/path/in/backend/brick' and 'getfattr -d -m. -e hex >>>> parent/dir/to/file/path/in/backend/brick' >>>> >>>> Pranith >>>> >>>> >>>> >>>> And why wouldn't the file show up in split-brain before, anyway? >>>> >>>> Gfid split-brains are not shown in heal-info-split-brain yet. >>>> >>>> Pranith >>>> >>>> >>>> >>>> On Tue, Sep 9, 2014 at 7:46 AM, Pranith Kumar Karampuri < >>>> pkara...@redhat.com> wrote: >>>> >>>>> >>>>> On 09/09/2014 01:54 AM, Ilya Ivanov wrote: >>>>> >>>>> Hello. >>>>> >>>>> I've Gluster 3.5.2 on Centos 6. A primitive replicated volume, as >>>>> describe here >>>>> <https://www.digitalocean.com/community/tutorials/how-to-create-a-redundant-storage-pool-using-glusterfs-on-ubuntu-servers>. >>>>> I tried to simulate split-brain by temporarily disconnecting the nodes and >>>>> creating a file with the same name and different contents. That worked. >>>>> >>>>> The question is, how do I fix it now? All the tutorials suggest >>>>> deleting the file from one of the nodes. I can't do that, it reports >>>>> "Input/output error". The file won't even show up in "gluster volume heal >>>>> gv00 info split-brain". That shows 0 entries. >>>>> >>>>> The deletion needs to happen on one of the bricks, not from the mount >>>>> point. >>>>> >>>>> Pranith >>>>> >>>>> I can see the file in "gluster volume heal gv00 info heal-failed", >>>>> though. >>>>> >>>>> >>>>> -- >>>>> Ilya. >>>>> >>>>> >>>>> _______________________________________________ >>>>> Gluster-users mailing >>>>> listGluster-users@gluster.orghttp://supercolony.gluster.org/mailman/listinfo/gluster-users >>>>> >>>>> >>>>> >>>> >>>> >>>> -- >>>> Ilya. >>>> >>>> >>>> >>> >>> >>> -- >>> Ilya. >>> >> >> >> >> -- >> Ilya. >> >> >> > > > -- > Ilya. > > > -- Ilya.
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://supercolony.gluster.org/mailman/listinfo/gluster-users