Hi Anatoliy, The heal command is basically used to heal any mismatching contents between replica copies of the files. For the command "gluster volume heal <volname>" to succeed, you should have the self-heal-daemon running, which is true only if your volume is of type replicate/disperse. In your case you have a plain distribute volume where you do not store the replica of any files. So the volume heal will return you the error.
Regards, Karthik On Tue, Mar 13, 2018 at 7:53 PM, Anatoliy Dmytriyev <to...@tolid.eu.org> wrote: > Hi, > > > Maybe someone can point me to a documentation or explain this? I can't > find it myself. > Do we have any other useful resources except doc.gluster.org? As I see > many gluster options are not described there or there are no explanation > what is doing... > > > > On 2018-03-12 15:58, Anatoliy Dmytriyev wrote: > >> Hello, >> >> We have a very fresh gluster 3.10.10 installation. >> Our volume is created as distributed volume, 9 bricks 96TB in total >> (87TB after 10% of gluster disk space reservation) >> >> For some reasons I can’t “heal” the volume: >> # gluster volume heal gv0 >> Launching heal operation to perform index self heal on volume gv0 has >> been unsuccessful on bricks that are down. Please check if all brick >> processes are running. >> >> Which processes should be run on every brick for heal operation? >> >> # gluster volume status >> Status of volume: gv0 >> Gluster process TCP Port RDMA Port Online >> Pid >> ------------------------------------------------------------ >> ------------------ >> Brick cn01-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 70850 >> Brick cn02-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 102951 >> Brick cn03-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 57535 >> Brick cn04-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 56676 >> Brick cn05-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 56880 >> Brick cn06-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 56889 >> Brick cn07-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 56902 >> Brick cn08-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 94920 >> Brick cn09-ib:/gfs/gv0/brick1/brick 0 49152 Y >> 56542 >> >> Task Status of Volume gv0 >> ------------------------------------------------------------ >> ------------------ >> There are no active volume tasks >> >> >> # gluster volume info gv0 >> Volume Name: gv0 >> Type: Distribute >> Volume ID: 8becaf78-cf2d-4991-93bf-f2446688154f >> Status: Started >> Snapshot Count: 0 >> Number of Bricks: 9 >> Transport-type: rdma >> Bricks: >> Brick1: cn01-ib:/gfs/gv0/brick1/brick >> Brick2: cn02-ib:/gfs/gv0/brick1/brick >> Brick3: cn03-ib:/gfs/gv0/brick1/brick >> Brick4: cn04-ib:/gfs/gv0/brick1/brick >> Brick5: cn05-ib:/gfs/gv0/brick1/brick >> Brick6: cn06-ib:/gfs/gv0/brick1/brick >> Brick7: cn07-ib:/gfs/gv0/brick1/brick >> Brick8: cn08-ib:/gfs/gv0/brick1/brick >> Brick9: cn09-ib:/gfs/gv0/brick1/brick >> Options Reconfigured: >> client.event-threads: 8 >> performance.parallel-readdir: on >> performance.readdir-ahead: on >> cluster.nufa: on >> nfs.disable: on >> > > -- > Best regards, > Anatoliy > _______________________________________________ > Gluster-users mailing list > Gluster-users@gluster.org > http://lists.gluster.org/mailman/listinfo/gluster-users >
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://lists.gluster.org/mailman/listinfo/gluster-users