On 15 June 2016 at 06:48, Atin Mukherjee <amukh...@redhat.com> wrote:

>
>
> On 06/15/2016 11:06 AM, Gandalf Corvotempesta wrote:
> > Il 15 giu 2016 07:09, "Atin Mukherjee" <amukh...@redhat.com
> > <mailto:amukh...@redhat.com>> ha scritto:
> >> To get rid of this situation you'd need to stop all the running glusterd
> >> instances and go into /var/lib/glusterd/peers folder on all the nodes
> >> and manually correct the UUID file names and their content if required.
> >
> > If i understood properly the only way to fix this is by bringing the
> > whole cluster down? "you'd need to stop all the running glusterd
> instances"
> >
> > I hope you are referring to all instances on the failed node...
>
> No, since the configuration are synced across all the nodes, any
> incorrect data gets replicated through out. So in this case to be on the
> safer side and validate the correctness all glusterd instances on *all*
> the nodes should be brought down. Having said that, this doesn't impact
> I/O as the management path is different than I/O.
>
>
As a sanity, one of the things I did last night, was to reboot the whole
gluster system, when I had downtime arranged. I thought this is something
would be asked, as I had seen similar requests on the mailing list
previously

Unfortunately though, it didn't fix the problem.

Any other suggestions are welcome
_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to