Step 10 isn't really necessary. The changes should probably be monitored under the brick directory.
On Thu, Oct 6, 2016 at 10:25 PM, Sergei Gerasenko <sgerasenk...@gmail.com> wrote: > I've simulated the problem on 4 VMs in a distributed replicated setup with > a 2 replica-factor. I've repeatedly torn down and brought up a VM from a > snapshot in each of my tests. > > What has worked so far is this: > > > 1. Make a copy of /var/lib/glusterd from the affected machine, save it > elsewhere. > 2. Configure your new machine (in my case I reverted to a VM > snapshot). Assign the same ip and hostname! > 3. Install gluster. > 4. Stop the daemons if they are running. > 5. Nuke the /var/lib/glusterd directory and replace it with the saved > copy in step 1. > 6. Create the brick directory. > 7. Get the extended volume attribute from a healthy node like so: getfattr > -e base64 -n trusted.glusterfs.volume-id /data/brick_dir > 8. Apply the extended attribute volume id attribute like so: setfattr > -n trusted.glusterfs.volume-id -v 'the_value_you_got_in_7==' > /data/brick_dir > 9. Start the daemons. > 10. FUSE mount the gluster partition through the daemons running > locally. So the /etc/fstab would contain something like: > localhost:/gluster_volume /mnt/gluster glusterfs _netdev,defaults 0 0 > 11. On the healthy partner machine with another fuse mount point to > the same volume do something like: find /mnt/fuse | xargs stat. > 12. Step 8 will make files appear under the mount point on the new box > but the files are not going to be physically in the brick directory -- yet. > See 10. > 13. Run the heal command from the same host where you ran find. That > will finally sync the files to the brick. Run the heal info command > periodically and the number of files being healed should eventually go down > to 0. > > That's my experience with the VMs today. > > On Wed, Oct 5, 2016 at 4:46 PM, Joe Julian <j...@julianfamily.org> wrote: > >> What I always do is just shut it down, repair (or replace) the brick, >> then start it up again with "... start $volname force". >> >> On October 5, 2016 11:27:36 PM GMT+02:00, Sergei Gerasenko < >> sgerasenk...@gmail.com> wrote: >>> >>> Hi, sorry if this has been asked before but the documentation is a bit >>> conflicting in various sources on what to do exactly. >>> >>> I have an 6-node, distributed replicated cluster with a replica factor >>> of 2. So it's 3 pairs of servers. I need to remove a server from one of >>> those replica sets, rebuild it and put it back in. >>> >>> What's the tried and proven sequence of steps for this? Any pointers >>> would be very useful. >>> >>> Thanks! >>> Sergei >>> >>> ------------------------------ >>> >>> Gluster-users mailing list >>> Gluster-users@gluster.org >>> http://www.gluster.org/mailman/listinfo/gluster-users >>> >>> >> -- >> Sent from my Android device with K-9 Mail. Please excuse my brevity. >> >> _______________________________________________ >> Gluster-users mailing list >> Gluster-users@gluster.org >> http://www.gluster.org/mailman/listinfo/gluster-users >> > >
_______________________________________________ Gluster-users mailing list Gluster-users@gluster.org http://www.gluster.org/mailman/listinfo/gluster-users