Am 09.12.2015 um 17:17 schrieb Joe Julian:
A-1) shut down node #1 (the first that is about to be upgraded)
A-2) remove node #1 from the Proxmox cluster (/pvevm delnode "metal1"/)
A-3) remove node #1 from the Gluster volume/cluster (/gluster volume remove-brick ... && gluster peer detach "metal1"/) A-4) install Debian Jessie on node #1, overwriting all data on the HDD -*with same Network settings and hostname as before* A-5)install Proxmox 4.0 <https://pve.proxmox.com/wiki/Install_Proxmox_VE_on_Debian_Jessie>on node #1 A-6) install Gluster on node #1 and add it back to the Gluster volume (/gluster volume add-brick .../) => shared storage will be complete again (spanning 3.4 and 4.0 nodes) A-7) configure the Gluster volume as shared storage in Proxmox 4 (node #1)
A-8) configure the external Backup storage on node #1 (Proxmox 4)

Was the data on the gluster brick deleted as part of step 4?

Yes, all data on physical HDD was deleted (reformatted / repartitioned).


When you remove the brick, gluster will no longer track pending changes for that brick. If you add it back in with stale data but matching gfids, you would have two clean bricks with mismatching data. Did you have to use "add-brick...force"?

No, "force" was not necessary and the added directory "/data/gluster/systems" did not exist.

This were the commands executed on node #2 during step 6:

   gluster volume add-brick "systems" replica 3
   metal1:/data/gluster/systems
   gluster volume heal "systems" full   # to trigger sync


Then I waited for replication to finish before doing anything else (about 1 hour or maybe more), checking _gluster volume heal "systems" info_


Udo

_______________________________________________
Gluster-users mailing list
Gluster-users@gluster.org
http://www.gluster.org/mailman/listinfo/gluster-users

Reply via email to