Thanks for sharing! On Sun, Jul 8, 2018 at 1:42 PM, Hesham Ahmed <hsah...@gmail.com> wrote:
> Apparently this is known default behavior of LVM snapshots and in that > case maybe Cockpit in oVirt node should create mountpoints using > /dev/mapper path instead of UUID by default. The timeout issue > persists even after switching to /dev/mapper/devices in fstab > On Sun, Jul 8, 2018 at 12:59 PM Hesham Ahmed <hsah...@gmail.com> wrote: > > > > I also noticed that Gluster Snapshots have the SAME UUID as the main > > LV and if using UUID in fstab, the snapshot device is sometimes > > mounted instead of the primary LV > > > > For instance: > > /etc/fstab contains the following line: > > > > UUID=a0b85d33-7150-448a-9a70-6391750b90ad /gluster_bricks/gv01_data01 > > auto inode64,noatime,nodiratime,x-parent=dMeNGb-34lY-wFVL-WF42- > hlpE-TteI-lMhvvt > > 0 0 > > > > # lvdisplay gluster00/lv01_data01 > > --- Logical volume --- > > LV Path /dev/gluster00/lv01_data01 > > LV Name lv01_data01 > > VG Name gluster00 > > > > # mount > > /dev/mapper/gluster00-55e97e7412bf48db99bb389bb708edb8_0 on > > /gluster_bricks/gv01_data01 type xfs > > (rw,noatime,nodiratime,seclabel,attr2,inode64,sunit= > 1024,swidth=2048,noquota) > > > > Notice above the device mounted at the brick mountpoint is not > > /dev/gluster00/lv01_data01 and instead is one of the snapshot devices > > of that LV > > > > # blkid > > /dev/mapper/gluster00-lv01_shaker_com_sa: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-55e97e7412bf48db99bb389bb708edb8_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-4ca8eef409ec4932828279efb91339de_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-59992b6c14644f13b5531a054d2aa75c_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-362b50c994b04284b1664b2e2eb49d09_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-0b3cc414f4cb4cddb6e81f162cdb7efe_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-da98ce5efda549039cf45a18e4eacbaf_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > /dev/mapper/gluster00-4ea5cce4be704dd7b29986ae6698a666_0: > > UUID="a0b85d33-7150-448a-9a70-6391750b90ad" TYPE="xfs" > > > > Notice the UUID of LV and its snapshots is the same causing systemd to > > mount one of the snapshot devices instead of LV which results in the > > following gluster error: > > > > gluster> volume start gv01_data01 force > > volume start: gv01_data01: failed: Volume id mismatch for brick > > vhost03:/gluster_bricks/gv01_data01/gv. Expected volume id > > be6bc69b-c6ed-4329-b300-3b9044f375e1, volume id > > 55e97e74-12bf-48db-99bb-389bb708edb8 found > > > > On Sun, Jul 8, 2018 at 12:32 PM <hsah...@gmail.com> wrote: > > > > > > I am facing this trouble since version 4.1 up to the latest 4.2.4, once > > > we enable gluster snapshots and accumulate some snapshots (as few as 15 > > > snapshots per server) we start having trouble booting the server. The > > > server enters emergency shell upon boot after timing out waiting for > > > snapshot devices. Waiting a few minutes and pressing Control-D then > boots the server normally. In case of very large number of snapshots (600+) > it can take days before the sever will boot. Attaching journal > > > log, let me know if you need any other logs. > > > > > > Details of the setup: > > > > > > 3 node hyperconverged oVirt setup (64GB RAM, 8-Core E5 Xeon) > > > oVirt 4.2.4 > > > oVirt Node 4.2.4 > > > 10Gb Interface > > > > > > Thanks, > > > > > > Hesham S. Ahmed > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/site/privacy-policy/ > oVirt Code of Conduct: https://www.ovirt.org/community/about/community- > guidelines/ > List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/ > message/ZOHBIIIKKVZDWGWRXKZ5GEZOADLCSGJB/ >
_______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/site/privacy-policy/ oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/STAK2QO6Q3YDE4LAP7UBPNMEURV24WMP/