[ovirt-users] Re: [EXT] Re: v4.4.3 Node Cockpit Gluster deploy fails
Thank you Donald! Your and Strahil's suggested solutions regarding disabling multipath for the nvme drives were correct. The Gluster deployment progressed much further but stalled at TASK [gluster.features/roles/gluster_hci : Set granual-entry-heal on] ** task path: /etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/hci_volumes.yml:67 with "stdout": "One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals\nVolume heal failed." Specifically TASK [gluster.features/roles/gluster_hci : Set granual-entry-heal on] ** task path: /etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/hci_volumes.yml:67 failed: [fmov1n1.sn.dtcorp.com] (item={'volname': 'engine', 'brick': '/gluster_bricks/engine/engine', 'arbiter': 0}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["gluster", "volume", "heal", "engine", "granular-entry-heal", "enable"], "delta": "0:00:10.112451", "end": "2020-12-18 19:50:22.818741", "item": {"arbiter": 0, "brick": "/gluster_bricks/engine/engine", "volname": "engine"}, "msg": "non-zero return code", "rc": 107, "start": "2020-12-18 19:50:12.706290", "stderr": "", "stderr_lines": [], "stdout": "One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals\nVolume heal failed.", "stdout_lines": ["One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals", "Volume heal failed."]} failed: [fmov1n1.sn.dtcorp.com] (item={'volname': 'data', 'brick': '/gluster_bricks/data/data', 'arbiter': 0}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["gluster", "volume", "heal", "data", "granular-entry-heal", "enable"], "delta": "0:00:10.110165", "end": "2020-12-18 19:50:38.260277", "item": {"arbiter": 0, "brick": "/gluster_bricks/data/data", "volname": "data"}, "msg": "non-zero return code", "rc": 107, "start": "2020-12-18 19:50:28.150112", "stderr": "", "stderr_lines": [], "stdout": "One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals\nVolume heal failed.", "stdout_lines": ["One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals", "Volume heal failed."]} failed: [fmov1n1.sn.dtcorp.com] (item={'volname': 'vmstore', 'brick': '/gluster_bricks/vmstore/vmstore', 'arbiter': 0}) => {"ansible_loop_var": "item", "changed": true, "cmd": ["gluster", "volume", "heal", "vmstore", "granular-entry-heal", "enable"], "delta": "0:00:10.113203", "end": "2020-12-18 19:50:53.767864", "item": {"arbiter": 0, "brick": "/gluster_bricks/vmstore/vmstore", "volname": "vmstore"}, "msg": "non-zero return code", "rc": 107, "start": "2020-12-18 19:50:43.654661", "stderr": "", "stderr_lines": [], "stdout": "One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals\nVolume heal failed.", "stdout_lines": ["One or more bricks could be down. Please execute the command again after bringing all bricks online and finishing any pending heals", "Volume heal failed."]} As this is a different issue, I will post a new thread. Gratefully yours, Charles ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/EV6DND2FOX3RF2JBB37COW425ZCUVIHL/
[ovirt-users] Re: [EXT] Re: v4.4.3 Node Cockpit Gluster deploy fails
When I deployed my gluster hyperconverged setup using nvme drives I had to disable multipath for all my drives. I'm not sure if this is your issue but here are the instruction I followed to disable it. Create a custom multipath configuration file. # mkdir /etc/multipath/conf.d # touch /etc/multipath/conf.d/99-custom-multipath.conf Add the following content to the file, replacing with the name of the device to blacklist: blacklist { devnode "" } For example, to blacklist the /dev/sdb device, add the following: blacklist { devnode "sdb" } Restart multipathd. # systemctl restart multipathd -Original Message- From: Charles Lam Sent: Friday, December 18, 2020 11:51 AM To: users@ovirt.org Subject: [EXT] [ovirt-users] Re: v4.4.3 Node Cockpit Gluster deploy fails I have been asked if multipath has been disabled for the cluster's nvme drives. I have not enabled or disabled multipath for the nvme drives. In Gluster deploy Step 4 - Bricks I have checked "Multipath Configuration: Blacklist Gluster Devices." I have not performed any custom setup of nvme drives other than wiping them in between deployment attempts. Below is the output of lsscsi and multipath -ll on the first host after failed Gluster deployment and before cleanup. Thanks! Should I set up multipath? If so, if you could point me to documentation re setup for oVirt. I still have a lot to learn and appreciate any direction. [root@Host1 conf.d]# lsscsi [15:0:0:0] diskATA DELLBOSS VD 00-0 /dev/sda [17:0:0:0] process Marvell Console 1.01 - [N:0:33:1] diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme0n1 [N:1:33:1] diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme1n1 [N:2:33:1] diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme2n1 [N:3:33:1] diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme3n1 [N:4:33:1] diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme4n1 [root@Host1 conf.d]# multipath -ll eui.343756304d702022002538580004 dm-0 NVME,Dell Express Flash PM1725b 1.6TB SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw `-+- policy='service-time 0' prio=1 status=active `- 0:33:1:1 nvme0n1 259:1 active ready running eui.343756304d702054002538580004 dm-1 NVME,Dell Express Flash PM1725b 1.6TB SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw `-+- policy='service-time 0' prio=1 status=active `- 1:33:1:1 nvme1n1 259:0 active ready running eui.343756304d700763002538580004 dm-2 NVME,Dell Express Flash PM1725b 1.6TB SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw `-+- policy='service-time 0' prio=1 status=active `- 2:33:1:1 nvme2n1 259:3 active ready running eui.343756304d702047002538580004 dm-4 NVME,Dell Express Flash PM1725b 1.6TB SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw `-+- policy='service-time 0' prio=1 status=active `- 4:33:1:1 nvme4n1 259:4 active ready running eui.343756304d702046002538580004 dm-3 NVME,Dell Express Flash PM1725b 1.6TB SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw `-+- policy='service-time 0' prio=1 status=active `- 3:33:1:1 nvme3n1 259:2 active ready running ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/S42IKSHJ7NHFHZDIXYYNTZK7SNWZBJSJ/ ___ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/XSR3R5RN7YVCB5EBPLTY2JZCJHTAYLBB/