[ovirt-users] Re: [EXT] Re: v4.4.3 Node Cockpit Gluster deploy fails

2020-12-18 Thread Charles Lam
Thank you Donald!  Your and Strahil's suggested solutions regarding disabling 
multipath for the nvme drives were correct.  The Gluster deployment progressed 
much further but stalled at
TASK [gluster.features/roles/gluster_hci : Set granual-entry-heal on] 
**
task path: 
/etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/hci_volumes.yml:67
with
"stdout": "One or more bricks could be down. Please execute the command again 
after bringing all bricks online and finishing any pending heals\nVolume heal 
failed."

Specifically
TASK [gluster.features/roles/gluster_hci : Set granual-entry-heal on] **
task path: 
/etc/ansible/roles/gluster.features/roles/gluster_hci/tasks/hci_volumes.yml:67
failed: [fmov1n1.sn.dtcorp.com] (item={'volname': 'engine', 'brick': 
'/gluster_bricks/engine/engine', 'arbiter': 0}) => {"ansible_loop_var": "item", 
"changed": true, "cmd": ["gluster", "volume", "heal", "engine", 
"granular-entry-heal", "enable"], "delta": "0:00:10.112451", "end": "2020-12-18 
19:50:22.818741", "item": {"arbiter": 0, "brick": 
"/gluster_bricks/engine/engine", "volname": "engine"}, "msg": "non-zero return 
code", "rc": 107, "start": "2020-12-18 19:50:12.706290", "stderr": "", 
"stderr_lines": [], "stdout": "One or more bricks could be down. Please execute 
the command again after bringing all bricks online and finishing any pending 
heals\nVolume heal failed.", "stdout_lines": ["One or more bricks could be 
down. Please execute the command again after bringing all bricks online and 
finishing any pending heals", "Volume heal failed."]}
failed: [fmov1n1.sn.dtcorp.com] (item={'volname': 'data', 'brick': 
'/gluster_bricks/data/data', 'arbiter': 0}) => {"ansible_loop_var": "item", 
"changed": true, "cmd": ["gluster", "volume", "heal", "data", 
"granular-entry-heal", "enable"], "delta": "0:00:10.110165", "end": "2020-12-18 
19:50:38.260277", "item": {"arbiter": 0, "brick": "/gluster_bricks/data/data", 
"volname": "data"}, "msg": "non-zero return code", "rc": 107, "start": 
"2020-12-18 19:50:28.150112", "stderr": "", "stderr_lines": [], "stdout": "One 
or more bricks could be down. Please execute the command again after bringing 
all bricks online and finishing any pending heals\nVolume heal failed.", 
"stdout_lines": ["One or more bricks could be down. Please execute the command 
again after bringing all bricks online and finishing any pending heals", 
"Volume heal failed."]}
failed: [fmov1n1.sn.dtcorp.com] (item={'volname': 'vmstore', 'brick': 
'/gluster_bricks/vmstore/vmstore', 'arbiter': 0}) => {"ansible_loop_var": 
"item", "changed": true, "cmd": ["gluster", "volume", "heal", "vmstore", 
"granular-entry-heal", "enable"], "delta": "0:00:10.113203", "end": "2020-12-18 
19:50:53.767864", "item": {"arbiter": 0, "brick": 
"/gluster_bricks/vmstore/vmstore", "volname": "vmstore"}, "msg": "non-zero 
return code", "rc": 107, "start": "2020-12-18 19:50:43.654661", "stderr": "", 
"stderr_lines": [], "stdout": "One or more bricks could be down. Please execute 
the command again after bringing all bricks online and finishing any pending 
heals\nVolume heal failed.", "stdout_lines": ["One or more bricks could be 
down. Please execute the command again after bringing all bricks online and 
finishing any pending heals", "Volume heal failed."]}

As this is a different issue, I will post a new thread.

Gratefully yours,
Charles
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/EV6DND2FOX3RF2JBB37COW425ZCUVIHL/


[ovirt-users] Re: [EXT] Re: v4.4.3 Node Cockpit Gluster deploy fails

2020-12-18 Thread Franklin, Donald
When I deployed my gluster hyperconverged setup using nvme drives I had to 
disable multipath for all my drives.  I'm not sure if this is your issue but 
here are the instruction I followed to disable it.

Create a custom multipath configuration file.

# mkdir /etc/multipath/conf.d
# touch /etc/multipath/conf.d/99-custom-multipath.conf

Add the following content to the file, replacing  with the name of the 
device to blacklist:

blacklist {
  devnode ""
}

For example, to blacklist the /dev/sdb device, add the following:

blacklist {
  devnode "sdb"
}

Restart multipathd.

# systemctl restart multipathd




-Original Message-
From: Charles Lam  
Sent: Friday, December 18, 2020 11:51 AM
To: users@ovirt.org
Subject: [EXT] [ovirt-users] Re: v4.4.3 Node Cockpit Gluster deploy fails

I have been asked if multipath has been disabled for the cluster's nvme drives.

I have not enabled or disabled multipath for the nvme drives.  In Gluster 
deploy Step 4 - Bricks I have checked "Multipath Configuration: Blacklist 
Gluster Devices."  I have not performed any custom setup of nvme drives other 
than wiping them in between deployment attempts.  Below is the output of lsscsi 
and multipath -ll on the first host after failed Gluster deployment and before 
cleanup.

Thanks!  Should I set up multipath?  If so, if you could point me to 
documentation re setup for oVirt.  I still have a lot to learn and appreciate 
any direction.

[root@Host1 conf.d]# lsscsi
[15:0:0:0]   diskATA  DELLBOSS VD  00-0  /dev/sda
[17:0:0:0]   process Marvell  Console  1.01  -
[N:0:33:1]   diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme0n1
[N:1:33:1]   diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme1n1
[N:2:33:1]   diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme2n1
[N:3:33:1]   diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme3n1
[N:4:33:1]   diskDell Express Flash PM1725b 1.6TB SFF__1/dev/nvme4n1
[root@Host1 conf.d]# multipath -ll
eui.343756304d702022002538580004 dm-0 NVME,Dell Express Flash PM1725b 1.6TB 
SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
  `- 0:33:1:1 nvme0n1 259:1 active ready running
eui.343756304d702054002538580004 dm-1 NVME,Dell Express Flash PM1725b 1.6TB 
SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
  `- 1:33:1:1 nvme1n1 259:0 active ready running
eui.343756304d700763002538580004 dm-2 NVME,Dell Express Flash PM1725b 1.6TB 
SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
  `- 2:33:1:1 nvme2n1 259:3 active ready running
eui.343756304d702047002538580004 dm-4 NVME,Dell Express Flash PM1725b 1.6TB 
SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
  `- 4:33:1:1 nvme4n1 259:4 active ready running
eui.343756304d702046002538580004 dm-3 NVME,Dell Express Flash PM1725b 1.6TB 
SFF size=1.5T features='1 queue_if_no_path' hwhandler='0' wp=rw
`-+- policy='service-time 0' prio=1 status=active
  `- 3:33:1:1 nvme3n1 259:2 active ready running 
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: 
https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/S42IKSHJ7NHFHZDIXYYNTZK7SNWZBJSJ/
___
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/privacy-policy.html
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/XSR3R5RN7YVCB5EBPLTY2JZCJHTAYLBB/