thank you Stefan for your tips !!
i do not do activation manualy, i use rc-local.service (details follows)
but i will reconfigure multipath in the evening and reboot both nodes
will let you know results by tomorrow
have a nice day ;-)
Nada

root@mox11:~# cat /etc/systemd/system/multi-user.target.wants/rc-local.service
[Unit]
Description=/etc/rc.local Compatibility
Documentation=man:systemd-rc-local-generator(8)
ConditionFileIsExecutable=/etc/rc.local
After=network.target iscsid.service multipathd.service open-iscsi.service

[Service]
Type=forking
ExecStart=/etc/rc.local
TimeoutSec=0
RemainAfterExit=yes
GuessMainPID=no

[Install]
WantedBy=multi-user.target

root@mox11:~# cat /etc/rc.local
#!/bin/bash
# just to activate VGs from SAN
/bin/sleep 10
/sbin/vgchange -aly santest
/sbin/vgchange -aly san2020jan


On 2020-01-15 12:06, Stefan M. Radman wrote:
Hi Nada,

Unfortunately I don't have any first hand experience with PVE6 yet.

On the PVE5.4 cluster I am working with I had an issue that looked
very similar to yours:
LVM refused to activate iSCSI multipath volumes on boot, making the
lvm2-activation-net.service fail.
This only happened during boot of the host.
Restarting the lvm2-activation-net.service after boot activated the
volume with multipath working.
Suspecting a timing/dependency issue specific to my configuration I
took a pragmatic approach and added a custom systemd service template
to restart the lvm2-activation-net.service after multipath
initialization (see below).

# cat /etc/systemd/system/lvm2-after-multipath.service
[Unit]
Description=LVM2 after Multipath
After=multipathd.service lvm2-activation-net.service

[Service]
Type=oneshot
ExecStart=/bin/systemctl start lvm2-activation-net.service

[Install]
WantedBy=sysinit.target

Things on PVE6 seem to have changed a bit but your lvm2-pvescan
service failures indicate a similar problem ("failed LVM event
activation").
Disable your rc.local workaround and try to restart the two failed
services after reboot.
If that works you might want to take a similar approach instead of
activating the volumes manually.

The masked status of multipath-tools-boot.service is ok. The package
is only needed when booting from multipath devices.

Your mistake in multipath.conf seems to be in the multipaths section.
Each multipath device can only have one WWID. For two WWIDs you'll
need two multiparty subsections.
https://help.ubuntu.com/lts/serverguide/multipath-dm-multipath-config-file.html#multipath-config-multipath

Stefan

On Jan 15, 2020, at 10:55, nada
<n...@verdnatura.es<mailto:n...@verdnatura.es>> wrote:

On 2020-01-14 19:46, Stefan M. Radman via pve-user wrote:
Hi Nada
What's the output of "systemctl --failed" and "systemctl status
lvm2-activation-net.service".
Stefan

Hi Stefan
thank you for your response !
the output of "systemctl --failed" was claiming devices from SAN during boot,
which were activated by rc-local.service after boot
i do NOT have "lvm2-activation-net.service"
and find masked status of multipath-tools-boot.service < is this ok ???
but i find some mistake in multipath.conf eventhough it is running
i am going to reconfigure it and reboot this afternoon
following are details
Nada

root@mox11:~# systemctl --failed --all
 UNIT                      LOAD   ACTIVE SUB    DESCRIPTION
 lvm2-pvscan@253:7.service loaded failed failed LVM event activation
on device 253:7
 lvm2-pvscan@253:8.service loaded failed failed LVM event activation
on device 253:8

LOAD   = Reflects whether the unit definition was properly loaded.
ACTIVE = The high-level unit activation state, i.e. generalization of SUB. SUB = The low-level unit activation state, values depend on unit type.

2 loaded units listed.
To show all installed unit files use 'systemctl list-unit-files'.

root@mox11:~# dmsetup ls
san2020jan-vm--903--disk--0 (253:18)
santest-santestpool (253:12)
3600c0ff000195f8e7d0a855701000000 (253:7)
pve-data-tpool (253:4)
pve-data_tdata (253:3)
pve-zfs (253:6)
santest-santestpool-tpool (253:11)
santest-santestpool_tdata (253:10)
pve-data_tmeta (253:2)
san2020jan-san2020janpool (253:17)
santest-santestpool_tmeta (253:9)
pve-swap (253:0)
pve-root (253:1)
pve-data (253:5)
3600c0ff000195f8ec3f01d5e01000000 (253:8)
san2020jan-san2020janpool-tpool (253:16)
san2020jan-san2020janpool_tdata (253:15)
san2020jan-san2020janpool_tmeta (253:14)

root@mox11:~# pvs -a
 PV                                            VG         Fmt  Attr
PSize   PFree
 /dev/mapper/3600c0ff000195f8e7d0a855701000000 santest    lvm2 a--
<9.31g 292.00m
 /dev/mapper/3600c0ff000195f8ec3f01d5e01000000 san2020jan lvm2 a--
<93.13g  <2.95g
 /dev/mapper/san2020jan-vm--903--disk--0                       ---
  0       0
 /dev/sdb                                                      ---
  0       0
 /dev/sdc2                                                     ---
  0       0
 /dev/sdc3                                     pve        lvm2 a--
67.73g   6.97g
root@mox11:~# vgs -a
 VG         #PV #LV #SN Attr   VSize   VFree
 pve          1   4   0 wz--n-  67.73g   6.97g
 san2020jan   1   2   0 wz--n- <93.13g  <2.95g
 santest      1   1   0 wz--n-  <9.31g 292.00m
root@mox11:~# lvs -a
 LV                     VG         Attr       LSize  Pool
Origin Data%  Meta%  Move Log Cpy%Sync Convert
 data                   pve        twi-aotz--  9.98g
    0.00   10.61
 [data_tdata]           pve        Twi-ao----  9.98g
 [data_tmeta]           pve        ewi-ao---- 12.00m
 [lvol0_pmspare]        pve        ewi------- 12.00m
 root                   pve        -wi-ao---- 16.75g
 swap                   pve        -wi-ao----  4.00g
 zfs                    pve        -wi-ao---- 30.00g
 [lvol0_pmspare]        san2020jan ewi------- 92.00m
 san2020janpool         san2020jan twi-aotz-- 90.00g
    0.86   10.84
 [san2020janpool_tdata] san2020jan Twi-ao---- 90.00g
 [san2020janpool_tmeta] san2020jan ewi-ao---- 92.00m
vm-903-disk-0 san2020jan Vwi-aotz-- 2.50g san2020janpool 30.99
 [lvol0_pmspare]        santest    ewi------- 12.00m
 santestpool            santest    twi-aotz--  9.00g
    0.00   10.58
 [santestpool_tdata]    santest    Twi-ao----  9.00g
 [santestpool_tmeta]    santest    ewi-ao---- 12.00m

root@mox11:~# multipathd -k"show maps"
Jan 15 10:50:02 | /etc/multipath.conf line 24, duplicate keyword: wwid
name                              sysfs uuid
3600c0ff000195f8e7d0a855701000000 dm-7 3600c0ff000195f8e7d0a855701000000 3600c0ff000195f8ec3f01d5e01000000 dm-8 3600c0ff000195f8ec3f01d5e01000000

root@mox11:~# multipathd -k"show paths"
Jan 15 10:50:07 | /etc/multipath.conf line 24, duplicate keyword: wwid
hcil     dev dev_t pri dm_st  chk_st dev_st  next_check
6:0:0:3  sde 8:64  10  active ready  running XX........ 2/8
9:0:0:3  sdn 8:208 10  active ready  running XX........ 2/8
7:0:0:3  sdh 8:112 50  active ready  running XX........ 2/8
5:0:0:3  sdd 8:48  10  active ready  running XXXXXX.... 5/8
11:0:0:3 sdp 8:240 50  active ready  running X......... 1/8
10:0:0:3 sdl 8:176 50  active ready  running XXXXXXXXXX 8/8
8:0:0:6  sdk 8:160 50  active ready  running XXXXXX.... 5/8
8:0:0:3  sdj 8:144 50  active ready  running XXXXXXXXXX 8/8
9:0:0:6  sdo 8:224 10  active ready  running X......... 1/8
6:0:0:6  sdg 8:96  10  active ready  running XXXXXX.... 5/8
5:0:0:6  sdf 8:80  10  active ready  running XXXXXX.... 5/8
10:0:0:6 sdm 8:192 50  active ready  running XXXXXXX... 6/8
11:0:0:6 sdq 65:0  50  active ready  running XXXXXXXX.. 7/8
7:0:0:6  sdi 8:128 50  active ready  running XXXXXXX... 6/8

root@mox11:~# cat /etc/multipath.conf
defaults {
       polling_interval        2
       uid_attribute           ID_SERIAL
no_path_retry queue
find_multipaths yes
}
blacklist {
wwid .*
# BECAREFULL @mox11 blacklit sda,sdb,sdc
devnode "^sd[a-c]"
}
blacklist_exceptions {
# 25G v_multitest
# wwid "3600c0ff000195f8e2172de5d01000000"
# 10G prueba
wwid "3600c0ff000195f8e7d0a855701000000"
# 100G sanmox11
wwid "3600c0ff000195f8ec3f01d5e01000000"
}
multipaths {
 multipath {
       # wwid "3600c0ff000195f8e2172de5d01000000"
wwid "3600c0ff000195f8e7d0a855701000000"
wwid "3600c0ff000195f8ec3f01d5e01000000"
 }
}
devices {
 device {
#### the following 6 lines do NOT change
vendor "HP"
product "P2000 G3 FC|P2000G3 FC/iSCSI|P2000 G3 SAS|P2000 G3 iSCSI"
# getuid_callout "/lib/udev/scsi_id -g -u -s /block/%n"
path_grouping_policy "group_by_prio"
prio "alua"
failback "immediate"
no_path_retry 18
####
hardware_handler "0"
path_selector "round-robin 0"
rr_weight uniform
rr_min_io 100
path_checker tur
 }
}


root@mox11:~# systemctl status multipath-tools-boot
 multipath-tools-boot.service
  Loaded: masked (Reason: Unit multipath-tools-boot.service is masked.)
  Active: inactive (dead)

root@mox11:~# pveversion -V
proxmox-ve: 6.1-2 (running kernel: 5.3.13-1-pve)
pve-manager: 6.1-5 (running version: 6.1-5/9bf06119)
pve-kernel-5.3: 6.1-1
pve-kernel-helper: 6.1-1
pve-kernel-4.15: 5.4-12
pve-kernel-5.3.13-1-pve: 5.3.13-1
pve-kernel-4.15.18-24-pve: 4.15.18-52
pve-kernel-4.15.18-21-pve: 4.15.18-48
pve-kernel-4.15.18-11-pve: 4.15.18-34
ceph-fuse: 12.2.11+dfsg1-2.1+b1
corosync: 3.0.2-pve4
criu: 3.11-3
glusterfs-client: 5.5-3
ifupdown: 0.8.35+pve1
ksm-control-daemon: 1.3-1
libjs-extjs: 6.0.1-10
libknet1: 1.13-pve1
libpve-access-control: 6.0-5
libpve-apiclient-perl: 3.0-2
libpve-common-perl: 6.0-9
libpve-guest-common-perl: 3.0-3
libpve-http-server-perl: 3.0-3
libpve-storage-perl: 6.1-3
libqb0: 1.0.5-1
libspice-server1: 0.14.2-4~pve6+1
lvm2: 2.03.02-pve3
lxc-pve: 3.2.1-1
lxcfs: 3.0.3-pve60
novnc-pve: 1.1.0-1
proxmox-mini-journalreader: 1.1-1
proxmox-widget-toolkit: 2.1-1
pve-cluster: 6.1-2
pve-container: 3.0-15
pve-docs: 6.1-3
pve-edk2-firmware: 2.20191127-1
pve-firewall: 4.0-9
pve-firmware: 3.0-4
pve-ha-manager: 3.0-8
pve-i18n: 2.0-3
pve-qemu-kvm: 4.1.1-2
pve-xtermjs: 3.13.2-1
qemu-server: 6.1-4
smartmontools: 7.0-pve2
spiceterm: 3.1-1
vncterm: 1.6-1
zfsutils-linux: 0.8.2-pve2
_______________________________________________
pve-user mailing list
pve-user@pve.proxmox.com<mailto:pve-user@pve.proxmox.com>
https://nam04.safelinks.protection.outlook.com/?url=https%3A%2F%2Fpve.proxmox.com%2Fcgi-bin%2Fmailman%2Flistinfo%2Fpve-user&amp;data=02%7C01%7Csmr%40kmi.com%7C5197e0e1f4c64fb738cf08d799a1067b%7Cc2283768b8d34e008f3d85b1b4f03b33%7C0%7C1%7C637146789208877811&amp;sdata=JhYOtpzDpjbQ1g4yrV%2FUwB8d3d4vX08Pd9wISQUmGp8%3D&amp;reserved=0
_______________________________________________
pve-user mailing list
pve-user@pve.proxmox.com
https://pve.proxmox.com/cgi-bin/mailman/listinfo/pve-user

Reply via email to