By the way, why do you use  multipath for local storage  like this EVO  nvme  ?

Happy New Year !

Best Regards,
Strahil NikolovOn Dec 31, 2019 21:51, Strahil <hunter86...@yahoo.com> wrote:
>
> You can check https://access.redhat.com/solutions/2437991 &  
> https://access.redhat.com/solutions/3014361
>
> You have 2 options:
> 1. Set a udev rule like this one (replace NETAPP with your storage)
> ACTION!="add|change", GOTO="rule_end" ENV{ID_VENDOR}=="NETAPP*", 
> RUN+="/bin/sh -c 'echo 4096 > /sys%p/queue/max_sectors_kb'" LABEL="rule_end"
>
> 2. Set max_sectors_kb in devices  section of multipath.conf
> You will need to stop LVM and then flush the device map , so the new option 
> to take effect (faster  is to reboot)
>
> Good Luck  & Happy New Year.
>
> Best Regards,
> Strahil Nikolov
> On Dec 31, 2019 17:53, Stefan Wolf <shb...@gmail.com> wrote:
> >
> > hi all, 
> >
> > i ve 4 nodes running with current ovirt. 
> > I ve only a problem on one host even after a fresh installation . 
> > I ve installed the latest image. 
> > Than I add the node to the cluster 
> > Everything is working good. 
> > After this I configure the network. 
> > BUT, after a restart the host does not come up again. 
> > I got this error: blk_cloned_rq_check_limits: over max size limit 
> > every 5 seconds 
> >
> > I can continue with control-D 
> > or I can login with root password to fix the problem. but i dont know what 
> > is the problem and where does it came from 
> >
> > I ve also changed the sas disk to nvme storage, but I ve changed this on 
> > every host. And this problem exists only on one host 
> >
> > i found this 
> > https://lists.centos.org/pipermail/centos/2017-December/167727.html 
> > the output is 
> > [root@kvm380 ~]# ./test.sh 
> > Sys Block Node    : Device                                max_sectors_kb  
> > max_hw_sectors_kb 
> > /sys/block/dm-0   : onn_kvm380-pool00_tmeta               256             
> > 4096 
> > /sys/block/dm-1   : onn_kvm380-pool00_tdata               256             
> > 4096 
> > /sys/block/dm-10  : onn_kvm380-var                        256             
> > 4096 
> > /sys/block/dm-11  : onn_kvm380-tmp                        256             
> > 4096 
> > /sys/block/dm-12  : onn_kvm380-home                       256             
> > 4096 
> > /sys/block/dm-13  : onn_kvm380-var_crash                  256             
> > 4096 
> > /sys/block/dm-2   : onn_kvm380-pool00-tpool               256             
> > 4096 
> > /sys/block/dm-3   : onn_kvm380-ovirt--node--ng--4.3.7--0.20191121.0+1 256   
> >           4096 
> > /sys/block/dm-4   : onn_kvm380-swap                       256             
> > 4096 
> > /sys/block/dm-5   : eui.0025385991b1e27a                  512             
> > 2048 
> > /sys/block/dm-6   : eui.0025385991b1e27a1                 512             
> > 2048 
> > /sys/block/dm-7   : onn_kvm380-pool00                     256             
> > 4096 
> > /sys/block/dm-8   : onn_kvm380-var_log_audit              256             
> > 4096 
> > /sys/block/dm-9   : onn_kvm380-var_log                    256             
> > 4096 
> > cat: /sys/block/nvme0n1/device/vendor: Datei oder Verzeichnis nicht 
> > gefunden 
> > /sys/block/nvme0n1:  Samsung SSD 970 EVO 1TB                  512           
> >   2048 
> > /sys/block/sda    : HP       LOGICAL VOLUME               256             
> > 4096 
> >
> > is the nvme not starting correct 
> > [root@kvm380 ~]# systemctl status multipathd 
> > ● multipathd.service - Device-Mapper Multipath Device Controller 
> >    Loaded: loaded (/usr/lib/systemd/system/multipathd.service; enabled; 
> > vendor preset: enabled) 
> >    Active: active (running) since Di 2019-12-31 16:16:32 CET; 31min ago 
> >   Process: 1919 ExecStart=/sbin/multipathd (code=exited, status=0/SUCCESS) 
> >   Process: 1916 ExecStartPre=/sbin/multipath -A (code=exited, 
> > status=0/SUCCESS) 
> >   Process: 1911 ExecStartPre=/sbin/modprobe dm-multipath (code=exited, 
> > status=0/SUCCESS) 
> > Main PID: 1921 (multipathd) 
> >     Tasks: 7 
> >    CGroup: /system.slice/multipathd.service 
> >            └─1921 /sbin/multipathd 
> >
> > Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark 
> > as failed 
> > Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: Entering recovery mode: max_retries=4 
> > Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: remaining active paths: 0 
> > Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 259:0: 
> > reinstated 
> > Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: queue_if_no_path enabled 
> > Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: Recovered to normal mode 
> > Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: remaining active paths: 1 
> > Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark 
> > as failed 
> > Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: Entering recovery mode: max_retries=4 
> > Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: 
> > eui.0025385991b1e27a: remaining active paths: 0 
> >
> > why is it marked as failed? 
> >
> > if i create a new volume with cockpit and use it for bricks for gluster, 
> > every thing is fine. until reboot 
> >
> >
> > maybe some one can point me the direction
> > _______________________________________________
> > Users mailing list -- users@ovirt.org
> > To unsubscribe send an email to users-le...@ovirt.org
> > Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> > oVirt Code of Conduct: 
> > https://www.ovirt.org/community/about/community-guidelines/
> > List Archives: 
> > https://lists.ovirt.org/archives/list/users@ovirt.org/message/MHHFFWAY5T5BVTD5TAMW2R6PC6SMGGA2/
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/users@ovirt.org/message/LKCRBN3QMYELEDFCQQJQ7SQJDDZM24RA/

Reply via email to