----- Original Message ----- > From: "Oliver Albl" <oliver.a...@fabasoft.com> > To: d...@redhat.com > Cc: users@ovirt.org > Sent: Friday, January 3, 2014 4:56:33 PM > Subject: Re: [Users] Host cannot access storage domains > > Redirecting to /bin/systemctl reconfigure vdsmd.service > Unknown operation 'reconfigure'.
/usr/lib/systemd/systemd-vdsmd reconfigure force > > ... seems to me, I should get rid of the ovirt-node iso installation and move > to a rpm based install? > > Thanks, > Oliver > -----Ursprüngliche Nachricht----- > Von: Dafna Ron [mailto:d...@redhat.com] > Gesendet: Freitag, 03. Jänner 2014 15:51 > An: Albl, Oliver > Cc: users@ovirt.org > Betreff: Re: AW: AW: AW: AW: AW: [Users] Host cannot access storage domains > > can you run: > service vdsmd reconfigure on the second host? > > On 01/03/2014 02:43 PM, Albl, Oliver wrote: > > Dafna, > > > > yes, the VM starts on the first node, the issues are on the second node > > only. > > > > /etc/libvirt/qemu-sanlock.conf is identical on on both nodes: > > > > auto_disk_leases=0 > > require_lease_for_disks=0 > > > > yum updates reports "Using yum is not supported"... > > > > Thanks, > > Oliver > > > > -----Ursprüngliche Nachricht----- > > Von: Dafna Ron [mailto:d...@redhat.com] > > Gesendet: Freitag, 03. Jänner 2014 15:39 > > An: Albl, Oliver > > Cc: users@ovirt.org > > Betreff: Re: AW: AW: AW: AW: [Users] Host cannot access storage > > domains > > > > ok, let's try to zoom in on the issue... > > can you run vm's on the first host or do you have issues only on the second > > host you added? > > can you run on both hosts? > > # egrep -v ^# /etc/libvirt/qemu-sanlock.conf > > > > can you run yum update on one of the hosts and see if there are newer > > packages? > > > > Thanks, > > > > Dafna > > > > On 01/03/2014 02:30 PM, Albl, Oliver wrote: > >> I installed both hosts using the oVirt Node ISO image: > >> > >> OS Version: oVirt Node - 3.0.3 - 1.1.fc19 Kernel Version: 3.11.9 - > >> 200.fc19.x86_64 KVM Version: 1.6.1 - 2.fc19 LIBVIRT Version: > >> libvirt-1.1.3.1-2.fc19 VDSM Version: vdsm-4.13.0-11.fc19 > >> > >> Thanks, > >> Oliver > >> -----Ursprüngliche Nachricht----- > >> Von: Dafna Ron [mailto:d...@redhat.com] > >> Gesendet: Freitag, 03. Jänner 2014 15:24 > >> An: Albl, Oliver > >> Cc: users@ovirt.org > >> Betreff: Re: AW: AW: AW: [Users] Host cannot access storage domains > >> > >> ignore the link :) > >> > >> so searching for this error I hit an old bug and it seemed to be an issue > >> between libvirt/sanlock. > >> > >> https://bugzilla.redhat.com/show_bug.cgi?id=828633 > >> > >> are you using latest packages? > >> > >> > >> > >> > >> On 01/03/2014 02:15 PM, Albl, Oliver wrote: > >>> Dafna, > >>> > >>> Libvirtd.log shows no errors, but VM log shows the following: > >>> > >>> 2014-01-03 13:52:11.296+0000: starting up LC_ALL=C > >>> PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin > >>> QEMU_AUDIO_DRV=spice /usr/bin/qemu-kvm -name OATEST2 -S -machine > >>> pc-1.0,accel=kvm,usb=off -cpu SandyBridge -m 1024 -realtime > >>> mlock=off -smp 1,sockets=1,cores=1,threads=1 -uuid > >>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -smbios > >>> type=1,manufacturer=oVirt,product=oVirt > >>> Node,version=3.0.3-1.1.fc19,serial=30313436-3631-5A43-4A33-3332304C3 > >>> 8 > >>> 4 > >>> C,uuid=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 -no-user-config > >>> -nodefaults -chardev > >>> socket,id=charmonitor,path=/var/lib/libvirt/qemu/OATEST2.monitor,ser > >>> v e r,nowait -mon chardev=charmonitor,id=monitor,mode=control -rtc > >>> base=2014-01-03T13:52:11,driftfix=slew -no-shutdown -device > >>> piix3-usb-uhci,id=usb,bus=pci.0,addr=0x1.0x2 -device > >>> virtio-scsi-pci,id=scsi0,bus=pci.0,addr=0x4 -device > >>> virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x3 -drive > >>> if=none,id=drive-ide0-1-0,readonly=on,format=raw,serial= -device > >>> ide-cd,bus=ide.1,unit=0,drive=drive-ide0-1-0,id=ide0-1-0 -drive > >>> file=/rhev/data-center/mnt/blockSD/7841a1c0-181a-4d43-9a25-b707accb5 > >>> c > >>> 4 > >>> b/images/de7ca992-b1c1-4cb8-9470-2494304c9b69/cbf1f376-23e8-40f3-838 > >>> 7 > >>> - > >>> ed299ee62607,if=none,id=drive-virtio-disk0,format=raw,serial=de7ca99 > >>> 2 > >>> - > >>> b1c1-4cb8-9470-2494304c9b69,cache=none,werror=stop,rerror=stop,aio=n > >>> a > >>> t > >>> ive -device > >>> virtio-blk-pci,scsi=off,bus=pci.0,addr=0x6,drive=drive-virtio-disk0, > >>> i > >>> d > >>> =virtio-disk0,bootindex=1 -chardev > >>> socket,id=charchannel0,path=/var/lib/libvirt/qemu/channels/d2bddcdb- > >>> a > >>> 2 c8-4c77-b0cf-b83fa3c2a0b6.com.redhat.rhevm.vdsm,server,nowait > >>> -device > >>> virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=cha > >>> n n el0,name=com.redhat.rhevm.vdsm -chardev > >>> socket,id=charchannel1,path=/var/lib/libvirt/qemu/channels/d2bddcdb- > >>> a > >>> 2 c8-4c77-b0cf-b83fa3c2a0b6.org.qemu.guest_agent.0,server,nowait > >>> -device > >>> virtserialport,bus=virtio-serial0.0,nr=2,chardev=charchannel1,id=cha > >>> n > >>> n > >>> el1,name=org.qemu.guest_agent.0 -chardev > >>> spicevmc,id=charchannel2,name=vdagent -device > >>> virtserialport,bus=virtio-serial0.0,nr=3,chardev=charchannel2,id=cha > >>> n > >>> n > >>> el2,name=com.redhat.spice.0 -spice > >>> tls-port=5900,addr=0,x509-dir=/etc/pki/vdsm/libvirt-spice,tls-channe > >>> l > >>> = > >>> main,tls-channel=display,tls-channel=inputs,tls-channel=cursor,tls-c > >>> h > >>> a > >>> nnel=playback,tls-channel=record,tls-channel=smartcard,tls-channel=u > >>> s b redir,seamless-migration=on -k en-us -device > >>> qxl-vga,id=video0,ram_size=67108864,vram_size=33554432,bus=pci.0,add > >>> r > >>> = > >>> 0x2 -device virtio-balloon-pci,id=balloon0,bus=pci.0,addr=0x5 > >>> libvirt: Lock Driver error : unsupported configuration: Read/write, > >>> exclusive access, disks were present, but no leases specified > >>> 2014-01-03 13:52:11.306+0000: shutting down > >>> > >>> Not sure what you mean with this > >>> http://forums.opensuse.org/english/get-technical-help-here/virtualization/492483-cannot-start-libvert-kvm-guests-after-update-tumbleweed.html. > >>> Do you want me to update libvirt with these repos on the oVirt-Node > >>> based installation? > >>> > >>> Thanks, > >>> Oliver > >>> -----Ursprüngliche Nachricht----- > >>> Von: Dafna Ron [mailto:d...@redhat.com] > >>> Gesendet: Freitag, 03. Jänner 2014 15:10 > >>> An: Albl, Oliver > >>> Cc: users@ovirt.org > >>> Betreff: Re: AW: AW: [Users] Host cannot access storage domains > >>> > >>> actually, looking at this again, it's a libvirt error and it can be > >>> related to selinux or sasl. > >>> can you also, look at libvirt log and the vm log under /var/log/libvirt? > >>> > >>> On 01/03/2014 02:00 PM, Albl, Oliver wrote: > >>>> Dafna, > >>>> > >>>> please find the logs below: > >>>> > >>>> ERRORs in vdsm.log on host02: > >>>> > >>>> Thread-61::ERROR::2014-01-03 > >>>> 13:51:48,956::sdc::137::Storage.StorageDomainCache::(_findDomain) > >>>> looking for unfetched domain f404398a-97f9-474c-af2c-e8887f53f688 > >>>> Thread-61::ERROR::2014-01-03 > >>>> 13:51:48,959::sdc::154::Storage.StorageDomainCache::(_findUnfetched > >>>> D > >>>> o > >>>> m > >>>> ain) looking for domain f404398a-97f9-474c-af2c-e8887f53f688 > >>>> Thread-323::ERROR::2014-01-03 > >>>> 13:52:11,527::vm::2132::vm.Vm::(_startUnderlyingVm) > >>>> vmId=`d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6`::The vm start process > >>>> failed Traceback (most recent call last): > >>>> File "/usr/share/vdsm/vm.py", line 2092, in _startUnderlyingVm > >>>> self._run() > >>>> File "/usr/share/vdsm/vm.py", line 2959, in _run > >>>> self._connection.createXML(domxml, flags), > >>>> File > >>>> "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", > >>>> line 76, in wrapper > >>>> ret = f(*args, **kwargs) > >>>> File "/usr/lib64/python2.7/site-packages/libvirt.py", line > >>>> 2920, in createXML > >>>> libvirtError: Child quit during startup handshake: Input/output > >>>> error > >>>> Thread-60::ERROR::2014-01-03 > >>>> 13:52:23,111::sdc::137::Storage.StorageDomainCache::(_findDomain) > >>>> looking for unfetched domain 52cf84ce-6eda-4337-8c94-491d94f5a18d > >>>> Thread-60::ERROR::2014-01-03 > >>>> 13:52:23,111::sdc::154::Storage.StorageDomainCache::(_findUnfetched > >>>> D > >>>> o > >>>> m > >>>> ain) looking for domain 52cf84ce-6eda-4337-8c94-491d94f5a18d > >>>> Thread-62::ERROR::2014-01-03 > >>>> 13:52:26,353::sdc::137::Storage.StorageDomainCache::(_findDomain) > >>>> looking for unfetched domain 7841a1c0-181a-4d43-9a25-b707accb5c4b > >>>> Thread-62::ERROR::2014-01-03 > >>>> 13:52:26,355::sdc::154::Storage.StorageDomainCache::(_findUnfetched > >>>> D > >>>> o > >>>> m > >>>> ain) looking for domain 7841a1c0-181a-4d43-9a25-b707accb5c4b > >>>> > >>>> engine.log: > >>>> > >>>> 2014-01-03 14:52:06,976 INFO > >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, > >>>> IsVmDuringInitiatingVDSCommand( vmId = > >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 5940cf72 > >>>> 2014-01-03 14:52:06,976 INFO > >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, > >>>> IsVmDuringInitiatingVDSCommand, return: false, log id: 5940cf72 > >>>> 2014-01-03 14:52:07,057 INFO > >>>> [org.ovirt.engine.core.bll.RunVmOnceCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] Running command: RunVmOnceCommand > >>>> internal: false. > >>>> Entities affected : ID: d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 Type: VM, > >>>> ID: > >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6 Type: VM > >>>> 2014-01-03 14:52:07,151 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.IsoPrefixVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, > >>>> IsoPrefixVDSCommand(HostName = host02, HostId = > >>>> 6dc7fac6-149e-4445-ace1-3c334a24d52a, > >>>> storagePoolId=b33d1793-252b-44ac-9685-3fe56b83c4c9), log id: > >>>> 1705b611 > >>>> 2014-01-03 14:52:07,152 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.IsoPrefixVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, IsoPrefixVDSCommand, return: > >>>> /rhev/data-center/mnt/vmmgmt:_var_lib_exports_iso/f74f052e-0dc6-456 > >>>> d > >>>> - a f95-248c2227c2e5/images/11111111-1111-1111-1111-111111111111, > >>>> log > >>>> id: > >>>> 1705b611 > >>>> 2014-01-03 14:52:07,170 INFO > >>>> [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, > >>>> CreateVmVDSCommand(HostName = host02, HostId = > >>>> 6dc7fac6-149e-4445-ace1-3c334a24d52a, > >>>> vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, vm=VM [TEST2]), log id: > >>>> 27b504de > >>>> 2014-01-03 14:52:07,190 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] START, CreateVDSCommand(HostName > >>>> = host02, HostId = 6dc7fac6-149e-4445-ace1-3c334a24d52a, > >>>> vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, vm=VM [TEST2]), log id: > >>>> 6ad0220 > >>>> 2014-01-03 14:52:08,472 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] > >>>> org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand > >>>> spiceSslCipherSuite=DEFAULT,memSize=1024,kvmEnable=true,smp=1,vmTyp > >>>> e > >>>> = > >>>> k > >>>> vm,emulatedMachine=pc-1.0,keyboardLayout=en-us,memGuaranteedSize=10 > >>>> 2 > >>>> 4 > >>>> , > >>>> pitReinjection=false,nice=0,display=qxl,smartcardEnable=false,smpCo > >>>> r > >>>> e > >>>> s > >>>> PerSocket=1,spiceSecureChannels=smain,sinputs,scursor,splayback,sre > >>>> c > >>>> o > >>>> r > >>>> d,sdisplay,susbredir,ssmartcard,timeOffset=0,transparentHugePages=t > >>>> r > >>>> u > >>>> e > >>>> ,vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6,devices=[Ljava.util.Hash > >>>> M > >>>> a > >>>> p > >>>> ;@3692311a,acpiEnable=true,vmName=TEST2,cpuType=SandyBridge,custom= > >>>> { > >>>> } > >>>> 2014-01-03 14:52:08,476 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.CreateVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, CreateVDSCommand, log id: > >>>> 6ad0220 > >>>> 2014-01-03 14:52:08,484 INFO > >>>> [org.ovirt.engine.core.vdsbroker.CreateVmVDSCommand] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] FINISH, CreateVmVDSCommand, return: > >>>> WaitForLaunch, log id: 27b504de > >>>> 2014-01-03 14:52:08,497 INFO > >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > >>>> (ajp--127.0.0.1-8702-3) [2ab5cd2] Correlation ID: 2ab5cd2, Job ID: > >>>> 2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event > >>>> ID: -1, Message: VM TEST2 was started by oliver.albl (Host: host02). > >>>> 2014-01-03 14:52:14,728 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] > >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] START, > >>>> DestroyVDSCommand(HostName = host02, HostId = > >>>> 6dc7fac6-149e-4445-ace1-3c334a24d52a, > >>>> vmId=d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6, force=false, > >>>> secondsToWait=0, gracefully=false), log id: 6a95ffd5 > >>>> 2014-01-03 14:52:15,783 INFO > >>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.DestroyVDSCommand] > >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] FINISH, > >>>> DestroyVDSCommand, log id: 6a95ffd5 > >>>> 2014-01-03 14:52:15,804 INFO > >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] Correlation ID: null, Call > >>>> Stack: null, Custom Event ID: -1, Message: VM TEST2 is down. Exit > >>>> message: Child quit during startup handshake: Input/output error. > >>>> 2014-01-03 14:52:15,805 INFO > >>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] > >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] Running on vds during > >>>> rerun failed vm: null > >>>> 2014-01-03 14:52:15,805 INFO > >>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] > >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] vm TEST2 running in db > >>>> and not running in vds - add to rerun treatment. vds host02 > >>>> 2014-01-03 14:52:15,808 ERROR > >>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo] > >>>> (DefaultQuartzScheduler_Worker-7) [24696b3e] Rerun vm > >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6. Called from vds host02 > >>>> 2014-01-03 14:52:15,810 INFO > >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > >>>> (pool-6-thread-40) [24696b3e] Correlation ID: 2ab5cd2, Job ID: > >>>> 2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event > >>>> ID: -1, Message: Failed to run VM TEST2 on Host host02. > >>>> 2014-01-03 14:52:15,823 INFO > >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] > >>>> (pool-6-thread-40) [24696b3e] START, > >>>> IsVmDuringInitiatingVDSCommand( vmId = > >>>> d2bddcdb-a2c8-4c77-b0cf-b83fa3c2a0b6), log id: 35e1eec > >>>> 2014-01-03 14:52:15,824 INFO > >>>> [org.ovirt.engine.core.vdsbroker.IsVmDuringInitiatingVDSCommand] > >>>> (pool-6-thread-40) [24696b3e] FINISH, > >>>> IsVmDuringInitiatingVDSCommand, > >>>> return: false, log id: 35e1eec > >>>> 2014-01-03 14:52:15,858 WARN > >>>> [org.ovirt.engine.core.bll.RunVmOnceCommand] (pool-6-thread-40) > >>>> [24696b3e] CanDoAction of action RunVmOnce failed. > >>>> Reasons:VAR__ACTION__RUN,VAR__TYPE__VM,VAR__ACTION__RUN,VAR__TYPE__ > >>>> V > >>>> M > >>>> , > >>>> SCHEDULING_ALL_HOSTS_FILTERED_OUT > >>>> 2014-01-03 14:52:15,862 INFO > >>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] > >>>> (pool-6-thread-40) [24696b3e] Correlation ID: 2ab5cd2, Job ID: > >>>> 2913133b-1301-484e-9887-b110841c8078, Call Stack: null, Custom Event > >>>> ID: -1, Message: Failed to run VM TEST2 (User: oliver.albl). > >>>> > >>>> Thanks, > >>>> Oliver > >>>> -----Ursprüngliche Nachricht----- > >>>> Von: Dafna Ron [mailto:d...@redhat.com] > >>>> Gesendet: Freitag, 03. Jänner 2014 14:51 > >>>> An: Albl, Oliver > >>>> Cc: users@ovirt.org > >>>> Betreff: Re: AW: [Users] Host cannot access storage domains > >>>> > >>>> Thanks for reporting the issue :) > >>>> > >>>> As for the vm, can you please find the error in vdsm.log and in engine > >>>> and paste it? > >>>> > >>>> Thanks, > >>>> > >>>> Dafna > >>>> > >>>> > >>>> On 01/03/2014 01:49 PM, Albl, Oliver wrote: > >>>>> Dafna, > >>>>> > >>>>> you were right, it seems to be a caching issue. Rebooting the > >>>>> host did the job: > >>>>> > >>>>> Before Reboot: > >>>>> > >>>>> [root@host01 log]# vdsClient -s 0 getStorageDomainsList > >>>>> 52cf84ce-6eda-4337-8c94-491d94f5a18d > >>>>> f404398a-97f9-474c-af2c-e8887f53f688 > >>>>> 7841a1c0-181a-4d43-9a25-b707accb5c4b > >>>>> > >>>>> [root@host02 log]# vdsClient -s 0 getStorageDomainsList > >>>>> 52cf84ce-6eda-4337-8c94-491d94f5a18d > >>>>> f404398a-97f9-474c-af2c-e8887f53f688 > >>>>> 7841a1c0-181a-4d43-9a25-b707accb5c4b > >>>>> 925ee53a-69b5-440f-b145-138ada5b452e > >>>>> > >>>>> After Reboot: > >>>>> > >>>>> [root@host02 admin]# vdsClient -s 0 getStorageDomainsList > >>>>> 52cf84ce-6eda-4337-8c94-491d94f5a18d > >>>>> f404398a-97f9-474c-af2c-e8887f53f688 > >>>>> 7841a1c0-181a-4d43-9a25-b707accb5c4b > >>>>> > >>>>> So now I have both hosts up and running but when I try to start a VM on > >>>>> the second host, I receive the following messages in the events pane: > >>>>> > >>>>> VM TEST2 was started by oliver.albl (Host: host02) VM TEST2 is down. > >>>>> Exit message: Child quit during startup handshake: Input/output error. > >>>>> > >>>>> Thanks again for your help! > >>>>> Oliver > >>>>> > >>>>> -----Ursprüngliche Nachricht----- > >>>>> Von: Dafna Ron [mailto:d...@redhat.com] > >>>>> Gesendet: Freitag, 03. Jänner 2014 14:22 > >>>>> An: Albl, Oliver > >>>>> Cc: users@ovirt.org > >>>>> Betreff: Re: [Users] Host cannot access storage domains > >>>>> > >>>>> yes, please attach the vdsm log > >>>>> also, can you run vdsClient 0 getStorageDomainsList and vdsClient 0 > >>>>> getDeviceList on both hosts? > >>>>> > >>>>> It might be a cache issue, so can you please restart the host and if it > >>>>> helps attach output before and after the reboot? > >>>>> > >>>>> Thanks, > >>>>> > >>>>> Dafna > >>>>> > >>>>> > >>>>> On 01/03/2014 01:12 PM, Albl, Oliver wrote: > >>>>>> Hi, > >>>>>> > >>>>>> I am starting with oVirt 3.3.2 and I have an issue adding a host > >>>>>> to a cluster. > >>>>>> > >>>>>> I am using oVirt Engine Version 3.3.2-1.el6 > >>>>>> > >>>>>> There is a cluster with one host (installed with oVirt Node - > >>>>>> 3.0.3 > >>>>>> - > >>>>>> 1.1.fc19 ISO image) up and running. > >>>>>> > >>>>>> I installed a second host using the same ISO image. > >>>>>> > >>>>>> I approved the host in the cluster. > >>>>>> > >>>>>> When I try to activate the second host, I receive the following > >>>>>> messages in the events pane: > >>>>>> > >>>>>> State was set to Up for host host02. > >>>>>> > >>>>>> Host host02 reports about one of the Active Storage Domains as > >>>>>> Problematic. > >>>>>> > >>>>>> Host host02 cannot access one of the Storage Domains attached to > >>>>>> the Data Center Test303. Stetting Host state to Non-Operational. > >>>>>> > >>>>>> Failed to connect Host host02 to Storage Pool Test303 > >>>>>> > >>>>>> There are 3 FC Storage Domains configured and visible to both hosts. > >>>>>> > >>>>>> multipath -ll shows all LUNs on both hosts. > >>>>>> > >>>>>> The engine.log reports the following about every five minutes: > >>>>>> > >>>>>> 2014-01-03 13:50:15,408 ERROR > >>>>>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] > >>>>>> (pool-6-thread-44) Domain 7841a1c0-181a-4d43-9a25-b707accb5c4b: > >>>>>> LUN_105 check timeot 69.7 is too big > >>>>>> > >>>>>> 2014-01-03 13:50:15,409 ERROR > >>>>>> [org.ovirt.engine.core.vdsbroker.irsbroker.IrsBrokerCommand] > >>>>>> (pool-6-thread-44) Domain 52cf84ce-6eda-4337-8c94-491d94f5a18d: > >>>>>> LUN_103 check timeot 59.6 is too big > >>>>>> > >>>>>> 2014-01-03 13:50:15,410 ERROR > >>>>>> [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44) > >>>>>> Storage Domain LUN_105 of pool Test303 is in problem in host > >>>>>> host02 > >>>>>> > >>>>>> 2014-01-03 13:50:15,411 ERROR > >>>>>> [org.ovirt.engine.core.bll.InitVdsOnUpCommand] (pool-6-thread-44) > >>>>>> Storage Domain LUN_103 of pool Test030 is in problem in host > >>>>>> host02 > >>>>>> > >>>>>> Please let me know if there are any log files I should attach. > >>>>>> > >>>>>> Thank you for your help! > >>>>>> > >>>>>> All the best, > >>>>>> > >>>>>> Oliver Albl > >>>>>> > >>>>>> > >>>>>> > >>>>>> _______________________________________________ > >>>>>> Users mailing list > >>>>>> Users@ovirt.org > >>>>>> http://lists.ovirt.org/mailman/listinfo/users > >>>>> -- > >>>>> Dafna Ron > >>>> -- > >>>> Dafna Ron > >>> -- > >>> Dafna Ron > >> -- > >> Dafna Ron > >> > > > > -- > > Dafna Ron > > > -- > Dafna Ron > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users