On Wed, Jan 20, 2021 at 7:36 PM Matt Snow <matts...@gmail.com> wrote: > > > > On Wed, Jan 20, 2021 at 2:46 AM Nir Soffer <nsof...@redhat.com> wrote: >> >> On Mon, Jan 18, 2021 at 8:58 AM Matt Snow <matts...@gmail.com> wrote: >> > >> > I installed ovirt node 4.4.4 as well as 4.4.5-pre and experience the same >> > problem with both versions. The issue occurs in both cockpit UI and tmux'd >> > CLI of ovirt-hosted-engine-setup. I get passed the point where the VM is >> > created and running. >> > I tried to do some debugging on my own before reaching out to this list. >> > Any help is much appreciated! >> > >> > ovirt node hardware: NUC format Jetway w/ Intel N3160 (Braswell 4 >> > cores/4threads), 8GB RAM, 64GB SSD. I understand this is underspec'd, but >> > I believe it meets the minimum requirements. >> > >> > NFS server: >> > * Ubuntu 19.10 w/ ZFS share w/ 17TB available space. >> >> We don't test ZFS, but since you tried also non-ZFS server with same issue, >> the >> issue is probably not ZFS. >> >> > * NFS share settings are just 'rw=@172.16.1.0/24' but have also tried >> > 'rw,sec=sys,anon=0' and '@172.16.1.0/24,insecure' >> >> You are missing anonuid=36,anongid=36. This should not affect sanlock, but >> you will have issues with libvirt without this. > > > I found this setting in another email thread and have applied it > > root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs > > tanker/ovirt sharenfs rw,anonuid=36,anongid=36 local > > root@stumpy:/tanker/ovirt/host_storage# >> >> >> Here is working export from my test system: >> >> /export/00 *(rw,sync,anonuid=36,anongid=36,no_subtree_check) >> > I have updated both ZFS server and the Ubuntu 20.04 NFS server to use the > above settings. > # ZFS Server > > root@stumpy:/tanker/ovirt/host_storage# zfs set > sharenfs="rw,sync,anonuid=36,anongid=36,no_subtree_check" tanker/ovirt > > root@stumpy:/tanker/ovirt/host_storage# zfs get all tanker/ovirt| grep nfs > > tanker/ovirt sharenfs > rw,sync,anonuid=36,anongid=36,no_subtree_check local > > root@stumpy:/tanker/ovirt/host_storage# > > # Ubuntu laptop > > ls -ld /exports/host_storage > > drwxr-xr-x 2 36 36 4096 Jan 16 19:42 /exports/host_storage > > root@msnowubntlt:/exports/host_storage# showmount -e dongle > > Export list for dongle: > > /exports/host_storage * > > root@msnowubntlt:/exports/host_storage# grep host_storage /etc/exports > > /exports/host_storage *(rw,sync,anonuid=36,anongid=36,no_subtree_check) > > root@msnowubntlt:/exports/host_storage# > > > > Interesting point - upon being re-prompted to configure storage after initial > failure I provided *different* NFS server information (the ubuntu laptop) > **snip** > > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] > > [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail > is "[]". HTTP response code is 400. > > [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault > reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is > 400."} > > Please specify the storage you would like to use (glusterfs, iscsi, > fc, nfs)[nfs]: nfs > > Please specify the nfs version you would like to use (auto, v3, v4, > v4_0, v4_1, v4_2)[auto]: > > Please specify the full shared storage connection path to use > (example: host:/path): dongle:/exports/host_storage > > If needed, specify additional mount options for the connection to > the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: > > [ INFO ] Creating Storage Domain > > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific set > of steps] > > **snip** > > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] > > [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail > is "[]". HTTP response code is 400. > > [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault > reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code is > 400."} > > Please specify the storage you would like to use (glusterfs, iscsi, > fc, nfs)[nfs]: > > **snip** > Upon checking mounts I see that the original server is still being used. > > [root@brick ~]# mount | grep nfs > > sunrpc on /var/lib/nfs/rpc_pipefs type rpc_pipefs (rw,relatime) > > stumpy:/tanker/ovirt/host_storage on > /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage type nfs4 > (rw,relatime,vers=4.2,rsize=1048576,wsize=1048576,namlen=255,soft,nosharecache,proto=tcp,timeo=100,retrans=3,sec=sys,clientaddr=172.16.1.49,local_lock=none,addr=172.16.1.50)
Using node and hosted engine makes everything more complicated and harder to debug. I would create an engine vm for testing on some machine, and add this host as regular host. It will be easier to debug this way, since you don't have to reinstall everything from scratch on every failure. Once we sort out the issue with the NFS/ZFS storage, using node and hosted engine makes sense for better availability and saving hardware resources. Also, can you explain how to build and configure identical ubuntu storage server? I can try to test a vm if you give detailed description. Maybe this an issue with AppArmor? https://linuxconfig.org/how-to-disable-enable-selinux-on-ubuntu-20-04-focal-fossa-linux Did you try to disable it on the storage server, or enable selinux? > > [root@brick ~]# > > >> >> > * The target directory is always empty and chown'd 36:36 with 0755 >> > permissions. >> >> Correct. >> >> > * I have tried using both IP and DNS names. forward and reverse DNS works >> > from ovirt host and other systems on the network. >> > * The NFS share always gets mounted successfully on the ovirt node system. >> > * I have tried auto and v3 NFS versions in other various combinations. >> > * I have also tried setting up an NFS server on a non-ZFS backed storage >> > system that is open to any host and get the same errors as shown below. >> > * I ran nfs-check.py script without issue against both NFS servers and >> > followed other verification steps listed on >> > https://www.ovirt.org/develop/troubleshooting-nfs-storage-issues.html >> > >> > ***Snip from ovirt-hosted-engine-setup*** >> > Please specify the storage you would like to use (glusterfs, >> > iscsi, fc, nfs)[nfs]: nfs >> > Please specify the nfs version you would like to use (auto, v3, >> > v4, v4_0, v4_1, v4_2)[auto]: >> > Please specify the full shared storage connection path to use >> > (example: host:/path): stumpy.mydomain.com:/tanker/ovirt/host_storage >> > If needed, specify additional mount options for the connection >> > to the hosted-engine storagedomain (example: rsize=32768,wsize=32768) []: >> > rw >> > [ INFO ] Creating Storage Domain >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Execute just a specific >> > set of steps] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Force facts gathering] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Wait for the storage >> > interface to be up] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check local VM dir stat] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Enforce local VM dir >> > existence] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : include_tasks] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Obtain SSO token using >> > username/password credentials] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch host facts] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster ID] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch cluster facts] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter facts] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter ID] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Fetch Datacenter name] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add NFS storage domain] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add glusterfs storage >> > domain] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add iSCSI storage domain] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Add Fibre Channel >> > storage domain] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get storage domain >> > details] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Find the appliance OVF] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Parse OVF] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Get required size] >> > [ INFO ] ok: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Remove unsuitable >> > storage domain] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Check storage domain >> > free space] >> > [ INFO ] skipping: [localhost] >> > [ INFO ] TASK [ovirt.ovirt.hosted_engine_setup : Activate storage domain] >> > [ ERROR ] ovirtsdk4.Error: Fault reason is "Operation Failed". Fault >> > detail is "[]". HTTP response code is 400. >> > [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false, "msg": "Fault >> > reason is \"Operation Failed\". Fault detail is \"[]\". HTTP response code >> > is 400."} >> > Please specify the storage you would like to use (glusterfs, >> > iscsi, fc, nfs)[nfs]: >> > >> > ***End snippet** >> > >> > The relevant section from >> > /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-ansible-create_storage_domain-20210116195220-bnfg1w.log: >> > **Snip*** >> > 2021-01-16 19:53:56,010-0700 DEBUG ansible on_any args >> > <ansible.executor.task_result.TaskResult object at 0x7f11a04c8320> kwargs >> > 2021-01-16 19:53:57,219-0700 INFO ansible task start {'status': 'OK', >> > 'ansible_type': 'task', 'ansible_playbook': >> > '/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml', >> > 'ansible_task': 'ovirt.ovirt.hosted_engine_setup : Activate storage >> > domain'} >> > 2021-01-16 19:53:57,220-0700 DEBUG ansible on_any args TASK: >> > ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs >> > is_conditional:False >> > 2021-01-16 19:53:57,221-0700 DEBUG ansible on_any args localhost TASK: >> > ovirt.ovirt.hosted_engine_setup : Activate storage domain kwargs >> > 2021-01-16 19:54:00,346-0700 DEBUG var changed: host "localhost" var >> > "ansible_play_hosts" type "<class 'list'>" value: "[]" >> > 2021-01-16 19:54:00,347-0700 DEBUG var changed: host "localhost" var >> > "ansible_play_batch" type "<class 'list'>" value: "[]" >> > 2021-01-16 19:54:00,348-0700 DEBUG var changed: host "localhost" var >> > "play_hosts" type "<class 'list'>" value: "[]" >> > 2021-01-16 19:54:00,349-0700 ERROR ansible failed { >> > "ansible_host": "localhost", >> > "ansible_playbook": >> > "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", >> > "ansible_result": { >> > "_ansible_no_log": false, >> > "changed": false, >> > "exception": "Traceback (most recent call last):\n File >> > \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", >> > line 783, in main\n File >> > \"/tmp/ansible_ovirt_storage_domain_payload_f38n25ab/ansible_ovirt_storage_domain_payload.zip/ansible_collections/ovirt/ovirt/plugins/modules/ovirt_storage_domain.py\", >> > line 638, in post_create_check\n File >> > \"/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py\", line 3647, >> > in add\n return self._internal_add(storage_domain, headers, query, >> > wait)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", >> > line 232, in _internal_add\n return future.wait() if wait else future\n >> > File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line >> > 55, in wait\n return self._code(response)\n File >> > \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 229, in >> > callback\n self._check_fault(respon >> > se)\n File \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", >> > line 132, in _check_fault\n self._raise_error(response, body)\n File >> > \"/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py\", line 118, in >> > _raise_error\n raise error\novirtsdk4.Error: Fault reason is >> > \"Operation Failed\". Fault detail is \"[]\". HTTP response code is >> > 400.\n", >> > "invocation": { >> > "module_args": { >> > "backup": null, >> > "comment": null, >> > "critical_space_action_blocker": null, >> > "data_center": "Default", >> > "description": null, >> > "destroy": null, >> > "discard_after_delete": null, >> > "domain_function": "data", >> > "fcp": null, >> > "fetch_nested": false, >> > "format": null, >> > "glusterfs": null, >> > "host": "brick.mydomain.com", >> > "id": null, >> > "iscsi": null, >> > "localfs": null, >> > "managed_block_storage": null, >> > "name": "hosted_storage", >> > "nested_attributes": [], >> > "nfs": null, >> > "poll_interval": 3, >> > "posixfs": null, >> > "state": "present", >> > "timeout": 180, >> > "wait": true, >> > "warning_low_space": null, >> > "wipe_after_delete": null >> > } >> > }, >> > "msg": "Fault reason is \"Operation Failed\". Fault detail is >> > \"[]\". HTTP response code is 400." >> > }, >> > "ansible_task": "Activate storage domain", >> > "ansible_type": "task", >> > "status": "FAILED", >> > "task_duration": 4 >> > } >> > 2021-01-16 19:54:00,350-0700 DEBUG ansible on_any args >> > <ansible.executor.task_result.TaskResult object at 0x7f11a05f2d30> kwargs >> > ignore_errors:None >> > 2021-01-16 19:54:00,358-0700 INFO ansible stats { >> > "ansible_playbook": >> > "/usr/share/ovirt-hosted-engine-setup/ansible/trigger_role.yml", >> > "ansible_playbook_duration": "01:35 Minutes", >> > "ansible_result": "type: <class 'dict'>\nstr: {'localhost': {'ok': 23, >> > 'failures': 1, 'unreachable': 0, 'changed': 0, 'skipped': 7, 'rescued': 0, >> > 'ignored': 0}}", >> > "ansible_type": "finish", >> > "status": "FAILED" >> > } >> > 2021-01-16 19:54:00,359-0700 INFO SUMMARY: >> > Duration Task Name >> > -------- -------- >> > [ < 1 sec ] Execute just a specific set of steps >> > [ 00:05 ] Force facts gathering >> > [ 00:03 ] Check local VM dir stat >> > [ 00:04 ] Obtain SSO token using username/password credentials >> > [ 00:04 ] Fetch host facts >> > [ 00:02 ] Fetch cluster ID >> > [ 00:04 ] Fetch cluster facts >> > [ 00:04 ] Fetch Datacenter facts >> > [ 00:02 ] Fetch Datacenter ID >> > [ 00:02 ] Fetch Datacenter name >> > [ 00:04 ] Add NFS storage domain >> > [ 00:04 ] Get storage domain details >> > [ 00:03 ] Find the appliance OVF >> > [ 00:03 ] Parse OVF >> > [ 00:02 ] Get required size >> > [ FAILED ] Activate storage domain >> > 2021-01-16 19:54:00,359-0700 DEBUG ansible on_any args >> > <ansible.executor.stats.AggregateStats object at 0x7f11a2e3b8d0> kwargs >> > **End snip** >> > >> > The relevantsection from /var/log/vdsm/vdsm.log: >> > ***begin snip*** >> > 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] START >> > getConnectedStoragePoolsList(options=None) from=internal, >> > task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:48) >> > 2021-01-16 19:53:58,439-0700 INFO (vmrecovery) [vdsm.api] FINISH >> > getConnectedStoragePoolsList return={'poollist': []} from=internal, >> > task_id=b8b21668-189e-4b68-a7f0-c2d2ebf14546 (api:54) >> > 2021-01-16 19:53:58,440-0700 INFO (vmrecovery) [vds] recovery: waiting >> > for storage pool to go up (clientIF:726) >> > 2021-01-16 19:53:58,885-0700 INFO (jsonrpc/3) [vdsm.api] START >> > connectStorageServer(domType=1, >> > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': >> > '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', >> > 'connection': 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': >> > 'false', 'id': '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'user': '', 'tpgt': >> > '1'}], options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, >> > task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:48) >> > 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) >> > [storage.StorageDomainCache] Invalidating storage domain cache (sdc:74) >> > 2021-01-16 19:53:58,892-0700 INFO (jsonrpc/3) [vdsm.api] FINISH >> > connectStorageServer return={'statuslist': [{'id': >> > '3ffd1e3b-168e-4248-a2af-b28fbdf49eef', 'status': 0}]} >> > from=::ffff:192.168.222.53,41192, flow_id=592e278f, >> > task_id=5bd52fa3-f790-4ed3-826d-c1f51e5f2291 (api:54) >> > 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [vdsm.api] START >> > getStorageDomainInfo(sdUUID='54532dd4-3e5b-4885-b88e-599c81efb146', >> > options=None) from=::ffff:192.168.222.53,41192, flow_id=592e278f, >> > task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:48) >> > 2021-01-16 19:53:58,914-0700 INFO (jsonrpc/6) [storage.StorageDomain] >> > sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) >> > 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [vdsm.api] FINISH >> > getStorageDomainInfo error=Domain is either partially accessible or >> > entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > from=::ffff:192.168.222.53,41192, flow_id=592e278f, >> > task_id=4d5a352d-6096-45b7-a4ee-b6e08ac02f7b (api:52) >> > 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.TaskManager.Task] >> > (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') Unexpected error (task:880) >> > Traceback (most recent call last): >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, >> > in _run >> > return fn(*args, **kargs) >> > File "<decorator-gen-131>", line 2, in getStorageDomainInfo >> > File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in >> > method >> > ret = func(*args, **kwargs) >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2796, >> > in getStorageDomainInfo >> > dom = self.validateSdUUID(sdUUID) >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 312, >> > in validateSdUUID >> > sdDom.validate() >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line >> > 538, in validate >> > raise se.StorageDomainAccessError(self.sdUUID) >> > vdsm.storage.exception.StorageDomainAccessError: Domain is either >> > partially accessible or entirely inaccessible: >> > ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > 2021-01-16 19:53:58,918-0700 INFO (jsonrpc/6) [storage.TaskManager.Task] >> > (Task='4d5a352d-6096-45b7-a4ee-b6e08ac02f7b') aborting: Task is aborted: >> > "value=Domain is either partially accessible or entirely inaccessible: >> > ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) >> > 2021-01-16 19:53:58,918-0700 ERROR (jsonrpc/6) [storage.Dispatcher] FINISH >> > getStorageDomainInfo error=Domain is either partially accessible or >> > entirely inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > (dispatcher:83) >> > 2021-01-16 19:53:58,919-0700 INFO (jsonrpc/6) [jsonrpc.JsonRpcServer] RPC >> > call StorageDomain.getInfo failed (error 379) in 0.00 seconds >> > (__init__:312) >> > 2021-01-16 19:53:58,956-0700 INFO (jsonrpc/1) [vdsm.api] START >> > createStoragePool(poolType=None, >> > spUUID='91fb3f22-5795-11eb-ad9f-00163e3f4683', poolName='Default', >> > masterDom='54532dd4-3e5b-4885-b88e-599c81efb146', >> > domList=['54532dd4-3e5b-4885-b88e-599c81efb146'], masterVersion=7, >> > lockPolicy=None, lockRenewalIntervalSec=5, leaseTimeSec=60, >> > ioOpTimeoutSec=10, leaseRetries=3, options=None) >> > from=::ffff:192.168.222.53,41192, flow_id=592e278f, >> > task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:48) >> > 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] >> > updating pool 91fb3f22-5795-11eb-ad9f-00163e3f4683 backend from type >> > NoneType instance 0x7f86e45f89d0 to type StoragePoolDiskBackend instance >> > 0x7f8680738408 (sp:157) >> > 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StoragePool] >> > spUUID=91fb3f22-5795-11eb-ad9f-00163e3f4683 poolName=Default >> > master_sd=54532dd4-3e5b-4885-b88e-599c81efb146 >> > domList=['54532dd4-3e5b-4885-b88e-599c81efb146'] masterVersion=7 >> > {'LEASERETRIES': 3, 'LEASETIMESEC': 60, 'LOCKRENEWALINTERVALSEC': 5, >> > 'IOOPTIMEOUTSEC': 10} (sp:602) >> > 2021-01-16 19:53:58,958-0700 INFO (jsonrpc/1) [storage.StorageDomain] >> > sdUUID=54532dd4-3e5b-4885-b88e-599c81efb146 (fileSD:535) >> > 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.StoragePool] >> > Unexpected error (sp:618) >> > Traceback (most recent call last): >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in >> > create >> > domain.validate() >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line >> > 538, in validate >> > raise se.StorageDomainAccessError(self.sdUUID) >> > vdsm.storage.exception.StorageDomainAccessError: Domain is either >> > partially accessible or entirely inaccessible: >> > ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > 2021-01-16 19:53:58,963-0700 INFO (jsonrpc/1) [vdsm.api] FINISH >> > createStoragePool error=Domain is either partially accessible or entirely >> > inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > from=::ffff:192.168.222.53,41192, flow_id=592e278f, >> > task_id=a946357f-537a-4c9c-9040-e70a98ee5643 (api:52) >> > 2021-01-16 19:53:58,963-0700 ERROR (jsonrpc/1) [storage.TaskManager.Task] >> > (Task='a946357f-537a-4c9c-9040-e70a98ee5643') Unexpected error (task:880) >> > Traceback (most recent call last): >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 613, in >> > create >> > domain.validate() >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/fileSD.py", line >> > 538, in validate >> > raise se.StorageDomainAccessError(self.sdUUID) >> > vdsm.storage.exception.StorageDomainAccessError: Domain is either >> > partially accessible or entirely inaccessible: >> > ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > >> > During handling of the above exception, another exception occurred: >> > >> > Traceback (most recent call last): >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, >> > in _run >> > return fn(*args, **kargs) >> > File "<decorator-gen-31>", line 2, in createStoragePool >> > File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in >> > method >> > ret = func(*args, **kwargs) >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 1027, >> > in createStoragePool >> > leaseParams) >> > File "/usr/lib/python3.6/site-packages/vdsm/storage/sp.py", line 619, in >> > create >> > raise se.StorageDomainAccessError(sdUUID) >> > vdsm.storage.exception.StorageDomainAccessError: Domain is either >> > partially accessible or entirely inaccessible: >> > ('54532dd4-3e5b-4885-b88e-599c81efb146',) >> > 2021-01-16 19:53:58,964-0700 INFO (jsonrpc/1) [storage.TaskManager.Task] >> > (Task='a946357f-537a-4c9c-9040-e70a98ee5643') aborting: Task is aborted: >> > "value=Domain is either partially accessible or entirely inaccessible: >> > ('54532dd4-3e5b-4885-b88e-599c81efb146',) abortedcode=379" (task:1190) >> > 2021-01-16 19:53:58,964-0700 ERROR (jsonrpc/1) [storage.Dispatcher] FINISH >> > createStoragePool error=Domain is either partially accessible or entirely >> > inaccessible: ('54532dd4-3e5b-4885-b88e-599c81efb146',) (dispatcher:83) >> > 2021-01-16 19:53:58,965-0700 INFO (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC >> > call StoragePool.create failed (error 379) in 0.01 seconds (__init__:312) >> > 2021-01-16 19:54:01,923-0700 INFO (jsonrpc/5) [api.host] START getStats() >> > from=::ffff:192.168.222.53,41192 (api:48) >> > 2021-01-16 19:54:01,964-0700 INFO (jsonrpc/7) [api.host] START >> > getAllVmStats() from=::ffff:192.168.222.53,41192 (api:48) >> > 2021-01-16 19:54:01,966-0700 INFO (jsonrpc/7) [api.host] FINISH >> > getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, >> > 'statsList': (suppressed)} from=::ffff:192.168.222.53,41192 (api:54) >> > 2021-01-16 19:54:01,977-0700 INFO (jsonrpc/5) [vdsm.api] START >> > repoStats(domains=()) from=::ffff:192.168.222.53,41192, >> > task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:48) >> > 2021-01-16 19:54:01,978-0700 INFO (jsonrpc/5) [vdsm.api] FINISH repoStats >> > return={} from=::ffff:192.168.222.53,41192, >> > task_id=7a6c8536-3f1e-41b3-9ab7-bb58a5e75fac (api:54) >> > 2021-01-16 19:54:01,980-0700 INFO (jsonrpc/5) [vdsm.api] START >> > multipath_health() from=::ffff:192.168.222.53,41192, >> > task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:48) >> > 2021-01-16 19:54:01,981-0700 INFO (jsonrpc/5) [vdsm.api] FINISH >> > multipath_health return={} from=::ffff:192.168.222.53,41192, >> > task_id=b8af45b5-6488-4940-b7d0-1d11b5f76db4 (api:54) >> > 2021-01-16 19:54:01,996-0700 INFO (jsonrpc/5) [api.host] FINISH getStats >> > return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} >> > from=::ffff:192.168.222.53,41192 (api:54) >> > ***end snip*** >> > _______________________________________________ >> > Users mailing list -- users@ovirt.org >> > To unsubscribe send an email to users-le...@ovirt.org >> > Privacy Statement: https://www.ovirt.org/privacy-policy.html >> > oVirt Code of Conduct: >> > https://www.ovirt.org/community/about/community-guidelines/ >> > List Archives: >> > https://lists.ovirt.org/archives/list/users@ovirt.org/message/G2TQC6XTJTIBAIOG7BWAFJ3YW3XOMNXF/ >> _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/QF2QHQEWDFJDCHUCTVK4R4X2MVQSN7TA/