On Mon, Jan 18, 2021 at 6:01 PM Matt Snow <matts...@gmail.com> wrote: > > Hi Didi, > I did log clean up and am re-running ovirt-hosted-engine-cleanup && > ovirt-hosted-engine-setup to get you cleaner log files. > > searching for host_storage in vdsm.log... > **snip** > 2021-01-18 08:43:18,842-0700 INFO (jsonrpc/3) [api.host] FINISH getStats > return={'status': {'code': 0, 'message': 'Done'}, 'info': (suppressed)} > from=::ffff:192.168.222.53,39612 (api:54) > 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] START > getConnectedStoragePoolsList(options=None) from=internal, > task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:48) > 2021-01-18 08:43:19,963-0700 INFO (vmrecovery) [vdsm.api] FINISH > getConnectedStoragePoolsList return={'poollist': []} from=internal, > task_id=fb80c883-2447-4ed2-b344-aa0c0fb65809 (api:54) > 2021-01-18 08:43:19,964-0700 INFO (vmrecovery) [vds] recovery: waiting for > storage pool to go up (clientIF:726) > 2021-01-18 08:43:20,441-0700 INFO (jsonrpc/4) [vdsm.api] START > connectStorageServer(domType=1, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': > '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': > 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': > '00000000-0000-0000-0000-000000000000', 'user': '', 'tpgt': '1'}], > options=None) from=::ffff:192.168.222.53,39612, > flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, > task_id=032afa50-381a-44af-a067-d25bcc224355 (api:48) > 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) > [storage.StorageServer.MountConnection] Creating directory > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage' (storageServer:167) > 2021-01-18 08:43:20,446-0700 INFO (jsonrpc/4) [storage.fileUtils] Creating > directory: /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage mode: > None (fileUtils:201) > 2021-01-18 08:43:20,447-0700 INFO (jsonrpc/4) [storage.Mount] mounting > stumpy:/tanker/ovirt/host_storage at > /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:207) > 2021-01-18 08:43:21,271-0700 INFO (jsonrpc/4) [IOProcessClient] (Global) > Starting client (__init__:340) > 2021-01-18 08:43:21,313-0700 INFO (ioprocess/51124) [IOProcess] (Global) > Starting ioprocess (__init__:465) > 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [storage.StorageDomainCache] > Invalidating storage domain cache (sdc:74) > 2021-01-18 08:43:21,373-0700 INFO (jsonrpc/4) [vdsm.api] FINISH > connectStorageServer return={'statuslist': [{'id': > '00000000-0000-0000-0000-000000000000', 'status': 0}]} > from=::ffff:192.168.222.53,39612, > flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, > task_id=032afa50-381a-44af-a067-d25bcc224355 (api:54) > 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [vdsm.api] START > getStorageDomainsList(spUUID='00000000-0000-0000-0000-000000000000', > domainClass=1, storageType='', > remotePath='stumpy:/tanker/ovirt/host_storage', options=None) > from=::ffff:192.168.222.53,39612, > flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, > task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:48) > 2021-01-18 08:43:21,497-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] > Refreshing storage domain cache (resize=True) (sdc:80) > 2021-01-18 08:43:21,498-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI > devices (iscsi:442) > 2021-01-18 08:43:21,628-0700 INFO (jsonrpc/5) [storage.ISCSI] Scanning iSCSI > devices: 0.13 seconds (utils:390) > 2021-01-18 08:43:21,629-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC > devices (hba:60) > 2021-01-18 08:43:21,908-0700 INFO (jsonrpc/5) [storage.HBA] Scanning FC > devices: 0.28 seconds (utils:390) > 2021-01-18 08:43:21,969-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing > multipath devices (multipath:104) > 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.Multipath] Resizing > multipath devices: 0.01 seconds (utils:390) > 2021-01-18 08:43:21,975-0700 INFO (jsonrpc/5) [storage.StorageDomainCache] > Refreshing storage domain cache: 0.48 seconds (utils:390) > 2021-01-18 08:43:22,167-0700 INFO (tmap-0/0) [IOProcessClient] > (stumpy:_tanker_ovirt_host__storage) Starting client (__init__:340) > 2021-01-18 08:43:22,204-0700 INFO (ioprocess/51144) [IOProcess] > (stumpy:_tanker_ovirt_host__storage) Starting ioprocess (__init__:465) > 2021-01-18 08:43:22,208-0700 INFO (jsonrpc/5) [vdsm.api] FINISH > getStorageDomainsList return={'domlist': []} > from=::ffff:192.168.222.53,39612, > flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, > task_id=e37eb000-13da-440f-9197-07495e53ce52 (api:54) > 2021-01-18 08:43:22,999-0700 INFO (jsonrpc/7) [vdsm.api] START > connectStorageServer(domType=1, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': > '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': > 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': > 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}], > options=None) from=::ffff:192.168.222.53,39612, flow_id=5618fb28, > task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:48) > 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [storage.StorageDomainCache] > Invalidating storage domain cache (sdc:74) > 2021-01-18 08:43:23,007-0700 INFO (jsonrpc/7) [vdsm.api] FINISH > connectStorageServer return={'statuslist': [{'id': > 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} > from=::ffff:192.168.222.53,39612, flow_id=5618fb28, > task_id=51daa36a-e1cf-479d-a93c-1c87f21ce934 (api:54) > 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [vdsm.api] START > createStorageDomain(storageType=1, > sdUUID='4b3fb9a9-6975-4b80-a2c1-af4e30865088', domainName='hosted_storage', > typeSpecificArg='stumpy:/tanker/ovirt/host_storage', domClass=1, > domVersion='5', block_size=512, max_hosts=250, options=None) > from=::ffff:192.168.222.53,39612, flow_id=5618fb28, > task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:48) > 2021-01-18 08:43:23,130-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] > Refreshing storage domain cache (resize=True) (sdc:80) > 2021-01-18 08:43:23,131-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI > devices (iscsi:442) > 2021-01-18 08:43:23,238-0700 INFO (jsonrpc/0) [storage.ISCSI] Scanning iSCSI > devices: 0.11 seconds (utils:390) > 2021-01-18 08:43:23,239-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC > devices (hba:60) > 2021-01-18 08:43:23,522-0700 INFO (jsonrpc/0) [storage.HBA] Scanning FC > devices: 0.28 seconds (utils:390) > 2021-01-18 08:43:23,576-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing > multipath devices (multipath:104) > 2021-01-18 08:43:23,582-0700 INFO (jsonrpc/0) [storage.Multipath] Resizing > multipath devices: 0.01 seconds (utils:390) > 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] > Refreshing storage domain cache: 0.45 seconds (utils:390) > 2021-01-18 08:43:23,583-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] > Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 (sdc:171) > 2021-01-18 08:43:23,760-0700 WARN (jsonrpc/0) [storage.LVM] All 1 tries have > failed: cmd=['/sbin/lvm', 'vgs', '--config', 'devices { > preferred_names=["^/dev/mapper/"] ignore_suspended_devices=1 > write_cache_state=0 disable_after_error_count=3 filter=["r|.*|"] > hints="none" obtain_device_list_from_udev=0 } global { locking_type=1 > prioritise_write_locks=1 wait_for_locks=1 use_lvmetad=0 } backup { > retain_min=50 retain_days=0 }', '--noheadings', '--units', 'b', > '--nosuffix', '--separator', '|', '--ignoreskippedcluster', '-o', > 'uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name', > '4b3fb9a9-6975-4b80-a2c1-af4e30865088'] rc=5 err=[' Volume group > "4b3fb9a9-6975-4b80-a2c1-af4e30865088" not found', ' Cannot process volume > group 4b3fb9a9-6975-4b80-a2c1-af4e30865088'] (lvm:538) > 2021-01-18 08:43:23,771-0700 INFO (jsonrpc/0) [storage.StorageDomainCache] > Looking up domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088: 0.19 seconds > (utils:390) > 2021-01-18 08:43:23,818-0700 INFO (jsonrpc/0) [IOProcessClient] > (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting client (__init__:340) > 2021-01-18 08:43:23,856-0700 INFO (ioprocess/51165) [IOProcess] > (4b3fb9a9-6975-4b80-a2c1-af4e30865088) Starting ioprocess (__init__:465) > 2021-01-18 08:43:24,166-0700 INFO (jsonrpc/0) [storage.StorageDomain] > Creating domain metadata directory > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md' > (fileSD:441) > 2021-01-18 08:43:24,970-0700 INFO (vmrecovery) [vdsm.api] START > getConnectedStoragePoolsList(options=None) from=internal, > task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:48) > 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vdsm.api] FINISH > getConnectedStoragePoolsList return={'poollist': []} from=internal, > task_id=5b920619-cd95-40d0-a253-c39d41d71245 (api:54) > 2021-01-18 08:43:24,971-0700 INFO (vmrecovery) [vds] recovery: waiting for > storage pool to go up (clientIF:726) > 2021-01-18 08:43:25,285-0700 INFO (jsonrpc/0) [storage.xlease] Formatting > index for lockspace '4b3fb9a9-6975-4b80-a2c1-af4e30865088' (version=1) > (xlease:701) > 2021-01-18 08:43:25,441-0700 INFO (jsonrpc/0) [storage.StorageDomain] > Creating domain images directory > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/images' > (nfsSD:110) > 2021-01-18 08:43:25,522-0700 INFO (jsonrpc/0) [storage.StorageDomain] > Removing remnants of deleted images [] (fileSD:735) > 2021-01-18 08:43:25,524-0700 INFO (jsonrpc/0) [storage.SANLock] Initializing > sanlock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 > path=/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage/4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids > alignment=1048576 block_size=512 io_timeout=10 (clusterlock:286) > 2021-01-18 08:43:25,533-0700 ERROR (jsonrpc/0) [storage.SANLock] Cannot > initialize lock for domain 4b3fb9a9-6975-4b80-a2c1-af4e30865088 > (clusterlock:305) > Traceback (most recent call last): > File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line > 295, in initLock > sector=self._block_size) > sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such > device') > 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [vdsm.api] FINISH > createStorageDomain error=Could not initialize cluster lock: () > from=::ffff:192.168.222.53,39612, flow_id=5618fb28, > task_id=49a1bc04-91d0-4d8f-b847-b6461d980495 (api:52) > 2021-01-18 08:43:25,534-0700 ERROR (jsonrpc/0) [storage.TaskManager.Task] > (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') Unexpected error (task:880) > Traceback (most recent call last): > File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line > 295, in initLock > sector=self._block_size) > sanlock.SanlockException: (19, 'Sanlock lockspace write failure', 'No such > device')
I think this ^^ is the issue. Can you please check /var/log/sanlock.log? Adding Nir. Best regards, > > During handling of the above exception, another exception occurred: > > Traceback (most recent call last): > File "/usr/lib/python3.6/site-packages/vdsm/storage/task.py", line 887, in > _run > return fn(*args, **kargs) > File "<decorator-gen-123>", line 2, in createStorageDomain > File "/usr/lib/python3.6/site-packages/vdsm/common/api.py", line 50, in > method > ret = func(*args, **kwargs) > File "/usr/lib/python3.6/site-packages/vdsm/storage/hsm.py", line 2669, in > createStorageDomain > max_hosts=max_hosts) > File "/usr/lib/python3.6/site-packages/vdsm/storage/nfsSD.py", line 120, in > create > fsd.initSPMlease() > File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 1019, in > initSPMlease > return self._manifest.initDomainLock() > File "/usr/lib/python3.6/site-packages/vdsm/storage/sd.py", line 627, in > initDomainLock > self._domainLock.initLock(self.getDomainLease()) > File "/usr/lib/python3.6/site-packages/vdsm/storage/clusterlock.py", line > 306, in initLock > raise se.ClusterLockInitError() > vdsm.storage.exception.ClusterLockInitError: Could not initialize cluster > lock: () > 2021-01-18 08:43:25,534-0700 INFO (jsonrpc/0) [storage.TaskManager.Task] > (Task='49a1bc04-91d0-4d8f-b847-b6461d980495') aborting: Task is aborted: > 'value=Could not initialize cluster lock: () abortedcode=701' (task:1190) > 2021-01-18 08:43:25,535-0700 ERROR (jsonrpc/0) [storage.Dispatcher] FINISH > createStorageDomain error=Could not initialize cluster lock: () > (dispatcher:83) > 2021-01-18 08:43:25,536-0700 INFO (jsonrpc/0) [jsonrpc.JsonRpcServer] RPC > call StorageDomain.create failed (error 701) in 2.41 seconds (__init__:312) > 2021-01-18 08:43:26,086-0700 INFO (jsonrpc/6) [vdsm.api] START > disconnectStorageServer(domType=1, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'password': > '********', 'protocol_version': 'auto', 'port': '', 'iqn': '', 'connection': > 'stumpy:/tanker/ovirt/host_storage', 'ipv6_enabled': 'false', 'id': > 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'user': '', 'tpgt': '1'}], > options=None) from=::ffff:192.168.222.53,39612, > flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, > task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:48) > 2021-01-18 08:43:26,087-0700 INFO (jsonrpc/6) [storage.Mount] unmounting > /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (mount:215) > 2021-01-18 08:43:26,209-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] > Refreshing storage domain cache (resize=False) (sdc:80) > 2021-01-18 08:43:26,210-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI > devices (iscsi:442) > 2021-01-18 08:43:26,346-0700 INFO (jsonrpc/6) [storage.ISCSI] Scanning iSCSI > devices: 0.14 seconds (utils:390) > 2021-01-18 08:43:26,347-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC > devices (hba:60) > 2021-01-18 08:43:26,635-0700 INFO (jsonrpc/6) [storage.HBA] Scanning FC > devices: 0.29 seconds (utils:390) > 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [storage.StorageDomainCache] > Refreshing storage domain cache: 0.48 seconds (utils:390) > 2021-01-18 08:43:26,693-0700 INFO (jsonrpc/6) [vdsm.api] FINISH > disconnectStorageServer return={'statuslist': [{'id': > 'bc87e1a4-004e-41b4-b569-9e9413e9c027', 'status': 0}]} > from=::ffff:192.168.222.53,39612, > flow_id=2227465c-5040-4199-b1f9-f5305b10b5e5, > task_id=5e35a85a-4010-425e-81b5-5db3ad6d63f3 (api:54) > 2021-01-18 08:43:27,353-0700 INFO (periodic/0) [vdsm.api] START > repoStats(domains=()) from=internal, > task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:48) > 2021-01-18 08:43:27,355-0700 INFO (periodic/0) [vdsm.api] FINISH repoStats > return={} from=internal, task_id=236f3956-28f6-407e-b893-e018f90fcd08 (api:54) > 2021-01-18 08:43:27,694-0700 INFO (jsonrpc/1) [api.host] START > getAllVmStats() from=::ffff:192.168.222.53,39612 (api:48) > 2021-01-18 08:43:27,697-0700 INFO (jsonrpc/1) [api.host] FINISH > getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': > (suppressed)} from=::ffff:192.168.222.53,39612 (api:54) > 2021-01-18 08:43:27,800-0700 INFO (jsonrpc/2) [api.host] START > getAllVmStats() from=::1,55776 (api:48) > 2021-01-18 08:43:27,802-0700 INFO (jsonrpc/2) [api.host] FINISH > getAllVmStats return={'status': {'code': 0, 'message': 'Done'}, 'statsList': > (suppressed)} from=::1,55776 (api:54) > 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] START > getAllVmIoTunePolicies() from=::1,55776 (api:48) > 2021-01-18 08:43:27,826-0700 INFO (jsonrpc/3) [api.host] FINISH > getAllVmIoTunePolicies return={'status': {'code': 0, 'message': 'Done'}, > 'io_tune_policies_dict': {'ddceefdd-57d3-432f-816e-5dc013249bff': {'policy': > [], 'current_values': []}}} from=::1,55776 (api:54) > 2021-01-18 08:43:29,975-0700 INFO (vmrecovery) [vdsm.api] START > getConnectedStoragePoolsList(options=None) from=internal, > task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:48) > 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vdsm.api] FINISH > getConnectedStoragePoolsList return={'poollist': []} from=internal, > task_id=821ce12a-ab6e-4db5-ba22-bd2630f2730c (api:54) > 2021-01-18 08:43:29,976-0700 INFO (vmrecovery) [vds] recovery: waiting for > storage pool to go up (clientIF:726) > 2021-01-18 08:43:33,826-0700 INFO (jsonrpc/4) [api.host] START getStats() > from=::ffff:192.168.222.53,39612 (api:48) > > **end snip** > > > > > from supervdsm.log. This is the only section where 'host_storage' shows up. > > **snip** > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:42:55,447::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> > = 0 > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:42:55,448::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return dmsetup_run_status with b'' > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:05,450::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > dmsetup_run_status with ('multipath',) {} > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:05,450::commands::153::common.commands::(start) /usr/bin/taskset > --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:05,483::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> > = 0 > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:05,484::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return dmsetup_run_status with b'' > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:15,485::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > dmsetup_run_status with ('multipath',) {} > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:15,485::logutils::354::root::(_report_stats) ThreadedHandler is ok in > the last 60 seconds (max pending: 2) > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:15,485::commands::153::common.commands::(start) /usr/bin/taskset > --cpu-list 0-3 /usr/sbin/dmsetup status --target multipath (cwd None) > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:15,515::commands::98::common.commands::(run) SUCCESS: <err> = b''; <rc> > = 0 > MainProcess|mpathhealth::DEBUG::2021-01-18 > 08:43:15,516::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return dmsetup_run_status with b'' > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:20,448::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > mount with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>, > 'stumpy:/tanker/ovirt/host_storage', > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage') {'mntOpts': > 'soft,nosharecache,timeo=100,retrans=3,nolock', 'vfstype': 'nfs', 'cgroup': > None} > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:20,448::commands::211::root::(execCmd) /usr/bin/taskset --cpu-list 0-3 > /usr/bin/mount -t nfs -o soft,nosharecache,timeo=100,retrans=3,nolock > stumpy:/tanker/ovirt/host_storage > /rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage (cwd None) > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:21,198::commands::224::root::(execCmd) SUCCESS: <err> = b''; <rc> = 0 > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:21,198::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return mount with None > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:21,320::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at > 0x7f19293acc88>, 'vdsm', ('kvm',), > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:21,340::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return validateAccess with None > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:21,341::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at > 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} > MainProcess|jsonrpc/4::DEBUG::2021-01-18 > 08:43:21,368::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return validateAccess with None > MainProcess|jsonrpc/5::DEBUG::2021-01-18 > 08:43:21,630::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) > {} > MainProcess|jsonrpc/5::DEBUG::2021-01-18 > 08:43:21,630::commands::153::common.commands::(start) /usr/bin/taskset > --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) > MainProcess|jsonrpc/5::DEBUG::2021-01-18 > 08:43:21,907::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return hbaRescan with None > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,246::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > hbaRescan with (<vdsm.supervdsm_server._SuperVdsm object at 0x7f19293acc88>,) > {} > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,246::commands::153::common.commands::(start) /usr/bin/taskset > --cpu-list 0-3 /usr/libexec/vdsm/fc-scan (cwd None) > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,521::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return hbaRescan with None > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,778::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at > 0x7f19293acc88>, 'vdsm', ('kvm',), > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 7) {} > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,796::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return validateAccess with None > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,798::supervdsm_server::93::SuperVdsm.ServerCallback::(wrapper) call > validateAccess with (<vdsm.supervdsm_server._SuperVdsm object at > 0x7f19293acc88>, 'qemu', ('qemu', 'kvm'), > '/rhev/data-center/mnt/stumpy:_tanker_ovirt_host__storage', 5) {} > MainProcess|jsonrpc/0::DEBUG::2021-01-18 > 08:43:23,816::supervdsm_server::100::SuperVdsm.ServerCallback::(wrapper) > return validateAccess with None > **end snip** > > As for the mount area, it does get mounted and unmounted. Files and > directory structure is also created. > > root@stumpy:/tanker/ovirt/host_storage# ls -l > total 1 > drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 > root@stumpy:/tanker/ovirt/host_storage# find . > . > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/images > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox > root@stumpy:/tanker/ovirt/host_storage# find . -exec ls -ltrh '{}' \; > total 512 > drwxrwsr-x 4 vdsm 36 4 Jan 18 08:43 4b3fb9a9-6975-4b80-a2c1-af4e30865088 > total 12K > drwxrwsr-x 2 vdsm 36 8 Jan 18 08:43 dom_md > drwxrwsr-x 2 vdsm 36 2 Jan 18 08:43 images > total 0 > total 275K > -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 ids > -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 inbox > -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 outbox > -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 leases > -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 xleases > -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 metadata > -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/inbox > -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/leases > -rw-rw-r-- 1 vdsm 36 342 Jan 18 08:43 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/metadata > -rw-rw---- 1 vdsm 36 0 Jan 18 08:43 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/ids > -rw-rw---- 1 vdsm 36 1.3M Jan 18 08:43 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/xleases > -rw-rw---- 1 vdsm 36 16M Jan 18 08:43 > ./4b3fb9a9-6975-4b80-a2c1-af4e30865088/dom_md/outbox > root@stumpy:/tanker/ovirt/host_storage# > > Thank you, > > ..Matt > _______________________________________________ > Users mailing list -- users@ovirt.org > To unsubscribe send an email to users-le...@ovirt.org > Privacy Statement: https://www.ovirt.org/privacy-policy.html > oVirt Code of Conduct: > https://www.ovirt.org/community/about/community-guidelines/ > List Archives: > https://lists.ovirt.org/archives/list/users@ovirt.org/message/BORUNFGGNC6TKTRDOBDB4HOGS2OL532J/ -- Didi _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-le...@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/XU2OY3AASCFTMLKK7RD5I6VLBMP4FDBJ/