Looks like the "vers" option is not recognized at conenctStorageServer. >From browsing engine code, I see that we potentially send (on NFS storage >domain, should not be sent on PosixFS) "protocol_version", but not "vers"
----- Original Message ----- > From: "Daniel Rowe" <daniel.fatho...@gmail.com> > To: users@ovirt.org > Sent: Monday, October 29, 2012 8:16:34 AM > Subject: [Users] Issue adding a gluster storage domain. > > Hi > > I can't seem to get a gluster storage domain added I am using Fedora > 17 on both the nodes management machine. I have the gluster volumes > showing in ovirt and I can manually mount the gluster volume both > locally on the node and on the management machine. > > If I manually mount the volume on the node with /usr/bin/mount -t > glusterfs -o vers=3 localhost:/bsdvol1 > /rhev/data-center/mnt/localhost:_bsdvol1 as root then add the domain > in the web interface it works. > > Although I have the domain done with the manual mounting, I am > wondering what is going on. > > Thread-1934::INFO::2012-10-29 > 13:27:09,793::logUtils::37::dispatcher::(wrapper) Run and protect: > validateStorageServerConnection(domType=6, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', > 'connection': 'localhost:/bsdvol1', 'iqn': '', 'portal': '', 'user': > '', 'vfs_type': 'glusterfs', 'password': '******', 'id': > '00000000-0000-0000-0000-000000000000'}], options=None) > Thread-1934::INFO::2012-10-29 > 13:27:09,793::logUtils::39::dispatcher::(wrapper) Run and protect: > validateStorageServerConnection, Return response: {'statuslist': > [{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]} > Thread-1934::DEBUG::2012-10-29 > 13:27:09,793::task::1172::TaskManager.Task::(prepare) > Task=`4f9131eb-45cd-4e82-bc35-e29ed14f0818`::finished: {'statuslist': > [{'status': 0, 'id': '00000000-0000-0000-0000-000000000000'}]} > Thread-1934::DEBUG::2012-10-29 > 13:27:09,793::task::588::TaskManager.Task::(_updateState) > Task=`4f9131eb-45cd-4e82-bc35-e29ed14f0818`::moving from state > preparing -> state finished > Thread-1934::DEBUG::2012-10-29 > 13:27:09,794::resourceManager::809::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-1934::DEBUG::2012-10-29 > 13:27:09,794::resourceManager::844::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-1934::DEBUG::2012-10-29 > 13:27:09,794::task::978::TaskManager.Task::(_decref) > Task=`4f9131eb-45cd-4e82-bc35-e29ed14f0818`::ref 0 aborting False > Thread-1935::DEBUG::2012-10-29 > 13:27:09,842::BindingXMLRPC::156::vds::(wrapper) [192.168.1.10] > Thread-1935::DEBUG::2012-10-29 > 13:27:09,842::task::588::TaskManager.Task::(_updateState) > Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::moving from state init > -> > state preparing > Thread-1935::INFO::2012-10-29 > 13:27:09,843::logUtils::37::dispatcher::(wrapper) Run and protect: > connectStorageServer(domType=6, > spUUID='00000000-0000-0000-0000-000000000000', conList=[{'port': '', > 'connection': 'localhost:/bsdvol1', 'mnt_options': 'vers=3', > 'portal': > '', 'user': '', 'iqn': '', 'vfs_type': 'glusterfs', 'password': > '******', 'id': 'bbe1bbc5-62b8-4115-b409-6ddea910a688'}], > options=None) > Thread-1935::DEBUG::2012-10-29 > 13:27:09,851::__init__::1249::Storage.Misc.excCmd::(_log) > '/usr/bin/sudo -n /usr/bin/mount -t glusterfs -o vers=3 > localhost:/bsdvol1 /rhev/data-center/mnt/localhost:_bsdvol1' (cwd > None) > Thread-1935::ERROR::2012-10-29 > 13:27:09,930::hsm::1932::Storage.HSM::(connectStorageServer) Could > not > connect to storageServer > Traceback (most recent call last): > File "/usr/share/vdsm/storage/hsm.py", line 1929, in > connectStorageServer > conObj.connect() > File "/usr/share/vdsm/storage/storageServer.py", line 179, in > connect > self._mount.mount(self.options, self._vfsType) > File "/usr/share/vdsm/storage/mount.py", line 190, in mount > return self._runcmd(cmd, timeout) > File "/usr/share/vdsm/storage/mount.py", line 206, in _runcmd > raise MountError(rc, ";".join((out, err))) > MountError: (1, 'unknown option vers (ignored)\nMount failed. Please > check the log file for more details.\n;ERROR: failed to create > logfile > "/var/log/glusterfs/rhev-data-center-mnt-localhost:_bsdvol1.log" > (Permission denied)\nERROR: failed to open logfile > /var/log/glusterfs/rhev-data-center-mnt-localhost:_bsdvol1.log\n') > Thread-1935::DEBUG::2012-10-29 > 13:27:09,931::lvm::457::OperationMutex::(_invalidateAllPvs) Operation > 'lvm invalidate operation' got the operation mutex > Thread-1935::DEBUG::2012-10-29 > 13:27:09,932::lvm::459::OperationMutex::(_invalidateAllPvs) Operation > 'lvm invalidate operation' released the operation mutex > Thread-1935::DEBUG::2012-10-29 > 13:27:09,932::lvm::469::OperationMutex::(_invalidateAllVgs) Operation > 'lvm invalidate operation' got the operation mutex > Thread-1935::DEBUG::2012-10-29 > 13:27:09,932::lvm::471::OperationMutex::(_invalidateAllVgs) Operation > 'lvm invalidate operation' released the operation mutex > Thread-1935::DEBUG::2012-10-29 > 13:27:09,933::lvm::490::OperationMutex::(_invalidateAllLvs) Operation > 'lvm invalidate operation' got the operation mutex > Thread-1935::DEBUG::2012-10-29 > 13:27:09,933::lvm::492::OperationMutex::(_invalidateAllLvs) Operation > 'lvm invalidate operation' released the operation mutex > Thread-1935::INFO::2012-10-29 > 13:27:09,934::logUtils::39::dispatcher::(wrapper) Run and protect: > connectStorageServer, Return response: {'statuslist': [{'status': > 477, > 'id': 'bbe1bbc5-62b8-4115-b409-6ddea910a688'}]} > Thread-1935::DEBUG::2012-10-29 > 13:27:09,934::task::1172::TaskManager.Task::(prepare) > Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::finished: {'statuslist': > [{'status': 477, 'id': 'bbe1bbc5-62b8-4115-b409-6ddea910a688'}]} > Thread-1935::DEBUG::2012-10-29 > 13:27:09,934::task::588::TaskManager.Task::(_updateState) > Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::moving from state > preparing -> state finished > Thread-1935::DEBUG::2012-10-29 > 13:27:09,934::resourceManager::809::ResourceManager.Owner::(releaseAll) > Owner.releaseAll requests {} resources {} > Thread-1935::DEBUG::2012-10-29 > 13:27:09,935::resourceManager::844::ResourceManager.Owner::(cancelAll) > Owner.cancelAll requests {} > Thread-1935::DEBUG::2012-10-29 > 13:27:09,935::task::978::TaskManager.Task::(_decref) > Task=`ef920d52-3082-495f-bfbc-2eb508c3de18`::ref 0 aborting False > Thread-1937::DEBUG::2012-10-29 > 13:27:10,073::BindingXMLRPC::156::vds::(wrapper) [192.168.1.10] > Thread-1937::DEBUG::2012-10-29 > 13:27:10,073::task::588::TaskManager.Task::(_updateState) > Task=`322f7d1b-0e8d-458b-baf2-5c7f8c5f4303`::moving from state init > -> > state preparing > Thread-1937::INFO::2012-10-29 > 13:27:10,073::logUtils::37::dispatcher::(wrapper) Run and protect: > createStorageDomain(storageType=6, > sdUUID='d8591d07-1611-46cc-9122-fc7badea0f4c', > domainName='DCS_Bis_Sys_gluster', > typeSpecificArg='localhost:/bsdvol1', domClass=1, domVersion='0', > options=None) > Thread-1937::DEBUG::2012-10-29 > 13:27:10,074::misc::1053::SamplingMethod::(__call__) Trying to enter > sampling method (storage.sdc.refreshStorage) > Thread-1937::DEBUG::2012-10-29 > 13:27:10,074::misc::1055::SamplingMethod::(__call__) Got in to > sampling method > Thread-1937::DEBUG::2012-10-29 > 13:27:10,074::misc::1053::SamplingMethod::(__call__) Trying to enter > sampling method (storage.iscsi.rescan) > Thread-1937::DEBUG::2012-10-29 > 13:27:10,074::misc::1055::SamplingMethod::(__call__) Got in to > sampling method > Thread-1937::DEBUG::2012-10-29 > 13:27:10,075::__init__::1249::Storage.Misc.excCmd::(_log) > '/usr/bin/sudo -n /sbin/iscsiadm -m session -R' (cwd None) > Thread-1937::DEBUG::2012-10-29 > 13:27:10,092::__init__::1249::Storage.Misc.excCmd::(_log) FAILED: > <err> = 'iscsiadm: No session found.\n'; <rc> = 21 > Thread-1937::DEBUG::2012-10-29 > 13:27:10,092::misc::1063::SamplingMethod::(__call__) Returning last > result > Thread-1937::DEBUG::2012-10-29 > 13:27:15,382::__init__::1249::Storage.Misc.excCmd::(_log) > '/usr/bin/sudo -n /sbin/multipath' (cwd None) > Thread-1937::DEBUG::2012-10-29 > 13:27:15,414::__init__::1249::Storage.Misc.excCmd::(_log) SUCCESS: > <err> = ''; <rc> = 0 > Thread-1937::DEBUG::2012-10-29 > 13:27:15,415::lvm::457::OperationMutex::(_invalidateAllPvs) Operation > 'lvm invalidate operation' got the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,415::lvm::459::OperationMutex::(_invalidateAllPvs) Operation > 'lvm invalidate operation' released the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,415::lvm::469::OperationMutex::(_invalidateAllVgs) Operation > 'lvm invalidate operation' got the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,416::lvm::471::OperationMutex::(_invalidateAllVgs) Operation > 'lvm invalidate operation' released the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,416::lvm::490::OperationMutex::(_invalidateAllLvs) Operation > 'lvm invalidate operation' got the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,416::lvm::492::OperationMutex::(_invalidateAllLvs) Operation > 'lvm invalidate operation' released the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,416::misc::1063::SamplingMethod::(__call__) Returning last > result > Thread-1937::DEBUG::2012-10-29 > 13:27:15,417::lvm::349::OperationMutex::(_reloadvgs) Operation 'lvm > reload operation' got the operation mutex > Thread-1937::DEBUG::2012-10-29 > 13:27:15,419::__init__::1249::Storage.Misc.excCmd::(_log) > '/usr/bin/sudo -n /sbin/lvm vgs --config " devices { preferred_names > = > [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0 > disable_after_error_count=3 filter = [ > \\"a%36848f690e0b7860017e62e4257089a6e|36848f690e0b78600181b77200a8142b8%\\", > \\"r%.*%\\" ] } global { locking_type=1 prioritise_write_locks=1 > wait_for_locks=1 } backup { retain_min = 50 retain_days = 0 } " > --noheadings --units b --nosuffix --separator | -o > uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free > d8591d07-1611-46cc-9122-fc7badea0f4c' (cwd None) > Thread-1937::DEBUG::2012-10-29 > 13:27:15,536::__init__::1249::Storage.Misc.excCmd::(_log) FAILED: > <err> = ' Volume group "d8591d07-1611-46cc-9122-fc7badea0f4c" not > found\n'; <rc> = 5 > Thread-1937::WARNING::2012-10-29 > 13:27:15,538::lvm::353::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 > [] > [' Volume group "d8591d07-1611-46cc-9122-fc7badea0f4c" not found'] > Thread-1937::DEBUG::2012-10-29 > 13:27:15,538::lvm::376::OperationMutex::(_reloadvgs) Operation 'lvm > reload operation' released the operation mutex > Thread-1937::INFO::2012-10-29 > 13:27:15,542::nfsSD::64::Storage.StorageDomain::(create) > sdUUID=d8591d07-1611-46cc-9122-fc7badea0f4c > domainName=DCS_Bis_Sys_gluster remotePath=localhost:/bsdvol1 > domClass=1 > Thread-1937::ERROR::2012-10-29 > 13:27:15,550::task::853::TaskManager.Task::(_setError) > Task=`322f7d1b-0e8d-458b-baf2-5c7f8c5f4303`::Unexpected error > Traceback (most recent call last): > File "/usr/share/vdsm/storage/task.py", line 861, in _run > return fn(*args, **kargs) > File "/usr/share/vdsm/logUtils.py", line 38, in wrapper > res = f(*args, **kwargs) > File "/usr/share/vdsm/storage/hsm.py", line 2136, in > createStorageDomain > typeSpecificArg, storageType, domVersion) > File "/usr/share/vdsm/storage/nfsSD.py", line 75, in create > cls._preCreateValidation(sdUUID, mntPoint, remotePath, version) > File "/usr/share/vdsm/storage/nfsSD.py", line 44, in > _preCreateValidation > raise se.StorageDomainFSNotMounted(domPath) > StorageDomainFSNotMounted: Storage domain remote path not mounted: > ('/rhev/data-center/mnt/localhost:_bsdvol1',) > > [root@bsdvmhvnode01 mnt]# cat > /var/log/glusterfs/rhev-data-center-mnt-localhost:_bsdvol1.log > [2012-10-29 13:27:42.311022] I [glusterfsd.c:1666:main] > 0-/usr/sbin/glusterfs: Started running /usr/sbin/glusterfs version > 3.3.1 > [2012-10-29 13:27:42.328054] I > [io-cache.c:1549:check_cache_ users@ovirt.org. size_ok] > 0-bsdvol1-quick-read: Max cache size is 50636935168 > [2012-10-29 13:27:42.328157] I [io-cache.c:1549:check_cache_size_ok] > 0-bsdvol1-io-cache: Max cache size is 50636935168 > [2012-10-29 13:27:42.329037] I [client.c:2142:notify] > 0-bsdvol1-client-0: parent translators are ready, attempting connect > on transport > Given volfile: > +------------------------------------------------------------------------------+ > 1: volume bsdvol1-client-0 > 2: type protocol/client > 3: option remote-host 192.168.1.11 > 4: option remote-subvolume /data/bsdvol1 > 5: option transport-type tcp > 6: option username ed9b899c-26c1-4d8b-a2f8-ffd2d0b10fed > 7: option password 37a8f9ed-e27c-41cb-a049-da6a65042920 > 8: end-volume > 9: > 10: volume bsdvol1-write-behind > 11: type performance/write-behind > 12: subvolumes bsdvol1-client-0 > 13: end-volume > 14: > 15: volume bsdvol1-read-ahead > 16: type performance/read-ahead > 17: subvolumes bsdvol1-write-behind > 18: end-volume > 19: > 20: volume bsdvol1-io-cache > 21: type performance/io-cache > 22: subvolumes bsdvol1-read-ahead > 23: end-volume > 24: > 25: volume bsdvol1-quick-read > 26: type performance/quick-read > 27: subvolumes bsdvol1-io-cache > 28: end-volume > 29: > 30: volume bsdvol1-md-cache > 31: type performance/md-cache > 32: subvolumes bsdvol1-quick-read > 33: end-volume > 34: > 35: volume bsdvol1 > 36: type debug/io-stats > 37: option latency-measurement off > 38: option count-fop-hits off > 39: subvolumes bsdvol1-md-cache > 40: end-volume > > +------------------------------------------------------------------------------+ > [2012-10-29 13:27:42.334438] I [rpc-clnt.c:1657:rpc_clnt_reconfig] > 0-bsdvol1-client-0: changing port to 24009 (from 0) > [2012-10-29 13:27:46.331768] I > [client-handshake.c:1636:select_server_supported_programs] > 0-bsdvol1-client-0: Using Program GlusterFS 3.3.1, Num (1298437), > Version (330) > [2012-10-29 13:27:46.332216] I > [client-handshake.c:1433:client_setvolume_cbk] 0-bsdvol1-client-0: > Connected to 192.168.1.11:24009, attached to remote volume > '/data/bsdvol1'. > [2012-10-29 13:27:46.332249] I > [client-handshake.c:1445:client_setvolume_cbk] 0-bsdvol1-client-0: > Server and Client lk-version numbers are not same, reopening the fds > [2012-10-29 13:27:46.337041] I [fuse-bridge.c:4191:fuse_graph_setup] > 0-fuse: switched to graph 0 > [2012-10-29 13:27:46.337157] I > [client-handshake.c:453:client_set_lk_version_cbk] > 0-bsdvol1-client-0: > Server lk version = 1 > [2012-10-29 13:27:46.337466] I [fuse-bridge.c:3376:fuse_init] > 0-glusterfs-fuse: FUSE inited with protocol versions: glusterfs 7.13 > kernel 7.20 > _______________________________________________ > Users mailing list > Users@ovirt.org > http://lists.ovirt.org/mailman/listinfo/users > _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users