On Wed, Apr 17, 2013 at 11:56 AM, Gianluca Cecchi wrote: > On Wed, Apr 17, 2013 at 11:30 AM, Yeela Kaplan wrote: >> Gianluca, >> You need to first put the domain to maintenance, >> then detach the storage domain from the data center, >> and then the 'remove' option will be available. >> >> -- >> Yeela > > I supposed that, but I didn't find in the gui the place where to put > the SD into maintenance.... > > following the guide for rhev 3.1 it should something like this, correct? > > Procedure 7.12. Removing a Storage Domain > 1. Use the Storage resource tab, tree mode, or the search function to > find and select the > appropriate storage domain in the results list. > 2. Move the domain into maintenance mode to deactivate it. > 3. Detach the domain from the data center. > 4. Click Remove to open the Rem ove Storage confirmation window. > 5. Select a host from the list. > 6. Click OK to remove the storage domain and close the window. > > Possibly I selected the wrong place in 1. because I don't remember to > have seen a "maintenance" option there. > I'm going to re-check.... > Or perhaps it was greyed out... > > Gianluca
Ok, I found it.... not so intuitive in my opinion... 1. Use the Storage resource tab, tree mode 2. Select the storage domain in the main tab 3. In the details pane at bottom select datacenter tab 4. Select the DC where the storage domain is active 5. Select Maintenance link 6. Select Detach link to detach from the data center. Now you don't see any more the storage domain in storage resource tab. 7. Go to system --> Storage and the storage domain now in in an unlinked state 8. Select the storage domain and choose "remove" 9. Select a host from the list. 10. Click OK to remove the storage domain and close the window. So this seems the detailed workflow. In my case after step 10 I receive: Error while executing action Remove Storage Domain: Volume Group remove error In vdsm.log: Thread-2043307::DEBUG::2013-04-18 09:08:16,909::task::1151::TaskManager.Task::(prepare) Task=`c2c45a63-6361-4d68-b93d-b03a 6825cf5f`::finished: {u'be882de7-6c79-413b-85aa-f0b8b77cb59e': {'delay': '0.0529000759125', 'lastCheck': '4.7', 'code': 0, 'valid': True}, u'3fb66ba1-cfcb-4341-8960-46f0e8cf6e83': {'delay': '0.0538139343262', 'lastCheck': '8.6', 'code': 0, 'val id': True}, u'8573d237-f86f-4b27-be80-479281a53645': {'delay': '0.0613639354706', 'lastCheck': '5.6', 'code': 0, 'valid': True}, u'596a3408-67d7-4b26-b482-e3a7554a5897': {'delay': '0.0538330078125', 'lastCheck': '5.1', 'code': 0, 'valid': True} , u'e3251723-08e1-4b4b-bde4-c10d6372074b': {'delay': '0.0559990406036', 'lastCheck': '0.2', 'code': 0, 'valid': True}, u'2 aff7dc6-e25b-433b-9681-5541a29bb07c': {'delay': '0.0598528385162', 'lastCheck': '0.2', 'code': 0, 'valid': True}, u'14b516 7c-5883-4920-8236-e8905456b01f': {'delay': '0.0535669326782', 'lastCheck': '4.6', 'code': 0, 'valid': True}} Thread-2043307::DEBUG::2013-04-18 09:08:16,909::task::568::TaskManager.Task::(_updateState) Task=`c2c45a63-6361-4d68-b93d-b03a6825cf5f`::moving from state preparing -> state finished Thread-2043307::DEBUG::2013-04-18 09:08:16,909::resourceManager::830::ResourceManager.Owner::(releaseAll) Owner.releaseAll requests {} resources {} Thread-2043307::DEBUG::2013-04-18 09:08:16,910::resourceManager::864::ResourceManager.Owner::(cancelAll) Owner.cancelAll requests {} Thread-2043307::DEBUG::2013-04-18 09:08:16,910::task::957::TaskManager.Task::(_decref) Task=`c2c45a63-6361-4d68-b93d-b03a6825cf5f`::ref 0 aborting False VM Channels Listener::DEBUG::2013-04-18 09:08:17,722::vmChannels::61::vds::(_handle_timeouts) Timeout on fileno 60. Dummy-1143::DEBUG::2013-04-18 09:08:17,969::misc::84::Storage.Misc.excCmd::(<lambda>) 'dd if=/rhev/data-center/5849b030-626e-47cb-ad90-3ce782d831b3/mastersd/dom_md/inbox iflag=direct,fullblock count=1 bs=1024000' (cwd None) Dummy-1143::DEBUG::2013-04-18 09:08:18,069::misc::84::Storage.Misc.excCmd::(<lambda>) SUCCESS: <err> = '1+0 records in\n1+0 records out\n1024000 bytes (1.0 MB) copied, 0.031044 s, 33.0 MB/s\n'; <rc> = 0 Thread-1928975::ERROR::2013-04-18 09:08:18,136::utils::416::vm.Vm::(collect) vmId=`c0a43bef-7c9d-4170-bd9c-63497e61d3fc`::Stats function failed: <AdvancedStatsFunction _highWrite at 0x16f45d0> Traceback (most recent call last): File "/usr/lib64/python2.7/site-packages/vdsm/utils.py", line 412, in collect statsFunction() File "/usr/lib64/python2.7/site-packages/vdsm/utils.py", line 287, in __call__ retValue = self._function(*args, **kwargs) File "/usr/share/vdsm/libvirtvm.py", line 134, in _highWrite self._vm._dom.blockInfo(vmDrive.path, 0) File "/usr/share/vdsm/libvirtvm.py", line 541, in f ret = attr(*args, **kwargs) File "/usr/lib64/python2.7/site-packages/vdsm/libvirtconnection.py", line 111, in wrapper ret = f(*args, **kwargs) File "/usr/lib64/python2.7/site-packages/libvirt.py", line 1781, in blockInfo if ret is None: raise libvirtError ('virDomainGetBlockInfo() failed', dom=self) libvirtError: failed to open path '/rhev/data-center/5849b030-626e-47cb-ad90-3ce782d831b3/013bcc40-5f3d-4394-bd3b-971b14852654/images/01488698-6420-4a32-9095-cfed1ff8f4bf/d477fcba-2110-403e-93fe-15565aae5304': No such file or directory Thread-23::DEBUG::2013-04-18 09:08:18,351::misc::84::Storage.Misc.excCmd::(<lambda>) '/usr/bin/dd iflag=direct if=/dev/3fb66ba1-cfcb-4341-8960-46f0e8cf6e83/metadata bs=4096 count=1' (cwd None) Thread-23::DEBUG::2013-04-18 09:08:18,397::misc::84::Storage.Misc.excCmd::(<lambda>) SUCCESS: <err> = '1+0 records in\n1+0 records out\n4096 bytes (4.1 kB) copied, 0.000349033 s, 11.7 MB/s\n'; <rc> = 0 Any hint? _______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users