Hi Gabriel, Thank you very much for your help, it was exactly the same problem that you had.
I set the host variable on my rbd backend in the cinder config file and restarted the cinder-volume processes on my controllers. Then I manually updated the database “cinder” to reflect the new cinder hostname “UPDATE volumes SET host=my_rbd_backend';” It fixed all my inconsistent volumes and images, I could deleted them again. Thank you again you definitely made my day. +++ Romain From: Staicu Gabriel [mailto:gabriel_sta...@yahoo.com] Sent: vendredi 5 juin 2015 10:31 To: Aviolat Romain Subject: Re: [Openstack] VMs not terminated correctly and undeletable, with cinder and CEPH Hi Romain, I remember having a similar problem some time ago on havana. Some volumes were undeleteble. The issue in my case was the parameter at line 570 from this document: cinder.conf - OpenStack Configuration Reference - kilo<http://docs.openstack.org/kilo/config-reference/content/section_cinder.conf.html> cinder.conf - OpenStack Configuration Reference - kilo<http://docs.openstack.org/kilo/config-reference/content/section_cinder.conf.html> cinder.conf The cinder.conf file is installed in /etc/cinder by default. View on docs.openstack.org<http://docs.openstack.org/kilo/config-reference/content/section_cinder.conf.html> Preview by Yahoo The idea was that when I was having the problem this parameter was set on the default value which was the host of the machine. So each volume was bond the the cinder worker that created it. After I have modified it to an arbitrary string I did not see the problem with deleting volumes. Also that made the installation HA as any cinder worked could take care of any volume in ceph cluster. Hope it gives you some hint for your issue. Have a nice day, Gabriel On Friday, June 5, 2015 10:11 AM, Aviolat Romain <romain.avio...@nagra.com<mailto:romain.avio...@nagra.com>> wrote: Dear Openstack community I have a strange intermittent problem with cinder when VMs with volumes attached are terminated. Sometimes the process doesn't work and the VMs are stuck into and undeletable state. Some info about my setup: ***************** * Openstack Juno from ubuntu-cloud repo * Ubuntu 14.04 LTS on every machines * CEPH (giant) as storage backend on the storage nodes * 3x controllers in HA (HA-proxy, Galera, CEPH-monitors...) * RabbitMQ as messaging system ***************** I put my whole setup in debug mode and here's what happened for a problematic VM: Info about the VM and its volume: ***************** Vm: suricata-bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc Volume attached: 821006f1-c655-4589-ba3e-7c6804c8e120 on /dev/vda Volume size: 500GB Running on hypervisor: cloudcompute2 Libvirt instance ID: instance-000002e3 ***************** To begin here's the VM action log from Horizon, we can see that the user deleted the VM at 1.23 PM: ***************** Request ID Action Start Time User ID Message req-22f33ed5-a231-4ea2-95fa-0022bf731079 delete June 4, 2015, 1:23 p.m. 27cf3aaa0d7942009b03eabf7f686849 Error ***************** Here's the corresponding action on the controller that received the request: ***************** 2015-06-04 13:23:16.693 32302 DEBUG nova.compute.api [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] [instance: bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc] Going to try to terminate instance delete /usr/lib/python2.7/dist-packages/nova/compute/api.py:1802 2015-06-04 13:23:16.774 32302 DEBUG nova.quota [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] Created reservations ['2221d555-1db0-4c36-9bd6-6dc815c22fc9', '3b70b444-7633-4a6b-b8b2-16128ee1469c', '6b8b2421-cfc1-4568-8bd8-430a982d977f'] reserve /usr/lib/python2.7/dist-packages/nova/quota.py:1310 2015-06-04 13:23:16.799 32302 INFO nova.osapi_compute.wsgi.server [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] 172.24.1.17 "DELETE /v2/c9324c924a5049e7922882aff55c3813/servers/bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc HTTP/1.1" status: 204 len: 179 time: 0.1563590 2015-06-04 13:23:16.883 32296 DEBUG nova.api.openstack.wsgi [req-903c3ea2-151a-4f8c-a48a-6991721c8c18 None] Calling method '<bound method Controller.detail of <nova.api.openstack.compute.servers.Controller object at 0x7f869ec56790>>' _process_stack /usr/lib/python2.7/dist-packages/nova/api/openstack/wsgi.py:937 2015-06-04 13:23:16.884 32296 DEBUG nova.api.openstack.compute.servers [req-903c3ea2-151a-4f8c-a48a-6991721c8c18 None] Removing options 'project_id, limit' from query remove_invalid_options /usr/lib/python2.7/dist-packages/nova/api/openstack/compute/servers.py:1533 ***************** Then on nova-compute on the compute node hosting the VM, we receive the DELETE-instance request: ***************** 2015-06-04 13:23:16.847 1971 AUDIT nova.compute.manager [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] [instance: bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc] Terminating instance 2015-06-04 13:23:16.854 1971 INFO nova.virt.libvirt.driver [-] [instance: bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc] Instance destroyed successfully. ***************** Then the network device is removed: ***************** 2015-06-04 13:23:17.558 1971 DEBUG nova.network.linux_net [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] Net device removed: 'qvo20393401-b2' delete_net_dev /usr/lib/python2.7/dist-packages/nova/network/linux_net.py:1381 ***************** The the libvirt config file is removed: ***************** 2015-06-04 13:23:18.048 1971 INFO nova.virt.libvirt.driver [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] [instance: bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc] Deleting instance files /var/lib/nova/instances/bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc_del ***************** The next one is strange: ***************** 2015-06-04 13:23:19.303 1971 DEBUG nova.virt.libvirt.driver [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] [instance: bbd3c3f5-85fb-4c90-8e17-e7f083a4a0bc] Could not determine fibre channel world wide node names get_volume_connector /usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py:1280 2015-06-04 13:23:19.358 1971 DEBUG nova.volume.cinder [req-22f33ed5-a231-4ea2-95fa-0022bf731079 None] Cinderclient connection created using URL: http://haproxy-api:8776/v2/c9324c924a5049e7922882aff55c3813 get_cinder_client_version /usr/lib/python2.7/dist-packages/nova/volume/cinder.py:255 ***************** We can see that the instance was correctly removed from the computenode: ***************** [CloudCompute2:~] 1 $ sudo virsh list --all Id Name State ---------------------------------------------------- 31 instance-0000043b running 32 instance-000002d4 running 42 instance-00000288 running 47 instance-0000028f running 48 instance-00000465 running 52 instance-00000482 running ***************** Then on controller2 we have some strange errors, looks like cinder sent messages but those timed-out (cinder-api.log). ***************** 2015-06-04 13:23:19.512 17252 INFO cinder.api.openstack.wsgi [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] POST http://haproxy-api:8776/v2/c9324c924a5049e7922882aff55c3813/volumes/821006f1-c655-4589-ba3e-7c6804c8e120/action 2015-06-04 13:23:19.513 17252 DEBUG cinder.api.openstack.wsgi [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Action body: {"os-terminate_connection": {"connector": {"ip": "172.24.1.27", "host": "cloudcompute2.ksnet.nagra.com", "initiator": "iqn.1993-08.org.debian:01:73625b6efe77"}}} get_method /usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py:1010 2015-06-04 13:23:19.546 17252 INFO oslo.messaging._drivers.impl_rabbit [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Connecting to AMQP server on cloudcontroller1-api:5672 2015-06-04 13:23:19.556 17252 INFO oslo.messaging._drivers.impl_rabbit [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Connected to AMQP server on cloudcontroller1-api:5672 2015-06-04 13:23:19.567 17252 INFO oslo.messaging._drivers.impl_rabbit [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Connecting to AMQP server on cloudcontroller0-api:5672 2015-06-04 13:23:19.581 17252 INFO oslo.messaging._drivers.impl_rabbit [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Connected to AMQP server on cloudcontroller0-api:5672 2015-06-04 13:23:20.170 17247 INFO eventlet.wsgi.server [-] (17247) accepted ('172.24.1.19', 36820) 2015-06-04 13:23:20.171 17247 INFO cinder.api.openstack.wsgi [-] OPTIONS http://172.24.1.19:8776/ 2015-06-04 13:23:20.171 17247 DEBUG cinder.api.openstack.wsgi [-] Empty body provided in request get_body /usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py:789 2015-06-04 13:23:20.172 17247 INFO cinder.api.openstack.wsgi [-] http://172.24.1.19:8776/ returned with HTTP 200 2015-06-04 13:23:20.172 17247 INFO eventlet.wsgi.server [-] 172.24.1.19 - - [04/Jun/2015 13:23:20] "OPTIONS / HTTP/1.0" 200 281 0.001354 2015-06-04 13:23:21.157 17245 INFO eventlet.wsgi.server [-] (17245) accepted ('172.24.1.17', 34260) 2015-06-04 13:23:21.159 17245 INFO cinder.api.openstack.wsgi [-] OPTIONS http://172.24.1.19:8776/ 2015-06-04 13:23:21.159 17245 DEBUG cinder.api.openstack.wsgi [-] Empty body provided in request get_body /usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py:789 2015-06-04 13:23:21.160 17245 INFO cinder.api.openstack.wsgi [-] http://172.24.1.19:8776/ returned with HTTP 200 2015-06-04 13:23:21.161 17245 INFO eventlet.wsgi.server [-] 172.24.1.17 - - [04/Jun/2015 13:23:21] "OPTIONS / HTTP/1.0" 200 281 0.002750 2015-06-04 13:23:21.173 17251 INFO eventlet.wsgi.server [-] (17251) accepted ('172.24.1.18', 49259) 2015-06-04 13:23:21.174 17251 INFO cinder.api.openstack.wsgi [-] OPTIONS http://172.24.1.19:8776/ 2015-06-04 13:23:21.175 17251 DEBUG cinder.api.openstack.wsgi [-] Empty body provided in request get_body /usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py:789 2015-06-04 13:23:21.176 17251 INFO cinder.api.openstack.wsgi [-] http://172.24.1.19:8776/ returned with HTTP 200 ..... 2015-06-04 13:24:19.338 17254 INFO eventlet.wsgi.server [-] 172.24.1.18 - - [04/Jun/2015 13:24:19] "OPTIONS / HTTP/1.0" 200 281 0.001299 2015-06-04 13:24:19.585 17252 ERROR cinder.api.middleware.fault [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Caught error: Timed out waiting for a reply to message ID 27d7e0cc6f444391823d5d4d1c1c85c1 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault Traceback (most recent call last): 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/api/middleware/fault.py", line 76, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return req.get_response(self.application) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1320, in send 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault application, catch_exc_info=False) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1284, in call_application 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault app_iter = application(self.environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return resp(environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 130, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault resp = self.call_func(req, *args, **self.kwargs) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 195, in call_func 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return self.func(req, *args, **kwargs) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/osprofiler/web.py", line 99, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return request.get_response(self.application) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1320, in send 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault application, catch_exc_info=False) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/request.py", line 1284, in call_application 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault app_iter = application(self.environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/keystonemiddleware/auth_token.py", line 661, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return self._app(env, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return resp(environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return resp(environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/routes/middleware.py", line 131, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault response = self.app(environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 144, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return resp(environ, start_response) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 130, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault resp = self.call_func(req, *args, **self.kwargs) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/webob/dec.py", line 195, in call_func 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return self.func(req, *args, **kwargs) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", line 897, in __call__ 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault content_type, body, accept) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", line 945, in _process_stack 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault action_result = self.dispatch(meth, request, action_args) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/api/openstack/wsgi.py", line 1021, in dispatch 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return method(req=request, **action_args) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/api/contrib/volume_actions.py", line 218, in _terminate_connection 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault self.volume_api.terminate_connection(context, volume, connector) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/volume/api.py", line 87, in wrapped 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault return func(self, context, target_obj, *args, **kwargs) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/volume/api.py", line 519, in terminate_connection 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault force) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/cinder/volume/rpcapi.py", line 170, in terminate_connection 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault connector=connector, force=force) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/oslo/messaging/rpc/client.py", line 152, in call 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault retry=self.retry) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/oslo/messaging/transport.py", line 90, in _send 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault timeout=timeout, retry=retry) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/oslo/messaging/_drivers/amqpdriver.py", line 408, in send 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault retry=retry) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/oslo/messaging/_drivers/amqpdriver.py", line 397, in _send 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault result = self._waiter.wait(msg_id, timeout) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/oslo/messaging/_drivers/amqpdriver.py", line 285, in wait 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault reply, ending = self._poll_connection(msg_id, timeout) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault File "/usr/lib/python2.7/dist-packages/oslo/messaging/_drivers/amqpdriver.py", line 235, in _poll_connection 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault % msg_id) 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault MessagingTimeout: Timed out waiting for a reply to message ID 27d7e0cc6f444391823d5d4d1c1c85c1 2015-06-04 13:24:19.585 17252 TRACE cinder.api.middleware.fault 2015-06-04 13:24:19.596 17252 INFO cinder.api.middleware.fault [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] http://haproxy-api:8776/v2/c9324c924a5049e7922882aff55c3813/volumes/821006f1-c655-4589-ba3e-7c6804c8e120/action returned with HTTP 500 2015-06-04 13:24:19.598 17252 INFO eventlet.wsgi.server [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] Traceback (most recent call last): File "/usr/lib/python2.7/dist-packages/eventlet/wsgi.py", line 405, in handle_one_response write(''.join(towrite)) File "/usr/lib/python2.7/dist-packages/eventlet/wsgi.py", line 349, in write _writelines(towrite) File "/usr/lib/python2.7/socket.py", line 334, in writelines self.flush() File "/usr/lib/python2.7/socket.py", line 303, in flush self._sock.sendall(view[write_offset:write_offset+buffer_size]) File "/usr/lib/python2.7/dist-packages/eventlet/greenio.py", line 307, in sendall tail = self.send(data, flags) File "/usr/lib/python2.7/dist-packages/eventlet/greenio.py", line 293, in send total_sent += fd.send(data[total_sent:], flags) error: [Errno 104] Connection reset by peer 2015-06-04 13:24:19.598 17252 INFO eventlet.wsgi.server [req-ce2dcb7f-4bfd-4345-a199-165f13f7f11d 27cf3aaa0d7942009b03eabf7f686849 c9324c924a5049e7922882aff55c3813 - - -] 172.24.1.17 - - [04/Jun/2015 13:24:19] "POST /v2/c9324c924a5049e7922882aff55c3813/volumes/821006f1-c655-4589-ba3e-7c6804c8e120/action HTTP/1.1" 500 0 60.230434 2015-06-04 13:24:20.321 17254 INFO eventlet.wsgi.server [-] (17254) accepted ('172.24.1.19', 37634) 2015-06-04 13:24:20.321 17254 INFO cinder.api.openstack.wsgi [-] OPTIONS http://172.24.1.19:8776/ ***************** I checked in the CEPH cluster and the volume wasn't deleted. I also checked the state of the RabbitMQ cluster and everything is working as expected. I'm not able to reproduce the problem, it happen sometimes but I haven't found the trigger... When it happens the volume is un-deletable, it's marked as in-use by the VM and the VM is un-deletable because a volume is attached. The only way I found is to edit the database (nove, cinder, ...) and delete the volume in the CEPH cluster manually. You help will be very much appreciated ! Thanks, Romain Romain Aviolat Senior System Administrator - R&D and ops Infrastructure Kudelski Security - Kudelski Group rte de Genève 22-24, 1033 Cheseaux, SWITZERLAND +41 21 732 03 79 _______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org<mailto:openstack@lists.openstack.org> Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack