Hello again Razique, I was trying to migrate through the terminal and I'm getting this line from my controller node:
nova live-migration f18b6d2b-f685-4476-a7c0-e735db5113cf 192.168.3.3 ERROR: Compute service of 192.168.3.3 is unavailable at this time. (HTTP 400) (Request-ID: req-454960fb-6652-46ed-9080-6d23b384cbb3) Haven't found anything yet. Any ideas? Regards. Guilherme. 2013/8/19 Guilherme Russi <luisguilherme...@gmail.com> > I'm migrating it through the dashboard with the admin user. Is it wrong? > > Thank you. > > Guilherme. > > > 2013/8/19 Razique Mahroua <razique.mahr...@gmail.com> > >> quick question, how do you migrate the instance ? and does it belong to >> the tenant? >> >> >> Le 19 août 2013 à 15:36, Guilherme Russi <luisguilherme...@gmail.com> a >> écrit : >> >> The VM looks to be there: >> >> du -h instances/ >> 4,0K instances/keys >> 1,9G instances/_base >> 0 instances/locks >> 310M instances/edca1460-9c97-475e-997b-3266b235e797 >> 260M instances/0fe59b28-04a3-41bb-9926-b7d139a70548 >> 323M instances/aeda1c56-72ba-40a7-9857-01a622540505 >> 7,0M instances/f4af662d-ae55-4346-82cf-a997192706b5 >> 264M instances/483f98e3-8ef5-43e2-8c3a-def55abdabcd >> 268M instances/eeb81e32-32ce-49fa-8249-afc572f26233 >> 268M instances/c0224f0d-d7e8-4621-8c1a-af68740a3267 >> 188M instances/72ec37a3-b209-4729-b628-005fdcea5a3c >> 12M instances/2c7c123b-456a-4c70-a991-80efee541324 >> 11M instances/c9e1c5ed-a108-4196-bfbc-24495e2e71bd >> 6,5M instances/96d8f2de-a72d-40b5-b083-1f5e511f91ce >> 531M instances/1b35a4a6-e158-48d3-a804-44b05ec171d0 >> 529M instances/1a2b6146-6ef0-4d76-a603-e4715b4dfb62 >> 514M instances/f6cb9f81-9281-419f-9e20-adaa6a07cea3 >> 2,7G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4_resize >> 2,4G instances/75e479cf-37a4-4511-b2a5-d4f3027702d4 >> 11G instances/ >> >> >> It's still running, no Error messagem from dashboard (it is showing >> Status: Resize/Migrate - Task: Resizing or Migrating - Power State: >> Running). The logs are not showing that permission error anymore. Should >> I wait to see if it will migrate or returns another error status? One thing >> is true, the machine is really slow now :P >> >> Regards. >> >> Guilherme. >> >> >> 2013/8/19 Razique Mahroua <razique.mahr...@gmail.com> >> >>> Looks like the file doesn't exist. >>> Can you spawn a new instance and try to migrate it again? >>> thanks >>> >>> *Razique Mahroua** - **Nuage & Co* >>> razique.mahr...@gmail.com >>> Tel : +33 9 72 37 94 15 >>> >>> <NUAGECO-LOGO-Fblan_petit.jpg> >>> >>> Le 19 août 2013 à 13:41, Guilherme Russi <luisguilherme...@gmail.com> a >>> écrit : >>> >>> Hello Peter, >>> >>> Thank you for your help too :) >>> I've changed the permission and now I'm getting another error when I >>> try to migrate: >>> >>> cat /var/log/nova/nova-compute.log | grep -E ERROR >>> 2013-08-19 08:17:19.990 12667 ERROR nova.manager [-] Error during >>> ComputeManager._run_image_cache_manager_pass: Unexpected error while >>> running command. >>> 2013-08-19 08:23:40.309 12667 ERROR nova.network.quantumv2 [-] >>> _get_auth_token() failed >>> 2013-08-19 08:24:27.013 12667 ERROR nova.virt.libvirt.driver [-] Getting >>> disk size of instance-000000bc: [Errno 2] No such file or directory: >>> '/var/lib/nova/instances/75e479cf-37a4-4511-b2a5-d4f3027702d4/disk' >>> >>> >>> cat /var/log/nova/nova-compute.log | grep -E TRACE >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Traceback (most recent >>> call last): >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/manager.py", line 241, in >>> periodic_tasks >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager task(self, context) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 4020, in >>> _run_image_cache_manager_pass >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>> self.driver.manage_image_cache(context, filtered_instances) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py", line 3401, >>> in manage_image_cache >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>> self.image_cache_manager.verify_base_images(context, all_instances) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line >>> 588, in verify_base_images >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>> self._handle_base_image(img, base_file) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/imagecache.py", line >>> 553, in _handle_base_image >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>> virtutils.chown(base_file, os.getuid()) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 449, in >>> chown >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager execute('chown', >>> owner, path, run_as_root=True) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/utils.py", line 47, in >>> execute >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager return >>> utils.execute(*args, **kwargs) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager File >>> "/usr/lib/python2.7/dist-packages/nova/utils.py", line 239, in execute >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager cmd=' '.join(cmd)) >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager ProcessExecutionError: >>> Unexpected error while running command. >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Command: sudo >>> nova-rootwrap /etc/nova/rootwrap.conf chown 123 >>> /var/lib/nova/instances/_base/b0d90eb60f048f6de16c3efbc0eb4c16cbac77a6 >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Exit code: 1 >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stdout: '' >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager Stderr: 'Traceback >>> (most recent call last):\n File "/usr/bin/nova-rootwrap", line 108, in >>> <module>\n os.getlogin(), pwd.getpwuid(os.getuid())[0],\nOSError: [Errno >>> 22] Invalid argument\n' >>> 2013-08-19 08:17:19.990 12667 TRACE nova.manager >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Traceback >>> (most recent call last): >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File >>> "/usr/lib/python2.7/dist-packages/nova/network/quantumv2/__init__.py", line >>> 40, in _get_auth_token >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 >>> httpclient.authenticate() >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 File >>> "/usr/lib/python2.7/dist-packages/quantumclient/client.py", line 198, in >>> authenticate >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 raise >>> exceptions.Unauthorized(message=body) >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 Unauthorized: >>> Request Timeout >>> 2013-08-19 08:23:40.309 12667 TRACE nova.network.quantumv2 >>> >>> Any ideas? >>> >>> Regards. >>> >>> Guilherme. >>> >>> >>> 2013/8/19 Peter Funk <p...@artcom-gmbh.de> >>> >>>> Guilherme Russi wrote 16.08.2013 um 09:53: >>>> > Here is my directory configs: >>>> > >>>> > drwxrwxrwx 2 nova nova 76 Ago 15 14:54 .ssh >>>> > >>>> > I made chmod a+xwr at the folder at all computers, I made it because I >>>> > couldn't find the reason I was no allowed to migrate, can it be like >>>> that? >>>> >>>> For SSH to work the permissions of $HOME/.ssh should look like this: >>>> drwx------ 2 nova nova 76 Ago 15 14:54 .ssh >>>> >>>> Use: >>>> chmod -R go-rwx .ssh >>>> >>>> Regards, Peter Funk >>>> -- >>>> Peter Funk, home: ✉Oldenburger Str.86, D-27777 Ganderkesee >>>> mobile:+49-179-640-8878 phone:+49-421-20419-0 < >>>> http://www.artcom-gmbh.de/> >>>> office: ArtCom GmbH, ✉Haferwende 2, D-28357 Bremen, Germany >>>> >>>> _______________________________________________ >>>> Mailing list: >>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>> Post to : openstack@lists.openstack.org >>>> Unsubscribe : >>>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>>> >>> >>> _______________________________________________ >>> Mailing list: >>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>> Post to : openstack@lists.openstack.org >>> Unsubscribe : >>> http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack >>> >>> >>> >> >> >
_______________________________________________ Mailing list: http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack Post to : openstack@lists.openstack.org Unsubscribe : http://lists.openstack.org/cgi-bin/mailman/listinfo/openstack