Here is the relevant excerpt from nova-compute.log. note that the first line (at /usr/lib/python2.7/dist- packages/nova/virt/libvirt/connection.py:604)is two lines before the timer call which should wait until the container is booted.
2011-09-27 21:20:43,830 DEBUG nova.virt.libvirt_conn [-] instance instance-0000003d: is running from (pid=2100) spawn /usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py:604 2011-09-27 21:20:43,831 DEBUG nova.compute.manager [-] Checking state of instance-0000003d from (pid=2100) _get_power_state /usr/lib/python2.7/dist-packages/nova/compute/manager.py:188 2011-09-27 21:20:43,868 ERROR nova.exception [-] Uncaught exception (nova.exception): TRACE: Traceback (most recent call last): (nova.exception): TRACE: File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 98, in wrapped (nova.exception): TRACE: return f(*args, **kw) (nova.exception): TRACE: File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 454, in run_instance (nova.exception): TRACE: self._run_instance(context, instance_id, **kwargs) (nova.exception): TRACE: File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 432, in _run_instance (nova.exception): TRACE: current_power_state = self._get_power_state(context, instance) (nova.exception): TRACE: File "/usr/lib/python2.7/dist-packages/nova/compute/manager.py", line 190, in _get_power_state (nova.exception): TRACE: return self.driver.get_info(instance['name'])["state"] (nova.exception): TRACE: File "/usr/lib/python2.7/dist-packages/nova/virt/libvirt/connection.py", line 1204, in get_info (nova.exception): TRACE: (state, max_mem, mem, num_cpu, cpu_time) = virt_dom.info() (nova.exception): TRACE: File "/usr/lib/python2.7/dist-packages/libvirt.py", line 1059, in info (nova.exception): TRACE: if ret is None: raise libvirtError ('virDomainGetInfo() failed', dom=self) (nova.exception): TRACE: libvirtError: internal error Unable to get cgroup for instance-0000003d (nova.exception): TRACE: 2011-09-27 21:20:43,869 ERROR nova.rpc [-] Exception during message handling (nova.rpc): TRACE: Traceback (most recent call last): (nova.rpc): TRACE: File "/usr/lib/python2.7/dist-packages/nova/rpc/impl_kombu.py", line 620, in _process_data (nova.rpc): TRACE: rval = node_func(context=ctxt, **node_args) (nova.rpc): TRACE: File "/usr/lib/python2.7/dist-packages/nova/exception.py", line 129, in wrapped (nova.rpc): TRACE: raise Error(str(e)) (nova.rpc): TRACE: Error: internal error Unable to get cgroup for instance-0000003d (nova.rpc): TRACE: 2011-09-27 21:20:43,964 INFO nova.virt.libvirt_conn [-] Instance instance-0000003d spawned successfully. Note also that it says the container spawned successfully. And while the libvirt container is in fact running, euca-describe-instances shows: serge@stack:~$ euca-describe-instances RESERVATION r-lj3xi0v8 test-project default INSTANCE i-0000003d ami-00000003 11.0.0.5 11.0.0.5 pending mykey 3 m1.small 2011-09-28T02:20:25Z nova ami-00000000 ami-00000000 INSTANCE i-0000003c ami-00000003 11.0.0.4 11.0.0.4 running mykey 2 m1.small 2011-09-28T02:20:24Z nova ami-00000000 ami-00000000 INSTANCE i-0000003b ami-00000003 11.0.0.3 11.0.0.3 pending mykey 1 m1.small 2011-09-28T02:20:24Z nova ami-00000000 ami-00000000 INSTANCE i-0000003a ami-00000003 11.0.0.2 11.0.0.2 running mykey 0 m1.small 2011-09-28T02:20:23Z nova ami-00000000 ami-00000000 so i-0000003d is marked pending. -- You received this bug notification because you are a member of Ubuntu Bugs, which is subscribed to Ubuntu. https://bugs.launchpad.net/bugs/842845 Title: problems starting multiple lxc instances concurrently To manage notifications about this bug go to: https://bugs.launchpad.net/ubuntu/+source/libvirt/+bug/842845/+subscriptions -- ubuntu-bugs mailing list ubuntu-bugs@lists.ubuntu.com https://lists.ubuntu.com/mailman/listinfo/ubuntu-bugs