[Yahoo-eng-team] [Bug 1423654] Re: Nova rescue causes LVM timeouts after moving attachments

2016-07-20 Thread Andrea Rosa
from the last comment it seems to me that this bug is not going to be fixed in 
Nova but a workaround is in place in Cinder. Considering that and considering 
that the last comment was 1+ year ago I am going to close this ticket as "Won't 
fix".
Please feel free to comment here if you disagree.

** Changed in: nova
   Status: Confirmed => Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1423654

Title:
  Nova rescue causes LVM timeouts after moving attachments

Status in OpenStack Compute (nova):
  Won't Fix

Bug description:
  The Nova rescue feature powers off a running instance and, boots a
  rescue instance attaching the ephemeral disk of the original instance
  to it to allow an admin to try and recover the instance.  The problem
  is that if a Cinder Volume is attached to that instance when we do a
  rescue we don't do a detach or any sort of maintenance on the block
  mapping that we have set up for it.  We do check to see if we have it,
  and verify it's attached but that's it.

  The result is that after the rescue operation subsequent LVM calls to do 
things like lvs and vgs will attempt to open a device file that no longer 
exists which takes up to 60 seconds for each device.  An example is the current 
tempest test:
  
tempest.api.compute.servers.test_server_rescue_negative.ServerRescueNegativeTestJSON.test_rescued_vm_detach_volume[gate,negative,volume]

  Which if you look at tempest results you'll notice that test always
  takes in excess of 100 seconds, but it's not just because it's a long
  test, it's the blocking LVM calls.

  We should detach any cinder volumes that are attached to an instance during 
the rescue process.  One concern with this that came from folks on the Nova 
team was 'what about boot from volume'?  Rescue of a volume booted instance is 
currently an invalid case as is evident by the code that checks for it and 
fails here:
  https://github.com/openstack/nova/blob/master/nova/compute/api.py#L2822

  Probably no reason we can't automate this as part of rescue in the
  future but for now it's a separate enhancement independent of this
  bug.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1423654/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1604798] Re: Use DDT library to reduce code duplication

2016-07-20 Thread Andrea Rosa
I agree with Markus that this is not a bug. 
+1 on the discussion on the ML and if there will be any interest than please 
submit a blueprint and/or a spec to better discuss the proposal.
I am going to mark the bug as invalid.

** Changed in: nova
   Status: In Progress => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1604798

Title:
  Use DDT library to reduce code duplication

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  Use DDT library to reduce code duplication

  DDT can ease up error tracing and autogenerates tests on basis of different 
input data.
  It allows to multiply one test case by running it with different test data, 
and make it
  appear as multiple test cases. This will help to reduce code duplication.

  Please refer example use: 
  http://ddt.readthedocs.io/en/latest/example.html

  Currently DDT is implemented in openstack/cinder and openstack/rally.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1604798/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1588378] [NEW] Cancelled live migration are reported as in progress

2016-06-02 Thread Andrea Rosa
Public bug reported:

With the introduction of the API for aborting a running live-migration 
(https://review.openstack.org/277971) we have introduced a new status for the 
aborted live migration jobs. This new status called "cancelled" is not filtered 
by the sqlalchemy query used to return the list of migration in progress: 
https://github.com/openstack/nova/blob/87dc738763d6a7a10409e14b878f5cdd39ba805e/nova/db/sqlalchemy/api.py#L4851

** Affects: nova
 Importance: Low
 Assignee: Andrea Rosa (andrea-rosa-m)
 Status: In Progress


** Tags: libvirt live-migration

** Changed in: nova
 Assignee: (unassigned) => Andrea Rosa (andrea-rosa-m)

** Tags added: libvirt

** Tags added: live-migration

** Changed in: nova
   Status: New => In Progress

** Changed in: nova
   Importance: Undecided => Low

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1588378

Title:
  Cancelled live migration are reported as in progress

Status in OpenStack Compute (nova):
  In Progress

Bug description:
  With the introduction of the API for aborting a running live-migration 
(https://review.openstack.org/277971) we have introduced a new status for the 
aborted live migration jobs. This new status called "cancelled" is not filtered 
by the sqlalchemy query used to return the list of migration in progress: 
  
https://github.com/openstack/nova/blob/87dc738763d6a7a10409e14b878f5cdd39ba805e/nova/db/sqlalchemy/api.py#L4851

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1588378/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1582409] Re: nova-api performance issues on "openstack server list"

2016-05-17 Thread Andrea Rosa
based on your last comment I am going to close this bug as it is not 
"reproducible".
Please feel free to re-open it if you have a clear way to reproduce it again.
Thanks


** Changed in: nova
   Status: New => Incomplete

** Changed in: nova
   Status: Incomplete => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1582409

Title:
  nova-api performance issues on "openstack server list"

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  Relevant information TL;DR:
  RDO stable Mitaka deployed with OpenStack puppet modules.
  Keystone with fernet tokens, Nova with keystone_authtoken memcached caching 
set (also reproduced with sql tokens and nova without memcached caching for 
authtoken). nova-api is handled behind apache with mod wsgi.

  Problem TL;DR:
  In an empty new deployment on bare metal (no images, no virtual machines, 
etc.), "openstack server list" will intermittently take >5 seconds (up to ~10) 
to return. The only noticeable thing that happens when "openstack server list" 
is slow is that the logs will be plastered with a bunch of 
"_create_extension_schema", "_rebuild_extension_schema", "_register_resources" 
and so on.
  All other installed projects (glance, keystone, neutron) return command 
output consistently under 2 seconds.

  Version info:
  # rpm -qa |grep nova
  openstack-nova-api-13.0.0-1.el7.noarch
  python-novaclient-3.3.0-1.el7.noarch
  openstack-nova-scheduler-13.0.0-1.el7.noarch
  python-nova-13.0.0-1.el7.noarch
  openstack-nova-conductor-13.0.0-1.el7.noarch
  openstack-nova-common-13.0.0-1.el7.noarch

  /etc/nova/nova.conf:
  http://paste.fedoraproject.org/367251/34295211/raw/

  Example log of when things are fast: 
http://paste.fedoraproject.org/367272/32783146/raw/
  Example log of when things are slow: 
http://paste.fedoraproject.org/367271/34327341/raw/

  Happy to help provider further details if required.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1582409/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1582558] Re: Live-migration exception handling need improvement

2016-05-17 Thread Andrea Rosa
I am a bit skeptical in ignoring the exception in both cases you described. For 
example during the roll_back if we fail calling cinder the risk is that we 
leave some open connections between the host and the cinder server, which is 
not good.
The same thing could happen during the _post_live_migration.
I agree with you that would be good to not stop those two actions for errors 
which Nova can't control but ignoring those errors doesn't seem the right 
approach.
Probably we need a new design for implementing the rollback and the 
post_live_migration tasks in a more async way but with the ability to control 
the results and take correct actions to fix potential errors but I reckon that 
to do that it is a complicated task which, probably, require some help from 
other services (Cidner, for example). I am seeing this more than a new spec 
than a bug, I'd suggest to come up with a more structured proposal and put a 
spec where we can get a discussion involving a broader audience.

I am going to mark this bug as invalid as we do not have a real bug here.
if you are not happy about this decision please join the IRC meeting for the 
nova live migration sub-team:
https://wiki.openstack.org/wiki/Meetings/NovaLiveMigration



** Changed in: nova
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1582558

Title:
  Live-migration exception handling need improvement

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  Description:
  1.In /nova/compute/manager.py the function _rollback_live_migration, we 
should be catch exception of 
  'remove_volume_connection', let the rollback process continue to clean up 
other resources. Becaue,the
  'remove_volume_connection' will visit cinder, the probability of an exception 
is relatively large.
  2.In /nova/compute/manager.py the function _post_live_migration, we should be 
catch all exception of 
  source host clean up resources. Because the vm has been migrated the dest 
host, we should try to make sure the vm run normally.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1582558/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1581977] Re: Invalid input for dns_name when spawning instance with .number at the end

2016-05-16 Thread Andrea Rosa
I am not sure if this is a bug, we could replace the dot(s) in the host name 
with a different char, for example a "_" but then the user will lose the option 
to define a hostname as a FQDN.
And according to the rfc952:
' A "name" (Net, Host, Gateway, or Domain name) is a text string up
   to 24 characters drawn from the alphabet (A-Z), digits (0-9), minus
   sign (-), and period (.).  Note that periods are only allowed when
   they serve to delimit components of "domain style names"'
So if the user uses a period it should know that it is allowed only to delimit 
a domain name and it has to be a valid one.
I am going to mark it as invalid, please let me know if you are not happy about 
this decision.

** Changed in: nova
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1581977

Title:
  Invalid input for dns_name when spawning instance with .number at the
  end

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  When attempting to deploy an instance with a name which ends in dot
   (e.g. .123, as in an all-numeric TLD) or simply a name that,
  after conversion to dns_name, ends as ., nova conductor fails
  with the following error:

  2016-05-15 13:15:04.824 ERROR nova.scheduler.utils [req-4ce865cd-e75b-
  4de8-889a-ed7fc7fece18 admin demo] [instance:
  c4333432-f0f8-4413-82e8-7f12cdf3b5c8] Error from last host:
  silpixa00394065 (node silpixa00394065): [u'Traceback (most recent call
  last):\n', u'  File "/opt/stack/nova/nova/compute/manager.py", line
  1926, in _do_build_and_run_instance\nfilter_properties)\n', u'
  File "/opt/stack/nova/nova/compute/manager.py", line 2116, in
  _build_and_run_instance\ninstance_uuid=instance.uuid,
  reason=six.text_type(e))\n', u"RescheduledException: Build of instance
  c4333432-f0f8-4413-82e8-7f12cdf3b5c8 was re-scheduled: Invalid input
  for dns_name. Reason: 'networking-ovn-ubuntu-16.04' not a valid PQDN
  or FQDN. Reason: TLD '04' must not be all numeric.\nNeutron server
  returns request_ids: ['req-7317c3e3-2875-4073-8076-40e944845b69']\n"]

  This throws one instance of the infamous Horizon message: Error: No
  valid host was found. There are not enough hosts available.

  
  This issue was observed using stable/mitaka via DevStack (nova commit 
fb3f1706c68ea5b58f05ea810c6339f2449959de).

  In the above example, the instance name is "networking-ovn (Ubuntu
  16.04)", which resulted in an attempted dns_name="networking-ovn-
  ubuntu-16.04", where the 04 was interpreted as a TLD and,
  consequently, an invalid TLD.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1581977/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1550525] Re: Abort an ongoing live migration

2016-03-08 Thread Andrea Rosa
I added details Ito describe the new feature introduced in nova by the 2.24 
version, please let me know more information are required.
I think that the new feature needs to be documented in the API guide.

** Project changed: nova => openstack-api-site

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1550525

Title:
  Abort an ongoing live migration

Status in openstack-api-site:
  Confirmed

Bug description:
  https://review.openstack.org/277971
  Dear bug triager. This bug was created since a commit was marked with 
DOCIMPACT.
  Your project "openstack/nova" is set up so that we directly report the 
documentation bugs against it. If this needs changing, the docimpact-group 
option needs to be added for the project. You can ask the OpenStack infra team 
(#openstack-infra on freenode) for help if you need to.

  commit fa002925460e70d988d1b4dd1ea594c680a43740
  Author: Andrea Rosa <andrea.r...@hp.com>
  Date:   Fri Feb 5 08:31:06 2016 +

  Abort an ongoing live migration
  
  This change adds a DELETE call on the server-migrations object to cancel
  a running live migration of a specific instance.
  TO perform the cancellation the virtualization driver needs to support
  it, in case that the feature is not supported we return an error.
  We allow a cancellation of a migration only if the migration is
  running at the moment of the request and if the migration type is equal
  to 'live-migration'.
  In this change we implement this feature for the libvirt driver.
  When the cancellation of a live migration succeeded we rollback the live
  migration and we set the state of the Migration object equals to
  'cancelled'.
  The implementation of this change is based on the work done by the
  implementation of the feature called 'force live migration':
  https://review.openstack.org/245921
  
  DocImpact
  ApiImpact
  
  Implements blueprint: abort-live-migration
  Change-Id: I1ff861e54997a069894b542bd764ac3ef1b3dbb2

To manage notifications about this bug go to:
https://bugs.launchpad.net/openstack-api-site/+bug/1550525/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1499405] Re: Live block migration fails for instances with LVM ephemeral devices

2015-10-05 Thread Andrea Rosa
I had time to did more investigation and I found that the live migration for 
LVM backed instances is not supported:
https://github.com/openstack/nova/blob/master/nova/virt/libvirt/driver.py#L6665

This change https://review.openstack.org/73387 was for raising the 
NotImplementedError exception when we try LVM migration.
In my case I didn't see that exception probably because I hit a different 
exception before, we still need to check that the control about LVM is done at 
the right time, but that will be a different bug.

I am going to mark  this bug as Invalid as it is a duplicate of
https://bugs.launchpad.net/nova/+bug/1282643

** Changed in: nova
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1499405

Title:
  Live block migration fails for instances with LVM ephemeral devices

Status in OpenStack Compute (nova):
  Invalid

Bug description:
  On a multinode devstack installation I have created a new lvm device:
  sudo dd if=/dev/zero of=/loopy bs=1M count=5124 # 5gb
  sudo losetup /dev/loop2 /loopy

  sudo sfdisk /dev/loop2 << EOF
  ,,8e,,
  EOF
  sudo pvcreate /dev/loop2
  sudo vgcreate encrypted-ephemeral /dev/loop2
   
  Please note that the volume group name is "ecnrypted-ephemeral" but I am not 
using any encryption.
  I set up my nova compute to use lvm images for the ephemeral disks:
  [libvirt]
  images_type = lvm
  images_volume_group = encrypted-ephemeral

  Then I changed the live migration uri to use tcp instead of ssh:
  live_migration_uri = qemu+tcp://%s/system

  I did the same configuration on a second compute node.

  At this point I booted a vm which became ACTIVE with no error then I tried to 
migrate the instance:
  nova live-migration  devstackmn-2 --block-migrate

  and the migration fails and in the libivirt log file there is this error 
reported:
  2015-09-24 08:23:17.674+: 26629: error : 
virNetClientProgramDispatchError:175 : Failed to open file 
'/dev/encrypted-ephemeral/9c1c82cb-cf11-4b2b-a994-2d94d286bfb9_disk': No such 
file or directory

  and in the nova-compute log file of the source node I see a similar
  error:

  2015-09-24 14:33:19.023 ERROR nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] Live Migration failure: Failed to open 
file '/dev/encrypted-ephemeral/9c1c82cb-cf11-4b2b-a994-2d94d286bfb9_disk': No 
such file or directory
  2015-09-24 14:33:19.023 DEBUG nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] Migration operation thread notification 
from (pid=21098) thread_finished 
/opt/stack/nova/nova/virt/libvirt/driver.py:6041
  Traceback (most recent call last):
File "/usr/local/lib/python2.7/dist-packages/eventlet/hubs/hub.py", line 
457, in fire_timers
  timer()
File "/usr/local/lib/python2.7/dist-packages/eventlet/hubs/timer.py", line 
58, in __call__
  cb(*args, **kw)
File "/usr/local/lib/python2.7/dist-packages/eventlet/event.py", line 168, 
in _do_send
  waiter.switch(result)
File "/usr/local/lib/python2.7/dist-packages/eventlet/greenthread.py", line 
214, in main
  result = function(*args, **kwargs)
File "/opt/stack/nova/nova/utils.py", line 1154, in context_wrapper
  return func(*args, **kwargs)
File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 5643, in 
_live_migration_operation
  instance=instance)
File "/usr/local/lib/python2.7/dist-packages/oslo_utils/excutils.py", line 
195, in __exit__
  six.reraise(self.type_, self.value, self.tb)
File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 5611, in 
_live_migration_operation
  CONF.libvirt.live_migration_bandwidth)
File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 183, 
in doit
  result = proxy_call(self._autowrap, f, *args, **kwargs)
File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 141, 
in proxy_call
  rv = execute(f, *args, **kwargs)
File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 122, 
in execute
  six.reraise(c, e, tb)
File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 80, 
in tworker
  rv = meth(*args, **kwargs)
File "/usr/local/lib/python2.7/dist-packages/libvirt.py", line 1586, in 
migrateToURI2
  if ret == -1: raise libvirtError ('virDomainMigrateToURI2() failed', 
dom=self)
  libvirtError: Failed to open file 
'/dev/encrypted-ephemeral/9c1c82cb-cf11-4b2b-a994-2d94d286bfb9_disk': No such 
file or directory
  2015-09-24 14:33:19.224 DEBUG nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] VM running on src, migration failed from 
(pid=21098) _live_migration_monitor 

[Yahoo-eng-team] [Bug 1499405] [NEW] Live-migration fails for instances with LVM ephemeral devices

2015-09-24 Thread Andrea Rosa
Public bug reported:

On a multinode devstack installation I have created a new lvm device:
sudo dd if=/dev/zero of=/loopy bs=1M count=5124 # 5gb
sudo losetup /dev/loop2 /loopy

sudo sfdisk /dev/loop2 << EOF
,,8e,,
EOF
sudo pvcreate /dev/loop2
sudo vgcreate encrypted-ephemeral /dev/loop2
 
Please note that the volume group name is "ecnrypted-ephemeral" but I am not 
using any encryption.
I set up my nova compute to use lvm images for the ephemeral disks:
[libvirt]
images_type = lvm
images_volume_group = encrypted-ephemeral

Then I changed the live migration uri to use tcp instead of ssh:
live_migration_uri = qemu+tcp://%s/system

I did the same configuration on a second compute node.

At this point I booted a vm which became ACTIVE with no error then I tried to 
migrate the instance:
nova live-migration  devstackmn-2 --block-migrate

and the migration fails and in the libivirt log file there is this error 
reported:
2015-09-24 08:23:17.674+: 26629: error : 
virNetClientProgramDispatchError:175 : Failed to open file 
'/dev/encrypted-ephemeral/9c1c82cb-cf11-4b2b-a994-2d94d286bfb9_disk': No such 
file or directory

and in the nova-compute log file of the source node I see a similar
error:

2015-09-24 14:33:19.023 ERROR nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] Live Migration failure: Failed to open 
file '/dev/encrypted-ephemeral/9c1c82cb-cf11-4b2b-a994-2d94d286bfb9_disk': No 
such file or directory
2015-09-24 14:33:19.023 DEBUG nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] Migration operation thread notification 
from (pid=21098) thread_finished 
/opt/stack/nova/nova/virt/libvirt/driver.py:6041
Traceback (most recent call last):
  File "/usr/local/lib/python2.7/dist-packages/eventlet/hubs/hub.py", line 457, 
in fire_timers
timer()
  File "/usr/local/lib/python2.7/dist-packages/eventlet/hubs/timer.py", line 
58, in __call__
cb(*args, **kw)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/event.py", line 168, in 
_do_send
waiter.switch(result)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/greenthread.py", line 
214, in main
result = function(*args, **kwargs)
  File "/opt/stack/nova/nova/utils.py", line 1154, in context_wrapper
return func(*args, **kwargs)
  File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 5643, in 
_live_migration_operation
instance=instance)
  File "/usr/local/lib/python2.7/dist-packages/oslo_utils/excutils.py", line 
195, in __exit__
six.reraise(self.type_, self.value, self.tb)
  File "/opt/stack/nova/nova/virt/libvirt/driver.py", line 5611, in 
_live_migration_operation
CONF.libvirt.live_migration_bandwidth)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 183, in 
doit
result = proxy_call(self._autowrap, f, *args, **kwargs)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 141, in 
proxy_call
rv = execute(f, *args, **kwargs)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 122, in 
execute
six.reraise(c, e, tb)
  File "/usr/local/lib/python2.7/dist-packages/eventlet/tpool.py", line 80, in 
tworker
rv = meth(*args, **kwargs)
  File "/usr/local/lib/python2.7/dist-packages/libvirt.py", line 1586, in 
migrateToURI2
if ret == -1: raise libvirtError ('virDomainMigrateToURI2() failed', 
dom=self)
libvirtError: Failed to open file 
'/dev/encrypted-ephemeral/9c1c82cb-cf11-4b2b-a994-2d94d286bfb9_disk': No such 
file or directory
2015-09-24 14:33:19.224 DEBUG nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] VM running on src, migration failed from 
(pid=21098) _live_migration_monitor 
/opt/stack/nova/nova/virt/libvirt/driver.py:5847
2015-09-24 14:33:19.224 DEBUG nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] Fixed incorrect job type to be 4 from 
(pid=21098) _live_migration_monitor 
/opt/stack/nova/nova/virt/libvirt/driver.py:5867
2015-09-24 14:33:19.224 ERROR nova.virt.libvirt.driver 
[req-3f358ce5-7746-4532-9bb4-bf1189181ce8 admin admin] [instance: 
9c1c82cb-cf11-4b2b-a994-2d94d286bfb9] Migration operation has aborted


Nova version as per git log

commit 806dc025dbc9a102f7bace78be9c56f8e9af0929
Merge: 424aaf3 ae3d970
Author: Jenkins 
Date:   Sat Sep 19 05:03:38 2015 +

Merge "libvirt:Remove duplicated check code for config option
sysinfo_serial"

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1499405

Title:
  Live-migration fails for instances with LVM ephemeral devices

Status in OpenStack 

[Yahoo-eng-team] [Bug 1475353] [NEW] _get_host_sysinfo_serial_os fails if the machine-id file is empty

2015-07-16 Thread Andrea Rosa
Public bug reported:

the _get_host_sysinfo_serial_os method try to read the machine-id file to get 
an UUID for the host operating system.
If the file is there but it is empty the code will raise an exception while it 
tries to parse the content of the file.

To reproduce the issue just add this test to the
nova/tests/unit/virt/libvirt/test_driver.py

def test_get_guest_config_sysinfo_serial_os_empty_machine_id(self):
self.flags(sysinfo_serial=os, group=libvirt)

real_open = __builtin__.open
with contextlib.nested(
mock.patch.object(__builtin__, open),
) as (mock_open, ):
theuuid = 

def fake_open(filename, *args, **kwargs):
if filename == /etc/machine-id:
h = mock.MagicMock()
h.read.return_value = theuuid
h.__enter__.return_value = h
return h
return real_open(filename, *args, **kwargs)

mock_open.side_effect = fake_open

self._test_get_guest_config_sysinfo_serial(None)

** Affects: nova
 Importance: Undecided
 Assignee: Andrea Rosa (andrea-rosa-m)
 Status: In Progress


** Tags: low-hanging-fruit

** Changed in: nova
 Assignee: (unassigned) = Andrea Rosa (andrea-rosa-m)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1475353

Title:
  _get_host_sysinfo_serial_os fails if the machine-id file is empty

Status in OpenStack Compute (nova):
  In Progress

Bug description:
  the _get_host_sysinfo_serial_os method try to read the machine-id file to get 
an UUID for the host operating system.
  If the file is there but it is empty the code will raise an exception while 
it tries to parse the content of the file.

  To reproduce the issue just add this test to the
  nova/tests/unit/virt/libvirt/test_driver.py

  def test_get_guest_config_sysinfo_serial_os_empty_machine_id(self):
  self.flags(sysinfo_serial=os, group=libvirt)

  real_open = __builtin__.open
  with contextlib.nested(
  mock.patch.object(__builtin__, open),
  ) as (mock_open, ):
  theuuid = 

  def fake_open(filename, *args, **kwargs):
  if filename == /etc/machine-id:
  h = mock.MagicMock()
  h.read.return_value = theuuid
  h.__enter__.return_value = h
  return h
  return real_open(filename, *args, **kwargs)

  mock_open.side_effect = fake_open

  self._test_get_guest_config_sysinfo_serial(None)

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1475353/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1375857] [NEW] It's not possibile to pass the cacert to the swift store

2014-09-30 Thread Andrea Rosa
Public bug reported:

The swift store device defined in the glance store doesn't allow to pass the ca 
cert file. When the driver creates a connection via the swift client it is not 
possible to pass that value. 
That means that if we have swift running on TLS in some cases we have to set 
the insecure option equals to True as the client can't correctly complete the 
handshake as it fails on the verification of the cert.

The fix I'd like to propose is to add a new parameter to define the ca
cert file and pass this value when we create the connection via the
swift client.

** Affects: glance
 Importance: Undecided
 Assignee: Andrea Rosa (andrea-rosa-m)
 Status: New


** Tags: store swift

** Changed in: glance
 Assignee: (unassigned) = Andrea Rosa (andrea-rosa-m)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1375857

Title:
  It's not possibile to pass the cacert to the swift store

Status in OpenStack Image Registry and Delivery Service (Glance):
  New

Bug description:
  The swift store device defined in the glance store doesn't allow to pass the 
ca cert file. When the driver creates a connection via the swift client it is 
not possible to pass that value. 
  That means that if we have swift running on TLS in some cases we have to set 
the insecure option equals to True as the client can't correctly complete the 
handshake as it fails on the verification of the cert.

  The fix I'd like to propose is to add a new parameter to define the ca
  cert file and pass this value when we create the connection via the
  swift client.

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1375857/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1192749] Re: Cannot create over 1000 VM due to previous VM deletion

2013-06-20 Thread Andrea Rosa
i think this is not a bug but just a configuration issue.
There is a parameter to check the max number of items returned, something like 
that:
osapi_opts = [
cfg.IntOpt('osapi_max_limit',
   default=1000,
   help='the maximum number of items returned in a single '
'response from a collection resource'),

Can you check in your configuration if you have this parameter ?

** Changed in: nova
   Status: New = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1192749

Title:
  Cannot create over 1000 VM due to previous VM deletion

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  The system currently has 1000 active VM.  When I create the 1001 VM
  the oldest VM 1 gets deleted without warning.  I'm not sure if it is
  still running on the compute/volume node but I cannot find it using
  'nova show stca-00044'.  I have checked the quota for enough resources
  but then if the quota was exceeded for any of the resources (i.e
  vcpu/memory/etc) there would have been an error.  In this case the new
  1001 VM becomes active and 1 VM is not longer accessible.   The system
  will always stay at 1000vm due to this and cannot be exceeded.

  This setup was configured using Cisco Openstack Edition Folsom-multinode.  
  Link to configuration documentation: 
http://docwiki.cisco.com/wiki/OpenStack:Folsom

  The setup contains 1 control/network node and 43 compute/volume nodes.

  
  Here is an example:

  root@control-server:~# nova show stca-0044
  
+-+--+
  | Property| Value 
   |
  
+-+--+
  | status  | ACTIVE
   |
  | updated | 2013-06-07T18:42:41Z  
   |
  | OS-EXT-STS:task_state   | None  
   |
  | OS-EXT-SRV-ATTR:host| compute-server19  
   |
  | net16 network   | 16.0.0.52 
   |
  | key_name| dkp   
   |
  | image   | f16_STCa_4.24 
(b978b4bc-044a-4ccc-a14c-bc4ed9f7cb34) |
  | hostId  | 
d63142d9402fb3f64934f6a5aacd8876f7df2869cecb9f653aa0acd0 |
  | net15 network   | 15.0.0.52 
   |
  | OS-EXT-STS:vm_state | active
   |
  | OS-EXT-SRV-ATTR:instance_name   | instance-0041 
   |
  | OS-EXT-SRV-ATTR:hypervisor_hostname | compute-server19.cisco.com
   |
  | flavor  | STCa-vm (6)   
   |
  | id  | 26825b9c-edd5-477e-acb3-c58bd1a2d775  
   |
  | security_groups | [{u'name': u'default'}]   
   |
  | user_id | 2562b65ac23840d483647f6deba74b0d  
   |
  | name| stca-0044 
   |
  | created | 2013-06-07T18:42:22Z  
   |
  | tenant_id   | 0f330a6acb4f43ac99961ba1573ced10  
   |
  | OS-DCF:diskConfig   | MANUAL
   |
  | accessIPv4  |   
   |
  | accessIPv6  |   
   |
  | progress| 0 
   |
  | OS-EXT-STS:power_state  | 1 
   |
  | metadata| {}
   |
  | config_drive|   
   |
  
+-+--+

  root@control-server:~# nova show stca-0044
  
+-+--+
  | Property| Value 
   |