[Yahoo-eng-team] [Bug 1437992] Re: policy file in policy.d will be reloaded every rest api call

2015-03-30 Thread Eli Qiao
olso.policy patch.
https://review.openstack.org/#/c/168798/

** Also affects: oslo.policy
   Importance: Undecided
   Status: New

** Changed in: oslo.policy
   Status: New = Confirmed

** Changed in: oslo.policy
 Assignee: (unassigned) = Eli Qiao (taget-9)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1437992

Title:
  policy file in policy.d will be reloaded every rest api call

Status in OpenStack Compute (Nova):
  In Progress
Status in Oslo Policy:
  Confirmed

Bug description:
  the policy file in policy.d will be reloaded every time when do a rest
  api call.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1437992/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438050] [NEW] Remember used tenant

2015-03-30 Thread Miroslav Suchý
Public bug reported:

I have user, which have access to several tenants. E.g. qa, persistent, 
scratch. I can select my current tenant in scroll box on top of page of 
Horizon dashboard. It is kept under current session.
However when session timeouts and I have to re-login, then chosen tenants is 
reset to first one - in my case qa.
This is quite annoying and it would be nice if Horizon can remember the chosen 
tenant and return to tenant I used before session timeouted or before I logoff.

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438050

Title:
  Remember used tenant

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  I have user, which have access to several tenants. E.g. qa, persistent, 
scratch. I can select my current tenant in scroll box on top of page of 
Horizon dashboard. It is kept under current session.
  However when session timeouts and I have to re-login, then chosen tenants is 
reset to first one - in my case qa.
  This is quite annoying and it would be nice if Horizon can remember the 
chosen tenant and return to tenant I used before session timeouted or before I 
logoff.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438050/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438040] [NEW] fdb entries can't be removed when a VM is migrated

2015-03-30 Thread shihanzhang
Public bug reported:

this problem can be reprodeced as bellow:
1. vm A in computeA, vm B in computeB, l2 pop enable;
2. vmB continue ping vmA 
3. live migrate vmA to computeB 
4. when live-migrate finish, vmB ping vmA failed

the reason is bellow, in l2pop driver, when vmA migrate to computeB, port 
status change form BUILD to ACTIVE,
it add the port to  self.migrated_ports when port status is ACTIVE, but 
'remove_fdb_entries' in port status is BUILD :
def update_port_postcommit(self, context):
...
...
elif (context.host != context.original_host
and context.status == const.PORT_STATUS_ACTIVE
and not self.migrated_ports.get(orig['id'])):
# The port has been migrated. We have to store the original
# binding to send appropriate fdb once the port will be set
# on the destination host
self.migrated_ports[orig['id']] = (
(orig, context.original_host))
elif context.status != context.original_status:
if context.status == const.PORT_STATUS_ACTIVE:
self._update_port_up(context)
elif context.status == const.PORT_STATUS_DOWN:
fdb_entries = self._update_port_down(
context, port, context.host)
self.L2populationAgentNotify.remove_fdb_entries(
self.rpc_ctx, fdb_entries)
elif context.status == const.PORT_STATUS_BUILD:
orig = self.migrated_ports.pop(port['id'], None)
if orig:
original_port = orig[0]
original_host = orig[1]
# this port has been migrated: remove its entries from fdb
fdb_entries = self._update_port_down(
context, original_port, original_host)
self.L2populationAgentNotify.remove_fdb_entries(
self.rpc_ctx, fdb_entries)

** Affects: neutron
 Importance: Undecided
 Assignee: shihanzhang (shihanzhang)
 Status: New

** Changed in: neutron
 Assignee: (unassigned) = shihanzhang (shihanzhang)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438040

Title:
  fdb entries can't be removed when a VM is migrated

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  this problem can be reprodeced as bellow:
  1. vm A in computeA, vm B in computeB, l2 pop enable;
  2. vmB continue ping vmA 
  3. live migrate vmA to computeB 
  4. when live-migrate finish, vmB ping vmA failed

  the reason is bellow, in l2pop driver, when vmA migrate to computeB, port 
status change form BUILD to ACTIVE,
  it add the port to  self.migrated_ports when port status is ACTIVE, but 
'remove_fdb_entries' in port status is BUILD :
  def update_port_postcommit(self, context):
  ...
  ...
  elif (context.host != context.original_host
  and context.status == const.PORT_STATUS_ACTIVE
  and not self.migrated_ports.get(orig['id'])):
  # The port has been migrated. We have to store the original
  # binding to send appropriate fdb once the port will be set
  # on the destination host
  self.migrated_ports[orig['id']] = (
  (orig, context.original_host))
  elif context.status != context.original_status:
  if context.status == const.PORT_STATUS_ACTIVE:
  self._update_port_up(context)
  elif context.status == const.PORT_STATUS_DOWN:
  fdb_entries = self._update_port_down(
  context, port, context.host)
  self.L2populationAgentNotify.remove_fdb_entries(
  self.rpc_ctx, fdb_entries)
  elif context.status == const.PORT_STATUS_BUILD:
  orig = self.migrated_ports.pop(port['id'], None)
  if orig:
  original_port = orig[0]
  original_host = orig[1]
  # this port has been migrated: remove its entries from fdb
  fdb_entries = self._update_port_down(
  context, original_port, original_host)
  self.L2populationAgentNotify.remove_fdb_entries(
  self.rpc_ctx, fdb_entries)

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438040/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438113] [NEW] Use plain HTTP listeners in the conductor

2015-03-30 Thread Attila Fazekas
Public bug reported:

The conductor is consuming messages form single queue which has performance 
limitation due to various reasons.:
- per queue lock
- Some broker also limiting same part of the message handling to single CPU 
thread/queue
- Multiple broker instances needs to synchronise to queue content, which causes 
additional delays die to the tcp request/response times

The single queue limitation is much greater than the limits getting by
single mysql server, the rate is even worse when you consider slave
reads.

This can be workarounded by explicitly or implicit distributing the rpc
calls to multiple different queue.

The message broker provides additional message durability properties which is 
not needed just for an rpc_call,
we spend resource on what we actually do not need.

For TCP/HTTP traffic load balancing we have many-many tools even hardware 
assisted options are available providing virtually unlimited scalability.
At TCP level also possible to exclude the loadbalancer node(s) form the 
response traffic.

Why HTTP?
Basically any protocol which can do request/response `thing` with arbitrary  
type and size of data with keep-alive connection and with ssl option, could be 
used.
HTTP is a simple and well know protocol, with already existing many-many load 
balancing tool.

Why not have the agents to do a regular API call?
The regular API calls needs to do policy check, which in this case is not 
required, every authenticated user can be considered as admin.  

The  the conductor clients needs to use at least a single shared key configured 
in every nova host.
It has similar security as openstack used with the brokers, basically all nova 
node had credentials in one rabbitmq virtual host,
configured in the /etc/nova/nova.conf . If any of those credentials stolen it 
provided access to the whole virtual host. 

NOTE.: HTTPs can be used with certificate or kerberos based
authentication as well.


I think the for `rpc_calls` which are served by the agents using AMQP is still 
better option,  this bug is just about the situation when the conductor itself 
serves  rpc_call(s). 

NOTE.: The 1 Million msq/sec rabbitmq benchmark is done 186 queues, in
way which does not hits the single queue limitations.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438113

Title:
  Use plain HTTP listeners in the conductor

Status in OpenStack Compute (Nova):
  New

Bug description:
  The conductor is consuming messages form single queue which has performance 
limitation due to various reasons.:
  - per queue lock
  - Some broker also limiting same part of the message handling to single CPU 
thread/queue
  - Multiple broker instances needs to synchronise to queue content, which 
causes additional delays die to the tcp request/response times

  The single queue limitation is much greater than the limits getting by
  single mysql server, the rate is even worse when you consider slave
  reads.

  This can be workarounded by explicitly or implicit distributing the
  rpc calls to multiple different queue.

  The message broker provides additional message durability properties which is 
not needed just for an rpc_call,
  we spend resource on what we actually do not need.

  For TCP/HTTP traffic load balancing we have many-many tools even hardware 
assisted options are available providing virtually unlimited scalability.
  At TCP level also possible to exclude the loadbalancer node(s) form the 
response traffic.

  Why HTTP?
  Basically any protocol which can do request/response `thing` with arbitrary  
type and size of data with keep-alive connection and with ssl option, could be 
used.
  HTTP is a simple and well know protocol, with already existing many-many load 
balancing tool.

  Why not have the agents to do a regular API call?
  The regular API calls needs to do policy check, which in this case is not 
required, every authenticated user can be considered as admin.  

  The  the conductor clients needs to use at least a single shared key 
configured in every nova host.
  It has similar security as openstack used with the brokers, basically all 
nova node had credentials in one rabbitmq virtual host,
  configured in the /etc/nova/nova.conf . If any of those credentials stolen it 
provided access to the whole virtual host. 

  NOTE.: HTTPs can be used with certificate or kerberos based
  authentication as well.

  
  I think the for `rpc_calls` which are served by the agents using AMQP is 
still better option,  this bug is just about the situation when the conductor 
itself serves  rpc_call(s). 

  NOTE.: The 1 Million msq/sec rabbitmq benchmark is done 186 queues, in
  way which does not hits the single queue limitations.

To manage notifications about this bug go to:

[Yahoo-eng-team] [Bug 1190559] Re: qcow2 images size should match the virtual image size

2015-03-30 Thread Flavio Percoco
@yatin Hey, I think you might be confused with what this bug aimed to
do. The bug proposed that the `size` field should match the
`virtual_size` instead of the `size`. What we did to fix this was adding
a new `virtual_size` field that would represent the `virtual_size` value
in some of the images.

I don't think nova uses the `virtual_size` field but I should double
check. That said, the `virtual_size` value is an independent field that
has nothing to do with the `size` of the image. Therefore, I do not
believe these 2 fields should match. If anything, the virtual_size
should be as big or bigger than the image size.

** Changed in: glance
   Status: Confirmed = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1190559

Title:
  qcow2 images size should match the virtual image size

Status in OpenStack Image Registry and Delivery Service (Glance):
  Invalid

Bug description:
  the image size of qcow2 images should match the virtual size reported
  by qemu-img not the actual data size

  # glance image-show ff8f8fb3-76a3-47c0-bfc2-79e46950
  [...]
  | size | 699592704   |
  [...]

  # qemu-img info /var/lib/glance/images/ff8f8fb3-76a3-47c0-bfc2-79e46950 
  image: /var/lib/glance/images/ff8f8fb3-76a3-47c0-bfc2-79e46950
  file format: qcow2
  virtual size: 6.0G (6442450944 bytes)
  disk size: 667M
  cluster_size: 65536

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1190559/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438063] [NEW] Remove repeated '__init__' from api/fwaas.py

2015-03-30 Thread Rob Cresswell
Public bug reported:

https://github.com/openstack/horizon/blob/master/openstack_dashboard/api/fwaas.py#L52

Duplicated init function from what appears to be a bad merge, but
unsure.

** Affects: horizon
 Importance: Undecided
 Assignee: Rob Cresswell (robcresswell)
 Status: In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438063

Title:
  Remove repeated '__init__' from api/fwaas.py

Status in OpenStack Dashboard (Horizon):
  In Progress

Bug description:
  
https://github.com/openstack/horizon/blob/master/openstack_dashboard/api/fwaas.py#L52

  Duplicated init function from what appears to be a bad merge, but
  unsure.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438063/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1437992] Re: policy file in policy.d will be reloaded every rest api call

2015-03-30 Thread Eli Qiao
** Also affects: oslo-incubator
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1437992

Title:
  policy file in policy.d will be reloaded every rest api call

Status in OpenStack Compute (Nova):
  In Progress
Status in The Oslo library incubator:
  New
Status in Oslo Policy:
  Confirmed

Bug description:
  the policy file in policy.d will be reloaded every time when do a rest
  api call.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1437992/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438088] [NEW] vpn row is not updating when the status changes

2015-03-30 Thread Masco Kaliyamoorthy
Public bug reported:

when the status changes the corresponding row is not updating in vpn table.
needs manual page reload.

** Affects: horizon
 Importance: Undecided
 Assignee: Masco Kaliyamoorthy (masco)
 Status: New

** Changed in: horizon
 Assignee: (unassigned) = Masco Kaliyamoorthy (masco)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438088

Title:
  vpn row is not updating when the status changes

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  when the status changes the corresponding row is not updating in vpn table.
  needs manual page reload.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438088/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438093] [NEW] Redundant method _set_vm_state in conductor _live_migrate

2015-03-30 Thread Bartosz Fic
Public bug reported:

In this file: 
https://github.com/openstack/nova/blob/master/nova/conductor/manager.py#L600
Implementing _set_vm_state() method in _live_migrate() method is unnecessary.

It could be used method from line #590:

def _set_vm_state_and_notify(self, context, instance_uuid, method, updates, ex, 
request_spec):
scheduler_utils.set_vm_state_and_notify(
context, instance_uuid, 'compute_task', 
 method, updates,ex, request_spec, self.db)

** Affects: nova
 Importance: Undecided
 Status: New


** Tags: low-hanging-fruit

** Tags added: low-hanging-fruit

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438093

Title:
  Redundant method _set_vm_state in conductor _live_migrate

Status in OpenStack Compute (Nova):
  New

Bug description:
  In this file: 
https://github.com/openstack/nova/blob/master/nova/conductor/manager.py#L600
  Implementing _set_vm_state() method in _live_migrate() method is unnecessary.

  It could be used method from line #590:

  def _set_vm_state_and_notify(self, context, instance_uuid, method, updates, 
ex, request_spec):
  scheduler_utils.set_vm_state_and_notify(
  context, instance_uuid, 'compute_task', 
   method, updates,ex, request_spec, self.db)

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438093/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260617] Re: Provide the ability to attach volumes in the read-only mode

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260617

Title:
  Provide the ability to attach volumes in the read-only mode

Status in OpenStack Dashboard (Horizon):
  Confirmed
Status in OpenStack Compute (Nova):
  Opinion
Status in Python client library for Nova:
  Triaged

Bug description:
  Cinder now support the ability to attach volumes in the read-only
  mode, this should be exposed through horizon. Read-only mode could be
  ensured by hypervisor configuration during the attachment. Libvirt,
  Xen, VMware and Hyper-V support R/O volumes.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1260617/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1227653] Re: xenapi: snapshot_attached_here should wait for vhd coalesce after deleting the snapshot

2015-03-30 Thread Sean Dague
Marking as Opinion as it is a bug by a core contributor without much
context.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1227653

Title:
  xenapi: snapshot_attached_here should wait for vhd coalesce after
  deleting the snapshot

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  When we take a snapshot, upload it, then delete the snapshot, we
  should really wait to make sure the VHD chain gets restored to its
  original state.

  If the VHD chain did not collapse:
  * snapshots contain more and more VHDs over time
  * longer chains mean longer possible max size of the chain

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1227653/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1276154] Re: suspending a paused instance

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1276154

Title:
  suspending a paused instance

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Is there a compelling reason why we don't support suspending a paused
  instance? At the moment we only allow 'active' and 'rescued' states to
  be suspended.

  In compute/api:

  @check_instance_state(vm_state=[vm_states.ACTIVE, vm_states.RESCUED])

  Trying to suspend a paused instance results in:

  nova suspend b10f8175-1663-41b2-8533-0b0606f369ff
  ERROR: Cannot 'suspend' while instance is in vm_state paused (HTTP 409) 
(Request-ID: req-7349d554-ff48-4155-a62b-967f0813c59c)

  I haven't tested with other hypervisors, but as far as libvirt/kvm is
  concerned it suspends (virsh managedsave) both states ('running',
  'paused') and resumes (virsh start) a suspended instance to the pre-
  suspended state.

  A use-case for this is to be able to quickly pause a bunch of
  instances for security reasons and then suspend one of them
  conveniently using the API rather than virsh. Taking a suspicious vm
  to ‘active’ state so that it can be suspended doesn’t seem like a good
  idea.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1276154/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438183] [NEW] Graceful shutdown of nova-compute service fails

2015-03-30 Thread Roman Podoliaka
Public bug reported:

nova-compute doesn't shutdown gracefully on SIGTERM, e.g. booting a VM
fails with:

09:29:18 AUDIT nova.compute.manager [req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 
None] [instance: 7ea3e761-6b85-49db-8dcc-79f6f2286
df8] Starting instance...
09:29:18 INFO nova.openstack.common.service [-] Caught SIGTERM, exiting
...
09:29:37 INFO nova.compute.manager [-] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VM Started (Lifecycle Event)
09:29:37 INFO nova.compute.manager [-] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VM Paused (Lifecycle Event)
...
09:34:37 WARNING nova.virt.libvirt.driver 
[req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 None] Timeout waiting for vif 
plugging callback for instance 7ea3e761-6b85-49db-8dcc-79f6f2286df8
09:34:37 INFO nova.compute.manager [-] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VM Stopped (Lifecycle Event)
09:34:38 INFO nova.virt.libvirt.driver 
[req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 None] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] Deleting instance files 
/var/lib/nova/instances/7ea3e761-6b85-49db-8dcc-79f6f2286df8
09:34:38 ERROR nova.compute.manager [req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 
None] [instance: 7ea3e761-6b85-49db-8dcc-79f6f2286df8] Instance failed to spawn
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] Traceback (most recent call last):
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] File 
/usr/lib/python2.7/dist-packages/nova/compute/manager.py, line 1773, in _spawn
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] block_device_info)
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] File 
/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py, line 2299, in 
spawn
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] block_device_info)
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] File 
/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py, line 3745, in 
_create_domain_and_network
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] raise 
exception.VirtualInterfaceCreateException()
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VirtualInterfaceCreateException: Virtual 
Interface creation failed
09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8]

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438183

Title:
  Graceful shutdown of nova-compute service fails

Status in OpenStack Compute (Nova):
  New

Bug description:
  nova-compute doesn't shutdown gracefully on SIGTERM, e.g. booting a VM
  fails with:

  09:29:18 AUDIT nova.compute.manager [req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 
None] [instance: 7ea3e761-6b85-49db-8dcc-79f6f2286
  df8] Starting instance...
  09:29:18 INFO nova.openstack.common.service [-] Caught SIGTERM, exiting
  ...
  09:29:37 INFO nova.compute.manager [-] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VM Started (Lifecycle Event)
  09:29:37 INFO nova.compute.manager [-] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VM Paused (Lifecycle Event)
  ...
  09:34:37 WARNING nova.virt.libvirt.driver 
[req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 None] Timeout waiting for vif 
plugging callback for instance 7ea3e761-6b85-49db-8dcc-79f6f2286df8
  09:34:37 INFO nova.compute.manager [-] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] VM Stopped (Lifecycle Event)
  09:34:38 INFO nova.virt.libvirt.driver 
[req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 None] [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] Deleting instance files 
/var/lib/nova/instances/7ea3e761-6b85-49db-8dcc-79f6f2286df8
  09:34:38 ERROR nova.compute.manager [req-9cdbba9c-af3b-4845-9deb-c68bffe63d75 
None] [instance: 7ea3e761-6b85-49db-8dcc-79f6f2286df8] Instance failed to spawn
  09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] Traceback (most recent call last):
  09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] File 
/usr/lib/python2.7/dist-packages/nova/compute/manager.py, line 1773, in _spawn
  09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] block_device_info)
  09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] File 
/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py, line 2299, in 
spawn
  09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] block_device_info)
  09:34:38 TRACE nova.compute.manager [instance: 
7ea3e761-6b85-49db-8dcc-79f6f2286df8] File 
/usr/lib/python2.7/dist-packages/nova/virt/libvirt/driver.py, line 3745, 

[Yahoo-eng-team] [Bug 1206877] Re: xenapi: iso config drive not visible on windows 8 or 2012

2015-03-30 Thread Sean Dague
@john has said it was a feature, marking as Opinion therefore

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1206877

Title:
  xenapi: iso config drive not visible on windows  8 or 2012

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  This was reported for libvirt:
  https://bugs.launchpad.net/nova/+bug/1155842

  While Linux can mount a disk with an ISO format, windows wants the ISO
  added as a CD rom device.

  This should be fixed in the XenAPI driver.

  The latest cloud-init is able to read the config drive if its on
  Windows 2012 or later, and Windows 8 or later.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1206877/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438138] [NEW] [Heat] Initial values aren't displaying properly in Launch Stack form

2015-03-30 Thread Tatiana Ovchinnikova
Public bug reported:

Using json type default data in a template causes annoying
inconsistencies in the second step of creating a stack. For example for
this template:


heat_template_version: 2013-05-23
parameters:
  key:
type: string
default: key
description: This is keypair to access into your stack.
  db_password:
type: string
default: secrete
hidden: true
  metadata:
type: json
default: {keyvalue: 5}


on the second step there would be metadata field with initial value 
{u'keyvalue': 5}. That is not valid JSON, and pressing Launch button would 
show: Expecting property name enclosed in double quotes.

** Affects: horizon
 Importance: Medium
 Assignee: Tatiana Ovchinnikova (tmazur)
 Status: New

** Changed in: horizon
 Assignee: (unassigned) = Tatiana Ovchinnikova (tmazur)

** Changed in: horizon
   Importance: Undecided = Medium

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438138

Title:
  [Heat] Initial values aren't displaying properly in Launch Stack
  form

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  Using json type default data in a template causes annoying
  inconsistencies in the second step of creating a stack. For example
  for this template:

  
  heat_template_version: 2013-05-23
  parameters:
key:
  type: string
  default: key
  description: This is keypair to access into your stack.
db_password:
  type: string
  default: secrete
  hidden: true
metadata:
  type: json
  default: {keyvalue: 5}

  
  on the second step there would be metadata field with initial value 
{u'keyvalue': 5}. That is not valid JSON, and pressing Launch button would 
show: Expecting property name enclosed in double quotes.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438138/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1437141] Re: nova floating-ip-delete should warn or refuse if an IP is still associated

2015-03-30 Thread Sean Dague
This is actually working as designed. Manually assigned floating IPs can
be deleted in a single go, auto assigned ones must be disassociated
first. Changing this default would probably *break* lots of users, so I
don't think we should.

** Changed in: nova
   Status: New = Opinion

** Changed in: nova
   Importance: Undecided = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1437141

Title:
  nova floating-ip-delete should warn or refuse if an IP is still
  associated

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Running 'nova floating-ip-delete IP' will happily delete IP
  addresses that are associated with currently-running instances.

  It would be safer if the user were required to either disassociate
  them from the instance before being permitted to delete them, or to
  pass a --force flag or similar.

  Version installed is python-novaclient 1:2.17.0-0ubuntu1~cloud0

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1437141/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438159] [NEW] Made neutron agents silent by using AMQP

2015-03-30 Thread Attila Fazekas
Public bug reported:

Problem.: Neutron agents does a lot of periodic task which leads  an rpc call + 
database transaction, which does not even provide a new information, because 
nothing changed.
This behaviour in scale can be called as `DDOS attack`, generally this kind of 
architecture is bad at scaling and can be considered as an any-pattern.

Instead of periodic poll, we can leverage the AMQP brokers bind capabilities. 
Neutron has many situation like security group rule change or dvr related 
changes which needs to be communicated with multiple agents, but usually not 
with all agent.

The agent at startup needs to synchronise the as usual, but during the
sync the agent can subscribe to the interesting events to avoid the
periodic tasks. (Note.: After the first subscribe loop a second one is
needed to do not miss changes during the subscribe process ).

The AMQP queues with 'auto-delete' can be considered as a reliable source of 
information which does not miss any event notification. 
On connection loss or full broker cluster die the agent needs to re sync 
everything guarded in this way,
in these cases, the queue will disappear so the situation easily detectable.

1. Create a Direct exchange for all kind of resourcestype what needs
to be synchronised in this way, for example.: 'neutron.securitygroups' .
The exchange declaration needs to happen at q-svc start-up time or at
full broker cluster die (not-found exception will tell it). The exchange
SHOULD NOT be redeclared or verified at every message publish.

2. Every agent creates a dedicated per agent queue with auto-delete flag, if 
the agent already maintains a queue with this property he MAY reuse that one. 
The agents SHOULD avoid to creating multiple queues per resource type. The 
messages MUST contain a type information. 
3. All agent creates a binding between his queue and the resource type queue 
with he realise he depends on the resource, for example it maintains at least 
one port with the given security-group. (The agents needs to remove the 
binding. when they stop using it)
4. The q-svc publishes just a single message  when the resource related change 
happened. The routing key is the uuid.

Alternatively a topic exchange can be used, with a single  exchange.
In this case the routing keys MUST contains the resource type like: 
neutron.resource_type.uuid ,
this type exchange is generally more expensive than a direct exchange (pattern 
matching), and only useful if you have agents which needs to listens to ALL 
event related to a type, but others just interested just in a few of them.

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438159

Title:
  Made neutron agents silent by using AMQP

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  Problem.: Neutron agents does a lot of periodic task which leads  an rpc call 
+ database transaction, which does not even provide a new information, because 
nothing changed.
  This behaviour in scale can be called as `DDOS attack`, generally this kind 
of architecture is bad at scaling and can be considered as an any-pattern.

  Instead of periodic poll, we can leverage the AMQP brokers bind capabilities. 
  Neutron has many situation like security group rule change or dvr related 
changes which needs to be communicated with multiple agents, but usually not 
with all agent.

  The agent at startup needs to synchronise the as usual, but during the
  sync the agent can subscribe to the interesting events to avoid the
  periodic tasks. (Note.: After the first subscribe loop a second one is
  needed to do not miss changes during the subscribe process ).

  The AMQP queues with 'auto-delete' can be considered as a reliable source of 
information which does not miss any event notification. 
  On connection loss or full broker cluster die the agent needs to re sync 
everything guarded in this way,
  in these cases, the queue will disappear so the situation easily detectable.

  1. Create a Direct exchange for all kind of resourcestype what needs
  to be synchronised in this way, for example.: 'neutron.securitygroups'
  . The exchange declaration needs to happen at q-svc start-up time or
  at full broker cluster die (not-found exception will tell it). The
  exchange SHOULD NOT be redeclared or verified at every message
  publish.

  2. Every agent creates a dedicated per agent queue with auto-delete flag, if 
the agent already maintains a queue with this property he MAY reuse that one. 
The agents SHOULD avoid to creating multiple queues per resource type. The 
messages MUST contain a type information. 
  3. All agent creates a binding between his queue and the resource type queue 
with he realise he depends on the resource, for example it maintains at least 
one port with the given security-group. 

[Yahoo-eng-team] [Bug 1078080] Re: libvirt rescue doesn't respect image_meta passed in

2015-03-30 Thread Sean Dague
Looks fixed in code:

libvirt/driver.py

def rescue(self, context, instance, network_info, image_meta,
   rescue_password):
Loads a VM using rescue images.

A rescue is normally performed when something goes wrong with the
primary images and data needs to be corrected/recovered. Rescuing
should not edit or over-ride the original image, only allow for
data recovery.


instance_dir = libvirt_utils.get_instance_path(instance)
unrescue_xml = self._get_existing_domain_xml(instance, network_info)
unrescue_xml_path = os.path.join(instance_dir, 'unrescue.xml')
libvirt_utils.write_to_file(unrescue_xml_path, unrescue_xml)

if image_meta is not None:
rescue_image_id = image_meta.get('id')
else:
rescue_image_id = None

** Changed in: nova
   Status: Triaged = Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1078080

Title:
  libvirt rescue doesn't respect image_meta passed in

Status in OpenStack Compute (Nova):
  Fix Released

Bug description:
  The libvirt driver currently ignores the `image_meta` field passed in
  on the rescue method and instead uses it's driver-specific
  'rescue_image_id` flag or the instance's current image.

  We should probably two things here to make this code more similar
  across hypervisors:

  1. We should add a generic `default_rescue_image_id` flag that can
  specific default rescue images for any driver. This would obsolete the
  libvirt specific flag.

  2. We should modify the libvirt `rescue` code to respect the
  `image_meta` passed in, delegating the logic of which image to use for
  rescue to the compute manager.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1078080/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1089877] Re: Extend used_limits extension to cover non-reservable resources

2015-03-30 Thread Sean Dague
this looks fixed upstream, no idea when, but quota show now includes
keypairs.

os1:~ nova quota-show
+-+---+
| Quota   | Limit |
+-+---+
| instances   | 10|
| cores   | 20|
| ram | 51200 |
| floating_ips| 10|
| fixed_ips   | -1|
| metadata_items  | 128   |
| injected_files  | 5 |
| injected_file_content_bytes | 10240 |
| injected_file_path_bytes| 255   |
| key_pairs   | 100   |
| security_groups | 10|
| security_group_rules| 20|
| server_groups   | 10|
| server_group_members| 10|
+-+---+


** Changed in: nova
   Status: Triaged = Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1089877

Title:
  Extend used_limits extension to cover non-reservable resources

Status in OpenStack Compute (Nova):
  Fix Released

Bug description:
  From bug #1062049

  The used_limits extension looks in the quota_usages table to see
  what's used for a given resource

  That corresponds to the ReservableResource resources in nova.quota -
  i.e. instances, cores, ram, floating IPs and security groups

  Resources like key pairs don't record their usage info in the
  quota_usages table, so we don't currently expose them in the
  used_limits output

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1089877/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1204583] Re: xenapi: in finish_migration should clean up move_disks if not used

2015-03-30 Thread Sean Dague
This is really old, owned by a core contributor. Marking as an invalid
bug that was probably used for self tracking.

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1204583

Title:
  xenapi: in finish_migration should clean up move_disks if not used

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  This bug fixed the majority of the finish_migration cleanup:
  https://bugs.launchpad.net/nova/+bug/1073303

  There is one remaining bit of cleanup that is required:
  * look at move_disks
  * if this is not run, we need to make sure the temp dir is cleaned up

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1204583/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1260274] Re: NoVNC Console not showing in Internet Explorer

2015-03-30 Thread Sean Dague
This is a novnc bug, not a nova one

** Changed in: nova
   Status: Triaged = Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1260274

Title:
  NoVNC Console not showing in Internet Explorer

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  When accessing the NoVNC console through Internet Explorer (tested
  with IE9 and IE10) the HTML5 Canvas never renders, instead showing
  'Canvas not supported'.

  Environment:
   - OS: Ubuntu 12.04 LTS
   - Platform: Openstack Grizzly
   - Packages:
    nova-novncproxy 1:2013.1.3-0ubuntu1~cloud0
    novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0
    python-novnc 2012.2~20120906+dfsg-0ubuntu4~cloud0

  According to https://github.com/kanaka/noVNC/wiki/Browser-support
  NoVNC should work with IE9 and above.

  [edit 6 Jan 2014] It would appear that it works properly with IE11

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1260274/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1101222] Re: Detect vncserver_proxyclient_address mis-configuration

2015-03-30 Thread Sean Dague
Triage indicates the bug is basically fixable with the information
provided in the bug. This is a nice to have enhancement, but not
triaged.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1101222

Title:
  Detect vncserver_proxyclient_address mis-configuration

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Say you have two compute nodes, compute1 and compute 2

  vncserver_proxyclient_address on compute1 is set to the public address
  of compute1 - i.e. the address by which the proxy should connect to
  the VNC server on compute1

  If an admin gets this wrong - e.g. copies nova.conf to compute2 and
  forgets to change the config option - it has pretty disastrous
  security implications. Without any warning or sign that there's a
  problem, users attempting to connect to the VNC console of their VM
  will be sent to the VNC console of some other VM which happens to be
  running on the same port on a different compute node.

  I've seen and debugged this in a real deployment :)

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1101222/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1161557] Re: Race condition in handling of udev generated symlinks

2015-03-30 Thread Sean Dague
This is unclear that it's a real bug, as it was disabled in the past.
Definitely not in a triaged state.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1161557

Title:
  Race condition in handling of udev generated symlinks

Status in Cinder:
  Triaged
Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  In several components in both Cinder and Nova udev provided symlinks are 
expected to appear immediately.
  This is not the case as udev rules run async of device plugging.
  Volume drivers in Cinder and the libvirt driver in Nova seem to be the 
primary culprits.

  To solve this we should wait for udevd to finish processing events, this can 
be done with udevadm by calling settle. This is done by simply calling udevadm 
settle. Additionally it can be called with a timeout (probably a good idea).
  Excerpt from udevadm settle --help:

  Usage: udevadm settle OPTIONS
    --timeout=seconds maximum time to wait for events
    --seq-start=seqnumfirst seqnum to wait for
    --seq-end=seqnum  last seqnum to wait for
    --exit-if-exists=file stop waiting if file exists
    --quiet do not print list after timeout
    --help

  For more intelligent use we could wrap this in a function that can use
  the --exit-if-exists behavior.

  This will ensure the symlink actually exists before we try use it.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cinder/+bug/1161557/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1173430] Re: cleanup_running_deleted_instances should not be needed

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1173430

Title:
  cleanup_running_deleted_instances should not be needed

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  cleanup_running_deleted_instances should likely be refactored so that
  said function is not needed in the first place, since in a ideal world
  an instance does not have the capability to be running when it is
  deleted (said deletion should fail in that case).

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1173430/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1195073] Re: pxe deploy timeout defaults to unset

2015-03-30 Thread Sean Dague
Can't see where this code even exists in Nova at this point. The bug
report also doesn't provide enough context to figure out where it might
have existed.

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1195073

Title:
  pxe deploy timeout defaults to unset

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  This is not suitable for production : failed deploys never recover.
  We're running with 1200 in TripleO.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1195073/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1215383] Re: XenAPI: Consider removing safe_copy_vdi

2015-03-30 Thread Sean Dague
Marking as Opinion as this bug discusses internal code structure changes
which are partially out of date. In future bugs that have a reproduce
would be more valuable.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1215383

Title:
  XenAPI: Consider removing safe_copy_vdi

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  safe_copy_vdi depends on ext3 and is not using supported interfaces,
  so should be considered for removal.

  It was added because the vdi.copy operation wasn't considered to be
  atomic.  However, VDI.copy is definitely atomic and will not return
  until the copy has finished.  It's therefore possible that the
  rationale for this function was based on looking at the SR during a
  copy operation, or use of async VDI copy.

  The reasons for this function need to be reviewed.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1215383/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1190076] Re: virt xenapi driver does not throw InstancePowerOffFailure exception on a failure while destroying an instance

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1190076

Title:
  virt xenapi driver does not throw InstancePowerOffFailure exception on
  a failure while destroying an instance

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  _shutdown_instance in compute manager expects InstancePowerOffFailure 
exception and reraises it.
  But the xenapi driver does not raise it. On the other hand it just returns a 
boolean value( False ) on a failure.

  Its potential consequences are that terminate_instance goes ahead and
  marks the instance as deleted, leaving behind orphaned instances.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1190076/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1076119] Re: VM can't connect to itself on it's own floating IP

2015-03-30 Thread Sean Dague
Very old bug that seems to not have a current reproduce. Putting to
invalid for now, it's definitely not triaged.

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1076119

Title:
  VM can't connect to itself on it's own floating IP

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  using nova.network.manager.FlatDHCPManager:

  fixed pool 192.168.0.0/24
  floating-ip 9.3.11.240

  i create 2 VMs, both are assigned IPs from the fixed IP pool.  i
  assign one VM the 9.3.11.240 address from the floating pool.  from the
  host machine ((and the other VM)), i can ping and SSH to the VM on the
  9.3.11.240 address.  However, on the machine which is assigned the
  9.3.11.240 address I can't ping or ssh back to itself on that address.

  this is a problem for some of the Bosh and Cloudfoundry applications
  who run services that may want to talk-back to themselves on the
  assigned floating ip.

  
  I'm running folsom on ubuntu 12.04.  i've attached my nova.conf.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1076119/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1097605] Re: osapi_compute_unique_server_name_scope is a strange little option

2015-03-30 Thread Sean Dague
The constraint checking kind of needs to be in the db layer because
that's the only place we have access to it. Possibly a future clean up.
But I think Opinion marks the current real state.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1097605

Title:
  osapi_compute_unique_server_name_scope is a strange little option

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  See https://review.openstack.org/19174

  osapi_compute_unique_server_name_scope sounds like its a detail of the
  openstack API layer, but the option and its usage lives in the DB
  layer

  It would be nice if we could figure out some way of moving it into the
  API layer

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1097605/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1161557] Re: Race condition in handling of udev generated symlinks

2015-03-30 Thread John Griffith
** Changed in: cinder
   Status: Triaged = Incomplete

** Changed in: cinder
   Status: Incomplete = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1161557

Title:
  Race condition in handling of udev generated symlinks

Status in Cinder:
  Invalid
Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  In several components in both Cinder and Nova udev provided symlinks are 
expected to appear immediately.
  This is not the case as udev rules run async of device plugging.
  Volume drivers in Cinder and the libvirt driver in Nova seem to be the 
primary culprits.

  To solve this we should wait for udevd to finish processing events, this can 
be done with udevadm by calling settle. This is done by simply calling udevadm 
settle. Additionally it can be called with a timeout (probably a good idea).
  Excerpt from udevadm settle --help:

  Usage: udevadm settle OPTIONS
    --timeout=seconds maximum time to wait for events
    --seq-start=seqnumfirst seqnum to wait for
    --seq-end=seqnum  last seqnum to wait for
    --exit-if-exists=file stop waiting if file exists
    --quiet do not print list after timeout
    --help

  For more intelligent use we could wrap this in a function that can use
  the --exit-if-exists behavior.

  This will ensure the symlink actually exists before we try use it.

To manage notifications about this bug go to:
https://bugs.launchpad.net/cinder/+bug/1161557/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1335554] Re: Documentation: incorrect use of weighted

2015-03-30 Thread Sean Dague
doc patches welcomed, but I'm not convinced a bug is needed for this.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1335554

Title:
  Documentation: incorrect use of weighted

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  A recent review made me wonder about how we are using weight in the
  documentation. When OpenStack talks about weight, it doesn't mean mass
  but a statistical property.

  Such properties are used when weighting the members of a class, when
  producing a weighted average or a weighted index. But some of the
  OpenStack documentation talks about weighing.

  Here are some samples from nova docs:

  Correct:
  # If we speak about `filtering` and `weighting`, their work is quite
  # flexible in the Filter Scheduler.

  Incorrect:
  # During its work Filter Scheduler firstly makes dictionary of unfiltered
  # hosts, then filters them using filter properties and finally chooses hosts
  # for the requested number of instances (each time it chooses the most weighed
  # host and appends it to the list of selected hosts).

  Unfortunately, the API docs are sometimes rather unhelpful, containing
  explanations like this gem:

  # def weight_multiplier(self):
  # How weighted this weigher should be.

  Related (in openstack-manuals): bug 1335423

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1335554/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1160850] Re: ec2 (DescribeKeyPairs) does not supports filtering by fingerprint

2015-03-30 Thread Sean Dague
ec2 in nova is deprecated, this should probably be addressed in the out
of tree code

** Changed in: nova
   Status: Triaged = Opinion

** Changed in: nova
   Importance: Low = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1160850

Title:
  ec2 (DescribeKeyPairs) does not supports filtering by fingerprint

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  euca-describe-keypairs  --filter 
fingerprint=c0:40:d9:8b:e2:09:11:2b:59:a8:86:3b:03:87:5d:e5
  KEYPAIR   testc0:40:d9:8b:e2:09:11:2b:59:a8:86:3b:03:87:5d:e5
  KEYPAIR   test2   1d:25:e8:42:1a:59:ce:59:6c:b1:2c:76:d6:4b:cd:de

  The listing shows me all keypairs  even if  I define a filter.

  nova version:  9a90f6ccdb88a22ff17b3f48d26378b4fa613ede

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1160850/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1312796] Re: we are able to destroy an instance while taking a snapshot

2015-03-30 Thread Sean Dague
There was a reproduce ask over 6 months ago which was not done. Marking
as Invalid.

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1312796

Title:
  we are able to destroy an instance while taking a snapshot

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  we are able to destroy an instance when taking a snapshot.

  the new image status would depend if it was already created and
  uploaded to /var/lib/glance/images

  I think that if we allow to destroy the instance when taking the
  snapshot we run the risk of data corruption on the new snapshot or the
  snapshot not being created at all.

  so I think that to destroy the instance while taking the snapshot we
  should have a --force in while the admin user knowingly destroys the
  instance.

  
  [root@puma31 ~(keystone_admin)]# nova list 
  
nov+--+--++--+-+--+
  | ID   | Name | Status | Task State   
| Power State | Networks |
  
+--+--++--+-+--+
  | e00ae899-e285-4f09-8cda-2c2680799bba | from | ACTIVE | image_pending_upload 
| Running | novanetwork=192.168.32.2 |
  
+--+--++--+-+--+
  [root@puma31 ~(keystone_admin)]# nova delete 
e00ae899-e285-4f09-8cda-2c2680799bba
  [root@puma31 ~(keystone_admin)]# nova list 
  
+--+--+++-+--+
  | ID   | Name | Status | Task State | Power 
State | Networks |
  
+--+--+++-+--+
  | e00ae899-e285-4f09-8cda-2c2680799bba | from | ACTIVE | deleting   | Running 
| novanetwork=192.168.32.2 |
  
+--+--+++-+--+

  [root@puma31 ~(keystone_admin)]# nova image-create 
e00ae899-e285-4f09-8cda-2c2680799bba destroy_test --poll
  Server snapshotting... 50% complete
  Server snapshotting... 50% complete
  Server snapshotting... 100% complete
  Finished
  ERROR: Instance could not be found (HTTP 404) (Request-ID: 
req-b6b7b066-0da8-441a-8788-b6969d7b1527)
  [root@puma31 ~(keystone_admin)]# 
  [root@puma31 ~(keystone_admin)]# 
  [root@puma31 ~(keystone_admin)]# glance image-list 
  
+--+--+-+--+++
  | ID   | Name | Disk Format | 
Container Format | Size   | Status |
  
+--+--+-+--+++
  | 6aa2362c-a1bb-490a-aeeb-3786ad7b9312 | destroy_test | qcow2   | bare
 | 3629645824 | active |
  | 73f92385-3080-4a4e-a100-76de38a3a569 | new_snap | qcow2   | bare
 | 3628728320 | active |
  | deddabea-475f-4c2f-88e3-0c76612e529c | poll-test1   | qcow2   | bare
 | 3629383680 | active |
  | df06e227-0d6a-4e2c-90c1-13cd32721360 | rhel | qcow2   | bare
 | 3628990464 | active |
  | 6175a441-8cb2-4d35-9b7d-241d51eaa270 | rhel1| qcow2   | bare
 | 3629383680 | active |
  
+--+--+-+--+++

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1312796/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1199475] Re: xenapi: bad error when trying to attach a volume to a non-pv guest

2015-03-30 Thread Sean Dague
This really isn't addressable in current API, need new APIs to address
this.

** Changed in: nova
   Status: Triaged = Opinion

** Changed in: nova
   Importance: Low = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1199475

Title:
  xenapi: bad error when trying to attach a volume to a non-pv guest

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  We get the following error from XenAPI when trying to do VBD.plug on a 
volume on a system without PV tools:
  VM_MISSING_PV_DRIVERS

  Really we need a way for these users to find out about this issue, and
  be able to attach the volume somehow.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1199475/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438189] [NEW] ComputeNode and Service objects out of sync with db schema

2015-03-30 Thread Hans Lindgren
Public bug reported:

A recent  commit [1] removed the relationship between compute_node and
service tables and in addition also made the compute_node.service_id
column nullable. These changes were never replicated to the object
counterparts.

[1] 551be2c52a29cb2755de4825a3fcb2c8f7d7b3f1

** Affects: nova
 Importance: Medium
 Assignee: Hans Lindgren (hanlind)
 Status: New


** Tags: unified-objects

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438189

Title:
  ComputeNode and Service objects out of sync with db schema

Status in OpenStack Compute (Nova):
  New

Bug description:
  A recent  commit [1] removed the relationship between compute_node and
  service tables and in addition also made the compute_node.service_id
  column nullable. These changes were never replicated to the object
  counterparts.

  [1] 551be2c52a29cb2755de4825a3fcb2c8f7d7b3f1

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438189/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1158424] Re: Nova doesnt validate parameters while using euca-run-instances

2015-03-30 Thread Sean Dague
ec2 apis in tree are deprecated, probably should be fixed out of tree
instead

** Changed in: nova
   Status: Triaged = Opinion

** Changed in: nova
   Importance: Low = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1158424

Title:
  Nova doesnt validate parameters while using euca-run-instances

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Using euca-run-instances -n 0 should run exactly instances instead of
  1 instance according to:

  http://docs.aws.amazon.com/AWSEC2/latest/APIReference/api-error-
  codes.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1158424/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1435386] Re: VMs are being taken over through a VNC proxy exploit

2015-03-30 Thread Thierry Carrez
That said, documentation recommends setting up 0.0.0.0 so that live
migration can work:

http://docs.openstack.org/admin-guide-cloud/content/section_configuring-compute-migrations.html
You must specify vncserver_listen=0.0.0.0 or live migration will not work 
correctly.

http://docs.openstack.org/admin-guide-cloud/content/nova-vncproxy-replaced-with-nova-novncproxy.html
To connect the service to your Compute deployment, add the following 
configuration options to your nova.conf file:
vncserver_listen=0.0.0.0
To use live migration, use the 0.0.0.0 address.

** Information type changed from Private Security to Public Security

** Also affects: openstack-manuals
   Importance: Undecided
   Status: New

** Summary changed:

- VMs are being taken over through a VNC proxy exploit
+ Specific config setting may result in VMs being taken over through VNC

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1435386

Title:
  Specific config setting may result in VMs being taken over through VNC

Status in OpenStack Compute (Nova):
  Invalid
Status in OpenStack Manuals:
  New
Status in OpenStack Security Advisories:
  Incomplete

Bug description:
  Jonathan Hogg from Chargebox reports (edited):

  On a single-machine cloud running OpenStack Icehouse and over the last
  week we have seen compromises of all of the Ubuntu 14.04 VMs running
  on the machine. Scenario shows the attacker gaining access through VNC
  (via controlled reboot to reset root password).

  QEMU instances are running with -vnc 0.0.0.0:1, which may or may not
  be the issue.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1435386/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1173893] Re: Grizzly: Task instance_usage_audit is not running when it is enabled and nova-compute is running

2015-03-30 Thread Sean Dague
qpid is used by nearly no one. Marking the nova bug as invalid, this
should be the upstream fix

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1173893

Title:
  Grizzly: Task instance_usage_audit is not running when it is enabled
  and nova-compute is running

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  When using qpid as message queue backend and set instance_usage_audit = True
  the instance usage audit function doesn't work correct.

  2013-02-26 17:34:38.727 ERROR nova.openstack.common.rpc.amqp 
[req-ff067345-7fba-4529-819c-ed8e5b1d6682 None None] Exception during message 
handling
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp Traceback 
(most recent call last):
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp File 
/usr/lib/python2.6/site-packages/nova/openstack/common/rpc/amqp.py, line 276, 
in _process_data
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp rval = 
self.proxy.dispatch(ctxt, version, method, **args)
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp File 
/usr/lib/python2.6/site-packages/nova/openstack/common/rpc/dispatcher.py, 
line 133, in dispatch
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp return 
getattr(proxyobj, method)(ctxt, **kwargs)
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp File 
/usr/lib/python2.6/site-packages/nova/conductor/manager.py, line 348, in 
task_log_end_task
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp begin, end, 
host, errors, message)
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp File 
/usr/lib/python2.6/site-packages/nova/db/api.py, line 1689, in 
task_log_end_task
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp message)
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp File 
/usr/lib/python2.6/site-packages/nova/db/sqlalchemy/api.py, line 118, in 
wrapper
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp return 
f(*args, **kwargs)
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp File 
/usr/lib/python2.6/site-packages/nova/db/sqlalchemy/api.py, line 4791, in 
task_log_end_task
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp raise 
exception.TaskNotRunning(task_name=task_name, host=host)
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp 
TaskNotRunning: Task instance_usage_audit is not running on host WIN-CJR1366L5AH
  2013-02-26 17:34:38.727 1399 TRACE nova.openstack.common.rpc.amqp
  2013-02-26 17:34:38.757 ERROR nova.openstack.common.rpc.common 
[req-ff067345-7fba-4529-819c-ed8e5b1d6682 None None] Returning exception Task 
instance_usage_audit is not running on host WIN-CJR1366L5AH to
  caller

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1173893/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1357437] Re: nova.tests.virt.libvirt.test_driver should use constants from fakelibvirt

2015-03-30 Thread Matt Riedemann
*** This bug is a duplicate of bug 1414708 ***
https://bugs.launchpad.net/bugs/1414708

We removed the dependency on libvirt in unit tests here:

https://github.com/openstack/nova/commit/50861d3221f625a5cc589d903054bb2c440b4026

** This bug has been marked a duplicate of bug 1414708
   libvirt unit tests are failing if no real libvirt installed

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1357437

Title:
  nova.tests.virt.libvirt.test_driver should use constants from
  fakelibvirt

Status in OpenStack Compute (Nova):
  Confirmed

Bug description:
  Commit f0883800660ab546f5667b973f339c4df4c5c458 adds some tests and
  new constants for _swap_volume and _live_snapshot, one of which is
  VIR_DOMAIN_BLOCK_REBASE_COPY.  It adds the required constants to the
  fakelibvirt module but doesn't use them when making assertions in the
  test, which can fail if you're not using fakelibvirt but not using a
  new enough version of libvirt on your system.

  I realize that we now require libvirt-python = 1.2.5 for testing, but
  that requires libvirt = 0.9.11 and if you're on 0.9.11 it doesn't
  have VIR_DOMAIN_BLOCK_REBASE_COPY  defined in libvirt.h.in:

  
http://libvirt.org/git/?p=libvirt.git;a=blob;f=include/libvirt/libvirt.h.in;h=499dcd4514bf793b531e53496c56237fb055e1ba;hb=782afa98e4a5fa9a0927a9e32f9cf36082a2e8e7

  It seems kind of strange to me that we use constants directly from the
  libvirt module at all in test_driver given the import conditional
  check at the top of the module:

  
http://git.openstack.org/cgit/openstack/nova/tree/nova/tests/virt/libvirt/test_driver.py#n88

  try:
  import libvirt
  except ImportError:
  libvirt = fakelibvirt
  libvirt_driver.libvirt = libvirt

  Given that, any change that requires testing new constants from
  libvirt should be in fakelibvirt, and we should use fakelibvirt when
  using those constants in the tests, otherwise the unit tests are
  really dependent on your environment.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1357437/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438238] [NEW] Several concurent scheduling requests for CPU pinning may fail due to racy host_state handling

2015-03-30 Thread Nikola Đipanov
Public bug reported:

The issue happens when multiple scheduling attempts that request CPU pinning 
are done in parallel.
 

015-03-25T14:18:00.222 controller-0 nova-scheduler err Exception during
message handling: Cannot pin/unpin cpus [4] from the following pinned
set [3, 4, 5, 6, 7, 8, 9]

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
Traceback (most recent call last):

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
/usr/lib64/python2.7/site-packages/oslo/messaging/rpc/dispatcher.py,
line 134, in _dispatch_and_reply

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
incoming.message))

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
/usr/lib64/python2.7/site-packages/oslo/messaging/rpc/dispatcher.py,
line 177, in _dispatch

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
return self._do_dispatch(endpoint, method, ctxt, args)

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
/usr/lib64/python2.7/site-packages/oslo/messaging/rpc/dispatcher.py,
line 123, in _do_dispatch

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
result = getattr(endpoint, method)(ctxt, **new_args)

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
/usr/lib64/python2.7/site-packages/oslo/messaging/rpc/server.py, line
139, in inner

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
return func(*args, **kwargs)

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-packages/nova/scheduler/manager.py, line
86, in select_destinations

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-
packages/nova/scheduler/filter_scheduler.py, line 80, in
select_destinations

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-
packages/nova/scheduler/filter_scheduler.py, line 241, in _schedule

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-packages/nova/scheduler/host_manager.py,
line 266, in consume_from_instance

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-packages/nova/virt/hardware.py, line 1472,
in get_host_numa_usage_from_instance

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-packages/nova/virt/hardware.py, line 1344,
in numa_usage_from_instances

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher   File
./usr/lib64/python2.7/site-packages/nova/objects/numa.py, line 91, in
pin_cpus

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
CPUPinningInvalid: Cannot pin/unpin cpus [4] from the following pinned
set [3, 4, 5, 6, 7, 8, 9]

2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher

What is likely happening is:

* nova-scheduler is handling several RPC calls to select_destinations at
the same time, in multiple greenthreads

* greenthread 1 runs the NUMATopologyFilter and selects a cpu on a
particular compute node, updating host_state.instance_numa_topology

* greenthread 1 then blocks for some reason

* greenthread 2 runs the NUMATopologyFilter and selects the same cpu on
the same compute node, updating host_state.instance_numa_topology. This
also seems like an issue if a different cpu was selected, as it would be
overwriting the instance_numa_topology selected by greenthread 1.

* greenthread 2 then blocks for some reason

* greenthread 1 gets scheduled and calls consume_from_instance, which
consumes the numa resources based on what is in
host_state.instance_numa_topology

*  greenthread 1 completes the scheduling operation

* greenthread 2 gets scheduled and calls consume_from_instance, which
consumes the numa resources based on what is in
host_state.instance_numa_topology - since the resources were already
consumed by greenthread 1, we get the exception above

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438238

Title:
  Several concurent scheduling requests for CPU pinning may fail due to
  racy host_state handling

Status in OpenStack Compute (Nova):
  New

Bug description:
  The issue happens when multiple scheduling attempts that request CPU pinning 
are done in parallel.
   

  015-03-25T14:18:00.222 controller-0 nova-scheduler err Exception
  during message handling: Cannot pin/unpin cpus [4] from the following
  pinned set [3, 4, 5, 6, 7, 8, 9]

  2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
  Traceback (most recent call last):

  2015-03-25 14:18:00.221 34127 TRACE oslo.messaging.rpc.dispatcher
  File /usr/lib64/python2.7/site-
  packages/oslo/messaging/rpc/dispatcher.py, line 134, in
  

[Yahoo-eng-team] [Bug 1194106] Re: db.utils.create_shadow_table() doesn't work if we have table with Enum column

2015-03-30 Thread Sean Dague
This code isn't in nova any more, moving to oslo.db

** Also affects: oslo.db
   Importance: Undecided
   Status: New

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1194106

Title:
  db.utils.create_shadow_table() doesn't work if we have table with Enum
  column

Status in OpenStack Compute (Nova):
  Invalid
Status in Oslo Database library:
  New

Bug description:
  
  In postgres we are not able to create two times Enum with the same name.

  So If we try to make a shadow table for table we will get exception
  that Enum with the same name already exists.

  So we should rename our Enum. 
  I think that shadow_prefix + original_name will be good convention.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1194106/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438226] [NEW] nova libvirt driver assumes libvirt support for CPU pinning

2015-03-30 Thread Stephen Finucane
Public bug reported:

CPU pinning support was implemented as part of this blueprint:

http://specs.openstack.org/openstack/nova-specs/specs/juno/approved
/virt-driver-cpu-pinning.html

However, CPU pinning support is broken in some libvirt versions
(summarized below), resulting in exceptions when attempting to schedule
instances with the 'hw:cpu_policy' flavor key.

We should add a libvirt version test against known broken versions and
use that to determine whether or not to support the flavor keys.

This is somewhat related to #1422775 (nova libvirt driver assumes qemu
support for NUMA pinning).

---

# Testing Configuration

Testing was conducted in a container which provided a single-node,
Fedora 21-based (3.17.8-300.fc21.x86_64) OpenStack instance (built with
devstack). The yum-provided libvirt and its dependencies were removed
and libvirt and libvirt-python were built and installed from source.

# Results

The results are as follows:

versions  status
  --
1.2.9 ok
1.2.9.1   ok
1.2.9.2   fail
1.2.10fail
1.2.11ok
1.2.12ok

v1.2.9.2 is broken by this (backported) patch:

https://www.redhat.com/archives/libvir-
list/2014-November/msg00275.html

This can be seen as commit

e226772 (qemu: fix domain startup failing with 'strict' mode in
numatune)

v1.2.10 inherits is broken at checkout but can be fixed by applying
these three patches (yes, one of these broke v1.2.9.2 - the irony is not
lost on me):

[0/3] 
https://www.redhat.com/archives/libvir-list/2014-November/msg00274.html
 - [1/3] 
https://www.redhat.com/archives/libvir-list/2014-November/msg00273.html
 - [2/3] 
https://www.redhat.com/archives/libvir-list/2014-November/msg00276.html
 - [3/3] 
https://www.redhat.com/archives/libvir-list/2014-November/msg00275.html

# Error logs

v1.2.9.2 produces the following exception:

Traceback (most recent call last):
  File /opt/stack/nova/nova/compute/manager.py, line 2301, in 
_build_resources
yield resources
  File /opt/stack/nova/nova/compute/manager.py, line 2171, in 
_build_and_run_instance
flavor=flavor)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 2357, in spawn
block_device_info=block_device_info)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 4376, in 
_create_domain_and_network
power_on=power_on)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 4307, in 
_create_domain
LOG.error(err)
  File /usr/lib/python2.7/site-packages/oslo_utils/excutils.py, line 82, 
in __exit__
six.reraise(self.type_, self.value, self.tb)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 4297, in 
_create_domain
domain.createWithFlags(launch_flags)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 183, in 
doit
result = proxy_call(self._autowrap, f, *args, **kwargs)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 141, in 
proxy_call
rv = execute(f, *args, **kwargs)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 122, in 
execute
six.reraise(c, e, tb)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 80, in 
tworker
rv = meth(*args, **kwargs)
  File /usr/lib64/python2.7/site-packages/libvirt.py, line 1029, in 
createWithFlags
if ret == -1: raise libvirtError ('virDomainCreateWithFlags() failed', 
dom=self)
libvirtError: Failed to create controller cpu for group: No such file or 
directory

v1.2.10 produces the following exception:

Traceback (most recent call last):
  File /opt/stack/nova/nova/compute/manager.py, line 2342, in 
_build_resources
yield resources
  File /opt/stack/nova/nova/compute/manager.py, line 2215, in 
_build_and_run_instance
block_device_info=block_device_info)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 2356, in spawn
block_device_info=block_device_info)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 4375, in 
_create_domain_and_network
power_on=power_on)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 4306, in 
_create_domain
LOG.error(err)
  File /usr/lib/python2.7/site-packages/oslo_utils/excutils.py, line 85, 
in __exit__
six.reraise(self.type_, self.value, self.tb)
  File /opt/stack/nova/nova/virt/libvirt/driver.py, line 4296, in 
_create_domain
domain.createWithFlags(launch_flags)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 183, in 
doit
result = proxy_call(self._autowrap, f, *args, **kwargs)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 141, in 
proxy_call
rv = execute(f, *args, **kwargs)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 122, in 
execute
six.reraise(c, e, tb)
  File /usr/lib/python2.7/site-packages/eventlet/tpool.py, line 80, in 
tworker

[Yahoo-eng-team] [Bug 1349515] Re: Commit 0e98f5a breaks db migrations with DB2

2015-03-30 Thread Matt Riedemann
This is being handled with a backend-specific condition in the 247
migration:

https://review.openstack.org/#/c/69047/45/nova/db/sqlalchemy/migrate_repo/versions/247_nullable_mismatch.py

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1349515

Title:
  Commit 0e98f5a breaks db migrations with DB2

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  This change:

  
https://github.com/openstack/nova/commit/0e98f5a522c08b17c98ed108459a179d14eacd4a

  Sets pci_devices.deleted.nullable=True which breaks with DB2 because
  that column is in a UniqueConstraint:

  
http://git.openstack.org/cgit/openstack/nova/tree/nova/db/sqlalchemy/models.py#n1374

  And DB2 requires that columns in a UniqueConstraint are not nullable.

  We get a trace like this:

  2014-07-28 03:40:33.511 7720 INFO migrate.versioning.api [-] 246 - 247...
  2014-07-28 03:40:37.073 7720 CRITICAL nova [-] ProgrammingError: 
(ProgrammingError) ibm_db_dbi::ProgrammingError: Statement Execute Failed: 
[IBM][CLI Driver][DB2/LINUXX8664] SQL0542N  The column named DELETED cannot 
be a column of a primary key or unique key constraint because it can contain 
null values.  SQLSTATE=42831 SQLCODE=-542 '\nALTER TABLE pci_devices ALTER 
COLUMN deleted DROP NOT NULL' ()
  2014-07-28 03:40:37.073 7720 TRACE nova Traceback (most recent call last):
  2014-07-28 03:40:37.073 7720 TRACE nova   File /usr/bin/nova-manage, line 
10, in module
  2014-07-28 03:40:37.073 7720 TRACE nova sys.exit(main())
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/nova/cmd/manage.py, line 1401, in main
  2014-07-28 03:40:37.073 7720 TRACE nova ret = fn(*fn_args, **fn_kwargs)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/nova/cmd/manage.py, line 902, in sync
  2014-07-28 03:40:37.073 7720 TRACE nova return migration.db_sync(version)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/nova/db/migration.py, line 29, in db_sync
  2014-07-28 03:40:37.073 7720 TRACE nova return 
IMPL.db_sync(version=version)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/nova/db/sqlalchemy/migration.py, line 44, in 
db_sync
  2014-07-28 03:40:37.073 7720 TRACE nova return 
versioning_api.upgrade(get_engine(), repository, version)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/versioning/api.py, line 186, in 
upgrade
  2014-07-28 03:40:37.073 7720 TRACE nova return _migrate(url, repository, 
version, upgrade=True, err=err, **opts)
  2014-07-28 03:40:37.073 7720 TRACE nova   File string, line 2, in _migrate
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/versioning/util/__init__.py, line 
160, in with_engine
  2014-07-28 03:40:37.073 7720 TRACE nova return f(*a, **kw)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/versioning/api.py, line 366, in 
_migrate
  2014-07-28 03:40:37.073 7720 TRACE nova schema.runchange(ver, change, 
changeset.step)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/versioning/schema.py, line 93, in 
runchange
  2014-07-28 03:40:37.073 7720 TRACE nova change.run(self.engine, step)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/versioning/script/py.py, line 148, 
in run
  2014-07-28 03:40:37.073 7720 TRACE nova script_func(engine)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/nova/db/sqlalchemy/migrate_repo/versions/247_nullable_mismatch.py,
 line 27, in upgrade
  2014-07-28 03:40:37.073 7720 TRACE nova 
pci_devices.c.deleted.alter(nullable=True)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/changeset/schema.py, line 534, in 
alter
  2014-07-28 03:40:37.073 7720 TRACE nova return alter_column(self, *p, **k)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/changeset/schema.py, line 141, in 
alter_column
  2014-07-28 03:40:37.073 7720 TRACE nova 
engine._run_visitor(visitorcallable, delta)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib64/python2.6/site-packages/sqlalchemy/engine/base.py, line 1479, in 
_run_visitor
  2014-07-28 03:40:37.073 7720 TRACE nova 
conn._run_visitor(visitorcallable, element, **kwargs)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib64/python2.6/site-packages/sqlalchemy/engine/base.py, line 1122, in 
_run_visitor
  2014-07-28 03:40:37.073 7720 TRACE nova **kwargs).traverse_single(element)
  2014-07-28 03:40:37.073 7720 TRACE nova   File 
/usr/lib/python2.6/site-packages/migrate/changeset/ansisql.py, line 56, in 

[Yahoo-eng-team] [Bug 1014647] Re: Tempest has no test for soft reboot

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1014647

Title:
  Tempest has no test for soft reboot

Status in OpenStack Compute (Nova):
  Invalid
Status in Tempest:
  Confirmed

Bug description:
  1. soft reboot requires support from the guest to operate. The current nova 
implementation tells the guest to reboot and
  then waits. If the soft reboot did not happen, it triggers a hard reboot but 
after a default wait of 2 minutes.

  Solution: Provide a new soft_reboot_image_ref, defaults to None, that
  is used for soft reboot tests which. If the value is None then the
  test is skipped.

  2. Because of (1), we should only use soft reboot when we are actually
  testing that feature.

  3. The current soft reboot test does not check that a soft reboot was
  done rather than hard. It should check for the server state of REBOOT.
  Same issue for the hard reboot test.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1014647/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1237683] Re: inconsitent virtual size in qcow base image after block-migration

2015-03-30 Thread Sean Dague
Addressed in Havana

** Changed in: nova
   Status: Triaged = Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1237683

Title:
  inconsitent virtual size in qcow base image after block-migration

Status in OpenStack Compute (Nova):
  Fix Released

Bug description:
  We're running a Grizzly node using KVM (1.0 from cloud-archive) with
  local ephemeral instance storage.

  Since approximately the time we upgraded to Grizzly we've been
  receiving complaints from particular users about secondary disk
  corruption issues. These users in particular are noticing the issue
  because they are relying on the secondary drive and also because hey
  are using CentOS, which drops to an interactive prompt before
  completing boot if it cannot mount all filesystems (Ubuntu does not).

  We've since discovered that this is specifically linked to block-
  migration of such disks which were created and formatted automatically
  by Nova. I.e., if we launch a new instance, log in and then reformat
  the drive internally (even as ext3), we don't encounter corruption
  issues after live-migration. If we change the virt_mkfs config option
  to use mkfs.ext4 then we also don't have the problem. Unfortunately
  that's not a simple fix for an active production cloud because all
  existing backing files must be removed in order to force their
  recreation.

  In investigating the problem we noticed a behaviour that might be 
interrelated - after block-migration the instances secondary disk has a 
generic backing file instances/_base/ephemeral, as opposed to the backing 
file it was created with on the origin host, e.g., 
instances/_base/ephemeral_30_default.
  These backing files have different virtual sizes(!):
  $ qemu-img info _base/ephemeral
  image: _base/ephemeral
  file format: raw
  virtual size: 2.0G (2147483648 bytes)
  disk size: 778M
  $ qemu-img info _base/ephemeral_30_default
  image: _base/ephemeral_30_default
  file format: raw
  virtual size: 30G (32212254720 bytes)
  disk size: 614M

  We're no experts on qcow, but this looks like it could be problematic
  and may explain the corruption issues we're seeing - I can imagine
  there would be problems for a migrated guest that attempts to read a
  previously untouched sector beyond the size of the new backing file.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1237683/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1257405] Re: Not checking image format produces lots of useless messages

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1257405

Title:
  Not checking image format produces lots of useless messages

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  The code for resizing partitionless images goes with the tell, don't
  ask idea and attempts to run extfs / mount utilities on an image even
  though they may fail. This produces lots of useless messages during
  the instance preparation, like these:

  2013-11-21 06:45:07 20902 DEBUG nova.virt.disk.mount.api 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] Failed 
to mount filesystem: Unexpected error while running command.
  Command: sudo nova-rootwrap /etc/nova/rootwrap.conf mount /dev/nbd8 
/tmp/openstack-vfs-localfsSz6ylg
  Exit code: 32
  Stdout: ''
  Stderr: 'mount: you must specify the filesystem type\n' mnt_dev 
/usr/lib/python2.7/dist-packages/nova/virt/disk/mount/api.py:198
  2013-11-21 06:45:07 20902 DEBUG nova.virt.disk.mount.api 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] Fail 
to mount, tearing back down do_mount 
/usr/lib/python2.7/dist-packages/nova/virt/disk/mount/api.py:219
  2013-11-21 06:45:07 20902 DEBUG nova.virt.disk.mount.api 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] Unmap 
dev /dev/nbd8 unmap_dev 
/usr/lib/python2.7/dist-packages/nova/virt/disk/mount/api.py:184
  2013-11-21 06:45:07 20902 DEBUG nova.virt.disk.mount.nbd 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] 
Release nbd device /dev/nbd8 unget_dev 
/usr/lib/python2.7/dist-packages/nova/virt/disk/mount/nbd.py:128
  2013-11-21 06:45:07 20902 DEBUG nova.openstack.common.processutils 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] 
Running cmd (subprocess): sudo nova-rootwrap /etc/nova/rootwrap.conf qemu-nbd 
-d /dev/nbd8 execute 
/usr/lib/python2.7/dist-packages/nova/openstack/common/processutils.py:147
  2013-11-21 06:45:07 20902 DEBUG nova.virt.disk.vfs.localfs 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] Failed 
to mount image Failed to mount filesystem: Unexpected error while running 
command.
  Command: sudo nova-rootwrap /etc/nova/rootwrap.conf mount /dev/nbd8 
/tmp/openstack-vfs-localfsSz6ylg
  Exit code: 32
  Stdout: ''
  Stderr: 'mount: you must specify the filesystem type\n') setup 
/usr/lib/python2.7/dist-packages/nova/virt/disk/vfs/localfs.py:83
  2013-11-21 06:45:07 20902 DEBUG nova.virt.disk.api 
[req-939d5d50-25ea-4f7a-8882-a880b9671e47 10873781609182 10816527907643] Unable 
to mount image 
/var/lib/nova/instances/5159cff0-91f6-4521-a0be-d74ce4f81fad/disk with error 
Failed to mount filesystem: Unexpected error while running command.
  Command: sudo nova-rootwrap /etc/nova/rootwrap.conf mount /dev/nbd8 
/tmp/openstack-vfs-localfsSz6ylg
  Exit code: 32
  Stdout: ''
  Stderr: 'mount: you must specify the filesystem type\n'. Cannot resize. 
is_image_partitionless 
/usr/lib/python2.7/dist-packages/nova/virt/disk/api.py:191

  This could be fixed by doing a simple check on the the image itself to
  pick up the magic signature. This would allow to skip e2resize on non-
  extfs files and simple mounting on partitioned images without the
  error messages.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1257405/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1437199] Re: zookeeper driver used with O(n^2) complexity by the scheduler or the api

2015-03-30 Thread Sean Dague
Attilla, please don't file specs kinds of issues as bugs, this is really
a specs level rearchitecture.

** Changed in: nova
   Status: New = Invalid

** Changed in: nova
   Importance: Undecided = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1437199

Title:
  zookeeper driver used with O(n^2) complexity  by the scheduler or the
  api

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  (Loop1) 
https://github.com/openstack/nova/blob/af2d6c9576b1ac5f3b3768870bb15d9b5cf1610b/nova/scheduler/driver.py#L55
  (Loop2) 
https://github.com/openstack/nova/blob/af2d6c9576b1ac5f3b3768870bb15d9b5cf1610b/nova/servicegroup/drivers/zk.py#L177

  Iterating the hosts through  the  ComputeFilter also has this issue,
  ComputeFilter usage in a loop has other performance issues .

  The API loop1 is here:
  
https://github.com/openstack/nova/blob/e5d0531d8ed4efcd612c0597557e5651c16294b5/nova/api/openstack/compute/contrib/services.py#L81

  The zk driver issue can be mitigated by doing the testing `filtering`
  in the is_up instead of the get_all , by reorganizing the code.

  However better solution would be to have the scheduler to use the get_all,
  or redesigning the servicegroup management.

  A better design would be to use the DB even with the zk,mc drvier, but
  do update ONLY when the service actually came up or dies, in this case
  the sg drivers MAY need dedicated service processes.

  NOTE: The servicegroup driver concept was introduced to avoid doing 10_000 DB 
update/sec @100_000 host (10/sec  update freq),
  if your servers are bad and every server has 1:1000 chance to die on the 
given day,  it would lead only to 0.001 UPDATE/sec (100/day) @100_000 host.

  NOTE: If the up/down is knowable just form the DB, the scheduler could
  eliminate the dead hosts at the first DB query, without using
  ComputeFilter as it is used now. (The plugins SHOULD be able to extend
  the  base hosts query)

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1437199/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1359136] Re: vmware: display incorrect host_ip for hypervisor show

2015-03-30 Thread Davanum Srinivas (DIMS)
** Changed in: nova
   Status: Confirmed = Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1359136

Title:
  vmware: display incorrect host_ip for hypervisor show

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  From VCenter driver,  with compute node configured to be cluster01. And 
VCenter IP is 10.9.1.43
  But with hypervisor show command, it will display the host IP for the 
controller node. It should be pointing to 
  the really VCenter IP.

  [root@dhcp-10-9-3-83 ~]# nova hypervisor-show 29
  
+---+--+
  | Property  | Value   
 |
  
+---+--+
  | cpu_info_model| [Intel(R) Xeon(R) CPU E5-2630 0 @ 2.30GHz, 
Intel(R) Xeon(R) CPU E5-2630 0 @ 2.30GHz] |
  | cpu_info_topology_cores   | 24  
 |
  | cpu_info_topology_threads | 48  
 |
  | cpu_info_vendor   | [IBM, IBM]  
 |
  | current_workload  | 0   
 |
  | disk_available_least  | -   
 |
  | free_disk_gb  | 763 
 |
  | free_ram_mb   | 354728  
 |
  | host_ip   | 10.9.3.83   
 |
  | hypervisor_hostname   | domain-c17(cluster01)   
 |
  | hypervisor_type   | VMware vCenter Server   
 |
  | hypervisor_version| 5005000 
 |
  | id| 29  
 |
  | local_gb  | 794 
 |
  | local_gb_used | 31  
 |
  | memory_mb | 362336  
 |
  | memory_mb_used| 7608
 |
  | running_vms   | 3   
 |
  | service_host  | dhcp-10-9-3-83.sce.cn.ibm.com   
 |
  | service_id| 6   
 |
  | vcpus | 48  
 |
  | vcpus_used| 5   
 |
  
+---+--+

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1359136/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1102714] Re: nova-network related fixed ip cleanup is needed

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

** Changed in: nova
   Importance: Low = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1102714

Title:
  nova-network related fixed ip cleanup is needed

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  overall description: 
  If nova-network crashes (and then restarts) during a nova delete operation, 
then the ip associated with the deleted VM instance may remain in the 
allocated state. This may affect the subsequent network related operations. 

  concrete example:
  step 1: create a fixed ip network with only 4 ips. 
  the fixed_ips table looks like the follwoing at this stage:
  
  mysql select id,address,instance_id,allocated,leased,reserved from fixed_ips;
  ++---+-+---++--+
  | id | address   | instance_id | allocated | leased | reserved |
  ++---+-+---++--+
  |  1 | 192.199.196.0 |NULL | 0 |  0 |1 |
  |  2 | 192.199.196.1 |NULL | 0 |  0 |1 |
  |  3 | 192.199.196.2 |NULL | 0 |  0 |0 |
  |  4 | 192.199.196.3 |NULL | 0 |  0 |1 |
  ++---+-+---++--+
  

  step 2: create a VM and then delete it.

  during the execution of nova delete command, nova-compute sends a
  rpc call and then a rpc cast to nova-network. At the time of the rpc
  cast, before the rpc is sent, we stop nova-network service (to emulate
  a service crash). The rpc cast is lost if the rpc queue related to
  nova-network is auto-deleted with the stopped service.

  the nova delete command returns successfully. the VM is marked as
  deleted. but the ip associated with the VM is still marked as
  allocated due to the lost rpc message.

  the fixed_ips table looks like the following at this stage:
  
  mysql select id,address,instance_id,allocated,leased,reserved from fixed_ips;
  ++---+-+---++--+
  | id | address   | instance_id | allocated | leased | reserved |
  ++---+-+---++--+
  |  1 | 192.199.196.0 |NULL | 0 |  0 |1 |
  |  2 | 192.199.196.1 |NULL | 0 |  0 |1 |
  |  3 | 192.199.196.2 |   1 | 1 |  0 |0 |
  |  4 | 192.199.196.3 |NULL | 0 |  0 |1 |
  ++---+-+---++--+
  

  step 3: restart the nova-network service. try to create another VM via
  nova boot. this time we will get an exception due to zero fixed ips
  available.

  thought:
  from a user's perspective, the VM in step 3 should be created. that operation 
cannot succeed due to the existence of an orphan ip related to a previously 
deleted instance. Admittedly, the situation in the above example can be avoided 
by configuring the rpc queue related to nova-network to be non-auto-deleted. 
But a periodical orphan ip cleanup logic may help in general in such cases. 

  Is there such a logic/module in OpenStack? If there is, should it be
  triggered when the fixed_ips table is about to be exhausted (as in
  the above example)?

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1102714/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1433805] Re: Absence of python-ironicclient in nova requirements.txt making upgrades awkward, python-ironicclient features diffucult

2015-03-30 Thread Sean Dague
Can this just be handled with changing the required upgrade order? Just
make ironic go first?

** Changed in: nova
   Status: New = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1433805

Title:
  Absence of python-ironicclient in nova requirements.txt making
  upgrades awkward, python-ironicclient features diffucult

Status in OpenStack Bare Metal Provisioning Service (Ironic):
  Won't Fix
Status in OpenStack Compute (Nova):
  Opinion
Status in OpenStack Compute (nova) juno series:
  In Progress

Bug description:
  Nova's requirements.txt does not list python-ironicclient, meaning a
  stable/juno nova deployment (at least in our gate) will be running
  with the most recent release of python-ironicclient.

  Many new features have been added to Ironic since juno and have been
  introduced incrementally via API micro-versions.   The client library
  released at the time of stable/juno did not send any API version
  header. The current (kilo) server recognizes this and defaults to the
  lowest API version (v1.1) it supports. The desired behavior of python-
  ironicclient is for it to request the greatest API version it
  understands (presently 1.6) [3].

  The nova.virt.ironic driver in juno/stable depends on node states only
  available in the corresponding version [1] of Ironic.  These have
  changed since then and the new node states are exposed via new API
  micro-versions [2]. Using a new client library with a new server
  release will result in the new states being returned to Nova. In
  particular, the state of a node that is available for use, as returned
  by the v1.1 API is NOSTATE, and as returned by the current Kilo API,
  is AVAILABLE.

  The goal is to make the client transparently negotiate which version
  to use with the Ironic server if the latest version is not supported.
  This is a feature that would be introduced in a future python-
  ironicclient release.

  However, since Nova is not listing python-ironicclient in its
  requirements, during upgrades we can end up with a stable/juno Nova
  using this new client version to speak to a Kilo Ironic server via the
  most recent API micro versions. This would result in nova driver
  errors as the Ironic server would be returning node states that
  stable/juno driver [1] does not understand [2].

  We either need to introduce python-ironicclient as a listed
  requirement of Nova (at least in stable), or explicitly declare that
  the driver use the older API version in its client interactions, or
  require that operators upgrade Nova (and python-ironicclient) to Kilo
  before upgrading Ironic.

  [1] 
https://git.openstack.org/cgit/openstack/nova/tree/nova/virt/ironic/ironic_states.py?h=stable%2Fjuno
  [2] 
https://git.openstack.org/cgit/openstack/nova/tree/nova/virt/ironic/ironic_states.py
  [3] 
http://specs.openstack.org/openstack/ironic-specs/specs/kilo/api-microversions.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/ironic/+bug/1433805/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1414437] Re: conf file with spaces before data cause error

2015-03-30 Thread Davanum Srinivas (DIMS)
Working as designed. it's reporting that you have a problem with the ini
file...no?

** Changed in: nova
   Status: Confirmed = Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1414437

Title:
  conf file with spaces before data cause error

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  Add a space before a conf line cause failure at restarting the service.
  example via nova.conf:

  rabbit_hosts=10.35.186.132:5672
  ==
   rabbit_hosts=10.35.186.132:5672

  cause:
  [root@controller ~(keystone_admin)]# openstack-service restart
  Job for openstack-nova-cert.service failed. See 'systemctl status 
openstack-nova-cert.service' and 'journalctl -xn' for details.
  Job for openstack-nova-api.service failed. See 'systemctl status 
openstack-nova-api.service' and 'journalctl -xn' for details.
  Job for openstack-nova-consoleauth.service failed. See 'systemctl status 
openstack-nova-consoleauth.service' and 'journalctl -xn' for details.
  Job for openstack-nova-conductor.service failed. See 'systemctl status 
openstack-nova-conductor.service' and 'journalctl -xn' for details.
  Job for openstack-nova-scheduler.service failed. See 'systemctl status 
openstack-nova-scheduler.service' and 'journalctl -xn' for details.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1414437/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1430521] Re: nova network-create is ignoring --bridge when using the VlanManager

2015-03-30 Thread Sean Dague
There are other bits of the nova-network code which assume bridge # maps
to vlan #, so I'm not sure this is reasonably changable above the driver
layer. I suspect a warning that vlan and bridge are incompatibile
options in the client would be the near term fix.

** Changed in: nova
   Importance: Undecided = Low

** Also affects: python-novaclient
   Importance: Undecided
   Status: New

** Changed in: nova
   Status: New = Incomplete

** Changed in: python-novaclient
   Status: New = Confirmed

** Changed in: python-novaclient
   Importance: Undecided = Low

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1430521

Title:
  nova network-create is ignoring --bridge when using the VlanManager

Status in OpenStack Compute (Nova):
  Incomplete
Status in Python client library for Nova:
  Confirmed

Bug description:
  The `nova network-create` is *silently* ignoring the --bridge when
  using the VlanManager. The bridge name is hardcoded to br%s % vlan.

  
https://github.com/openstack/nova/blob/stable/juno/nova/network/manager.py#L1361

  [root@jhenner-vmware ~(keystone_admin)]# nova network-create novanetwork 
--fixed-range-v4 192.168.32.0/22 --bridge br101 --bridge-interface br101 --uuid 
874ab2f1-c57d-464a-8a3f-5dda76ac7613 --vlan 102
  [root@jhenner-vmware ~(keystone_admin)]# nova network-show novanetwork
  +-+--+
  | Property| Value|
  +-+--+
  | bridge  | br102|
  | bridge_interface| br101|
  | broadcast   | 192.168.35.255   |
  | cidr| 192.168.32.0/22  |
  | cidr_v6 | -|
  | created_at  | 2015-03-10T20:25:41.00   |
  | deleted | False|
  | deleted_at  | -|
  | dhcp_server | 192.168.32.1 |
  | dhcp_start  | 192.168.32.3 |
  | dns1| 8.8.4.4  |
  | dns2| -|
  | enable_dhcp | True |
  | gateway | 192.168.32.1 |
  | gateway_v6  | -|
  | host| -|
  | id  | ca6b6f02-9f40-4578-89c6-9c8ca6d29bee |
  | injected| False|
  | label   | novanetwork  |
  | mtu | -|
  | multi_host  | True |
  | netmask | 255.255.252.0|
  | netmask_v6  | -|
  | priority| -|
  | project_id  | -|
  | rxtx_base   | -|
  | share_address   | False|
  | updated_at  | -|
  | vlan| 102  |
  | vpn_private_address | 192.168.32.2 |
  | vpn_public_address  | -|
  | vpn_public_port | 1000 |
  +-+--+


  It would be good to have the bridge name configurable when using
  VlanManager because when the nova vcenter driver is used, nova is
  checking whether there is a port group in the vcenter of same name
  as bridge used in nova-network.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1430521/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438020] Re: replace hpy.host with hyp['host']

2015-03-30 Thread Sean Dague
Not an actual bug

** Changed in: nova
   Status: In Progress = Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438020

Title:
  replace hpy.host with hyp['host']

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  class ComputeNode in 
https://github.com/openstack/nova/blob/master/nova/objects/compute_node.py#L27
  is extend from NovaObjectDictCompat in 
https://github.com/openstack/nova/blob/master/nova/objects/base.py#L652

  def __getitem__(self, name):
  For backwards-compatibility with dict-based objects.
  NOTE(danms): May be removed in the future.
  
  return getattr(self, name)

  as we can see above,function __getitem__ may be removed in the
  future.and now some codes use this function to get attribute from
  computenode object,

  hpy.host,
  
https://github.com/openstack/nova/blob/master/nova/api/openstack/compute/contrib/hypervisors.py#L92

  but the unit test data not,data in unit test is only a pure dict.

  
https://github.com/openstack/nova/blob/8b29068398b1217579aca07f2596ae88682d5055/nova/tests/unit/objects/test_compute_node.py

  In order to make it more consistent,change codes from obj.attr to
  obj['attr'].

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438020/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1394462] Re: nova show output should display sec-group id instead of name

2015-03-30 Thread Sean Dague
** No longer affects: nova

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1394462

Title:
  nova show output should display sec-group id instead of name

Status in Python client library for Nova:
  Confirmed

Bug description:

  As nova security-group-create allowing more than one group with the
  same name. showing security-group name in nova show is not correct

  I have two security groups with name default.

  ram@ubuntu:~$ neutron security-group-list
  +--+-+-+
  | id   | name| description |
  +--+-+-+
  | 2d3d1914-32d1-451f-b4bc-ed5eeda398ee | default | default |
  | c2d19dea-0863-40d5-872c-543f97b00bd4 | default | default |
  +--+-+-+

  In this case, I did not specify any security group while spawning the
  instance. I do not know how nova boot picks the default security group
  but nova show displays default in the security group.

  ram@ubuntu:~$ nova show first
  
+--++
  | Property | Value
  |
  
+--++
  | OS-DCF:diskConfig| MANUAL   
  |
  | OS-EXT-AZ:availability_zone  | nova 
  |
  | OS-EXT-SRV-ATTR:host | ubuntu   
  |
  | OS-EXT-SRV-ATTR:hypervisor_hostname  | ubuntu   
  |
  | OS-EXT-SRV-ATTR:instance_name| instance-0001
  |
  | OS-EXT-STS:power_state   | 1
  |
  | OS-EXT-STS:task_state| -
  |
  | OS-EXT-STS:vm_state  | active   
  |
  | OS-SRV-USG:launched_at   | 2014-11-19T18:42:00.00   
  |
  | OS-SRV-USG:terminated_at | -
  |
  | accessIPv4   |  
  |
  | accessIPv6   |  
  |
  | config_drive |  
  |
  | created  | 2014-11-19T18:41:15Z 
  |
  | flavor   | m1.tiny (1)  
  |
  | hostId   | 
4ac39bb970bb90f0aca2efaca1f43cc2997f6550a1449f08ade677af   |
  | id   | 8f7eb319-8f5d-46c2-bb1f-6a16838ef7b1 
  |
  | image| cirros-0.3.2-x86_64-uec 
(90aa74a4-138a-4a1a-a530-aa1cd4ee5e05) |
  | key_name | -
  |
  | metadata | {}   
  |
  | name | first
  |
  | os-extended-volumes:volumes_attached | []   
  |
  | private network  | 10.0.0.2 
  |
  | progress | 0
  |
  | security_groups  | default  
  | --- security group 
  | status   | ACTIVE   
  |
  | tenant_id| e747bc1a96ea4d88a0ddf7b2df8e0ad8 
  |
  | updated  | 2014-11-19T18:42:01Z 
  |
  | user_id  | 31a22dcf6b0a437294cb6c10f2996e08 
  |
  
+--++

  
  It leads to confusion in getting the security group id used by the instance..

To manage notifications about this bug go to:

[Yahoo-eng-team] [Bug 1434172] Re: security group create errors without description

2015-03-30 Thread Sean Dague
Right, optional means you are allowed to not provide the parameter. Not
that you can provide it as null. There is a python-novaclient bug here.

** Changed in: nova
   Status: Incomplete = Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1434172

Title:
  security group create errors without description

Status in OpenStack Compute (Nova):
  Won't Fix
Status in Python client library for Nova:
  In Progress
Status in OpenStack Command Line Client:
  In Progress

Bug description:
  security group create returns an error without --description supplied.
  This appears to be the server rejecting the request so we should set a
  default value rather than sending None.

    $ openstack security group create qaz
    ERROR: openstack Security group description is not a string or unicode 
(HTTP 400) (Request-ID: req-dee03de3-893a-4d58-bc3d-de87d09c3fb8)

  Sent body:

    {security_group: {name: qaz2, description: null}}

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1434172/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1326345] Re: Legacy networking (nova-network) in OpenStack Icehouse Installation results in no internet access for instances

2015-03-30 Thread Sean Dague
No longer reproducable, was fixed by the user

** Changed in: nova
   Status: Incomplete = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1326345

Title:
  Legacy networking (nova-network) in OpenStack Icehouse Installation
  results in no internet access for instances

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  I have performed installation of Openstack Icehouse on two CentOS6.5 machines 
from scratch (both the OS and Openstack).
  The problem is that the instances do not get internet access unless the 
following rule is added:

  iptables -t nat -I POSTROUTING -o  eth1 -j MASQUERADE

  where eth1 is the public interface. This rule has to be re-inserted
  every time an instance is launched!

  A few words regarding the setup of the compute node.

  -It has two ethernet interfaces eth0 (internal network) and eth1(public 
network).
  -I have added a bridge interface (br100) which has been bridged with eth0.
  -Both eth0 and eth1 are running in promiscuous mode
  -In sysctl.conf ip forwarding has been enabled
  -In nova.conf I have declared the flat_interface to be eth0 while the 
public_interface is eth1
  -I have followed the guide from here : 
http://docs.openstack.org/icehouse/install-guide/install/yum/content/   for 
both compute and controller nodes.
  -I have added the secgroup-rules for ICMP and SSH to the default group

  -I am launching a cirrOS instance at the default security group which is 
assigned the IP 10.0.0.2.
  -I can ping the cirrOS instance
  -I can ssh to the cirrOS
  -While logged in to the instance I can resolve external addresses (e.g. 
nslookup openstack.org produces a correct output)
  -While logged in to the instance I can ping 10.0.0.1
  -While logged in to the instance I can ping  the compute nodes external IP
  -While logged in to the instance I CANNOT ping the compute nodes gateway or 
anything else outside that!!!

  If I insert the afforementioned rule then pinging works as expected!

  In the case I have a second instance all the above apply but also I
  can ping and ssh from one instance to another. Unfortunately the
  problem of pinging outside the compute node from both instances
  remains until I re-insert the afforementioned rule.

  Tests have shown that every time I start an instance the rule has to
  be added at the top of the chain otherwise none of them works.

  The above are also the case when floating IPs are assigned to the
  instances. They cannot reach an external network unless the rule has
  been added. When it has instances can reach and be reached from an
  external network.

  Furthermore, comparison with an Havana installation with the same
  network configuration and topology (two nodes, two ethernet nics each
  etc.) revealed that the following chains are not populated correctly
  in Icehouse

   - Chain nova-network-OUTPUT
   - Chain nova-network-POSTROUTING
   - Chain nova-network-PREROUTING
   - Chain nova-network-float-snat

  
  I have been trying to find a solution to this through the mailing-lists:

  http://lists.openstack.org/pipermail/openstack/2014-May/007454.html
  http://lists.openstack.org/pipermail/openstack/2014-May/007515.html

  and in the IRC channel but without success.

  
  I have performed the above installations twice and every time the same 
problem appears leading me to think that it's a bug.

  Please check it!

  Best,

  G.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1326345/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438133] [NEW] django-admin.py collectstatic failing to find static/themes/default

2015-03-30 Thread Derek Higgins
Public bug reported:

At some stage over the last 10 days tripleo has stared to to fail to run

DJANGO_SETTINGS_MODULE=openstack_dashboard.settings django-admin.py
collectstatic --noinput

Exiting with the following error

os-collect-config: dib-run-parts Fri Mar 27 03:56:44 UTC 2015 Running 
/usr/libexec/os-refresh-config/post-configure.d/14-horizon
os-collect-config: WARNING:root:dashboards and default_dashboard in 
(local_)settings is DEPRECATED now and may be unsupported in some future 
release. The preferred way to specify the order of dashboards and the default 
dashboard is the pluggable dashboard mechanism (in 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/openstack_dashboard/enabled,
 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/openstack_dashboard/local/enabled).
os-collect-config: WARNING:py.warnings:DeprecationWarning: The oslo namespace 
package is deprecated. Please use oslo_config instead.
os-collect-config: WARNING:py.warnings:DeprecationWarning: The oslo namespace 
package is deprecated. Please use oslo_serialization instead. 
os-collect-config: Traceback (most recent call last):
os-collect-config:   File /opt/stack/venvs/openstack/bin/django-admin.py, 
line 5, in module
os-collect-config: management.execute_from_command_line()
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/core/management/__init__.py,
 line 399, in execute_from_command_line
os-collect-config: utility.execute()
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/core/management/__init__.py,
 line 392, in execute
os-collect-config: self.fetch_command(subcommand).run_from_argv(self.argv)
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/core/management/base.py,
 line 242, in run_from_argv
os-collect-config: self.execute(*args, **options.__dict__)
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/core/management/base.py,
 line 285, in execute
os-collect-config: output = self.handle(*args, **options)
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/core/management/base.py,
 line 415, in handle
os-collect-config: return self.handle_noargs(**options)
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/contrib/staticfiles/management/commands/collectstatic.py,
 line 173, in handle_noargs
os-collect-config: collected = self.collect()
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/contrib/staticfiles/management/commands/collectstatic.py,
 line 103, in collect
os-collect-config: for path, storage in finder.list(self.ignore_patterns):
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/contrib/staticfiles/finders.py,
 line 106, in list
os-collect-config: for path in utils.get_files(storage, ignore_patterns):
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/contrib/staticfiles/utils.py,
 line 25, in get_files
os-collect-config: directories, files = storage.listdir(location)   
os-collect-config:   File 
/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/django/core/files/storage.py,
 line 249, in listdir
os-collect-config: for entry in os.listdir(path):
os-collect-config: OSError: [Errno 2] No such file or directory: 
'/opt/stack/venvs/openstack/local/lib/python2.7/site-packages/openstack_dashboard/local/static/themes/default'
os-collect-config: [2015-03-27 03:56:45,548] (os-refresh-config) [ERROR] during 
post-configure phase. [Command '['dib-run-parts', 
'/usr/libexec/os-refresh-config/post-configure.d']' returned non-zero exit 
status


example here
http://logs.openstack.org/05/168205/1/check-tripleo/check-tripleo-ironic-undercloud-precise-nonha/3f384bd/logs/undercloud-undercloud_logs/os-collect-config.txt.gz#_Mar_27_03_56_45

** Affects: horizon
 Importance: Undecided
 Status: New

** Affects: tripleo
 Importance: Critical
 Status: Triaged


** Tags: ci

** Also affects: horizon
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438133

Title:
  django-admin.py collectstatic failing to find static/themes/default

Status in OpenStack Dashboard (Horizon):
  New
Status in tripleo - openstack on openstack:
  Triaged

Bug description:
  At some stage over the last 10 days tripleo has stared to to fail to
  run

  DJANGO_SETTINGS_MODULE=openstack_dashboard.settings django-
  admin.py collectstatic --noinput

  Exiting with the following error

  os-collect-config: dib-run-parts Fri Mar 27 03:56:44 UTC 2015 Running 

[Yahoo-eng-team] [Bug 1438113] Re: Use plain HTTP listeners in the conductor

2015-03-30 Thread Sean Dague
This is definitely way beyond a bug, it should be done as a nova-spec

** Changed in: nova
   Status: New = Invalid

** Changed in: nova
   Importance: Undecided = Wishlist

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438113

Title:
  Use plain HTTP listeners in the conductor

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  The conductor is consuming messages form single queue which has performance 
limitation due to various reasons.:
  - per queue lock
  - Some broker also limiting same part of the message handling to single CPU 
thread/queue
  - Multiple broker instances needs to synchronise to queue content, which 
causes additional delays die to the tcp request/response times

  The single queue limitation is much greater than the limits getting by
  single mysql server, the rate is even worse when you consider slave
  reads.

  This can be workarounded by explicitly or implicit distributing the
  rpc calls to multiple different queue.

  The message broker provides additional message durability properties which is 
not needed just for an rpc_call,
  we spend resource on what we actually do not need.

  For TCP/HTTP traffic load balancing we have many-many tools even hardware 
assisted options are available providing virtually unlimited scalability.
  At TCP level also possible to exclude the loadbalancer node(s) form the 
response traffic.

  Why HTTP?
  Basically any protocol which can do request/response `thing` with arbitrary  
type and size of data with keep-alive connection and with ssl option, could be 
used.
  HTTP is a simple and well know protocol, with already existing many-many load 
balancing tool.

  Why not have the agents to do a regular API call?
  The regular API calls needs to do policy check, which in this case is not 
required, every authenticated user can be considered as admin.

  The  the conductor clients needs to use at least a single shared key 
configured in every nova host.
  It has similar security as openstack used with the brokers, basically all 
nova node had credentials in one rabbitmq virtual host,
  configured in the /etc/nova/nova.conf . If any of those credentials stolen it 
provided access to the whole virtual host.

  NOTE.: HTTPs can be used with certificate or kerberos based
  authentication as well.

  I think the for `rpc_calls` which are served by the agents using
  something like AMQP is still the better option, this bug is just about
  the situation when the conductor itself serves  rpc_call(s).

  NOTE.: The 1 Million msq/sec rabbitmq benchmark is done 186 queues, in
  a way which does not hits the single queue limitations.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438113/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438154] [NEW] Creating metadef object without properties raises 500 error

2015-03-30 Thread Kamil Rykowski
Public bug reported:

When you try to create a metadef object without any properties
associated with it you get an Internal Server Error 500:

glance md-object-create --name obj_name --schema '{description: some
desc}' OS::Software::DBMS

AttributeError: 'NoneType' object has no attribute 'iteritems'

The object itself is created successfully, but the notification is not
fired.

** Affects: glance
 Importance: Undecided
 Assignee: Kamil Rykowski (kamil-rykowski)
 Status: In Progress


** Tags: metadef

** Changed in: glance
 Assignee: (unassigned) = Kamil Rykowski (kamil-rykowski)

** Changed in: glance
   Status: New = In Progress

** Tags added: metadef

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1438154

Title:
  Creating metadef object without properties raises 500 error

Status in OpenStack Image Registry and Delivery Service (Glance):
  In Progress

Bug description:
  When you try to create a metadef object without any properties
  associated with it you get an Internal Server Error 500:

  glance md-object-create --name obj_name --schema '{description:
  some desc}' OS::Software::DBMS

  AttributeError: 'NoneType' object has no attribute 'iteritems'

  The object itself is created successfully, but the notification is not
  fired.

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1438154/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1026621] Re: nova-network gets release_fixed_ip events from someplace, but the database still keeps them associated with instances

2015-03-30 Thread Sean Dague
*** This bug is a duplicate of bug 1231254 ***
https://bugs.launchpad.net/bugs/1231254

** This bug has been marked a duplicate of bug 1231254
   Windows instances need timezone to be localtime

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1026621

Title:
  nova-network gets release_fixed_ip events from someplace, but the
  database still keeps them associated with instances

Status in OpenStack Compute (Nova):
  Triaged
Status in nova package in Ubuntu:
  Triaged

Bug description:
  We've been seeing a lot of instances simply vanish from the network.
  Usually people have been willing to work around this by simply
  rebooting or re-creating their instances, but it's troubling for long-
  running instances (especially those that have volumes associated).

  Here's the relevant bit of nova-network.log for one of these:

  2012-07-16 14:06:32 DEBUG nova.rpc.amqp [-] received {u'_context_roles': 
[u'admin'], u'_context_request_id': 
u'req-d0905711-c4d1-4452-a3b2-46815d1983d7', u'_context_read_deleted': u'no', 
u'args': {u'address': u'10.55.60.141'}, u'_context_auth_token': 'SANITIZED', 
u'_context_is_admin': True, u'_context_project_id': None, 
u'_context_timestamp': u'2012-07-16T14:06:32.169100', u'_context_user_id': 
None, u'method': u'release_fixed_ip', u'_context_remote_address': None} from 
(pid=493) _safe_log /usr/lib/python2.7/dist-packages/nova/rpc/common.py:160
  2012-07-16 14:06:32 DEBUG nova.rpc.amqp 
[req-d0905711-c4d1-4452-a3b2-46815d1983d7 None None] unpacked context: 
{'user_id': None, 'roles': [u'admin'], 'timestamp': 
'2012-07-16T14:06:32.169100', 'auth_token': 'SANITIZED', 'remote_address': 
None, 'is_admin': True, 'request_id': 
u'req-d0905711-c4d1-4452-a3b2-46815d1983d7', 'project_id': None, 
'read_deleted': u'no'} from (pid=493) _safe_log 
/usr/lib/python2.7/dist-packages/nova/rpc/common.py:160
  2012-07-16 14:06:32 DEBUG nova.network.manager 
[req-d0905711-c4d1-4452-a3b2-46815d1983d7 None None] Released IP |10.55.60.141| 
from (pid=493) release_fixed_ip 
/usr/lib/python2.7/dist-packages/nova/network/manager.py:1260

  Then the dhcpbridge shows it being revoked:

  2012-07-16 14:04:29 DEBUG nova.dhcpbridge [-] Called 'old' for mac 
'fa:16:3e:11:c5:37' with ip '10.55.60.141' from (pid=23699) main 
/usr/bin/nova-dhcpbridge:113
  2012-07-16 14:06:32 DEBUG nova.dhcpbridge [-] Called 'del' for mac 
'fa:16:3e:11:c5:37' with ip '10.55.60.141' from (pid=24946) main 
/usr/bin/nova-dhcpbridge:113

  Is there any way we can find out what might have placed the
  release_fixed_ip event on the message queue?  There doesn't seeem to
  be any other mention of the IP in the nova logs on any of our systems.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1026621/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1398361] Re: Add support for EC2 API : ec2-modify-snapshot-attribute

2015-03-30 Thread Sean Dague
Moving to Opinion as EC2 in tree API is deprecated.

** Changed in: nova
   Status: Triaged = Opinion

** Tags added: ec2

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1398361

Title:
  Add support for EC2 API : ec2-modify-snapshot-attribute

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  The various attributes of the snapshots created from volumes can be
  modified for making them accessible to public or to a particular user
  account.

  The response parameters and supported attribute sets must be according the 
AWS specification:
  
http://docs.aws.amazon.com/AWSEC2/latest/CommandLineReference/ApiReference-cmd-ModifySnapshotAttribute.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1398361/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1398357] Re: Add support for EC2 API : ec2-reset-snapshot-attribute

2015-03-30 Thread Sean Dague
Moving to Opinion as EC2 in tree API is deprecated.

** Tags added: ec2

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1398357

Title:
  Add support for EC2 API : ec2-reset-snapshot-attribute

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Provide the implementation similar to the Amazon EC2 API

  ec2-reset-snapshot-attribute

  for making the volume based snapshot instances accessible only to the
  respective account holder.

  http://docs.aws.amazon.com/AWSEC2/latest/CommandLineReference
  /ApiReference-cmd-ResetSnapshotAttribute.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1398357/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1338679] Re: Some Neutron plugins might miss network-vif-* events

2015-03-30 Thread Sean Dague
I don't think this is really a nova side issue. Polling from nova seems
really odd here.

** Changed in: nova
   Status: Triaged = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1338679

Title:
  Some Neutron plugins might miss network-vif-* events

Status in OpenStack Neutron (virtual network service):
  Incomplete
Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  This has been observed with the VMware NSX plugin, but in theory this
  issue might occur with every Neutron plugin when the notifications to
  nova are enabled.

  In a nutshell the issue is that nova needs to receive a network-vif-
  plugged even from neutron in order to be able to boot an instance. On
  the other hand in some cases VIF unplug/plug events might be fairly
  quick, and thus the neutron side might not be aware at all of the
  status change and not send any event.

  As a consequence, the instance will not boot even if its VIF are
  correctly plugged.

  This bug is currently being assigned both neutron and nova because:
  - it might deemed a plugin problem. If the plugin is not smart enough to 
handle VIF plug/unplug notifications than it's not worth using the notification 
mechanism exposed by nova
  - the nova side might add an optional poll mechanism which could run 
alongside the mechanism waiting for an event and periodically poll neutron to 
get the VIF status.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1338679/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1279055] Re: performance issue on restart of nova compute when there are a large number of vms

2015-03-30 Thread Sean Dague
This is kind of working as designed. The concurency mode of n-cpu kind
of needs to work this way.

** Changed in: nova
   Status: Triaged = Won't Fix

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1279055

Title:
  performance issue on restart of nova compute when there are  a large
  number of vms

Status in OpenStack Compute (Nova):
  Won't Fix

Bug description:
  When there are large number of VMs in the cloud, there is a
  performance issue on restarting of nova compute. Compute server tries
  to sync states of all the vm and does not respond until sync is
  complete.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1279055/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1218994] Re: file based disk images do not get scrubbed on delete

2015-03-30 Thread Sean Dague
Marking as Opinion because the solution here really needs to be
encryption at rest so that a crashed nova compute doesn't leave customer
data out there.

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1218994

Title:
  file based disk images do not get scrubbed on delete

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Right now, LVM backed instances can be scrubbed (overwritten with
  zeros using dd) upon deletion.  However, there is no such option with
  file backed images.  While it is true that fallocate can handle some
  of this by returning 0s to the instance when reading any unwritten
  parts of the file, there are some cases where it is not desirable to
  enable fallocate.

  What would be preferred would be a similar the options cinder has
  implemented, so the operator can choose to shred or zero out the file,
  based on their organizations own internal data policies.   A zero out
  option satisfies those that must ensure they scrub tenant data upon
  deletion, and shred would satisfy those beholden to DoD 5220-22.

  This would of course make file backed disks vulnerable to
  https://bugs.launchpad.net/nova/+bug/889299 but that might not be a
  bad thing considering its quite old.

  Attached an initial patch for nova/virt/libvirt/driver.py that
  performs the same LVM zero scrub routine to disk backed files, however
  it lacks any flags to enable or disable it right now.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1218994/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1330873] Re: Instance name set in horizon is not set in VMware vCenter

2015-03-30 Thread Sean Dague
I think the point of using UUID for instance names is probably spot on.
Marking as Opinion, because it seems like for consistency it should be
done that way.

** Changed in: nova
   Status: In Progress = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1330873

Title:
  Instance name set in horizon is not set in VMware vCenter

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Instance name set in horizon is not set in VMware vCenter. In vCenter
  the name of the instance is the id (in UUID format) of the instance.
  This makes it difficult for the administrator to locate the instance
  in vCenter using the name displayed in horizon.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1330873/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438245] [NEW] Nova accesses image for resize even when resize is disabled

2015-03-30 Thread Alexander Schmidt
Public bug reported:

The resize_fs_using_block_device config option allows to configure if a
filesystem within an image is resized by nova during spawn or not. When
it's set to False (which is the default) nova still tries to access the
image to find out if it's extendable or not, although it's not resized
afterwards.

On system which don't have libguestfs or nbd installed, this causes the
following error messages repeatedly:

2015-03-30 16:49:40.326 ERROR nova.virt.disk.mount.nbd 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] nbd module not loaded
2015-03-30 16:49:40.327 INFO nova.virt.disk.mount.api 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] Device allocation failed. 
Will retry in 2 seconds.
2015-03-30 16:49:42.329 ERROR nova.virt.disk.mount.nbd 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] nbd module not loaded
2015-03-30 16:49:42.330 INFO nova.virt.disk.mount.api 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] Device allocation failed. 
Will retry in 2 seconds.
2015-03-30 16:49:44.332 ERROR nova.virt.disk.mount.nbd 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] nbd module not loaded
2015-03-30 16:49:44.333 INFO nova.virt.disk.mount.api 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] Device allocation failed. 
Will retry in 2 seconds.

So the instance spawn takes longer than necessary (retries are done for
~30 seconds).

I'll provide a patch to make sure the image is not accessed when
resizing is disabled.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438245

Title:
   Nova accesses image for resize even when resize is disabled

Status in OpenStack Compute (Nova):
  New

Bug description:
  The resize_fs_using_block_device config option allows to configure if
  a filesystem within an image is resized by nova during spawn or not.
  When it's set to False (which is the default) nova still tries to
  access the image to find out if it's extendable or not, although it's
  not resized afterwards.

  On system which don't have libguestfs or nbd installed, this causes
  the following error messages repeatedly:

  2015-03-30 16:49:40.326 ERROR nova.virt.disk.mount.nbd 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] nbd module not loaded
  2015-03-30 16:49:40.327 INFO nova.virt.disk.mount.api 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] Device allocation failed. 
Will retry in 2 seconds.
  2015-03-30 16:49:42.329 ERROR nova.virt.disk.mount.nbd 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] nbd module not loaded
  2015-03-30 16:49:42.330 INFO nova.virt.disk.mount.api 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] Device allocation failed. 
Will retry in 2 seconds.
  2015-03-30 16:49:44.332 ERROR nova.virt.disk.mount.nbd 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] nbd module not loaded
  2015-03-30 16:49:44.333 INFO nova.virt.disk.mount.api 
[req-3d3c5a50-6067-442a-9607-1f98141d795c None None] Device allocation failed. 
Will retry in 2 seconds.

  So the instance spawn takes longer than necessary (retries are done
  for ~30 seconds).

  I'll provide a patch to make sure the image is not accessed when
  resizing is disabled.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438245/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1379682] Re: instance_template_name does not work for vmware driver

2015-03-30 Thread Gary Kotton
*** This bug is a duplicate of bug 1330873 ***
https://bugs.launchpad.net/bugs/1330873

** This bug has been marked a duplicate of bug 1330873
   Instance name set in horizon is not set in VMware vCenter

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1379682

Title:
  instance_template_name does not work for vmware driver

Status in OpenStack Compute (Nova):
  Confirmed
Status in OpenStack Manuals:
  Fix Released

Bug description:
  Currently vmware driver will adopt uuid for instance names. This will
  lead two problems:

  1) instance name template will not apply for vmware driver.  But it
  will display instance name with nova show command. It will be
  misleading.

  [root@cmwo cmwo]# nova show temp-vm-host1-99
  
+--+---+
  | Property | Value
 |
  
+--+---+
  | OS-DCF:diskConfig| MANUAL   
 |
  | OS-EXT-AZ:availability_zone  | cluster01
 |
  | OS-EXT-SRV-ATTR:host | vcenter-cluster01
 |
  | OS-EXT-SRV-ATTR:hypervisor_hostname  | domain-c129(cluster01)   
 |
  | OS-EXT-SRV-ATTR:instance_name| instance-00ec
 |
  | OS-EXT-STS:power_state   | 1
 |
  | OS-EXT-STS:task_state| -
 |
  | OS-EXT-STS:vm_state  | active   
 |
  | OS-SRV-USG:launched_at   | 2014-10-09T10:02:15.00   
 |
  | OS-SRV-USG:terminated_at | -
 |
  | accessIPv4   |  
 |
  | accessIPv6   |  
 |
  | config_drive |  
 |
  | created  | 2014-10-09T09:59:15Z 
 |
  | demovlan network | 25.0.0.21
 |
  | flavor   | m1.tiny (1)  
 |
  | hostId   | 
0cfba386e1e5cad832d2fbc316c33ca6a124c3d0b386127a55707070  |
  | id   | 6d3e0f11-0a4c-46eb-a3a6-4e397d917228 
 |

  2) This uuid is not user friendly for VM names.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1379682/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438260] [NEW] Expensive Device allocation loop when NBD is absent

2015-03-30 Thread Davanum Srinivas (DIMS)
Public bug reported:

See log stash query:
Device allocation failed. Will retry in 2 seconds.

Example:
http://logs.openstack.org/10/163910/6/check/gate-rally-dsvm-cinder/46d89ab/logs/screen-n-cpu.txt.gz#_2015-03-30_14_55_56_941

Logs:
2015-03-30 14:55:56.940 ERROR nova.virt.disk.mount.nbd 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] nbd module not loaded
2015-03-30 14:55:56.940 DEBUG oslo_concurrency.lockutils 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Lock nbd-allocation-lock 
released by _inner_get_dev :: held 0.000s inner 
/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:456
2015-03-30 14:55:56.941 INFO nova.virt.disk.mount.api 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Device allocation failed. 
Will retry in 2 seconds.
2015-03-30 14:55:58.941 DEBUG oslo_concurrency.lockutils 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Lock nbd-allocation-lock 
acquired by _inner_get_dev :: waited 0.000s inner 
/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:444
2015-03-30 14:55:58.941 ERROR nova.virt.disk.mount.nbd 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] nbd module not loaded

We should not even enter the loop when /sys/block/nbd0 is clearly
absent.

** Affects: nova
 Importance: Undecided
 Assignee: Davanum Srinivas (DIMS) (dims-v)
 Status: In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438260

Title:
  Expensive Device allocation loop when NBD is absent

Status in OpenStack Compute (Nova):
  In Progress

Bug description:
  See log stash query:
  Device allocation failed. Will retry in 2 seconds.

  Example:
  
http://logs.openstack.org/10/163910/6/check/gate-rally-dsvm-cinder/46d89ab/logs/screen-n-cpu.txt.gz#_2015-03-30_14_55_56_941

  Logs:
  2015-03-30 14:55:56.940 ERROR nova.virt.disk.mount.nbd 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] nbd module not loaded
  2015-03-30 14:55:56.940 DEBUG oslo_concurrency.lockutils 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Lock nbd-allocation-lock 
released by _inner_get_dev :: held 0.000s inner 
/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:456
  2015-03-30 14:55:56.941 INFO nova.virt.disk.mount.api 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Device allocation failed. 
Will retry in 2 seconds.
  2015-03-30 14:55:58.941 DEBUG oslo_concurrency.lockutils 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Lock nbd-allocation-lock 
acquired by _inner_get_dev :: waited 0.000s inner 
/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:444
  2015-03-30 14:55:58.941 ERROR nova.virt.disk.mount.nbd 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] nbd module not loaded

  We should not even enter the loop when /sys/block/nbd0 is clearly
  absent.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438260/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1202965] Re: The name VCPUs (total) of Hypervisors is confusing

2015-03-30 Thread Sean Dague
This is really a UX feature addition

** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1202965

Title:
  The name VCPUs (total) of Hypervisors is confusing

Status in OpenStack Dashboard (Horizon):
  Invalid
Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  In Hypervisors panel, VCPUs(total) and VCPUs(used) fields causes
  confusion as used always bigger than total.

  Virtual CPU to Physical CPU allocation ratio is default to 16.0.  put
  Physical CPU total to VCPUS(total) is not correct.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1202965/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1257788] Re: nova-network doesn't handle large amount of concurrent booting instances well

2015-03-30 Thread Joe Gordon
This has been fixed by some of the work to move nova-net to objects.

** Changed in: nova
   Status: Triaged = Fix Released

** Changed in: nova
 Assignee: (unassigned) = Dan Smith (danms)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1257788

Title:
  nova-network doesn't handle large amount of concurrent booting
  instances well

Status in OpenStack Neutron (virtual network service):
  Invalid
Status in OpenStack Compute (Nova):
  Fix Released

Bug description:
  As seen by the large-ops test nova-network has trouble booting 150 VMs
  concurrently. Part of the problem is that setting up the network for
  each instance is relatively expensive with many locks and root-wrapped
  calls. Both of which significantly hurt performance.  One possible fix
  would be to do more bulk network operations when possible, so instead
  of 100 separate calls to a root-wrapped tool call it once.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1257788/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp



[Yahoo-eng-team] [Bug 1438259] [NEW] VPNaaS: remove dependency on Neutron in UT

2015-03-30 Thread Paul Michali
Public bug reported:

Some of the driver unit tests mock execute() from Neutron, which creates
a (brittle) dependency (as seen by a recent breakage).

The dependency can be removed, but mocking higher up, from within the
*Swan process _execute() method.

In addition, there are some unit tests that are not active, due to
missing test base class.

** Affects: neutron
 Importance: Undecided
 Assignee: Paul Michali (pcm)
 Status: In Progress


** Tags: vpnaas

** Changed in: neutron
 Assignee: (unassigned) = Paul Michali (pcm)

** Changed in: neutron
   Status: New = In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438259

Title:
  VPNaaS: remove dependency on Neutron in UT

Status in OpenStack Neutron (virtual network service):
  In Progress

Bug description:
  Some of the driver unit tests mock execute() from Neutron, which
  creates a (brittle) dependency (as seen by a recent breakage).

  The dependency can be removed, but mocking higher up, from within the
  *Swan process _execute() method.

  In addition, there are some unit tests that are not active, due to
  missing test base class.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438259/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438240] [NEW] nova libvirt driver does not handle resizing with 'hw:cpu_policy' flag

2015-03-30 Thread Stephen Finucane
*** This bug is a duplicate of bug 1417667 ***
https://bugs.launchpad.net/bugs/1417667

Public bug reported:

We conduced a series of resize checks on instances with/without the CPU
pinning parameters provided as part of this blueprint:

http://specs.openstack.org/openstack/nova-specs/specs/juno/approved
/virt-driver-cpu-pinning.html

This is related to 1417667
(https://bugs.launchpad.net/nova/+bug/1417667).

---

# Testing Configuration

Testing was conducted on a single-node, Fedora 21-based
(3.17.8-300.fc21.x86_64) OpenStack instance (built with devstack). Three
flavors were created:

* no-pinning   ('hw:cpu_policy=shared')
* pinning  ('hw:cpu_policy=dedicated' and 'hw:numa-nodes=1')
* pinning-no-numa  ('hw:cpu_policy=dedicated')

# Results

## no pinning - pinning

Doesn't work. Attempting to do this results in little difference in the
output XML.

Before:

vcpu placement='static' cpuset='10-19,30-39'10/vcpu

After:

vcpu placement='static' cpuset='0-9,20-29'10/vcpu

STEP: Create a VM with a 'no-pinning' flavor. Dump XML with virsh.
Resize and confirm to 'pinning' flavor. Dump XML with virsh.

## no pinning - pinning (without available CPUs)

Doesn't work (due to above). Pinning does not occur so there is no
reason for failure.

STEPS: Create three 10-vcpu VMs with a 'pinning' flavor. Create two
10-vcpu VMs with a 'no-pinning' flavor. Resize and confirm one of latter
to 'pinning' flavor. Ensure failure.

## pinning w/ no numa topology - pinning with numa topology

Doesn't work. There is only one (unrelated) change in the output XML:

Before:

disk type='file' device='disk'
...
backingStore type='file' index='1'
format type='raw'/
source 
file='/opt/stack/data/nova/instances/_base/598e2b66ae396bb852a504ea4a7471df6b847f8b'/
backingStore/
/backingStore

After:

disk type='file' device='disk'
...
backingStore/

STEPS: Create a VM with a 'pinning-no-numa' flavor. Dump XML with virsh.
Resize and confirm to 'pinning' flavor. Dump XML with virsh.

## pinning - no pinning

Doesn't work. Attempting to do this results in no serious change in the
output XML. Here's the important elements of the XML for a resized
(pinning - no pinning) instance:

vcpu placement='static'10/vcpu
cputune
shares10240/shares
vcpupin vcpu='0' cpuset='17'/
vcpupin vcpu='1' cpuset='37'/
vcpupin vcpu='2' cpuset='18'/
vcpupin vcpu='3' cpuset='38'/
vcpupin vcpu='4' cpuset='19'/
vcpupin vcpu='5' cpuset='39'/
vcpupin vcpu='6' cpuset='33'/
vcpupin vcpu='7' cpuset='13'/
vcpupin vcpu='8' cpuset='34'/
vcpupin vcpu='9' cpuset='14'/
emulatorpin cpuset='13-14,17-19,33-34,37-39'/
/cputune
numatune
memory mode='strict' nodeset='1'/
memnode cellid='0' mode='strict' nodeset='1'/
/numatune
cpu
topology sockets='5' cores='1' threads='2'/
numa
cell id='0' cpus='0-9' memory='4194304' unit='KiB'/
/numa
/cpu

Versus the unresized, no pinning instance:

  vcpu placement='static' cpuset='10-19,30-39'10/vcpu
  cputune
shares10240/shares
  /cputune
  cpu
topology sockets='10' cores='1' threads='1'/
  /cpu

** Affects: nova
 Importance: Undecided
 Status: New


** Tags: libvirt

** Description changed:

  We conduced a series of resize checks on instances with/without the CPU
  pinning parameters provided as part of this blueprint:
  
- http://specs.openstack.org/openstack/nova-specs/specs/juno/approved
+ http://specs.openstack.org/openstack/nova-specs/specs/juno/approved
  /virt-driver-cpu-pinning.html
  
  ---
  
  # Testing Configuration
  
  Testing was conducted on a single-node, Fedora 21-based
  (3.17.8-300.fc21.x86_64) OpenStack instance (built with devstack). Three
  flavors were created:
  
  * no-pinning   ('hw:cpu_policy=shared')
  * pinning  ('hw:cpu_policy=dedicated' and 'hw:numa-nodes=1')
  * pinning-no-numa  ('hw:cpu_policy=dedicated')
  
  # Results
  
  ## no pinning - pinning
  
  Doesn't work. Attempting to do this results in little difference in the
  output XML.
  
  Before:
  
- vcpu placement='static' cpuset='10-19,30-39'10/vcpu
+ vcpu placement='static' cpuset='10-19,30-39'10/vcpu
  
  After:
  
- vcpu placement='static' cpuset='0-9,20-29'10/vcpu
+ vcpu placement='static' cpuset='0-9,20-29'10/vcpu
  
  STEP: Create a VM with a 'no-pinning' flavor. Dump XML with virsh.
  Resize and confirm to 'pinning' flavor. Dump XML with virsh.
  
  ## no pinning - pinning (without available CPUs)
  
  Doesn't work (due to above). Pinning does not occur so there is no
  reason for failure.
  
  STEPS: Create three 10-vcpu VMs with a 'pinning' flavor. Create two
  10-vcpu VMs with a 'no-pinning' flavor. Resize and confirm one of latter
  to 'pinning' flavor. Ensure failure.
  
  ## pinning w/ no numa topology - pinning 

[Yahoo-eng-team] [Bug 1438253] [NEW] nova libvirt pinning won't work across numa nodes

2015-03-30 Thread Stephen Finucane
Public bug reported:

Using a CPU policy of dedicated ('hw:cpu_policy=dedicated') results in
vCPUs being pinned to pCPUs, per the original blueprint:

http://specs.openstack.org/openstack/nova-
specs/specs/kilo/implemented/virt-driver-cpu-pinning.html

When combined with the 'vcpu_pin_set' nova configuration, it should be
possible to get very good control over which pCPUs are used by
instances. This works as expected when creating a multi-vCPU instance
pinned to pCPUs on a single NUMA node. However, if these pCPUs are
spread across multiple NUMA nodes then it fails. This behavior is not
documented in the original blueprint and therefore should not function
like so.

---

# Testing Configuration

Testing was conducted on a single-node, Fedora 21-based
(3.17.8-300.fc21.x86_64) OpenStack instance (built with devstack). Two
flavors were used:

openstack flavor create --ram 4096 --disk 20 --vcpus 10 demo.no-
pinning

openstack flavor create --ram 4096 --disk 20 --vcpus 10 demo.pinning
nova flavor-key demo.pinning set hw:cpu_policy=dedicated 
hw:cpu_threads_policy=separate

# Results

Passing (same NUMA socket, contiguous)
==

vcpu_pin_set = 10-19

Pinned
--

$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 68150's current affinity list: 10-19
$ sudo -E virsh vcpucount 13
maximum  config10
maximum  live  10
current  config10
current  live  10

Unpinned


$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.no-pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 70249's current affinity list: 10-19

Passing (same NUMA socket, non-contiguous)
==

vcpu_pin_set = 5-9,25-29

Pinned
--

$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 91186's current affinity list: 5-9,25-29

Unpinned


$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.no-pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 92212's current affinity list: 5-9,25-29

Passing (multiple NUMA sockets, non-contiguous)
===

vcpu_pin_set = 5-10,25-29

Pinned
--

$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 93884's current affinity list: 5-9,25-29

Unpinned


$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.no-pinning test1 --wait
$ ps aux | grep qemu | awk '{print $2}' | head -1 | xargs taskset -pc
pid 94468's current affinity list: 5-9,25-29

Additional combinations
---

vcpu_pin_set = 5-14,25-29


Failing (different NUMA socket, contiguous)
===

vcpu_pin_set = 9-18

Pinned
--

$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.pinning test1 --wait
ERROR: openstackclient.compute.v2.server.CreateServer Error creating 
server: test1

Unpinned


$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.no-pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 71194's current affinity list: 9-18

Failing (different NUMA socket, non-contiguous)
===

vcpu_pin_set = 5-9,15-19

Pinned
--

$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.no-pinning test1 --wait
ERROR: openstackclient.compute.v2.server.CreateServer Error creating 
server: test1
$ openstack server show test1 | grep fault
| fault| {u'message': u'No valid host 
was found. There are not enough hosts available.', u'code': 500,

Unpinned


$ openstack server create --image Fedora-x86_64-20-20140618-sda 
--flavor demo.no-pinning test1 --wait
$ ps aux | grep qemu | awk '{print }' | head -1 | xargs taskset -pc
pid 88703's current affinity list: 5-9,15-19

** Affects: nova
 Importance: Undecided
 Status: New


** Tags: libvirt

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438253

Title:
  nova libvirt pinning won't work across numa 

[Yahoo-eng-team] [Bug 1438321] [NEW] Fix process management for neutron-server

2015-03-30 Thread Elena Ezhova
Public bug reported:

The following commit to oslo-incubator [1], that was supposed to optimize 
waiting for children processes
to exit, will break neutron-server behavior (i.e. signal handling).

1. In neutron-server eventlet monkey-patching (including patching os
module) is done in parent process. That is why os.waitpid(0, 0) in
_wait_child method also gets monkey-patched and eventlet goes crazy.
Connecting to parent process with strace shows that os.waitpid(0,
os.WNOHANG) is called, yet it is difficult to say what is really
happening because the process does not react on termination signals
(SIGTERM, SIGHUP, SIGINT).

2. Due to the fact that neutron-server initializes two instances of
ProcessLauncher in one parent process, calling
eventlet.greenthread.sleep(self.wait_interval) seems to be the only way
for the process to switch contexts and allow another instance of
ProcessLauncher to call _wait_child. It is important to mention that
ProcessLauncher is not supposed to be used in this way (2 instances in
one parent process) at all.

This bug is intended to track fixing the outlined problems on Neutron
side.

[1] https://github.com/openstack/oslo-
incubator/commit/bf92010cc9d4c2876eaf6092713aafa94dcc8b35

** Affects: neutron
 Importance: Undecided
 Assignee: Elena Ezhova (eezhova)
 Status: New

** Changed in: neutron
 Assignee: (unassigned) = Elena Ezhova (eezhova)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438321

Title:
  Fix process management for neutron-server

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  The following commit to oslo-incubator [1], that was supposed to optimize 
waiting for children processes
  to exit, will break neutron-server behavior (i.e. signal handling).

  1. In neutron-server eventlet monkey-patching (including patching os
  module) is done in parent process. That is why os.waitpid(0, 0) in
  _wait_child method also gets monkey-patched and eventlet goes
  crazy. Connecting to parent process with strace shows that
  os.waitpid(0, os.WNOHANG) is called, yet it is difficult to say what
  is really happening because the process does not react on termination
  signals (SIGTERM, SIGHUP, SIGINT).

  2. Due to the fact that neutron-server initializes two instances of
  ProcessLauncher in one parent process, calling
  eventlet.greenthread.sleep(self.wait_interval) seems to be the only
  way for the process to switch contexts and allow another instance of
  ProcessLauncher to call _wait_child. It is important to mention that
  ProcessLauncher is not supposed to be used in this way (2 instances in
  one parent process) at all.

  This bug is intended to track fixing the outlined problems on Neutron
  side.

  [1] https://github.com/openstack/oslo-
  incubator/commit/bf92010cc9d4c2876eaf6092713aafa94dcc8b35

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438321/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438320] [NEW] Subnet pool created should be blocked when allow_overlapping_ips=False

2015-03-30 Thread Ryan Tidwell
Public bug reported:

Creation of subnet pools should be blocked when
allow_overlapping_ips=False. This conflicts with the notion of subnet
pools and causes allocation of overlapping prefixes to be blocked, even
when allocating across different pools.  The simplest solution is to
declare subnet pools incompatible with allow_overlapping_ips=False and
block creation of subnet pools.

** Affects: neutron
 Importance: Undecided
 Assignee: Ryan Tidwell (ryan-tidwell)
 Status: New

** Changed in: neutron
 Assignee: (unassigned) = Ryan Tidwell (ryan-tidwell)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438320

Title:
  Subnet pool created should be blocked when allow_overlapping_ips=False

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  Creation of subnet pools should be blocked when
  allow_overlapping_ips=False. This conflicts with the notion of subnet
  pools and causes allocation of overlapping prefixes to be blocked,
  even when allocating across different pools.  The simplest solution is
  to declare subnet pools incompatible with allow_overlapping_ips=False
  and block creation of subnet pools.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438320/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438325] [NEW] Access and Security tab slowness

2015-03-30 Thread Aleksandr Shaposhnikov
Public bug reported:

Access and Security tab became very unresponsive and may time out on 
environments with more that 3k VMs.
Deep analisys shown that it because of Floating IPs tab even if there is no 
floating IPs assigned :).
It looks like code on this tab is suboptimal and trying to get informaton about 
all VMs/ports to use thoose further for assignment/associacion procedure. 
Anyway with more that 3k VMs Access and Security page becae very slow and 
horizon starts to throw timeouts very often.
I've checked that behavior on Havana and it looks like code didn't changed a 
lot since then so filing it for master.

** Affects: horizon
 Importance: Undecided
 Status: New

** Affects: mos
 Importance: Undecided
 Status: New

** Also affects: mos
   Importance: Undecided
   Status: New

** Description changed:

- Access and Security tab became very unresponcive and may time out on 
environments with more that 3k VMs.
- Deep analisys shown that it because of Floating IPs tab even if there is no 
floating IPs assigned :). 
+ Access and Security tab became very unresponsive and may time out on 
environments with more that 3k VMs.
+ Deep analisys shown that it because of Floating IPs tab even if there is no 
floating IPs assigned :).
  It looks like code on this tab is suboptimal and trying to get informaton 
about all VMs/ports to use thoose further for assignment/associacion procedure. 
Anyway with more that 3k VMs Access and Security page becae very slow and 
horizon starts to throw timeouts very often.
  I've checked that behavior on Havana and it looks like code didn't changed a 
lot since then so filing it for master.

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438325

Title:
  Access and Security tab slowness

Status in OpenStack Dashboard (Horizon):
  New
Status in Mirantis OpenStack:
  New

Bug description:
  Access and Security tab became very unresponsive and may time out on 
environments with more that 3k VMs.
  Deep analisys shown that it because of Floating IPs tab even if there is no 
floating IPs assigned :).
  It looks like code on this tab is suboptimal and trying to get informaton 
about all VMs/ports to use thoose further for assignment/associacion procedure. 
Anyway with more that 3k VMs Access and Security page becae very slow and 
horizon starts to throw timeouts very often.
  I've checked that behavior on Havana and it looks like code didn't changed a 
lot since then so filing it for master.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438325/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438331] [NEW] Nova fails to delete rbd image, puts guest in to ERROR state

2015-03-30 Thread Joe Talerico
Public bug reported:

When removing guests  that have been booted on Ceph, Nova will
occasionally put guests in to ERROR state with the following ...

Reported to the controller:

| fault| {message: error removing image, 
code: 500, details:   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 314, in 
decorated_function |
|  | return function(self, context, 
*args, **kwargs)
   |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 2525, in 
terminate_instance |
|  | do_terminate_instance(instance, 
bdms)   
  |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/openstack/common/lockutils.py\, line 
272, in inner|
|  | return f(*args, **kwargs)  

   |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 2523, in 
do_terminate_instance  |
|  | 
self._set_instance_error_state(context, instance)   
  |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/openstack/common/excutils.py\, line 
82, in __exit__   |
|  | six.reraise(self.type_, 
self.value, self.tb)
  |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 2513, in 
do_terminate_instance  |
|  | self._delete_instance(context, 
instance, bdms, quotas) 
   |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/hooks.py\, line 131, in inner  
   |
|  | rv = f(*args, **kwargs)

   |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 2482, in 
_delete_instance   |
|  | quotas.rollback()  

   |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/openstack/common/excutils.py\, line 
82, in __exit__   |
|  | six.reraise(self.type_, 
self.value, self.tb)
  |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 2459, in 
_delete_instance   |
|  | self._shutdown_instance(context, 
instance, bdms) 
 |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/compute/manager.py\, line 2389, in 
_shutdown_instance |
|  | requested_networks)

   |
|  |   File 
\/usr/lib/python2.7/site-packages/nova/openstack/common/excutils.py\, line 
82, in __exit__   |
|  | six.reraise(self.type_, 
self.value, 

[Yahoo-eng-team] [Bug 1438338] [NEW] servers api should return security group ids instead of names

2015-03-30 Thread melanie witt
Public bug reported:

Creating this from https://bugs.launchpad.net/python-
novaclient/+bug/1394462

In nova-network security group names can't be duplicated, but in
neutron, they can. For this reason, it would be nice to return server
security groups as ids instead of names.

Here's is a sample request and response showing the current state:
security_groups: [{name: default}],

DEBUG (connectionpool:415) GET 
/v2/038c717809174199a297f4ef774e6852/servers/d2b729b8-a626-4050-a756-d5a450c99811
 HTTP/1.1 200 1757
DEBUG (session:223) RESP: [200] date: Mon, 30 Mar 2015 17:57:50 GMT 
content-length: 1757 content-type: application/json x-compute-request-id: 
req-d6c33e18-cf62-4848-88ff-e57b64bd55e3
RESP BODY: {server: {status: ACTIVE, updated: 2015-03-25T19:04:47Z, 
hostId: cfca2250a844c76f4dd5ba369b2550ad3fb07d545e58e395b2271486, 
OS-EXT-SRV-ATTR:host: vagrant-ubuntu-trusty-64.localdomain, addresses: 
{private: [{OS-EXT-IPS-MAC:mac_addr: fa:16:3e:0d:0c:b3, version: 4, 
addr: 10.0.0.2, OS-EXT-IPS:type: fixed}]}, links: [{href: 
http://10.0.2.15:8774/v2/038c717809174199a297f4ef774e6852/servers/d2b729b8-a626-4050-a756-d5a450c99811;,
 rel: self}, {href: 
http://10.0.2.15:8774/038c717809174199a297f4ef774e6852/servers/d2b729b8-a626-4050-a756-d5a450c99811;,
 rel: bookmark}], key_name: null, image: {id: 
d0ddfda2-dbdc-48ae-b65e-27ca407d32ce, links: [{href: 
http://10.0.2.15:8774/038c717809174199a297f4ef774e6852/images/d0ddfda2-dbdc-48ae-b65e-27ca407d32ce;,
 rel: bookmark}]}, OS-EXT-STS:task_state: null, OS-EXT-STS:vm_state: 
active, OS-EXT-SRV-ATTR:instance_name: instance-0001, 
OS-SRV-USG:launched_at: 2015-03-25
 T19:04:47.00, OS-EXT-SRV-ATTR:hypervisor_hostname: 
vagrant-ubuntu-trusty-64.localdomain, flavor: {id: 42, links: 
[{href: http://10.0.2.15:8774/038c717809174199a297f4ef774e6852/flavors/42;, 
rel: bookmark}]}, id: d2b729b8-a626-4050-a756-d5a450c99811, 
security_groups: [{name: default}], OS-SRV-USG:terminated_at: null, 
OS-EXT-AZ:availability_zone: nova, user_id: 
0771ff29994b428fa15dfb4ec1b6bc7d, name: 
ServerActionsTestJSON-instance-367700261, created: 2015-03-25T19:04:35Z, 
tenant_id: a0c8d64c558b42d5a7b32a229c9f9a3e, OS-DCF:diskConfig: MANUAL, 
os-extended-volumes:volumes_attached: [], accessIPv4: , accessIPv6: , 
progress: 0, OS-EXT-STS:power_state: 1, config_drive: True, metadata: 
{}}}

** Affects: nova
 Importance: Wishlist
 Status: Confirmed


** Tags: api

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438338

Title:
  servers api should return security group ids instead of names

Status in OpenStack Compute (Nova):
  Confirmed

Bug description:
  Creating this from https://bugs.launchpad.net/python-
  novaclient/+bug/1394462

  In nova-network security group names can't be duplicated, but in
  neutron, they can. For this reason, it would be nice to return server
  security groups as ids instead of names.

  Here's is a sample request and response showing the current state:
  security_groups: [{name: default}],

  DEBUG (connectionpool:415) GET 
/v2/038c717809174199a297f4ef774e6852/servers/d2b729b8-a626-4050-a756-d5a450c99811
 HTTP/1.1 200 1757
  DEBUG (session:223) RESP: [200] date: Mon, 30 Mar 2015 17:57:50 GMT 
content-length: 1757 content-type: application/json x-compute-request-id: 
req-d6c33e18-cf62-4848-88ff-e57b64bd55e3
  RESP BODY: {server: {status: ACTIVE, updated: 2015-03-25T19:04:47Z, 
hostId: cfca2250a844c76f4dd5ba369b2550ad3fb07d545e58e395b2271486, 
OS-EXT-SRV-ATTR:host: vagrant-ubuntu-trusty-64.localdomain, addresses: 
{private: [{OS-EXT-IPS-MAC:mac_addr: fa:16:3e:0d:0c:b3, version: 4, 
addr: 10.0.0.2, OS-EXT-IPS:type: fixed}]}, links: [{href: 
http://10.0.2.15:8774/v2/038c717809174199a297f4ef774e6852/servers/d2b729b8-a626-4050-a756-d5a450c99811;,
 rel: self}, {href: 
http://10.0.2.15:8774/038c717809174199a297f4ef774e6852/servers/d2b729b8-a626-4050-a756-d5a450c99811;,
 rel: bookmark}], key_name: null, image: {id: 
d0ddfda2-dbdc-48ae-b65e-27ca407d32ce, links: [{href: 
http://10.0.2.15:8774/038c717809174199a297f4ef774e6852/images/d0ddfda2-dbdc-48ae-b65e-27ca407d32ce;,
 rel: bookmark}]}, OS-EXT-STS:task_state: null, OS-EXT-STS:vm_state: 
active, OS-EXT-SRV-ATTR:instance_name: instance-0001, 
OS-SRV-USG:launched_at: 2015-03-
 25T19:04:47.00, OS-EXT-SRV-ATTR:hypervisor_hostname: 
vagrant-ubuntu-trusty-64.localdomain, flavor: {id: 42, links: 
[{href: http://10.0.2.15:8774/038c717809174199a297f4ef774e6852/flavors/42;, 
rel: bookmark}]}, id: d2b729b8-a626-4050-a756-d5a450c99811, 
security_groups: [{name: default}], OS-SRV-USG:terminated_at: null, 
OS-EXT-AZ:availability_zone: nova, user_id: 
0771ff29994b428fa15dfb4ec1b6bc7d, name: 
ServerActionsTestJSON-instance-367700261, created: 2015-03-25T19:04:35Z, 
tenant_id: a0c8d64c558b42d5a7b32a229c9f9a3e, OS-DCF:diskConfig: MANUAL, 
os-extended-volumes:volumes_attached: [], accessIPv4: , 

[Yahoo-eng-team] [Bug 1437516] Re: Instances with delete_on_terminate volumes do not reschedule

2015-03-30 Thread Vish Ishaya
*** This bug is a duplicate of bug 1336127 ***
https://bugs.launchpad.net/bugs/1336127

Thanks andrew, marked this as a duplicate

** This bug has been marked a duplicate of bug 1336127
   The volumes will be deleted when  creating a virtual machine fails with the 
parameter delete_on_termination being set true, which causes that the 
rescheduling fails

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1437516

Title:
  Instances with delete_on_terminate volumes do not reschedule

Status in OpenStack Compute (Nova):
  Triaged

Bug description:
  If you create an instance with attached delete_on_terminate volumes
  and it fails to launch on a host (due to a hardware or hypervisor
  issue) it will be rescheduled but the volumes will have already been
  deleted. The volumes should only be deleted if the instance is not
  going to be rescheduled (or perhaps not at all).

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1437516/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438159] Re: Made neutron agents silent by using AMQP

2015-03-30 Thread Eugene Nikanorov
This looks like deep rework of current messaging strategy and probably
should be worked on within a scope of a blueprint rather than a bug.

So I would suggest to file a bp for this and provide a spec explaining
these ideas so spec review could be a better place to discuss these
ideas.

** Changed in: neutron
   Status: New = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438159

Title:
  Made neutron agents silent by using AMQP

Status in OpenStack Neutron (virtual network service):
  Opinion

Bug description:
  Problem.: Neutron agents does a lot of periodic task which leads  an rpc call 
+ database transaction, which does not even provide a new information, because 
nothing changed.
  This behaviour in scale can be called as `DDOS attack`, generally this kind 
of architecture is bad at scaling and can be considered as an any-pattern.

  Instead of periodic poll, we can leverage the AMQP brokers bind capabilities. 
  Neutron has many situation like security group rule change or dvr related 
changes which needs to be communicated with multiple agents, but usually not 
with all agent.

  The agent at startup needs to synchronise the as usual, but during the
  sync the agent can subscribe to the interesting events to avoid the
  periodic tasks. (Note.: After the first subscribe loop a second one is
  needed to do not miss changes during the subscribe process ).

  The AMQP queues with 'auto-delete' can be considered as a reliable source of 
information which does not miss any event notification. 
  On connection loss or full broker cluster die the agent needs to re sync 
everything guarded in this way,
  in these cases, the queue will disappear so the situation easily detectable.

  1. Create a Direct exchange for all kind of resourcestype what needs
  to be synchronised in this way, for example.: 'neutron.securitygroups'
  . The exchange declaration needs to happen at q-svc start-up time or
  at full broker cluster die (not-found exception will tell it). The
  exchange SHOULD NOT be redeclared or verified at every message
  publish.

  2. Every agent creates a dedicated per agent queue with auto-delete flag, if 
the agent already maintains a queue with this property he MAY reuse that one. 
The agents SHOULD avoid to creating multiple queues per resource type. The 
messages MUST contain a type information. 
  3. All agent creates a binding between his queue and the resource type queue 
with he realise he depends on the resource, for example it maintains at least 
one port with the given security-group. (The agents needs to remove the 
binding. when they stop using it)
  4. The q-svc publishes just a single message  when the resource related 
change happened. The routing key is the uuid.

  Alternatively a topic exchange can be used, with a single  exchange.
  In this case the routing keys MUST contains the resource type like: 
neutron.resource_type.uuid ,
  this type exchange is generally more expensive than a direct exchange 
(pattern matching), and only useful if you have agents which needs to listens 
to ALL event related to a type, but others just interested just in a few of 
them.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438159/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438329] [NEW] Example configuration files lack changes for Kilo

2015-03-30 Thread Matt Kassawara
Public bug reported:

The example configuration files in the source repository lack the
changes in the list of new, updated, and deprecated options in the Kilo
configuration reference guide [1]. For example, the example neutron.conf
file [2] lacks the [nova] and [oslo_messaging_*] sections and options.

[1] 
http://docs.openstack.org/trunk/config-reference/content/neutron-conf-changes-kilo.html
[2] https://github.com/openstack/neutron/blob/master/etc/neutron.conf

** Affects: neutron
 Importance: Undecided
 Status: New

** Description changed:

  The example configuration files in the source repository lack the
  changes in the list of new, updated, and deprecated options in the Kilo
- configuration reference guide [2]. For example, the example neutron.conf
- file [1] lacks the [nova] and [oslo_messaging_*] sections and options.
+ configuration reference guide [1]. For example, the example neutron.conf
+ file [2] lacks the [nova] and [oslo_messaging_*] sections and options.
  
- [1] https://github.com/openstack/neutron/blob/master/etc/neutron.conf
- [2] 
http://docs.openstack.org/trunk/config-reference/content/neutron-conf-changes-kilo.html
+ [1] 
http://docs.openstack.org/trunk/config-reference/content/neutron-conf-changes-kilo.html
+ [2] https://github.com/openstack/neutron/blob/master/etc/neutron.conf

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438329

Title:
  Example configuration files lack changes for Kilo

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  The example configuration files in the source repository lack the
  changes in the list of new, updated, and deprecated options in the
  Kilo configuration reference guide [1]. For example, the example
  neutron.conf file [2] lacks the [nova] and [oslo_messaging_*] sections
  and options.

  [1] 
http://docs.openstack.org/trunk/config-reference/content/neutron-conf-changes-kilo.html
  [2] https://github.com/openstack/neutron/blob/master/etc/neutron.conf

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438329/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1231226] Re: nova show shows wrong flavor name

2015-03-30 Thread Sean Dague
With the moving of instance_type information into the instance metadata
this should be addressed moving forward. There will be some legacy
instances that don't yet have this, but moving forward it's fixed.

** Changed in: nova
   Status: Triaged = Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1231226

Title:
  nova show shows wrong flavor name

Status in OpenStack Compute (Nova):
  Fix Released

Bug description:
  We created a flavor with flavor id 1, then deleted the flavor and
  created a flavor with the same id.  OK, then I create a instance with
  flavor 1, instance's resource is ok, but nova show shows the deleted
  flavor name.

  we can fix this by 3 ways: 1. reject create a flavor with flavor id
  which is same as one of deleted ones.  2. flavor show should not read
  deleted flavors.  3. actually, system meta has flavor name, so nova
  can return flavor name to client, and client use it directly.

  I am not sure which way will be accepted. Any one has idea on this?

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1231226/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438316] [NEW] Glance does not notify systemd when the services are up.

2015-03-30 Thread Alberto Murillo
Public bug reported:

Unlike keystone, glance does not notify systemd when the
api|registry|scrubber services are up.

root@controller:~# time systemctl start glance-api
Job for glance-api.service failed. See systemctl status glance-api.service 
and journalctl -xe for details.

real1m30.310s
user0m0.003s
sys 0m0.002s

The service actually starts but its status is not visible to systemd.

root@controller:~# systemctl status glance-api
● glance-api.service - OpenStack Image API Service
   Loaded: loaded (/usr/lib/systemd/system/glance-api.service; enabled; vendor 
preset: disabled)
   Active: activating (start) since Mon 2015-03-30 17:02:33 UTC; 41s ago
 Main PID: 810 (glance-api)
   CGroup: /system.slice/glance-api.service
   ├─810 /usr/bin/python2 /usr/bin/glance-api
   ├─829 /usr/bin/python2 /usr/bin/glance-api
   └─830 /usr/bin/python2 /usr/bin/glance-api

Mar 30 17:02:33 controller systemd[1]: Starting OpenStack Image API Service...
root@controller:~# echo $?
3

root@controller:~# ps aux |grep glance
glance 638  4.2  1.9 196948 78876 ?Rs   16:51   0:01 
/usr/bin/python2 /usr/bin/glance-api

Attached is a patch that resolves this issue, tested against 2015.1.0b3

** Affects: glance
 Importance: Undecided
 Status: New


** Tags: systemd

** Patch added: 0001-Enable-systemd-notification.patch
   
https://bugs.launchpad.net/bugs/1438316/+attachment/4361213/+files/0001-Enable-systemd-notification.patch

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1438316

Title:
  Glance does not notify systemd when the services are up.

Status in OpenStack Image Registry and Delivery Service (Glance):
  New

Bug description:
  Unlike keystone, glance does not notify systemd when the
  api|registry|scrubber services are up.

  root@controller:~# time systemctl start glance-api
  Job for glance-api.service failed. See systemctl status glance-api.service 
and journalctl -xe for details.

  real1m30.310s
  user0m0.003s
  sys 0m0.002s

  The service actually starts but its status is not visible to systemd.

  root@controller:~# systemctl status glance-api
  ● glance-api.service - OpenStack Image API Service
 Loaded: loaded (/usr/lib/systemd/system/glance-api.service; enabled; 
vendor preset: disabled)
 Active: activating (start) since Mon 2015-03-30 17:02:33 UTC; 41s ago
   Main PID: 810 (glance-api)
 CGroup: /system.slice/glance-api.service
 ├─810 /usr/bin/python2 /usr/bin/glance-api
 ├─829 /usr/bin/python2 /usr/bin/glance-api
 └─830 /usr/bin/python2 /usr/bin/glance-api

  Mar 30 17:02:33 controller systemd[1]: Starting OpenStack Image API Service...
  root@controller:~# echo $?
  3

  root@controller:~# ps aux |grep glance
  glance 638  4.2  1.9 196948 78876 ?Rs   16:51   0:01 
/usr/bin/python2 /usr/bin/glance-api

  Attached is a patch that resolves this issue, tested against
  2015.1.0b3

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1438316/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438260] Re: Expensive Device allocation loop when NBD is absent

2015-03-30 Thread Sean Dague
Nova is being asked to do a thing, which it can't, which is put files in
the guest. This retry is actually kind of important.

The bug really is exposing the fact that *no* working file injection
mechanism exists in these rally jobs, but they are asking guests to  do
something.

** Changed in: nova
   Status: In Progress = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438260

Title:
  Expensive Device allocation loop when NBD is absent

Status in OpenStack Compute (Nova):
  Invalid

Bug description:
  See log stash query:
  Device allocation failed. Will retry in 2 seconds.

  Example:
  
http://logs.openstack.org/10/163910/6/check/gate-rally-dsvm-cinder/46d89ab/logs/screen-n-cpu.txt.gz#_2015-03-30_14_55_56_941

  Logs:
  2015-03-30 14:55:56.940 ERROR nova.virt.disk.mount.nbd 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] nbd module not loaded
  2015-03-30 14:55:56.940 DEBUG oslo_concurrency.lockutils 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Lock nbd-allocation-lock 
released by _inner_get_dev :: held 0.000s inner 
/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:456
  2015-03-30 14:55:56.941 INFO nova.virt.disk.mount.api 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Device allocation failed. 
Will retry in 2 seconds.
  2015-03-30 14:55:58.941 DEBUG oslo_concurrency.lockutils 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] Lock nbd-allocation-lock 
acquired by _inner_get_dev :: waited 0.000s inner 
/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py:444
  2015-03-30 14:55:58.941 ERROR nova.virt.disk.mount.nbd 
[req-df13155c-0340-4e10-aec4-2fabea382eef None None] nbd module not loaded

  We should not even enter the loop when /sys/block/nbd0 is clearly
  absent.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438260/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1418635] Re: Neutron API reference is inconsistent and differs from implementation

2015-03-30 Thread Eugene Nikanorov
** Also affects: openstack-api-site
   Importance: Undecided
   Status: New

** No longer affects: neutron

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1418635

Title:
  Neutron API reference is inconsistent and differs from implementation

Status in OpenStack API documentation site:
  New

Bug description:
  I'm implementing modules for SaltStack (not yet on GitHub) to create entities 
in OpenStack Neutron and came across quite a few problems 
  with the API documentation:

  * Link Filtering and Column Selection to 
http://docs.openstack.org/api/openstack-network/2.0/content/filtering.html
    404s

  * Section http://developer.openstack.org/api-ref-
  networking-v2.html#subnets

    * POST /v2.0/subnets aka Create subnet
  - 'id' is listed to be an optional parameter but the Neutron-API in 
Icehouse refuses to set a user-selected ID
  - parameters 'dns_nameservers' and 'host_routes' missing from 
documentation (undocumented extensions?)

    * GET /v2.0/subnets aka List subnets
  - can't filter by allocation_pools or enable_dhcp

    * PUT /v2.0/subnets/​{subnet_id}​ aka Update subnet
  - parameters allocation_pools, network_id, tenant_id, id listed 
as optional request parameters but Neutron-API in Icehouse returns 
Cannot update read-only attribute $PARAMETER

  * Section http://developer.openstack.org/api-ref-
  networking-v2.html#networks

    * GET /v2.0/networks aka List networks
  - parameter shared is ignored as a filter

  * Section http://developer.openstack.org/api-ref-
  networking-v2.html#layer3

    * POST /v2.0/routers aka Create router
  - Description states router:external and external_gateway_info are 
valid request parameters but they're not listed in the table of 
request parameters
  - What's the parameter router described as A router object. supposed 
to be? A router object in JSON/XML notation inside a router object in 
JSON/XML notation?

  I'll probably add more when implementing functions for managing
  routers in Neutron.

To manage notifications about this bug go to:
https://bugs.launchpad.net/openstack-api-site/+bug/1418635/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1292999] Re: Agents APIv2 do not handle conflict errors

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Incomplete

** Changed in: nova
   Status: Incomplete = Opinion

** Changed in: nova
   Importance: Low = Wishlist

** Changed in: nova
 Assignee: Leandro Ignacio Costantino (leandro-i-costantino) = (unassigned)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1292999

Title:
  Agents APIv2 do not handle conflict errors

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  When trying to create an agent with duplicate name, a generic
  exception is raised instead of HttpConflict.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1292999/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1274732] Re: Instances table doesn't get cleaned up

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1274732

Title:
  Instances table doesn't get cleaned up

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  It seems like nova doesn't do any cleanup of the instances table. If
  you spin up and delete lots of VMs, this table eventually gets
  gigantic.

  Manually it can be cleaned using: `nova-manage db archive_deleted_rows
  1` but an auto-cleanup of stale entries would be useful too.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1274732/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1435655] Re: Can't manually assign a distributed router to a l3 agent

2015-03-30 Thread Armando Migliaccio
case 1: I think that creating a DVR router connecting Subnets that reach
VM's whose compute hosts are not DVR aware should lead to an error

case 2: I think that moving agents willy nilly should lead can't be
expected to work. Moving DVR routers from one agent to another has
always been a bit murky, but I believe that you can only move the SNAT
part; the DVR elements will have to stick to the hosts on which VM's
connected through the routers are.

** Changed in: neutron
   Status: In Progress = Invalid

** Changed in: neutron
   Status: Invalid = Incomplete

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1435655

Title:
  Can't manually assign a distributed router to a l3 agent

Status in OpenStack Neutron (virtual network service):
  Incomplete

Bug description:
  Now neutron does not  allow to manually assign a distributed router to a l3 
agent which is in 'dvr' mode, but in bellow use case, it does not work ok:
  1 case:
  (1)there are two computeA, B nodes which l3 agent are in legacy mode, l2 
agent 'enable_distributed_routing = False'
  (2)create a 'dvr' router, then add subnetA to this 'dvr' router
  (3)create VMs with subnetA  in computeA or B
  (4)modify  'agent_mode=dvr',  'enable_distributed_routing = True' in computeA
  (5)the VMs in  computeA  can't communicate with their gateway

  2 case:
  (1)there is a computeA,  it's 'agent_mode=dvr',  'enable_distributed_routing 
= True'
  (2)create a 'dvr' router, then add subnetA to this 'dvr' router
  (3)create VMs with subnetA  in computeA
  (4)use 'l3-agent-router-remove' remove l3 agent which in computeA from 'dvr' 
router
  (5)the VMs in computeA  can't communicate with their gateway, and can't 
manually assign it's l3 agent to dvr router

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1435655/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438325] Re: Access and Security tab slowness

2015-03-30 Thread Timur Sufiev
** Tags added: horizon

** No longer affects: horizon

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438325

Title:
  Access and Security tab slowness

Status in Mirantis OpenStack:
  New

Bug description:
  Access and Security tab became very unresponsive and may time out on 
environments with more that 3k VMs.
  Deep analisys shown that it because of Floating IPs tab even if there is no 
floating IPs assigned :).
  It looks like code on this tab is suboptimal and trying to get informaton 
about all VMs/ports to use thoose further for assignment/associacion procedure. 
Anyway with more that 3k VMs Access and Security page becae very slow and 
horizon starts to throw timeouts very often.
  I've checked that behavior on Havana and it looks like code didn't changed a 
lot since then so filing it for master.

To manage notifications about this bug go to:
https://bugs.launchpad.net/mos/+bug/1438325/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1089343] Re: Resize confirm should raise NotFound instead of BadRequest when confirming an invalid migration

2015-03-30 Thread Sean Dague
** Changed in: nova
   Status: Triaged = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1089343

Title:
  Resize confirm should raise NotFound instead of BadRequest when
  confirming an invalid migration

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  I believe that when the user tries to confirm an invalid migration NotFound 
(cause the migration is not found) should be raised instead of badRequest
  Happens in: 
https://github.com/openstack/nova/blob/master/nova/api/openstack/compute/servers.py#L1002

  If this bug is confirmed it should be merged only in API-V3

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1089343/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1248022] Re: Nova scheduler not updated immediately when a baremetal node is added or removed

2015-03-30 Thread Sean Dague
This is a much bigger architecture issue to address.

** Changed in: nova
   Status: Triaged = Invalid

** Changed in: nova
   Status: Invalid = Opinion

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1248022

Title:
  Nova scheduler not updated immediately when a baremetal node is added
  or removed

Status in OpenStack Bare Metal Provisioning Service (Ironic):
  Triaged
Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  With the Ironic driver, if a baremetal node is added/deleted, it is
  not removed from pool of available resources until the next run of
  update_available_resource(). During this window, the scheduler may
  continue to attempt to schedule instances on it (when deleted), or
  report NoValidHosts (when added) leading to unnecessary failures and
  scheduling retries.

  In compute manager, the update_available_resource() periodic task is
  responsible for updating the scheduler's knowledge of baremetal nodes:

  @periodic_task.periodic_task
  def update_available_resource(self, context):
  ...
  nodenames = set(self.driver.get_available_nodes())
  for nodename in nodenames:
  rt = self._get_resource_tracker(nodename)
  rt.update_available_resource(context)

  update_available_resource() is also called at service startup

  This means that you have to wait up to 60 seconds for a node to become
  available/no longer available.

To manage notifications about this bug go to:
https://bugs.launchpad.net/ironic/+bug/1248022/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438315] [NEW] add_instance_fault_from_exc expects object but isn't always passed one

2015-03-30 Thread Andrew Laski
Public bug reported:

2015-03-30 15:02:58.287 28424 DEBUG oslo_messaging._drivers.amqp [-] UNIQUE_ID 
is 9c2e286750e7407cb78e273becca4ab0. _add_unique_id 
/usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqp.py:224
Traceback (most recent call last):
  File /usr/local/lib/python2.7/dist-packages/eventlet/hubs/poll.py, line 
115, in wait
listener.cb(fileno)
  File /opt/stack/nova/nova/utils.py, line 888, in wrapper
return func(*args, **kwargs)
  File /opt/stack/nova/nova/conductor/manager.py, line 688, in build_instances
exc, request_spec)
  File /opt/stack/nova/nova/conductor/manager.py, line 594, in 
_set_vm_state_and_notify
ex, request_spec, self.db)
  File /opt/stack/nova/nova/scheduler/utils.py, line 107, in 
set_vm_state_and_notify
new_ref, ex, sys.exc_info())
  File /opt/stack/nova/nova/compute/utils.py, line 87, in 
add_instance_fault_from_exc
fault_obj.instance_uuid = instance.uuid
AttributeError: 'dict' object has no attribute 'uuid'

** Affects: nova
 Importance: High
 Status: New

** Changed in: nova
   Importance: Undecided = High

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1438315

Title:
  add_instance_fault_from_exc expects object but isn't always passed one

Status in OpenStack Compute (Nova):
  New

Bug description:
  2015-03-30 15:02:58.287 28424 DEBUG oslo_messaging._drivers.amqp [-] 
UNIQUE_ID is 9c2e286750e7407cb78e273becca4ab0. _add_unique_id 
/usr/local/lib/python2.7/dist-packages/oslo_messaging/_drivers/amqp.py:224
  Traceback (most recent call last):
File /usr/local/lib/python2.7/dist-packages/eventlet/hubs/poll.py, line 
115, in wait
  listener.cb(fileno)
File /opt/stack/nova/nova/utils.py, line 888, in wrapper
  return func(*args, **kwargs)
File /opt/stack/nova/nova/conductor/manager.py, line 688, in 
build_instances
  exc, request_spec)
File /opt/stack/nova/nova/conductor/manager.py, line 594, in 
_set_vm_state_and_notify
  ex, request_spec, self.db)
File /opt/stack/nova/nova/scheduler/utils.py, line 107, in 
set_vm_state_and_notify
  new_ref, ex, sys.exc_info())
File /opt/stack/nova/nova/compute/utils.py, line 87, in 
add_instance_fault_from_exc
  fault_obj.instance_uuid = instance.uuid
  AttributeError: 'dict' object has no attribute 'uuid'

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1438315/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1425639] Re: A functional floating ip stops working sometimes if another floating ip is created and deleted.

2015-03-30 Thread Rajeev Grover
The code structure has changed significantly and haven't heard of any
other citing of this specific issue for a while so marking it as invalid
for now.

** Changed in: neutron
   Status: In Progress = Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1425639

Title:
  A functional floating ip stops working sometimes if another floating
  ip is created and deleted.

Status in OpenStack Neutron (virtual network service):
  Invalid

Bug description:
  
  A test failure was noticed for test_cross_tenant_traffic at [1] . 
  A floating ip ‘a’ was instantiated on router A .
  The floating IP was functional
  A floating IP ‘b’ was instantiated on router B after about 30 seconds
  Floating IP on router B was functional
  Floating IP on router B was deleted
  Floating IP on A was not reachable anymore

  Root cause appears to be:

  Analyzing from the logs it appears Agent gateway port was being created at 
instantiation of both the floating IPs and deleted on deletion of Floating IP 
‘b’. 
  [2] , [3] , [4]
  Expected behavior was to have Agent Gateway port created only once and not 
deleted on deletion of Floating IP ‘b’. 
  Creation and deletion of Agent Gateway port is guarded by the is_first and 
is_last checks on the subscribers set.  
  The subscribers set is a member of the FipNamespace Class
  FipNamespace objects are stored as weakreferences 

  It appears the  weakreference to the FipNamespace object created at
  the time of instantiation of floating IP ‘a’ was lost/garbage
  collected before instantiation of floating IP ‘b’ and therefore a new
  object of class Fipnamespace was instantiated which came along with a
  new subscribers set member. This in turn led to passing of the
  is_first and is_last checks for router B thus leading to failures.

  [1] http://logs.openstack.org/22/153422/8/check/check-tempest-dsvm-
  neutron-dvr/3db7309/logs/testr_results.html.gz

  [2] http://logs.openstack.org/22/153422/8/check/check-tempest-dsvm-
  neutron-dvr/3db7309/logs/screen-q-vpn.txt.gz#_2015-02-24_02_20_16_355

  [3] http://logs.openstack.org/22/153422/8/check/check-tempest-dsvm-
  neutron-dvr/3db7309/logs/screen-q-vpn.txt.gz#_2015-02-24_02_22_10_735

  [4] http://logs.openstack.org/22/153422/8/check/check-tempest-dsvm-
  neutron-dvr/3db7309/logs/screen-q-vpn.txt.gz#_2015-02-24_02_21_57_960

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1425639/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438351] [NEW] Angular table hzSelectAll should select all visible checkboxes in table

2015-03-30 Thread Kelly Domico
Public bug reported:

Currently, if the select-all checkbox is clicked, all rows are selected
even if it's not visible. For example, when a table is paginated or
filtered.

The select-all checkbox should only select all the visible checkboxes in
the table.

** Affects: horizon
 Importance: Undecided
 Assignee: Kelly Domico (kelly-domico)
 Status: New

** Changed in: horizon
 Assignee: (unassigned) = Kelly Domico (kelly-domico)

** Description changed:

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438351

Title:
  Angular table hzSelectAll should select all visible checkboxes in
  table

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  Currently, if the select-all checkbox is clicked, all rows are
  selected even if it's not visible. For example, when a table is
  paginated or filtered.

  The select-all checkbox should only select all the visible checkboxes
  in the table.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438351/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1334923] Re: No need to backup instance when 'rotation' equals to 0

2015-03-30 Thread Sean Dague
This optimization has been awaiting someone to add unit tests for 4
months. Calling it Opinion not In Progress.

** Changed in: nova
   Status: In Progress = Confirmed

** Changed in: nova
   Status: Confirmed = Opinion

** Changed in: nova
 Assignee: wingwj (wingwj) = (unassigned)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1334923

Title:
  No need to backup instance when 'rotation' equals to 0

Status in OpenStack Compute (Nova):
  Opinion

Bug description:
  Nova will delete all backup images according the 'backup_type'
  parameter when 'rotation' equals to 0.

  But according the logic now, Nova will generate one new backup, and
  then delete it..

  So It's weird to snapshot a useless backup firstly IMO.
  We need to add one new branch here: if 'rotation' is equal to 0, no need to 
backup, just rotate it.

  The related discussion can be found here:
  http://lists.openstack.org/pipermail/openstack-dev/2014-June/038760.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1334923/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438426] [NEW] Functional job setup broken by devstack

2015-03-30 Thread Maru Newby
Public bug reported:

A recent change to devstack [1] removed lib/neutron which the functional
testing setup script depends on.  The script should be updated to refer
to lib/neutron-legacy instead.


1: https://review.openstack.org/#/c/168591/

** Affects: neutron
 Importance: Critical
 Assignee: Maru Newby (maru)
 Status: New

** Changed in: neutron
   Importance: Undecided = Critical

** Changed in: neutron
 Assignee: (unassigned) = Maru Newby (maru)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1438426

Title:
  Functional job setup broken by devstack

Status in OpenStack Neutron (virtual network service):
  New

Bug description:
  A recent change to devstack [1] removed lib/neutron which the
  functional testing setup script depends on.  The script should be
  updated to refer to lib/neutron-legacy instead.

  
  1: https://review.openstack.org/#/c/168591/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1438426/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438474] [NEW] [Launch Instance Fix] Adding unit tests to hz.dashboard

2015-03-30 Thread Shaoquan Chen
Public bug reported:

Angular module hz.dashboard should have unit tests.

** Affects: horizon
 Importance: Undecided
 Assignee: Shaoquan Chen (sean-chen2)
 Status: In Progress

** Changed in: horizon
 Assignee: (unassigned) = Shaoquan Chen (sean-chen2)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1438474

Title:
  [Launch Instance Fix] Adding unit tests to hz.dashboard

Status in OpenStack Dashboard (Horizon):
  In Progress

Bug description:
  Angular module hz.dashboard should have unit tests.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1438474/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1438469] [NEW] auth_version config not used

2015-03-30 Thread Tushar Kalra
Public bug reported:

When auth_version is set to be v3.0 in glance-api.conf, I see this in
the logs Auth Token proceeding with requested v3.0 apis but in the
very next debug log I see that the authentication request actually
appends v2.0 to the identity url, which seems incorrect.

2015-03-31 00:52:13.928 254 INFO keystonemiddleware.auth_token [-] Auth Token 
proceeding with requested v3.0 apis
2015-03-31 00:52:13.928 254 DEBUG keystoneclient.auth.identity.v2 [-] Making 
authentication request to https://keystone server:35357/v2.0/tokens 
get_auth_ref 
/usr/lib/python2.7/site-packages/keystoneclient/auth/identity/v2.py:76

Relevant config from glance-api.conf:

[keystone_authtoken]
identity_uri = https://keystone server:35357
admin_user = glance
admin_password =admin_password
auth_version = v3.0

** Affects: keystone
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Keystone.
https://bugs.launchpad.net/bugs/1438469

Title:
  auth_version config not used

Status in OpenStack Identity (Keystone):
  New

Bug description:
  When auth_version is set to be v3.0 in glance-api.conf, I see this in
  the logs Auth Token proceeding with requested v3.0 apis but in the
  very next debug log I see that the authentication request actually
  appends v2.0 to the identity url, which seems incorrect.

  2015-03-31 00:52:13.928 254 INFO keystonemiddleware.auth_token [-] Auth Token 
proceeding with requested v3.0 apis
  2015-03-31 00:52:13.928 254 DEBUG keystoneclient.auth.identity.v2 [-] Making 
authentication request to https://keystone server:35357/v2.0/tokens 
get_auth_ref 
/usr/lib/python2.7/site-packages/keystoneclient/auth/identity/v2.py:76

  Relevant config from glance-api.conf:

  [keystone_authtoken]
  identity_uri = https://keystone server:35357
  admin_user = glance
  admin_password =admin_password
  auth_version = v3.0

To manage notifications about this bug go to:
https://bugs.launchpad.net/keystone/+bug/1438469/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1362766] Re: ConnectionFailed: Connection to neutron failed: 'HTTPSConnectionPool' object has no attribute 'insecure'

2015-03-30 Thread Sam Morrison
** Also affects: keystonemiddleware
   Importance: Undecided
   Status: New

** Summary changed:

- ConnectionFailed: Connection to neutron failed: 'HTTPSConnectionPool' object 
has no attribute 'insecure'
+ ConnectionFailed: Connection to XX failed: 'HTTPSConnectionPool' object 
has no attribute 'insecure'

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1362766

Title:
  ConnectionFailed: Connection to XX failed: 'HTTPSConnectionPool'
  object has no attribute 'insecure'

Status in OpenStack Identity  (Keystone) Middleware:
  New
Status in OpenStack Compute (Nova):
  New
Status in Python client library for Glance:
  Fix Released
Status in Python client library for Neutron:
  Incomplete

Bug description:
  While compute manager was trying to authenticate with neutronclient,
  we see the following:

  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager Traceback 
(most recent call last):
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager   File 
/usr/lib/python2.7/site-packages/powervc_nova/compute/manager.py, line 672, 
in _populate_admin_context
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager 
nclient.authenticate()
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager   File 
/usr/lib/python2.7/site-packages/neutronclient/client.py, line 231, in 
authenticate
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager 
self._authenticate_keystone()
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager   File 
/usr/lib/python2.7/site-packages/neutronclient/client.py, line 209, in 
_authenticate_keystone
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager 
allow_redirects=True)
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager   File 
/usr/lib/python2.7/site-packages/neutronclient/client.py, line 113, in 
_cs_request
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager raise 
exceptions.ConnectionFailed(reason=e)
  2014-08-28 05:03:33.052 29982 TRACE powervc_nova.compute.manager 
ConnectionFailed: Connection to neutron failed: 'HTTPSConnectionPool' object 
has no attribute 'insecure'

  Setting a pdb breakpoint and stepping into the code, I see that the
  requests library is getting a connection object from a pool.  The
  interesting thing is that the connection object is actually from
  glanceclient.common.https.HTTPSConnectionPool.  It seems odd to me
  that neutronclient is using a connection object from glanceclient
  pool, but I do not know this requests code.  Here is the stack just
  before failure:

/usr/lib/python2.7/site-packages/neutronclient/client.py(234)authenticate()
  - self._authenticate_keystone()

/usr/lib/python2.7/site-packages/neutronclient/client.py(212)_authenticate_keystone()
  - allow_redirects=True)
/usr/lib/python2.7/site-packages/neutronclient/client.py(106)_cs_request()
  - resp, body = self.request(*args, **kargs)
/usr/lib/python2.7/site-packages/neutronclient/client.py(151)request()
  - **kwargs)
/usr/lib/python2.7/site-packages/requests/api.py(44)request()
  - return session.request(method=method, url=url, **kwargs)
/usr/lib/python2.7/site-packages/requests/sessions.py(335)request()
  - resp = self.send(prep, **send_kwargs)
/usr/lib/python2.7/site-packages/requests/sessions.py(438)send()
  - r = adapter.send(request, **kwargs)
/usr/lib/python2.7/site-packages/requests/adapters.py(292)send()
  - timeout=timeout
/usr/lib/python2.7/site-packages/urllib3/connectionpool.py(454)urlopen()
  - conn = self._get_conn(timeout=pool_timeout)
/usr/lib/python2.7/site-packages/urllib3/connectionpool.py(272)_get_conn()
  - return conn or self._new_conn()
   
/usr/lib/python2.7/site-packages/glanceclient/common/https.py(100)_new_conn()
  - return VerifiedHTTPSConnection(host=self.host,

  The code about to run there is this:

  class HTTPSConnectionPool(connectionpool.HTTPSConnectionPool):
  
  HTTPSConnectionPool will be instantiated when a new
  connection is requested to the HTTPSAdapter.This
  implementation overwrites the _new_conn method and
  returns an instances of glanceclient's VerifiedHTTPSConnection
  which handles no compression.

  ssl_compression is hard-coded to False because this will
  be used just when the user sets --no-ssl-compression.
  

  scheme = 'https'

  def _new_conn(self):
  self.num_connections += 1
  return VerifiedHTTPSConnection(host=self.host,
 port=self.port,
 key_file=self.key_file,
 cert_file=self.cert_file,
 cacert=self.ca_certs,
 insecure=self.insecure,
 

  1   2   >