[Yahoo-eng-team] [Bug 2039972] [NEW] [CI] TestVolumeBootPattern.test_volume_boot_pattern failed because test file wasn't found
Public bug reported: nova-ceph-multistore failed for one of my patches because TestVolumeBootPattern.test_volume_boot_pattern failed to find test file. test_volume_boot_pattern() [1] has quite simple logic: - Create in Cinder some bootable volume importing a Glance image - Boot an instance from the bootable volume - Write content to the volume - Delete an instance and Boot a new instance from the volume - Check written content in the instance # FAILED - Create a volume snapshot while the instance is running - Boot an additional instance from the new snapshot based volume - Check written content in the instance booted from snapshot For some reason test file in /tmp folder was missing after new instance was created with this volume attached. I have no idea what may cause it other than some cleanup operation against /tmp folder on OS level. Reporting this bug just in case, maybe there is some obvious problem. Logs: https://zuul.opendev.org/t/openstack/build/42a6c06ff5f545a5bb610d7c0d454032/logs [1] https://github.com/openstack/tempest/blob/d1742bad94e94ebabe875ed24415d1ff510a5ae7/tempest/scenario/test_volume_boot_pattern.py#L58 ** Affects: nova Importance: Undecided Status: New ** Tags: testing -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2039972 Title: [CI] TestVolumeBootPattern.test_volume_boot_pattern failed because test file wasn't found Status in OpenStack Compute (nova): New Bug description: nova-ceph-multistore failed for one of my patches because TestVolumeBootPattern.test_volume_boot_pattern failed to find test file. test_volume_boot_pattern() [1] has quite simple logic: - Create in Cinder some bootable volume importing a Glance image - Boot an instance from the bootable volume - Write content to the volume - Delete an instance and Boot a new instance from the volume - Check written content in the instance # FAILED - Create a volume snapshot while the instance is running - Boot an additional instance from the new snapshot based volume - Check written content in the instance booted from snapshot For some reason test file in /tmp folder was missing after new instance was created with this volume attached. I have no idea what may cause it other than some cleanup operation against /tmp folder on OS level. Reporting this bug just in case, maybe there is some obvious problem. Logs: https://zuul.opendev.org/t/openstack/build/42a6c06ff5f545a5bb610d7c0d454032/logs [1] https://github.com/openstack/tempest/blob/d1742bad94e94ebabe875ed24415d1ff510a5ae7/tempest/scenario/test_volume_boot_pattern.py#L58 To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/2039972/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 2020813] [NEW] OVS hardware offload for non admin users requires custom Neutron API policy
Public bug reported: OVS hardware offload was originally intended to be a feature used by normal users. But bugfix https://review.opendev.org/c/openstack/neutron/+/499203 for bug #1713590 removed the ability for non-admins to use OVS hardware offload without changing neutron default policy in a non-secure manner: "switchdev" capability must be added to port binding profile after port is created. At the same time, libvirt node device driver reports "switchdev" capability and we can translate it from NIC PCI device object to port binding profile when port is attached. ** Affects: nova Importance: Undecided Assignee: Alexey Stupnikov (astupnikov) Status: In Progress ** Tags: pci -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2020813 Title: OVS hardware offload for non admin users requires custom Neutron API policy Status in OpenStack Compute (nova): In Progress Bug description: OVS hardware offload was originally intended to be a feature used by normal users. But bugfix https://review.opendev.org/c/openstack/neutron/+/499203 for bug #1713590 removed the ability for non-admins to use OVS hardware offload without changing neutron default policy in a non-secure manner: "switchdev" capability must be added to port binding profile after port is created. At the same time, libvirt node device driver reports "switchdev" capability and we can translate it from NIC PCI device object to port binding profile when port is attached. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/2020813/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 2020358] [NEW] Allow to limit conntrack entries per tenant to avoid "nf_conntrack: table full, dropping packet"
Public bug reported: Description of problem: A tenant can cause network issues for other tenants: nf_conntrack: table full, dropping packet. In our cloud had a jmeter performance test running on two instances caused network issues for other tenants. In the /var/log/messages on the compute node we see the following message: "nf_conntrack: table full, dropping packet." This gerrit https://review.openstack.org/#/c/275769/ increases the limit to 500.000 but this is a workaround as a tenant can still increase usage up to this new limit. Neutron allows to limit bandwidth on a port, but you cannot limit the conntrack sessions for an instance, port or tenant. RHBZ: https://bugzilla.redhat.com/show_bug.cgi?id=1399987 ** Affects: neutron Importance: Undecided Status: New ** Tags: rfe -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/2020358 Title: Allow to limit conntrack entries per tenant to avoid "nf_conntrack: table full, dropping packet" Status in neutron: New Bug description: Description of problem: A tenant can cause network issues for other tenants: nf_conntrack: table full, dropping packet. In our cloud had a jmeter performance test running on two instances caused network issues for other tenants. In the /var/log/messages on the compute node we see the following message: "nf_conntrack: table full, dropping packet." This gerrit https://review.openstack.org/#/c/275769/ increases the limit to 500.000 but this is a workaround as a tenant can still increase usage up to this new limit. Neutron allows to limit bandwidth on a port, but you cannot limit the conntrack sessions for an instance, port or tenant. RHBZ: https://bugzilla.redhat.com/show_bug.cgi?id=1399987 To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/2020358/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 2017370] [NEW] stable/zed nova.tests.functional.wsgi.test_services.TestServicesAPI.test_resize_revert_after_deleted_source_compute sporadic failures
Public bug reported: A strange failure happens for one of my patches: nova.tests.functional.wsgi.test_services.TestServicesAPI.test_resize_revert_after_deleted_source_compute test fails because of OpenStackApiNotFoundException error. Logs and traces are available at https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_ab9/870053/2/gate/nova- tox-functional-py38/ab97c5a/job-output.txt 2023-04-21 06:27:36.066489 | ubuntu-focal | Captured traceback: 2023-04-21 06:27:36.066493 | ubuntu-focal | ~~~ 2023-04-21 06:27:36.066497 | ubuntu-focal | Traceback (most recent call last): 2023-04-21 06:27:36.066502 | ubuntu-focal | 2023-04-21 06:27:36.066506 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/wsgi/test_services.py", line 218, in test_resize_revert_after_deleted_source_compute 2023-04-21 06:27:36.066511 | ubuntu-focal | server = self._boot_and_check_allocations(flavor1, 'host1') 2023-04-21 06:27:36.066516 | ubuntu-focal | 2023-04-21 06:27:36.066520 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/integrated_helpers.py", line 932, in _boot_and_check_allocations 2023-04-21 06:27:36.066552 | ubuntu-focal | server = self._wait_for_state_change(created_server, 'ACTIVE') 2023-04-21 06:27:36.066556 | ubuntu-focal | 2023-04-21 06:27:36.066560 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/integrated_helpers.py", line 144, in _wait_for_state_change 2023-04-21 06:27:36.066564 | ubuntu-focal | return self._wait_for_server_parameter( 2023-04-21 06:27:36.066569 | ubuntu-focal | 2023-04-21 06:27:36.066573 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/integrated_helpers.py", line 130, in _wait_for_server_parameter 2023-04-21 06:27:36.066578 | ubuntu-focal | server = api.get_server(server['id']) 2023-04-21 06:27:36.066582 | ubuntu-focal | 2023-04-21 06:27:36.066586 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/api/client.py", line 248, in get_server 2023-04-21 06:27:36.066591 | ubuntu-focal | return self.api_get('/servers/%s' % server_id).body['server'] 2023-04-21 06:27:36.066595 | ubuntu-focal | 2023-04-21 06:27:36.066599 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/api/client.py", line 205, in api_get 2023-04-21 06:27:36.066603 | ubuntu-focal | return APIResponse(self.api_request(relative_uri, **kwargs)) 2023-04-21 06:27:36.066608 | ubuntu-focal | 2023-04-21 06:27:36.066612 | ubuntu-focal | File "/home/zuul/src/opendev.org/openstack/nova/nova/tests/functional/api/client.py", line 187, in api_request 2023-04-21 06:27:36.066616 | ubuntu-focal | raise OpenStackApiNotFoundException(response=response) 2023-04-21 06:27:36.066621 | ubuntu-focal | 2023-04-21 06:27:36.066625 | ubuntu-focal | nova.tests.functional.api.client.OpenStackApiNotFoundException: Item not found In Nova logs I can see that hypervisor and instance itself simply disappeared after successfull startup, so nova-api returned 404: 2023-04-21 06:26:17.760125 | ubuntu-focal | 2023-04-21 06:26:14,598 INFO [nova.api.openstack.requestlog] 127.0.0.1 "GET /v2.1/servers/6fe74316-77a0-4c99-826a-3c4b85613ee5" status: 200 len: 1770 microversion: 2.93 time: 0.026509 2023-04-21 06:26:17.760129 | ubuntu-focal | 2023-04-21 06:26:14,683 INFO [placement.requestlog] 127.0.0.1 "GET /placement/allocation_candidates?in_tree=7cf25e47-3e14-4944-91ea-ee3c5377f440=1000=DISK_GB%3A1%2CMEMORY_MB%3A512%2CVCPU%3A1_required=%21COMPUTE_STATUS_DISABLED" status: 200 len: 793 microversion: 1.36 2023-04-21 06:26:17.760136 | ubuntu-focal | 2023-04-21 06:26:14,711 INFO [nova.scheduler.host_manager] Host filter forcing available hosts to host1 2023-04-21 06:26:17.760140 | ubuntu-focal | 2023-04-21 06:26:14,718 INFO [placement.requestlog] 127.0.0.1 "GET /placement/allocations/6fe74316-77a0-4c99-826a-3c4b85613ee5" status: 200 len: 19 microversion: 1.28 2023-04-21 06:26:17.760148 | ubuntu-focal | 2023-04-21 06:26:14,764 INFO [placement.requestlog] 127.0.0.1 "PUT /placement/allocations/6fe74316-77a0-4c99-826a-3c4b85613ee5" status: 204 len: 0 microversion: 1.36 2023-04-21 06:26:17.760152 | ubuntu-focal | 2023-04-21 06:26:17,081 INFO [nova.api.openstack.requestlog] 127.0.0.1 "GET /v2.1/servers/6fe74316-77a0-4c99-826a-3c4b85613ee5" status: 200 len: 1787 microversion: 2.93 time: 0.504291 2023-04-21 06:26:17.760156 | ubuntu-focal | 2023-04-21 06:26:17,128 INFO [placement.requestlog] 127.0.0.1 "GET /placement/allocations/6fe74316-77a0-4c99-826a-3c4b85613ee5" status: 200 len: 230 microversion: 1.28 2023-04-21 06:26:17.760160 | ubuntu-focal | 2023-04-21 06:26:17,131 INFO [nova.compute.claims] Claim successful on node host1 2023-04-21 06:26:17.760167 | ubuntu-focal | 2023-04-21 06:26:17,169 INFO
[Yahoo-eng-team] [Bug 2015092] [NEW] Instance sticks in BUILD when scheduling with a port assigned invalid MAC
Public bug reported: This is a copy of downstream https://bugzilla.redhat.com/show_bug.cgi?id=1949703 caused by combination of Neutron bug (addressed in https://bugs.launchpad.net/neutron/+bug/1926273) and Nova inability to process ValueError raised by oslo_versionedobjects. Steps to reproduce (before Neutron bug was fixed): 1. Create bad port $ openstack port create --mac-address 40:28:0:00:2:6 --disable-port-security --network ess-dpdk-1370 test-port-bad-mac_user-dpdk2 2. Create server using bad port $ openstack server create --flavor m1.tiny --image cirros --nic port-id=023abee9-a896-4678-b6fe-d4abae9066f0 dpdk-instance-bad-mac2 Expected outcome: Instance is in ERROR state Actual outcome: Instance is sticks in BUILD state ** Affects: nova Importance: Undecided Assignee: Alexey Stupnikov (astupnikov) Status: New ** Tags: compute -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2015092 Title: Instance sticks in BUILD when scheduling with a port assigned invalid MAC Status in OpenStack Compute (nova): New Bug description: This is a copy of downstream https://bugzilla.redhat.com/show_bug.cgi?id=1949703 caused by combination of Neutron bug (addressed in https://bugs.launchpad.net/neutron/+bug/1926273) and Nova inability to process ValueError raised by oslo_versionedobjects. Steps to reproduce (before Neutron bug was fixed): 1. Create bad port $ openstack port create --mac-address 40:28:0:00:2:6 --disable-port-security --network ess-dpdk-1370 test-port-bad-mac_user-dpdk2 2. Create server using bad port $ openstack server create --flavor m1.tiny --image cirros --nic port-id=023abee9-a896-4678-b6fe-d4abae9066f0 dpdk-instance-bad-mac2 Expected outcome: Instance is in ERROR state Actual outcome: Instance is sticks in BUILD state To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/2015092/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 2009156] [NEW] stable/victoria tests are very unstable because of tempest test_live_block_migration_with_attached_volume
Public bug reported: stable/victoria merge process is effectively blocked because test_live_block_migration_with_attached_volume tempest test (called as tempest.api.compute.admin.test_live_migration.LiveMigrationTest.test_live_block_migration_with_attached_volume and tempest.api.compute.admin.test_live_migration.LiveAutoBlockMigrationV225Test.test_live_block_migration_with_attached_volume) is failing sporadically. As a result, nova-grenade-multinode or nova- live-migration are failing sporadically: one of these jobs fails almost every time. I have used failed tests for review https://review.opendev.org/c/openstack/nova/+/833436 to analyze the problem. Picture is always the same: 1. Tempest deletes volume attachment [1] 2. nova-compute fails to detach volume because of error [2] It looks like a communication problem between nova-compute and libvirt. [1] 2023-02-28 12:43:25.610324 | controller | 2023-02-28 12:40:08,991 6532 INFO [tempest.lib.common.rest_client] Request (LiveAutoBlockMigrationV225Test:_run_cleanups): 202 DELETE http://213.32.76.87/compute/v2.1/servers/e49ccfac-97fc-4b8d-a15a-6a319d8c9a9b/os-volume_attachments/e0d103c8-cc39-4ccc-a91e-a415ff68bdee 0.217s [2] Feb 28 12:45:26.707716 np0033291896 nova-compute[15752]: ERROR oslo_messaging.rpc.server File "/opt/stack/old/nova/nova/virt/libvirt/guest.py", line 472, in _do_wait_and_retry_detach Feb 28 12:45:26.707716 np0033291896 nova-compute[15752]: ERROR oslo_messaging.rpc.server device=alternative_device_name, reason=reason) Feb 28 12:45:26.707716 np0033291896 nova-compute[15752]: ERROR oslo_messaging.rpc.server nova.exception.DeviceDetachFailed: Device detach failed for vdb: Unable to detach the device from the live config. ** Affects: nova Importance: Undecided Status: New ** Tags: testing -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2009156 Title: stable/victoria tests are very unstable because of tempest test_live_block_migration_with_attached_volume Status in OpenStack Compute (nova): New Bug description: stable/victoria merge process is effectively blocked because test_live_block_migration_with_attached_volume tempest test (called as tempest.api.compute.admin.test_live_migration.LiveMigrationTest.test_live_block_migration_with_attached_volume and tempest.api.compute.admin.test_live_migration.LiveAutoBlockMigrationV225Test.test_live_block_migration_with_attached_volume) is failing sporadically. As a result, nova-grenade-multinode or nova- live-migration are failing sporadically: one of these jobs fails almost every time. I have used failed tests for review https://review.opendev.org/c/openstack/nova/+/833436 to analyze the problem. Picture is always the same: 1. Tempest deletes volume attachment [1] 2. nova-compute fails to detach volume because of error [2] It looks like a communication problem between nova-compute and libvirt. [1] 2023-02-28 12:43:25.610324 | controller | 2023-02-28 12:40:08,991 6532 INFO [tempest.lib.common.rest_client] Request (LiveAutoBlockMigrationV225Test:_run_cleanups): 202 DELETE http://213.32.76.87/compute/v2.1/servers/e49ccfac-97fc-4b8d-a15a-6a319d8c9a9b/os-volume_attachments/e0d103c8-cc39-4ccc-a91e-a415ff68bdee 0.217s [2] Feb 28 12:45:26.707716 np0033291896 nova-compute[15752]: ERROR oslo_messaging.rpc.server File "/opt/stack/old/nova/nova/virt/libvirt/guest.py", line 472, in _do_wait_and_retry_detach Feb 28 12:45:26.707716 np0033291896 nova-compute[15752]: ERROR oslo_messaging.rpc.server device=alternative_device_name, reason=reason) Feb 28 12:45:26.707716 np0033291896 nova-compute[15752]: ERROR oslo_messaging.rpc.server nova.exception.DeviceDetachFailed: Device detach failed for vdb: Unable to detach the device from the live config. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/2009156/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 2007860] [NEW] master nova-tox-functional-py310 fails because of VDPAServersTest.test_suspend_and_resume
Public bug reported: nova.tests.functional.libvirt.test_pci_sriov_servers.VDPAServersTest.test_suspend_and_resume failed for one of my patches. Failure itself looks sporadic, but may be a sign of underlying problems. Complete log: https://3597fc80eb5e70afafd6-8fe1ce567603c46a794844f7885416e9.ssl.cf1.rackcdn.com/873388/2/check/nova- tox-functional-py310/6921a72/job-output.txt My understanding: - "Original exception" [1] is logged when nova-conductor is started - This exception triggers calls, which in turn cause libvirt import and trigger ImportError exception: 2023-02-17 11:43:02.536761 | ubuntu-jammy | 2023-02-17 11:26:59,887 WARNING [nova.tests.fixtures.nova] !!!---!!! TestCase ID nova.tests.functional.libvirt.test_pci_sriov_servers.VDPAServersTest.test_suspend_and_resume hit the import poison while importing libvirt. If you see this in a failed functional test then please let #openstack-nova on IRC know about it. !!!---!!! It looks like original issue is caused by some RPC call related to live migration. This problem seem to be sporadic and is not a blocker, but it looks quite interesting and could affect CI in general, so I am reporting this bug. [1] 2023-02-17 11:43:02.529191 | ubuntu-jammy | 2023-02-17 11:26:59,782 ERROR [root] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/oslo_messaging/_drivers/impl_fake.py", line 207, in _send\nreply, failure = reply_q.get(timeout=timeout)\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/eventlet/queue.py", line 322, in get\nreturn waiter.wait()\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/eventlet/queue.py", line 141, in wait\nreturn get_hub().switch()\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 313, in switch\nreturn self.greenlet.switch()\n', '_queue.Empty\n', '\nDuring handling of the above exception, another exception occurred:\n\n', 'Tra ceback (most recent call last):\n', ' File "/home/zuul/src/opendev.org/openstack/nova/nova/compute/manager.py", line 203, in decorated_function\nreturn function(self, context, *args, **kwargs)\n', ' File "/home/zuul/src/opendev.org/openstack/nova/nova/compute/manager.py", line 9306, in _post_live_migration\n self._update_scheduler_instance_info(ctxt, instance)\n', ' File "/home/zuul/src/opendev.org/openstack/nova/nova/compute/manager.py", line 2225, in _update_scheduler_instance_info\n self.query_client.update_instance_info(context, self.host,\n', ' File "/home/zuul/src/opendev.org/openstack/nova/nova/scheduler/client/query.py", line 69, in update_instance_info\n self.scheduler_rpcapi.update_instance_info(context, host_name,\n', ' File "/home/zuul/src/opendev.org/openstack/nova/nova/scheduler/rpcapi.py", line 174, in update_instance_info\ncctxt.cast(ctxt, \'update_instance_info\', host_name=host_name,\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/ functional-py310/lib/python3.10/site-packages/fixtures/_fixtures/monkeypatch.py", line 86, in avoid_get\nreturn captured_method(*args, **kwargs)\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/oslo_messaging/rpc/client.py", line 190, in c 2023-02-17 11:43:02.529623 | ubuntu-jammy | all\nresult = self.transport._send(\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/oslo_messaging/transport.py", line 123, in _send\nreturn self._driver.send(target, ctxt, message,\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/oslo_messaging/_drivers/impl_fake.py", line 222, in send\nreturn self._send(target, ctxt, message, wait_for_reply, timeout,\n', ' File "/home/zuul/src/opendev.org/openstack/nova/.tox/functional-py310/lib/python3.10/site-packages/oslo_messaging/_drivers/impl_fake.py", line 213, in _send\nraise oslo_messaging.MessagingTimeout(\n', 'oslo_messaging.exceptions.MessagingTimeout: No reply on topic scheduler\n'] 2023-02-17 11:43:02.529678 | ubuntu-jammy | 2023-02-17 11:26:59,881 WARNING [nova.virt.libvirt.driver] Error monitoring migration: (sqlite3.OperationalError) no such table: compute_nodes ** Affects: nova Importance: Undecided Status: New ** Tags: testing -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2007860 Title: master nova-tox-functional-py310 fails because of VDPAServersTest.test_suspend_and_resume Status in OpenStack Compute (nova): New Bug description:
[Yahoo-eng-team] [Bug 2006952] [NEW] Ambigous error when trying to boot SEV based instances from volume
12:29:28.207 12 ERROR nova.api.openstack.wsgi validate_pci=validate_pci) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/nova/compute/api.py", line 747, in _validate_flavor_image_numa_pci 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi API._validate_flavor_image_mem_encryption(instance_type, image_meta) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/nova/compute/api.py", line 778, in _validate_flavor_image_mem_encryption 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi hardware.get_mem_encryption_constraint(instance_type, image) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/nova/virt/hardware.py", line 1273, in get_mem_encryption_constraint 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi _check_mem_encryption_uses_uefi_image(requesters, image_meta) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/nova/virt/hardware.py", line 1310, in _check_mem_encryption_uses_uefi_image 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi 'image_name': image_meta.name} 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/oslo_versionedobjects/base.py", line 67, in getter 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi self.obj_load_attr(name) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/oslo_versionedobjects/base.py", line 603, in obj_load_attr 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi _("Cannot load '%s' in the base class") % attrname) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi NotImplementedError: Cannot load 'name' in the base class 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi RH BZ: https://bugzilla.redhat.com/show_bug.cgi?id=1967606 ** Affects: nova Importance: Undecided Assignee: Alexey Stupnikov (astupnikov) Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/2006952 Title: Ambigous error when trying to boot SEV based instances from volume Status in OpenStack Compute (nova): New Bug description: More image_meta.name use in hardware.py causing failures when presented with an empty ImageMeta object: $ openstack flavor show m1.sev_med ++--+ | Field | Value| ++--+ | OS-FLV-DISABLED:disabled | False| | OS-FLV-EXT-DATA:ephemeral | 0| | access_project_ids | None | | description| None | | disk | 2| | extra_specs| {'hw:mem_encryption': 'True'}| | id | 3952db4d-e71a-4669-9bb7-666adaef6c36 | | name | m1.sev_med | | os-flavor-access:is_public | True | | properties | hw:mem_encryption='True' | | ram| 2048 | | rxtx_factor| 1.0 | | swap | 0| | vcpus | 4| ++--+ $ openstack volume create --bootable --size 1 blank $ openstack server create --volume blank --flavor m1.sev_med --network private test Unexpected API Error. Please report this at http://bugs.launchpad.net/nova/ and attach the Nova API log if possible. (HTTP 500) (Request-ID: req-c6d1a319-b4b0-4d1f-869c-dcbec2fd2554) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi [req-c6d1a319-b4b0-4d1f-869c-dcbec2fd2554 cd40fe796ff84e3a8ba5e473a6d61f05 025f8a0d412642f693782ae20ba415ec - default default] Unexpected exception in API method: NotImplementedError: Cannot load 'name' in the base class 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi Traceback (most recent call last): 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/nova/api/openstack/wsgi.py", line 671, in wrapped 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi return f(*args, **kwargs) 2021-06-03 12:29:28.207 12 ERROR nova.api.openstack.wsgi File "/usr/lib/python3.6/site-packages/nova/api/validation/__init__.p
[Yahoo-eng-team] [Bug 1994983] [NEW] Instance is stopped on destination node after evacuation if stop call was issued before evacuation
Public bug reported: Steps to reproduce: 1. Start an instance on a compute node (src compute). 2. Destroy src compute. Wait till nova-compute server will go down. 3. Run 'openstack server stop' for an instance. 4. Run 'server set --state error' to the instance to be able to evacuate the instance (no longer needed if fix for bug #1978983 presents) 5. Evacuate the instance to another compute node (dst compute). 6. Start src compute. 7. Confirm that the evacuated instance is stopped after src compute come online. This behavior is a bug because src compute could come up after few months and shutdown some important instance. It looks like this behavior is caused by cast RPC call for SRC compute to stop an instance which sits in the queue until compute is back online and then causes DB state change. RHBZ: https://bugzilla.redhat.com/show_bug.cgi?id=2130112 ** Affects: nova Importance: Undecided Assignee: Alexey Stupnikov (astupnikov) Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1994983 Title: Instance is stopped on destination node after evacuation if stop call was issued before evacuation Status in OpenStack Compute (nova): New Bug description: Steps to reproduce: 1. Start an instance on a compute node (src compute). 2. Destroy src compute. Wait till nova-compute server will go down. 3. Run 'openstack server stop' for an instance. 4. Run 'server set --state error' to the instance to be able to evacuate the instance (no longer needed if fix for bug #1978983 presents) 5. Evacuate the instance to another compute node (dst compute). 6. Start src compute. 7. Confirm that the evacuated instance is stopped after src compute come online. This behavior is a bug because src compute could come up after few months and shutdown some important instance. It looks like this behavior is caused by cast RPC call for SRC compute to stop an instance which sits in the queue until compute is back online and then causes DB state change. RHBZ: https://bugzilla.redhat.com/show_bug.cgi?id=2130112 To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1994983/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1960412] [NEW] Live migration artifacts are not cleaned up properly when queued live migration is aborted
Public bug reported: Bug #1949808 describes one of the problems affecting aborted queued live migrations: VM's status is not reverted back to ACTIVE and VM is left in MIGRATING state. However that's not the single problem with aborted queued live migrations: some changes (port bindings on destination host, resource allocations and possibly instance's pci_requests) are introduced by Nova control plane before live migration actually started by source compute host. Those left-overs should also be removed. ** Affects: nova Importance: Undecided Assignee: Alexey Stupnikov (astupnikov) Status: New ** Description changed: - Bug #1949808 describes one of the problems affecting aborted queued - live migrations: VM's status is not reverted back to ACTIVE and - VM is left in MIGRATING state. + Bug #1949808 describes one of the problems affecting aborted queued live + migrations: VM's status is not reverted back to ACTIVE and VM is left in + MIGRATING state. However that's not the single problem with aborted queued live - migrations: some changes (port bindings on destination host, - resource allocations and possibly instance's pci_requests) - are introduced by Nova control plane before live migration - actually started by source compute host. Those left-overs - should also be removed. + migrations: some changes (port bindings on destination host, resource + allocations and possibly instance's pci_requests) are introduced by Nova + control plane before live migration actually started by source compute + host. Those left-overs should also be removed. ** Changed in: nova Assignee: (unassigned) => Alexey Stupnikov (astupnikov) -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1960412 Title: Live migration artifacts are not cleaned up properly when queued live migration is aborted Status in OpenStack Compute (nova): New Bug description: Bug #1949808 describes one of the problems affecting aborted queued live migrations: VM's status is not reverted back to ACTIVE and VM is left in MIGRATING state. However that's not the single problem with aborted queued live migrations: some changes (port bindings on destination host, resource allocations and possibly instance's pci_requests) are introduced by Nova control plane before live migration actually started by source compute host. Those left-overs should also be removed. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1960412/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1949808] [NEW] After nova live-migration-abort canceled "queued" live-migration, instance status remains "MIGRATING"
Public bug reported: OpenStack supports cancelling "queued" live-migration by "nova live- migration-abort", but state of canceled instance remains "MIGRATING". ** Affects: nova Importance: Undecided Assignee: Alexey Stupnikov (astupnikov) Status: New ** Tags: live-migration ** Bug watch added: Red Hat Bugzilla #1729366 https://bugzilla.redhat.com/show_bug.cgi?id=1729366 -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1949808 Title: After nova live-migration-abort canceled "queued" live-migration, instance status remains "MIGRATING" Status in OpenStack Compute (nova): New Bug description: OpenStack supports cancelling "queued" live-migration by "nova live- migration-abort", but state of canceled instance remains "MIGRATING". To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1949808/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1849802] Re: Live-migration fails without a logged root cause
I have used my lab to understand original report and what is actually going on. Shortly speaking, Nova is unable to live migrate an instance because Neutron fails to create port binding on destination host (it already exists) when Nova Conductor tries to create port binding on destination host. There could be multiple causes of such behavior and this bug is not about solving a root cause. Instead, the problem is that it is hard to isolate the issue using Neutron Server logs. When debug is enabled Neutron generates logs [1] for such kinds of requests. It looks like for some reason the following exception raised by Neutron is not logged properly on Neutron side: we can see it in Nova logs, but not in Neutron Server logs. https://github.com/openstack/neutron/blob/master/neutron/plugins/ml2/plugin.py#L2463-L2466 As a result, it looks like this problem should be solved by improving logging for Neutron extensions in general or this particular function. With that being said, I am changing affected product to Neutron and unassigning this bug. [1] 2021-10-18 14:15:58.081 18 DEBUG neutron.api.v2.base [req-c237e714-0085-47c6-abab-e7e15cec7ea1 d79d0393fbb04564bc8fd3c62d290087 b8715d57125f4787a6701319d38f61e3 - default default] Request body: {'binding': {'host': 'compute-0.redhat.local', 'vnic_type': 'normal', 'profile': {}}} prepare_request_body /usr/lib/python3.6/site-packages/neutron/api/v2/base.py:719 2021-10-18 14:15:58.081 18 DEBUG neutron.api.v2.base [req-c237e714-0085-47c6-abab-e7e15cec7ea1 d79d0393fbb04564bc8fd3c62d290087 b8715d57125f4787a6701319d38f61e3 - default default] Unknown quota resources ['binding']. _create /usr/lib/python3.6/site-packages/neutron/api/v2/base.py:490 2021-10-18 14:15:58.135 18 INFO neutron.api.v2.resource [req-c237e714-0085-47c6-abab-e7e15cec7ea1 d79d0393fbb04564bc8fd3c62d290087 b8715d57125f4787a6701319d38f61e3 - default default] create failed (client error): There was a conflict when trying to complete your request. 2021-10-18 14:15:58.137 18 INFO neutron.wsgi [req-c237e714-0085-47c6-abab-e7e15cec7ea1 d79d0393fbb04564bc8fd3c62d290087 b8715d57125f4787a6701319d38f61e3 - default default] 172.17.1.17 "POST /v2.0/ports/7542a977-0586-423a-ae35-86e3ff791060/bindings HTTP/1.1" status: 409 len: 364 time: 0.3800023 2021-10-18 14:15:58.316 21 DEBUG neutron_lib.callbacks.manager [req-334ebddd-4e81-4c12-829c-64f3b0a278ff - - - - -] Notify callbacks ['neutron.services.segments.db._update_segment_host_mapping_for_agent-8793714910767', 'neutron.plugins.ml2.plugin.Ml2Plugin._retry_binding_revived_agents-16758855'] for agent, after_update _notify_loop /usr/lib/python3.6/site-packages/neutron_lib/callbacks/manager.py:193 ** Changed in: nova Assignee: Alexey Stupnikov (astupnikov) => (unassigned) ** Project changed: nova => neutron ** Changed in: neutron Status: Triaged => New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to neutron. https://bugs.launchpad.net/bugs/1849802 Title: Live-migration fails without a logged root cause Status in neutron: New Bug description: Operating system distribution and version: Ubuntu 18.04 Neutron package version: 2:14.0.2-0ubuntu1~cloud0 Nova package version: 2:19.0.1-0ubuntu2.1~cloud0 Cloud was deploying (using Juju charms) as rocky, then upgraded to stein. There are a number of instances that I need to migrate from one compute node to another, but: $ openstack server migrate --block-migration 8703d9db-81b0-4e86-a2ef-c4ba5250556c --live shinx --disk-overcommit Migration pre-check error: Binding failed for port 5a3c5d23-8727-47d2-af72-a53b495358d2, please check neutron logs for more information. (HTTP 400) (Request-ID: req-7c41ae70-6f5b-48a8-9d09-add2bbbe2b7e) $ However, even with debug logging enabled, all that shows up in the neutron-api logs is: 2019-10-25 09:34:12.147 1569534 INFO neutron.wsgi [req- ac358ed5-cfec-4618-b765-f2defd5a3aac 92e98c5c687a46d29ec28aca3025f3da 7555fff7e7eb4a0eb28149905b266a2b - 207337407e3647798c0f68a0a28a0f8b 207337407e3647798c0f68a0a28a0f8b] 10.x.y.z,127.0.0.1 "POST /v2.0/ports/5a3c5d23-8727-47d2-af72-a53b495358d2/bindings HTTP/1.1" status: 409 len: 371 time: 0.1632745 Which suggests that for some reason, the API call to retrieve port bindings is failing, but there's no further information in the logs for me to debug exactly why. To manage notifications about this bug go to: https://bugs.launchpad.net/neutron/+bug/1849802/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1718606] Re: doc: Outdated XenServer document
XenAPI driver was deprecated during Train release cycle and removed during Victoria cycle. This bug doesn't longer reflect actual situation and is invalid for master branch of Nova. References: https://review.opendev.org/c/openstack/nova/+/680732 https://review.opendev.org/c/openstack/nova/+/749304 https://docs.openstack.org/releasenotes/nova/victoria.html ** Changed in: nova Assignee: (unassigned) => Alexey Stupnikov (astupnikov) ** Changed in: nova Status: Confirmed => Invalid ** Tags removed: low-hanging-fruit -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1718606 Title: doc: Outdated XenServer document Status in OpenStack Compute (nova): Invalid Bug description: In the following document, os-xenapi version is 0.1.0. Currently it is 0.3.1. The guide needs a serious look at and probably it should be revised. https://docs.openstack.org/nova/latest/admin/configuration/hypervisor- xen-api.html To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1718606/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1747167] Re: Neutron driver handles incorrectly 404 errors
Changing bug's status to fix released as recent releases are no longer affected ** Changed in: nova Status: Confirmed => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1747167 Title: Neutron driver handles incorrectly 404 errors Status in OpenStack Compute (nova): Fix Released Bug description: Neutronclient transforms errors using exception_handler_v20[1]. In particular, it transforms 404 responses into: 1) [2] neutronclient.common.exceptions.Client if response body contains a type attribute (with value , type examples: NetworkNotFound, PortNotFound) and Client exists 2) [3] neutronclient.common.NotFound otherwise Nova neutron driver[4] supports mostly 1) but not 2). But Neutron in case of 404 returns the following body: {"message": "The resource could not be found.\n\n\n", "code": "404 Not Found", "title": "Not Found"} and we end in 2) and Nova Neutron driver doesn't catch the error correctly and raises a 500 You can reproduce it indirectly using: openstack server add fixed ip which queries: nova-base-url/os-networks/ assuming it would return the network or a 404 but it returns 500. This trouble can be solved by catching NotFound instead of NetworkNotFoundClient/PortNotFoundClient as there are subclasses of NotFound. [1] https://github.com/openstack/python-neutronclient/blob/6a2112e1824980af1bc9216441fa58a7dfd00988/neutronclient/v2_0/client.py#L47-L93 [2] https://github.com/openstack/python-neutronclient/blob/6a2112e1824980af1bc9216441fa58a7dfd00988/neutronclient/v2_0/client.py#L71 [3] https://github.com/openstack/python-neutronclient/blob/6a2112e1824980af1bc9216441fa58a7dfd00988/neutronclient/v2_0/client.py#L85 [4] nova.network.neutronv2.api To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1747167/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1747344] Re: nova-manage db online migration base message
This bug was actually fixed in https://review.opendev.org/#/c/420060/ Closing as Fix Released. ** Changed in: nova Status: Confirmed => Fix Released -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1747344 Title: nova-manage db online migration base message Status in OpenStack Compute (nova): Fix Released Bug description: env: Mitaka upgraded to Ocata Use command "su -s /bin/sh -c "nova-manage db sync" nova", the message was " Please run `nova-manage db online_migrations' on Newton code before continuing.There are still 6 unmigrated flavors. " The migration command was "nova-manage db online_data_migrations" not "nova-manage db online_migrations". To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1747344/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp
[Yahoo-eng-team] [Bug 1537062] Re: Fail to boot vm when set AggregateImagePropertiesIsolation filter and add custom metadata in the Host Aggregate
Removed MOS project from the list of affected products since I have opened separate bug #1608937. I will try to cherry-pick this issue to stable/mitaka and then sync it to MOS. @l-ivan, please check my comment above ^^^ ** No longer affects: mos ** Tags removed: customer-found -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1537062 Title: Fail to boot vm when set AggregateImagePropertiesIsolation filter and add custom metadata in the Host Aggregate Status in OpenStack Compute (nova): Fix Released Bug description: Image has no custom metadata, should not affect the AggregateImagePropertiesIsolation filter Reproduce steps: (1) add Host Aggregate with custom metadata ++---+---+--++ | Id | Name | Availability Zone | Hosts| Metadata | ++---+---+--++ | 1 | linux-agg | - | 'controller' | 'os=linux' | ++---+---+--++ (2) add AggregateImagePropertiesIsolation filter scheduler_default_filters = RetryFilter,AggregateImagePropertiesIsolation,AvailabilityZoneFilter,RamFilter,DiskFilter,ComputeFilter,ComputeCapabilitiesFilter,ImagePropertiesFilter,ServerGroupAntiAffinityFilter,ServerGroupAffinityFilter,SameHostFilter,DifferentHostFilter (3) boot vm and error log: 2016-01-22 21:00:10.834 ERROR oslo_messaging.rpc.dispatcher [req-1cded809-cfe6-4657-8e31-b494f1b3278d admin admin] Exception during messa ge handling: ImageMetaProps object has no attribute 'os' 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher Traceback (most recent call last): 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/dispatcher.py", l ine 143, in _dispatch_and_reply 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher executor_callback)) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/dispatcher.py", l ine 189, in _dispatch 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher executor_callback) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/dispatcher.py", l ine 130, in _do_dispatch 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher result = func(ctxt, **new_args) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/usr/lib/python2.7/site-packages/oslo_messaging/rpc/server.py", line 150, in inner 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher return func(*args, **kwargs) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/scheduler/manager.py", line 78, in select_destin ations 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher dests = self.driver.select_destinations(ctxt, spec_obj) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/scheduler/filter_scheduler.py", line 53, in sele ct_destinations 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher selected_hosts = self._schedule(context, spec_obj) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/scheduler/filter_scheduler.py", line 113, in _sc hedule 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher spec_obj, index=num) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/scheduler/host_manager.py", line 532, in get_fil tered_hosts 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher hosts, spec_obj, index) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/filters.py", line 89, in get_filtered_objects 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher list_objs = list(objs) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/filters.py", line 44, in filter_all 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher if self._filter_one(obj, spec_obj): 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/scheduler/filters/__init__.py", line 26, in _filter_one 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher return self.host_passes(obj, filter_properties) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File "/opt/stack/nova/nova/scheduler/filters/aggregate_image_properties_isolation.py", line 48, in host_passes 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher prop = image_props.get(key) 2016-01-22 21:00:10.834 TRACE oslo_messaging.rpc.dispatcher File
[Yahoo-eng-team] [Bug 1563419] Re: [UI] sahara uses UTC time instead of set timezone
This bug is also found in Horizon's kilo and liberty branches. ** Also affects: horizon Importance: Undecided Status: New ** Changed in: horizon Assignee: (unassigned) => Alexey Stupnikov (astupnikov) -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Dashboard (Horizon). https://bugs.launchpad.net/bugs/1563419 Title: [UI] sahara uses UTC time instead of set timezone Status in OpenStack Dashboard (Horizon): New Status in Sahara: Fix Released Bug description: All time values that are shown in sahara dashboard are in UTC no matter what kind of timezone we have set in settings. It affects the Data Sources, Job Execution detail views and Cluster provision steps table To manage notifications about this bug go to: https://bugs.launchpad.net/horizon/+bug/1563419/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : yahoo-eng-team@lists.launchpad.net Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp