Public bug reported: Description ===========
Shelve/unshelve operations fail when using "NFV flavors". This was reported on the mailing list initially. http://lists.openstack.org/pipermail/openstack-dev/2016-May/095631.html Steps to reproduce ================== 1. Create a flavor with 'hw:numa_nodes=2', 'hw:cpu_policy=dedicated' and 'hw:mempage_size=large' 2. Configure Tempest to use this new flavor 3. Run Tempest tests Expected result =============== All tests will pass. Actual result ============= The shelve/unshelve Tempest tests always result in a timeout exception being raised, looking similar to the following, from [1]: Traceback (most recent call last): File "tempest/api/compute/base.py", line 166, in server_check_teardown cls.server_id, 'ACTIVE') File "tempest/common/waiters.py", line 95, in wait_for_server_status raise exceptions.TimeoutException(message)2016-05-22 22:25:30.697 13974 ERROR tempest.api.compute.base TimeoutException: Request timed out Details: (ServerActionsTestJSON:tearDown) Server cae6fd47-0968-4922-a03e-3f2872e4eb52 failed to reach ACTIVE status and task state "None" within the required time (196 s). Current status: SHELVED_OFFLOADED. Current task state: None. The following errors are raised in the compute logs: Traceback (most recent call last): File "/opt/stack/new/nova/nova/compute/manager.py", line 4230, in _unshelve_instance with rt.instance_claim(context, instance, limits): File "/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py", line 271, in inner return f(*args, **kwargs) File "/opt/stack/new/nova/nova/compute/resource_tracker.py", line 151, in instance_claim self._update_usage_from_instance(context, instance_ref) File "/opt/stack/new/nova/nova/compute/resource_tracker.py", line 827, in _update_usage_from_instance self._update_usage(instance, sign=sign) File "/opt/stack/new/nova/nova/compute/resource_tracker.py", line 666, in _update_usage self.compute_node, usage, free) File "/opt/stack/new/nova/nova/virt/hardware.py", line 1482, in get_host_numa_usage_from_instance host_numa_topology, instance_numa_topology, free=free)) File "/opt/stack/new/nova/nova/virt/hardware.py", line 1348, in numa_usage_from_instances newcell.unpin_cpus(pinned_cpus) File "/opt/stack/new/nova/nova/objects/numa.py", line 94, in unpin_cpus pinned=list(self.pinned_cpus)) CPUPinningInvalid: Cannot pin/unpin cpus [6] from the following pinned set [0, 2, 4] [1] http://intel-openstack-ci-logs.ovh/86/319686/1/check/tempest-dsvm- full-nfv/b463722/testr_results.html.gz Environment =========== 1. Exact version of OpenStack you are running. See the following Commit '25fdf64'. ** Affects: nova Importance: Undecided Status: New -- You received this bug notification because you are a member of Yahoo! Engineering Team, which is subscribed to OpenStack Compute (nova). https://bugs.launchpad.net/bugs/1587386 Title: Unshelve results in duplicated resource deallocated Status in OpenStack Compute (nova): New Bug description: Description =========== Shelve/unshelve operations fail when using "NFV flavors". This was reported on the mailing list initially. http://lists.openstack.org/pipermail/openstack- dev/2016-May/095631.html Steps to reproduce ================== 1. Create a flavor with 'hw:numa_nodes=2', 'hw:cpu_policy=dedicated' and 'hw:mempage_size=large' 2. Configure Tempest to use this new flavor 3. Run Tempest tests Expected result =============== All tests will pass. Actual result ============= The shelve/unshelve Tempest tests always result in a timeout exception being raised, looking similar to the following, from [1]: Traceback (most recent call last): File "tempest/api/compute/base.py", line 166, in server_check_teardown cls.server_id, 'ACTIVE') File "tempest/common/waiters.py", line 95, in wait_for_server_status raise exceptions.TimeoutException(message)2016-05-22 22:25:30.697 13974 ERROR tempest.api.compute.base TimeoutException: Request timed out Details: (ServerActionsTestJSON:tearDown) Server cae6fd47-0968-4922-a03e-3f2872e4eb52 failed to reach ACTIVE status and task state "None" within the required time (196 s). Current status: SHELVED_OFFLOADED. Current task state: None. The following errors are raised in the compute logs: Traceback (most recent call last): File "/opt/stack/new/nova/nova/compute/manager.py", line 4230, in _unshelve_instance with rt.instance_claim(context, instance, limits): File "/usr/local/lib/python2.7/dist-packages/oslo_concurrency/lockutils.py", line 271, in inner return f(*args, **kwargs) File "/opt/stack/new/nova/nova/compute/resource_tracker.py", line 151, in instance_claim self._update_usage_from_instance(context, instance_ref) File "/opt/stack/new/nova/nova/compute/resource_tracker.py", line 827, in _update_usage_from_instance self._update_usage(instance, sign=sign) File "/opt/stack/new/nova/nova/compute/resource_tracker.py", line 666, in _update_usage self.compute_node, usage, free) File "/opt/stack/new/nova/nova/virt/hardware.py", line 1482, in get_host_numa_usage_from_instance host_numa_topology, instance_numa_topology, free=free)) File "/opt/stack/new/nova/nova/virt/hardware.py", line 1348, in numa_usage_from_instances newcell.unpin_cpus(pinned_cpus) File "/opt/stack/new/nova/nova/objects/numa.py", line 94, in unpin_cpus pinned=list(self.pinned_cpus)) CPUPinningInvalid: Cannot pin/unpin cpus [6] from the following pinned set [0, 2, 4] [1] http://intel-openstack-ci-logs.ovh/86/319686/1/check/tempest-dsvm- full-nfv/b463722/testr_results.html.gz Environment =========== 1. Exact version of OpenStack you are running. See the following Commit '25fdf64'. To manage notifications about this bug go to: https://bugs.launchpad.net/nova/+bug/1587386/+subscriptions -- Mailing list: https://launchpad.net/~yahoo-eng-team Post to : [email protected] Unsubscribe : https://launchpad.net/~yahoo-eng-team More help : https://help.launchpad.net/ListHelp

