Github user jburwell commented on a diff in the pull request: https://github.com/apache/cloudstack/pull/1642#discussion_r79104522 --- Diff: test/integration/plugins/solidfire/TestManagedSystemVMs.py --- @@ -0,0 +1,585 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. + +import logging +import random +import SignedAPICall +import XenAPI + +from util import sf_util + +from marvin.cloudstackAPI import destroySystemVm + +# All tests inherit from cloudstackTestCase +from marvin.cloudstackTestCase import cloudstackTestCase + +from nose.plugins.attrib import attr + +# Import Integration Libraries + +# base - contains all resources as entities and defines create, delete, list operations on them +from marvin.lib.base import Account, Router, ServiceOffering, StoragePool, User, VirtualMachine, Zone + +# common - commonly used methods for all tests are listed here +from marvin.lib.common import get_domain, get_template, get_zone, list_clusters, list_hosts, list_ssvms, list_routers + +# utils - utility classes for common cleanup, external library wrappers, etc. +from marvin.lib.utils import cleanup_resources, wait_until + +from solidfire import solidfire_element_api as sf_api + +from time import sleep as delay + +# Prerequisites: +# * Only use one SolidFire cluster for the two primary storages based on the "SolidFire" storage plug-in. +# * Do not run other workloads on the SolidFire cluster while running this test as this test checks at a certain +# point to make sure no active SolidFire volumes exist. +# * Only one zone +# * Only one secondary storage VM and one console proxy VM running on NFS (no virtual router or user VMs exist) +# * Only one pod +# * Only one cluster +# * Set storage.cleanup.enabled to true +# * Set storage.cleanup.interval to 150 +# * Set storage.cleanup.delay to 60 + + +class TestData(): + account = "account" + capacityBytes = "capacitybytes" + capacityIops = "capacityiops" + clusterId = "clusterid" + computeOffering = "computeoffering" + diskOffering = "diskoffering" + domainId = "domainid" + email = "email" + firstname = "firstname" + hypervisor = "hypervisor" + lastname = "lastname" + login = "login" + max_iops = "maxiops" + min_iops = "miniops" + mvip = "mvip" + name = "name" + password = "password" + port = "port" + primaryStorage = "primarystorage" + provider = "provider" + scope = "scope" + solidFire = "solidfire" + storageTag = "SolidFire_SAN_1" + systemOffering = "systemoffering" + systemOfferingFailure = "systemofferingFailure" + tags = "tags" + url = "url" + user = "user" + username = "username" + virtualMachine = "virtualmachine" + xenServer = "xenserver" + zoneId = "zoneid" + + def __init__(self): + self.testdata = { + TestData.solidFire: { + TestData.mvip: "192.168.139.112", + TestData.login: "admin", + TestData.password: "admin", + TestData.port: 443, + TestData.url: "https://192.168.139.112:443" + }, + TestData.xenServer: { + TestData.username: "root", + TestData.password: "solidfire" + }, + TestData.account: { + TestData.email: "t...@test.com", + TestData.firstname: "John", + TestData.lastname: "Doe", + TestData.username: "test", + TestData.password: "test" + }, + TestData.user: { + TestData.email: "u...@test.com", + TestData.firstname: "Jane", + TestData.lastname: "Doe", + TestData.username: "testuser", + TestData.password: "password" + }, + TestData.primaryStorage: { + TestData.name: TestData.get_name_for_solidfire_storage(), + TestData.scope: "ZONE", + TestData.url: "MVIP=192.168.139.112;SVIP=10.10.8.112;" + + "clusterAdminUsername=admin;clusterAdminPassword=admin;" + + "clusterDefaultMinIops=10000;clusterDefaultMaxIops=15000;" + + "clusterDefaultBurstIopsPercentOfMaxIops=1.5;", + TestData.provider: "SolidFire", + TestData.tags: TestData.storageTag, + TestData.capacityIops: 4500000, + TestData.capacityBytes: 2251799813685248, + TestData.hypervisor: "Any", + TestData.zoneId: 1 + }, + TestData.virtualMachine: { + TestData.name: "TestVM", + "displayname": "Test VM" + }, + TestData.computeOffering: { + TestData.name: "SF_CO_1", + "displaytext": "SF_CO_1 (Min IOPS = 10,000; Max IOPS = 15,000)", + "cpunumber": 1, + "cpuspeed": 100, + "memory": 128, + "storagetype": "shared", + "customizediops": False, + TestData.min_iops: 10000, + TestData.max_iops: 15000, + "hypervisorsnapshotreserve": 200, + TestData.tags: TestData.storageTag + }, + TestData.systemOffering: { + TestData.name: "SF_SO_1", + "displaytext": "Managed SO (Min IOPS = 4,000; Max IOPS = 8,000)", + "cpunumber": 1, + "cpuspeed": 100, + "memory": 128, + "storagetype": "shared", + TestData.min_iops: 4000, + TestData.max_iops: 8000, + TestData.tags: TestData.storageTag, + "issystem": True + }, + TestData.systemOfferingFailure: { + TestData.name: "SF_SO_2", + "displaytext": "Managed SO (Customized IOPS)", + "cpunumber": 1, + "cpuspeed": 100, + "memory": 128, + "storagetype": "shared", + "customizediops": True, + TestData.tags: TestData.storageTag, + "issystem": True + }, + TestData.zoneId: 1, + TestData.clusterId: 1, + TestData.domainId: 1, + TestData.url: "192.168.129.50" + } + + @staticmethod + def get_name_for_solidfire_storage(): + return "SolidFire-%d" % random.randint(0, 100) + + +class TestManagedSystemVMs(cloudstackTestCase): + _unique_name_suffix = "-Temp" + + _secondary_storage_unique_name = "Cloud.com-SecondaryStorage" + _secondary_storage_temp_unique_name = _secondary_storage_unique_name + _unique_name_suffix + + _console_proxy_unique_name = "Cloud.com-ConsoleProxy" + _console_proxy_temp_unique_name = _console_proxy_unique_name + _unique_name_suffix + + _virtual_router_unique_name = "Cloud.com-SoftwareRouter" + _virtual_router_temp_unique_name = _virtual_router_unique_name + _unique_name_suffix + + @classmethod + def setUpClass(cls): + # Set up API client + testclient = super(TestManagedSystemVMs, cls).getClsTestClient() + cls.apiClient = testclient.getApiClient() + cls.configData = testclient.getParsedTestDataConfig() + cls.dbConnection = testclient.getDbConnection() + + cls.testdata = TestData().testdata + + # Set up xenAPI connection + host_ip = "https://" + \ + list_hosts(cls.apiClient, clusterid=cls.testdata[TestData.clusterId], name="XenServer-6.5-1")[0].ipaddress + + # Set up XenAPI connection + cls.xen_session = XenAPI.Session(host_ip) + + xenserver = cls.testdata[TestData.xenServer] + + cls.xen_session.xenapi.login_with_password(xenserver[TestData.username], xenserver[TestData.password]) + + # Set up SolidFire connection + cls.sf_client = sf_api.SolidFireAPI(endpoint_dict=cls.testdata[TestData.solidFire]) + + # Get Resources from Cloud Infrastructure + cls.zone = Zone(get_zone(cls.apiClient, zone_id=cls.testdata[TestData.zoneId]).__dict__) + cls.cluster = list_clusters(cls.apiClient)[0] + cls.template = get_template(cls.apiClient, cls.zone.id, cls.configData["ostype"]) + cls.domain = get_domain(cls.apiClient, cls.testdata[TestData.domainId]) + + # Create test account + cls.account = Account.create( + cls.apiClient, + cls.testdata["account"], + admin=1 + ) + + # Set up connection to make customized API calls + cls.user = User.create( + cls.apiClient, + cls.testdata["user"], + account=cls.account.name, + domainid=cls.domain.id + ) + + url = cls.testdata[TestData.url] + + api_url = "http://" + url + ":8080/client/api" + userkeys = User.registerUserKeys(cls.apiClient, cls.user.id) + + cls.cs_api = SignedAPICall.CloudStack(api_url, userkeys.apikey, userkeys.secretkey) + + cls.compute_offering = ServiceOffering.create( + cls.apiClient, + cls.testdata[TestData.computeOffering] + ) + + systemoffering = cls.testdata[TestData.systemOffering] + + systemoffering[TestData.name] = "Managed SSVM" + systemoffering['systemvmtype'] = "secondarystoragevm" + + cls.secondary_storage_offering = ServiceOffering.create( + cls.apiClient, + systemoffering + ) + + systemoffering[TestData.name] = "Managed CPVM" + systemoffering['systemvmtype'] = "consoleproxy" + + cls.console_proxy_offering = ServiceOffering.create( + cls.apiClient, + systemoffering + ) + + systemoffering[TestData.name] = "Managed VR" + systemoffering['systemvmtype'] = "domainrouter" + + cls.virtual_router_offering = ServiceOffering.create( + cls.apiClient, + systemoffering + ) + + # Resources that are to be destroyed + cls._cleanup = [ + cls.secondary_storage_offering, + cls.console_proxy_offering, + cls.virtual_router_offering, + cls.compute_offering, + cls.user, + cls.account + ] + + @classmethod + def tearDownClass(cls): + try: + cleanup_resources(cls.apiClient, cls._cleanup) + except Exception as e: + logging.debug("Exception in tearDownClass(cls): %s" % e) + + def setUp(self): + self.cleanup = [] + + def tearDown(self): + try: + cleanup_resources(self.apiClient, self.cleanup) + + sf_util.purge_solidfire_volumes(self.sf_client) + except Exception as e: + logging.debug("Exception in tearDownClass(self): %s" % e) + + @attr(hypervisor='XenServer') + def test_01_create_system_vms_on_managed_storage(self): + self._disable_zone_and_delete_system_vms(None, False) + + primary_storage = self.testdata[TestData.primaryStorage] + + primary_storage_1 = StoragePool.create( + self.apiClient, + primary_storage + ) + + self._prepare_to_use_managed_storage_for_system_vms() + + enabled = "Enabled" + + self.zone.update(self.apiClient, id=self.zone.id, allocationstate=enabled) + + system_vms = self._wait_for_and_get_running_system_vms(2) + + virtual_machine = VirtualMachine.create( + self.apiClient, + self.testdata[TestData.virtualMachine], + accountid=self.account.name, + zoneid=self.zone.id, + serviceofferingid=self.compute_offering.id, + templateid=self.template.id, + domainid=self.domain.id, + startvm=True + ) + + # This virtual machine was only created and started so that the virtual router would be created and started. + # Just delete this virtual machine once it has been created and started. + virtual_machine.delete(self.apiClient, True) + + virtual_router = list_routers(self.apiClient, listall=True, state="Running")[0] + + system_vms.append(virtual_router) + + self._check_system_vms(system_vms, primary_storage_1.id) + + primary_storage[TestData.name] = TestData.get_name_for_solidfire_storage() + + primary_storage_2 = StoragePool.create( + self.apiClient, + primary_storage + ) + + StoragePool.enableMaintenance(self.apiClient, primary_storage_1.id) + + self._wait_for_storage_cleanup_thread(system_vms) + + sf_util.purge_solidfire_volumes(self.sf_client) + + system_vms = self._wait_for_and_get_running_system_vms(2) + + virtual_router = list_routers(self.apiClient, listall=True, state="Running")[0] + + system_vms.append(virtual_router) + + self._check_system_vms(system_vms, primary_storage_2.id) + + StoragePool.cancelMaintenance(self.apiClient, primary_storage_1.id) + + primary_storage_1.delete(self.apiClient) + + self._disable_zone_and_delete_system_vms(virtual_router) + + self._wait_for_storage_cleanup_thread(system_vms) + + sf_util.purge_solidfire_volumes(self.sf_client) + + primary_storage_2.delete(self.apiClient) + + self._verify_no_active_solidfire_volumes() + + self._prepare_to_stop_using_managed_storage_for_system_vms() + + self.zone.update(self.apiClient, id=self.zone.id, allocationstate=enabled) + + self._wait_for_and_get_running_system_vms(2) + + @attr(hypervisor='XenServer') + def test_02_failure_to_create_service_offering_with_customized_iops(self): + try: + ServiceOffering.create( + self.apiClient, + self.testdata[TestData.systemOfferingFailure] + ) + + self.assert_(True, "The service offering was created, but should not have been.") + except: + pass + + def _prepare_to_use_managed_storage_for_system_vms(self): + self._update_system_vm_unique_name(TestManagedSystemVMs._secondary_storage_unique_name, TestManagedSystemVMs._secondary_storage_temp_unique_name) + self._update_system_vm_unique_name(TestManagedSystemVMs._console_proxy_unique_name, TestManagedSystemVMs._console_proxy_temp_unique_name) + self._update_system_vm_unique_name(TestManagedSystemVMs._virtual_router_unique_name, TestManagedSystemVMs._virtual_router_temp_unique_name) + + self._update_system_vm_unique_name_based_on_uuid(self.secondary_storage_offering.id, TestManagedSystemVMs._secondary_storage_unique_name) + self._update_system_vm_unique_name_based_on_uuid(self.console_proxy_offering.id, TestManagedSystemVMs._console_proxy_unique_name) + self._update_system_vm_unique_name_based_on_uuid(self.virtual_router_offering.id, TestManagedSystemVMs._virtual_router_unique_name) + + def _prepare_to_stop_using_managed_storage_for_system_vms(self): + self._update_system_vm_unique_name_based_on_uuid(self.secondary_storage_offering.id, None) + self._update_system_vm_unique_name_based_on_uuid(self.console_proxy_offering.id, None) + self._update_system_vm_unique_name_based_on_uuid(self.virtual_router_offering.id, None) + + self._update_system_vm_unique_name(TestManagedSystemVMs._secondary_storage_temp_unique_name, TestManagedSystemVMs._secondary_storage_unique_name) + self._update_system_vm_unique_name(TestManagedSystemVMs._console_proxy_temp_unique_name, TestManagedSystemVMs._console_proxy_unique_name) + self._update_system_vm_unique_name(TestManagedSystemVMs._virtual_router_temp_unique_name, TestManagedSystemVMs._virtual_router_unique_name) + + def _wait_for_storage_cleanup_thread(self, system_vms): + retry_interval = 60 + num_tries = 10 + + wait_result, return_val = wait_until(retry_interval, num_tries, self._check_resource_state, system_vms) + + if not wait_result: + raise Exception(return_val) + + def _check_resource_state(self, system_vms): + try: + self._verify_system_vms_deleted(system_vms) + + return True, None + except: + return False, "The system is not in the necessary state." + + def _verify_system_vms_deleted(self, system_vms): + for system_vm in system_vms: + cs_root_volume = self._get_root_volume_for_system_vm(system_vm.id, 'Expunged') + + self._verify_managed_system_vm_deleted(cs_root_volume.name) + + def _disable_zone_and_delete_system_vms(self, virtual_router, verify_managed_system_vm_deleted=True): + self.zone.update(self.apiClient, id=self.zone.id, allocationstate="Disabled") + + if virtual_router is not None: + Router.destroy(self.apiClient, virtual_router.id) + + if verify_managed_system_vm_deleted: + cs_root_volume = self._get_root_volume_for_system_vm(virtual_router.id, 'Expunged') + + self._verify_managed_system_vm_deleted(cs_root_volume.name) + + # list_ssvms lists the secondary storage VM and the console proxy VM + system_vms = list_ssvms(self.apiClient) + + for system_vm in system_vms: + destroy_ssvm_cmd = destroySystemVm.destroySystemVmCmd() + + destroy_ssvm_cmd.id = system_vm.id + + self.apiClient.destroySystemVm(destroy_ssvm_cmd) + + if verify_managed_system_vm_deleted: + cs_root_volume = self._get_root_volume_for_system_vm(system_vm.id, 'Expunged') + + self._verify_managed_system_vm_deleted(cs_root_volume.name) + + def _verify_managed_system_vm_deleted(self, cs_root_volume_name): + sf_not_active_volumes = sf_util.get_not_active_sf_volumes(self.sf_client) + + sf_root_volume = sf_util.check_and_get_sf_volume(sf_not_active_volumes, cs_root_volume_name, self) + + self.assertEqual( + len(sf_root_volume['volumeAccessGroups']), + 0, + "The volume should not be in a volume access group." + ) + + sr_name = sf_util.format_iqn(sf_root_volume["iqn"]) + + sf_util.check_xen_sr(sr_name, self.xen_session, self, False) + + def _wait_for_and_get_running_system_vms(self, expected_number_of_system_vms): + retries = 10 + + while True: + # list_ssvms lists the secondary storage VM and the console proxy VM + system_vms = list_ssvms(self.apiClient, state="Running") + + if system_vms is not None and len(system_vms) == expected_number_of_system_vms: + return system_vms + + retries = retries - 1 + + if retries == 0: + raise Exception, "Timed out waiting for system VMs" + + delay(60) --- End diff -- Please consider replacing lines 485-499 with `utils.wait_until`
--- If your project is set up for it, you can reply to this email and have your reply appear on GitHub as well. If your project does not have this feature enabled and wishes so, or if the feature is enabled but not working, please contact infrastructure at infrastruct...@apache.org or file a JIRA ticket with INFRA. ---