CLOUDSTACK-8439:Automation for feature CPU/Memory overcommit CLOUDSTACK-8439:Automation for feature CPU/Memory overcommit This closes #226
Project: http://git-wip-us.apache.org/repos/asf/cloudstack/repo Commit: http://git-wip-us.apache.org/repos/asf/cloudstack/commit/896f6b1e Tree: http://git-wip-us.apache.org/repos/asf/cloudstack/tree/896f6b1e Diff: http://git-wip-us.apache.org/repos/asf/cloudstack/diff/896f6b1e Branch: refs/heads/reporter Commit: 896f6b1e5e2b23958886899334b09e71ffe38e6f Parents: 5d9f851 Author: nitt10prashant <nitt10.prash...@gmail.com> Authored: Mon May 4 12:13:52 2015 +0530 Committer: sanjeev <sanj...@apache.org> Committed: Mon Jul 6 09:45:26 2015 +0530 ---------------------------------------------------------------------- test/integration/component/test_overcommit.py | 458 +++++++++++++++++++++ 1 file changed, 458 insertions(+) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/cloudstack/blob/896f6b1e/test/integration/component/test_overcommit.py ---------------------------------------------------------------------- diff --git a/test/integration/component/test_overcommit.py b/test/integration/component/test_overcommit.py new file mode 100644 index 0000000..db04316 --- /dev/null +++ b/test/integration/component/test_overcommit.py @@ -0,0 +1,458 @@ +# Licensed to the Apache Software Foundation (ASF) under one +# or more contributor license agreements. See the NOTICE file +# distributed with this work for additional information +# regarding copyright ownership. The ASF licenses this file +# to you under the Apache License, Version 2.0 (the +# "License"); you may not use this file except in compliance +# with the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, +# software distributed under the License is distributed on an +# "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY +# KIND, either express or implied. See the License for the +# specific language governing permissions and limitations +# under the License. + +""" test for feature "CPU and MEM OVERCOMMIT" +""" + +from marvin.cloudstackTestCase import cloudstackTestCase +from marvin.lib.base import (Account, + Capacities, + Cluster, + Configurations, + FAILED, + Host, + PASS, + ServiceOffering, + time, + VirtualMachine,) +from marvin.lib.utils import cleanup_resources, validateList +from marvin.lib.common import (get_zone, + get_domain, + get_template) +from marvin.sshClient import SshClient +from nose.plugins.attrib import attr + + +def ssh_xen_host(password, ipaddr, instance_name): + """Ssh into xen host and get vm mem details""" + mem = [] + sshClient = SshClient( + ipaddr, + 22, + "root", + password + ) + command = "xe vm-list params=all name-label=%s" % instance_name + vm_detail = sshClient.execute(command) + max_str = vm_detail[17].split(":") + min_str = vm_detail[20].split(":") + max = int(max_str[1]) + min = int(min_str[1]) + mem.append(max) + mem.append(min) + return mem + + +def ssh_kvm_host(password, ipaddr, instance_name): + """Ssh into kvm host and get vm mem details""" + mem = [] + sshClient = SshClient( + ipaddr, + 22, + "root", + password + ) + + command = "virsh dominfo %s" % instance_name + vm_detail = sshClient.execute(command) + max = vm_detail[7].split() + min = vm_detail[8].split() + mem.append(int(max[2])) + mem.append(int(min[2])) + return mem + + +def capacity_parser(capacity): + cpu = [] + mem = [] + cpu.append(capacity[0].capacitytotal) + cpu.append(capacity[0].capacityused) + cpu.append(capacity[0].percentused) + mem.append(capacity[1].capacitytotal) + mem.append(capacity[1].capacityused) + mem.append(capacity[1].percentused) + return cpu, mem + + +class Overcommit (cloudstackTestCase): + + @classmethod + def setUpClass(cls): + testClient = super(Overcommit, cls).getClsTestClient() + cls.apiclient = testClient.getApiClient() + cls.testdata = testClient.getParsedTestDataConfig() + # Get Zone,Domain and templates + cls.domain = get_domain(cls.apiclient) + cls.zone = get_zone(cls.apiclient) + cls.testdata["mode"] = cls.zone.networktype + cls.testdata["configurableData"]["password"] = "xenroot" + cls.hypervisor = testClient.getHypervisorInfo() + + cls.template = get_template( + cls.apiclient, + cls.zone.id, + cls.testdata["ostype"]) + cls.testdata["template"]["ostypeid"] = cls.template.ostypeid + list_conf = Configurations.list(cls.apiclient, + name="capacity.check.period" + ) + cls.wait_time = 5 + int(list_conf[0].value) / 1000 + if cls.template == FAILED: + cls.fail( + "get_template() failed to return template with description \ + %s" % + cls.testdata["ostype"]) + cls._cleanup = [] + try: + cls.account = Account.create(cls.apiclient, + cls.testdata["account"], + domainid=cls.domain.id + ) + cls._cleanup.append(cls.account) + + cls.service_offering = ServiceOffering.create( + cls.apiclient, + cls.testdata["service_offerings"]["small"]) + + cls._cleanup.append(cls.service_offering) + + cls.deployVmResponse = VirtualMachine.create( + cls.apiclient, + services=cls.testdata["virtual_machine"], + accountid=cls.account.name, + domainid=cls.account.domainid, + serviceofferingid=cls.service_offering.id, + templateid=cls.template.id, + zoneid=cls.zone.id, + ) + + except Exception as e: + cls.tearDownClass() + raise e + + return + + @classmethod + def tearDownClass(cls): + try: + cleanup_resources(cls.apiclient, cls._cleanup) + except Exception as e: + raise Exception("Warning:Exception during cleanup: %s" % e) + + @attr( + tags=[ + "simulator", + "devcloud", + "basic", + "advanced"], + required_hardware="false") + def test_01_cluster_settings(self): + """change cpu/mem.overprovisioning.factor at cluster level and + verify the change """ + listHost = Host.list(self.apiclient, + id=self.deployVmResponse.hostid + ) + self.assertEqual( + validateList(listHost)[0], + PASS, + "check list host response for host id %s" % + self.deployVmResponse.hostid) + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="mem.overprovisioning.factor", + value=2) + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="cpu.overprovisioning.factor", + value=3) + + list_cluster = Cluster.list(self.apiclient, + id=listHost[0].clusterid) + self.assertEqual( + validateList(list_cluster)[0], + PASS, + "check list cluster response for cluster id %s" % + listHost[0].clusterid) + self.assertEqual(int(list_cluster[0].cpuovercommitratio), + 3, + "check the cpu overcommit value at cluster level ") + + self.assertEqual(int(list_cluster[0].memoryovercommitratio), + 2, + "check memory overcommit value at cluster level") + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="mem.overprovisioning.factor", + value=1) + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="cpu.overprovisioning.factor", + value=1) + list_cluster1 = Cluster.list(self.apiclient, + id=listHost[0].clusterid) + self.assertEqual( + validateList(list_cluster1)[0], + PASS, + "check the list cluster response for id %s" % + listHost[0].clusterid) + self.assertEqual(int(list_cluster1[0].cpuovercommitratio), + 1, + "check the cpu overcommit value at cluster level ") + + self.assertEqual(int(list_cluster1[0].memoryovercommitratio), + 1, + "check memory overcommit value at cluster level") + + @attr( + tags=["simulator", + "devcloud", + "basic", + "advanced"], + required_hardware="true") + def test_02_Overcommit_factor(self): + """change mem.overprovisioning.factor and verify vm memory """ + + listHost = Host.list(self.apiclient, + id=self.deployVmResponse.hostid + ) + self.assertEqual( + validateList(listHost)[0], + PASS, + "check list host for host id %s" % + self.deployVmResponse.hostid) + if listHost[0].hypervisor.lower() not in ['kvm', 'xenserver']: + self.skipTest( + "Skiping test because of not supported hypervisor type %s" % + listHost[0].hypervisor) + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="mem.overprovisioning.factor", + value=1) + + self.deployVmResponse.stop(self.apiclient) + self.deployVmResponse.start(self.apiclient) + + if listHost[0].hypervisor.lower() == 'xenserver': + + k = ssh_xen_host( + self.testdata["configurableData"]["password"], + listHost[0].ipaddress, + self.deployVmResponse.instancename) + + elif listHost[0].hypervisor.lower() == 'kvm': + + k = ssh_kvm_host( + self.testdata["configurableData"]["password"], + listHost[0].ipaddress, + self.deployVmResponse.instancename) + + self.assertEqual(k[0], + k[1], + "Check static max ,min on host for overcommit 1 ") + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="mem.overprovisioning.factor", + value=2) + + self.deployVmResponse.stop(self.apiclient) + self.deployVmResponse.start(self.apiclient) + + if listHost[0].hypervisor.lower() == 'xenserver': + k1 = ssh_xen_host( + self.testdata["configurableData"]["password"], + listHost[0].ipaddress, + self.deployVmResponse.instancename) + + elif listHost[0].hypervisor.lower() == 'kvm': + time.sleep(200) + k1 = ssh_kvm_host( + self.testdata["configurableData"]["password"], + listHost[0].ipaddress, + self.deployVmResponse.instancename) + self.assertEqual(k1[0], + 2 * k1[1], + "Check static max ,min on host for overcommit 2") + + @attr( + tags=[ + "simulator", + "devcloud", + "basic", + "advanced"], + required_hardware="false") + def test_03_cluste_capacity_check(self): + """change cpu/mem.overprovisioning.factor at cluster level and + verify cluster capacity """ + + listHost = Host.list(self.apiclient, + id=self.deployVmResponse.hostid + ) + self.assertEqual( + validateList(listHost)[0], + PASS, + "check list host for host id %s" % + self.deployVmResponse.hostid) + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="mem.overprovisioning.factor", + value=1) + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="cpu.overprovisioning.factor", + value=1) + + time.sleep(self.wait_time) + + capacity = Capacities.list(self.apiclient, + clusterid=listHost[0].clusterid) + self.assertEqual( + validateList(capacity)[0], + PASS, + "check list capacity response for cluster id %s" % + listHost[0].clusterid) + cpu, mem = capacity_parser(capacity) + + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="mem.overprovisioning.factor", + value=2) + Configurations.update(self.apiclient, + clusterid=listHost[0].clusterid, + name="cpu.overprovisioning.factor", + value=2) + + time.sleep(self.wait_time) + + capacity1 = Capacities.list(self.apiclient, + clusterid=listHost[0].clusterid) + self.assertEqual( + validateList(capacity1)[0], + PASS, + "check list capacity response for cluster id %s" % + listHost[0].clusterid) + cpu1, mem1 = capacity_parser(capacity1) + self.assertEqual(2 * cpu[0], + cpu1[0], + "check total capacity ") + self.assertEqual(2 * cpu[1], + cpu1[1], + "check capacity used") + self.assertEqual(cpu[2], + cpu1[2], + "check capacity % used") + + self.assertEqual(2 * mem[0], + mem1[0], + "check mem total capacity ") + self.assertEqual(2 * mem[1], + mem1[1], + "check mem capacity used") + self.assertEqual(mem[2], + mem1[2], + "check mem capacity % used") + + @attr( + tags=[ + "simulator", + "devcloud", + "basic", + "advanced"], + required_hardware="false") + def test_04_zone_capacity_check(self): + """change cpu/mem.overprovisioning.factor at cluster level for + all cluster in a zone and verify capacity at zone level """ + list_cluster = Cluster.list(self.apiclient, + zoneid=self.zone.id) + self.assertEqual( + validateList(list_cluster)[0], + PASS, + "check list cluster response for zone id %s" % + self.zone.id) + k = len(list_cluster) + for id in xrange(k): + Configurations.update(self.apiclient, + clusterid=list_cluster[id].id, + name="mem.overprovisioning.factor", + value=1) + Configurations.update(self.apiclient, + clusterid=list_cluster[id].id, + name="cpu.overprovisioning.factor", + value=1) + + time.sleep(self.wait_time) + + capacity = Capacities.list(self.apiclient, + zoneid=self.zone.id) + self.assertEqual( + validateList(capacity)[0], + PASS, + "check list capacity response for zone id %s" % + self.zone.id) + cpu, mem = capacity_parser(capacity) + for id in xrange(k): + Configurations.update(self.apiclient, + clusterid=list_cluster[id].id, + name="mem.overprovisioning.factor", + value=2) + Configurations.update(self.apiclient, + clusterid=list_cluster[id].id, + name="cpu.overprovisioning.factor", + value=2) + + time.sleep(self.wait_time) + + capacity1 = Capacities.list(self.apiclient, + zoneid=self.zone.id) + self.assertEqual(validateList(capacity1)[0], + PASS, + "check list capacity for zone id %s" % self.zone.id) + + cpu1, mem1 = capacity_parser(capacity1) + self.assertEqual(2 * cpu[0], + cpu1[0], + "check total capacity ") + self.assertEqual(2 * cpu[1], + cpu1[1], + "check capacity used") + self.assertEqual(cpu[2], + cpu1[2], + "check capacity % used") + + self.assertEqual(2 * mem[0], + mem1[0], + "check mem total capacity ") + self.assertEqual(2 * mem[1], + mem1[1], + "check mem capacity used") + self.assertEqual(mem[2], + mem1[2], + "check mem capacity % used") + for id in xrange(k): + Configurations.update(self.apiclient, + clusterid=list_cluster[id].id, + name="mem.overprovisioning.factor", + value=1) + Configurations.update(self.apiclient, + clusterid=list_cluster[id].id, + name="cpu.overprovisioning.factor", + value=1)