Repository: ambari Updated Branches: refs/heads/trunk 55747a5c2 -> 87fa076ec
Revert "AMBARI-19338 AM sizing for LLAP - min container size changes (dsen)" This reverts commit d256e82e628a6ba86acd17a21c2b925a7b3ae9c5. Project: http://git-wip-us.apache.org/repos/asf/ambari/repo Commit: http://git-wip-us.apache.org/repos/asf/ambari/commit/87fa076e Tree: http://git-wip-us.apache.org/repos/asf/ambari/tree/87fa076e Diff: http://git-wip-us.apache.org/repos/asf/ambari/diff/87fa076e Branch: refs/heads/trunk Commit: 87fa076ec017636a19ad209349c38708535f64f2 Parents: 55747a5 Author: Sumit Mohanty <[email protected]> Authored: Mon Jan 9 11:06:30 2017 -0800 Committer: Sumit Mohanty <[email protected]> Committed: Mon Jan 9 11:06:30 2017 -0800 ---------------------------------------------------------------------- .../stacks/HDP/2.0.6/services/stack_advisor.py | 15 ++--- .../stacks/HDP/2.2/services/stack_advisor.py | 2 +- .../stacks/2.0.6/common/test_stack_advisor.py | 43 ++++++-------- .../stacks/2.2/common/test_stack_advisor.py | 34 ++++------- .../stacks/2.3/common/test_stack_advisor.py | 2 +- .../stacks/2.5/common/test_stack_advisor.py | 60 +++++++------------- 6 files changed, 58 insertions(+), 98 deletions(-) ---------------------------------------------------------------------- http://git-wip-us.apache.org/repos/asf/ambari/blob/87fa076e/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py index 62f4e3e..0c02eab 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.0.6/services/stack_advisor.py @@ -126,7 +126,7 @@ class HDP206StackAdvisor(DefaultStackAdvisor): if "referenceNodeManagerHost" in clusterData: nodemanagerMinRam = min(clusterData["referenceNodeManagerHost"]["total_mem"]/1024, nodemanagerMinRam) putYarnProperty('yarn.nodemanager.resource.memory-mb', int(round(min(clusterData['containers'] * clusterData['ramPerContainer'], nodemanagerMinRam)))) - putYarnProperty('yarn.scheduler.minimum-allocation-mb', int(clusterData['minContainerRam'])) + putYarnProperty('yarn.scheduler.minimum-allocation-mb', int(clusterData['ramPerContainer'])) putYarnProperty('yarn.scheduler.maximum-allocation-mb', int(configurations["yarn-site"]["properties"]["yarn.nodemanager.resource.memory-mb"])) putYarnEnvProperty('min_user_id', self.get_system_min_uid()) @@ -1048,8 +1048,7 @@ class HDP206StackAdvisor(DefaultStackAdvisor): cluster["minContainerSize"] = { - cluster["ram"] <= 3: 128, - 3 < cluster["ram"] <= 4: 256, + cluster["ram"] <= 4: 256, 4 < cluster["ram"] <= 8: 512, 8 < cluster["ram"] <= 24: 1024, 24 < cluster["ram"]: 2048 @@ -1060,22 +1059,20 @@ class HDP206StackAdvisor(DefaultStackAdvisor): totalAvailableRam -= cluster["hbaseRam"] cluster["totalAvailableRam"] = max(512, totalAvailableRam * 1024) '''containers = max(3, min (2*cores,min (1.8*DISKS,(Total available RAM) / MIN_CONTAINER_SIZE))))''' - cluster["containers"] = int(round(max(3, + cluster["containers"] = round(max(3, min(2 * cluster["cpu"], min(ceil(1.8 * cluster["disk"]), - cluster["totalAvailableRam"] / cluster["minContainerSize"]))))) + cluster["totalAvailableRam"] / cluster["minContainerSize"])))) '''ramPerContainers = max(2GB, RAM - reservedRam - hBaseRam) / containers''' - cluster["ramPerContainer"] = int(abs(cluster["totalAvailableRam"] / cluster["containers"])) + cluster["ramPerContainer"] = abs(cluster["totalAvailableRam"] / cluster["containers"]) '''If greater than 1GB, value will be in multiples of 512.''' if cluster["ramPerContainer"] > 1024: cluster["ramPerContainer"] = int(cluster["ramPerContainer"] / 512) * 512 - cluster["minContainerRam"] = min(1024, cluster["ramPerContainer"]) - cluster["mapMemory"] = int(cluster["ramPerContainer"]) cluster["reduceMemory"] = cluster["ramPerContainer"] - cluster["amMemory"] = min(cluster["mapMemory"], cluster["minContainerRam"]) + cluster["amMemory"] = max(cluster["mapMemory"], cluster["reduceMemory"]) return cluster http://git-wip-us.apache.org/repos/asf/ambari/blob/87fa076e/ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py b/ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py index 63e4b20..cf7134e 100644 --- a/ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py +++ b/ambari-server/src/main/resources/stacks/HDP/2.2/services/stack_advisor.py @@ -818,7 +818,7 @@ class HDP22StackAdvisor(HDP21StackAdvisor): yarnMaxAllocationSize = min(30 * int(configurations["yarn-site"]["properties"]["yarn.scheduler.minimum-allocation-mb"]), int(configurations["yarn-site"]["properties"]["yarn.scheduler.maximum-allocation-mb"])) putTezProperty = self.putProperty(configurations, "tez-site", services) - putTezProperty("tez.am.resource.memory.mb", min(int(configurations["yarn-site"]["properties"]["yarn.scheduler.maximum-allocation-mb"]), int(clusterData['amMemory']))) + putTezProperty("tez.am.resource.memory.mb", min(int(configurations["yarn-site"]["properties"]["yarn.scheduler.maximum-allocation-mb"]), int(clusterData['amMemory']) * 2 if int(clusterData['amMemory']) < 3072 else int(clusterData['amMemory']))) taskResourceMemory = clusterData['mapMemory'] if clusterData['mapMemory'] > 2048 else int(clusterData['reduceMemory']) taskResourceMemory = min(clusterData['containers'] * clusterData['ramPerContainer'], taskResourceMemory, yarnMaxAllocationSize) http://git-wip-us.apache.org/repos/asf/ambari/blob/87fa076e/ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py b/ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py index ab7d53b..355e211 100644 --- a/ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py +++ b/ambari-server/src/test/python/stacks/2.0.6/common/test_stack_advisor.py @@ -258,7 +258,7 @@ class TestHDP206StackAdvisor(TestCase): result = self.stackAdvisor.validateConfigurations(services, hosts) expectedItems = [ - {"message": "Value is less than the recommended default of 510", "level": "WARN"}, + {"message": "Value is less than the recommended default of 512", "level": "WARN"}, {'message': 'Value should be set for yarn.nodemanager.linux-container-executor.group', 'level': 'ERROR'}, {"message": "Value should be integer", "level": "ERROR"}, {"message": "Value should be set", "level": "ERROR"} @@ -419,7 +419,6 @@ class TestHDP206StackAdvisor(TestCase): "reservedRam": 2, "hbaseRam": 1, "minContainerSize": 512, - "minContainerRam" : 512, "totalAvailableRam": 3072, "containers": 6, "ramPerContainer": 512, @@ -488,16 +487,15 @@ class TestHDP206StackAdvisor(TestCase): }) expected["referenceHost"] = hosts["items"][1]["Hosts"] expected["referenceNodeManagerHost"] = hosts["items"][1]["Hosts"] - expected["amMemory"] = 128 - expected["containers"] = 4 + expected["amMemory"] = 170.66666666666666 + expected["containers"] = 3.0 expected["cpu"] = 4 expected["totalAvailableRam"] = 512 - expected["mapMemory"] = 128 - expected["minContainerSize"] = 128 - expected["reduceMemory"] = 128 - expected["minContainerRam"] = 128 + expected["mapMemory"] = 170 + expected["minContainerSize"] = 256 + expected["reduceMemory"] = 170.66666666666666 expected["ram"] = 0 - expected["ramPerContainer"] = 128 + expected["ramPerContainer"] = 170.66666666666666 expected["reservedRam"] = 1 result = self.stackAdvisor.getConfigurationClusterSummary(servicesList, hosts, components, services) self.assertEquals(result, expected) @@ -538,8 +536,7 @@ class TestHDP206StackAdvisor(TestCase): "ramPerContainer": 3072, "mapMemory": 3072, "reduceMemory": 3072, - "amMemory": 1024, - "minContainerRam": 1024, + "amMemory": 3072, "referenceHost": hosts["items"][0]["Hosts"] } @@ -595,8 +592,7 @@ class TestHDP206StackAdvisor(TestCase): services = {"configurations": configurations, "services": []} clusterData = { "containers" : 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "yarn-env": { @@ -623,8 +619,7 @@ class TestHDP206StackAdvisor(TestCase): clusterData = { "mapMemory": 567, "reduceMemory": 345.6666666666666, - "amMemory": 123.54, - "minContainerRam": 123.54 + "amMemory": 123.54 } expected = { "mapred-site": { @@ -660,14 +655,13 @@ class TestHDP206StackAdvisor(TestCase): "ram": 0, "reservedRam": 1, "hbaseRam": 1, - "minContainerSize": 128, + "minContainerSize": 256, "totalAvailableRam": 512, "containers": 3, - "ramPerContainer": 170, + "ramPerContainer": 170.66666666666666, "mapMemory": 170, - "reduceMemory": 170, - "amMemory": 170, - "minContainerRam" : 170 + "reduceMemory": 170.66666666666666, + "amMemory": 170.66666666666666 } self.assertEquals(result, expected) @@ -1177,8 +1171,7 @@ class TestHDP206StackAdvisor(TestCase): } clusterData = { - "totalAvailableRam": 2048, - "totalAvailableRam": 256 + "totalAvailableRam": 2048 } ambariHostName = socket.getfqdn() expected = {'oozie-env': @@ -1217,7 +1210,7 @@ class TestHDP206StackAdvisor(TestCase): 'namenode_heapsize': '1024', 'proxyuser_group': 'users', 'namenode_opt_maxnewsize': '256', - 'namenode_opt_newsize': '128'}}} + 'namenode_opt_newsize': '256'}}} # Apart from testing other HDFS recommendations, also tests 'hadoop.proxyuser.hive.hosts' config value which includes both HiveServer # and Hive Server Interactive Host (installed on different host compared to HiveServer). @@ -1352,7 +1345,7 @@ class TestHDP206StackAdvisor(TestCase): 'namenode_heapsize': '1024', 'proxyuser_group': 'users', 'namenode_opt_maxnewsize': '256', - 'namenode_opt_newsize': '128'}}} + 'namenode_opt_newsize': '256'}}} # Apart from testing other HDFS recommendations, also tests 'hadoop.proxyuser.hive.hosts' config value which includes both HiveServer # and Hive Server Interactive Host (installed on same host compared to HiveServer). @@ -1485,7 +1478,7 @@ class TestHDP206StackAdvisor(TestCase): 'namenode_heapsize': '1024', 'proxyuser_group': 'users', 'namenode_opt_maxnewsize': '256', - 'namenode_opt_newsize': '128'}}} + 'namenode_opt_newsize': '256'}}} self.stackAdvisor.recommendHDFSConfigurations(configurations, clusterData, services3, hosts) self.assertEquals(configurations, expected) http://git-wip-us.apache.org/repos/asf/ambari/blob/87fa076e/ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py b/ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py index 2a61b34..db71ec8 100644 --- a/ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py +++ b/ambari-server/src/test/python/stacks/2.2/common/test_stack_advisor.py @@ -86,11 +86,10 @@ class TestHDP22StackAdvisor(TestCase): } clusterData = { "mapMemory": 3000, - "amMemory": 2048, + "amMemory": 2000, "reduceMemory": 2056, "containers": 3, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "tez-site": { @@ -184,7 +183,6 @@ class TestHDP22StackAdvisor(TestCase): "amMemory": 3100, "reduceMemory": 2056, "containers": 3, - "minContainerRam": 256, "ramPerContainer": 256 } expected = { @@ -267,11 +265,10 @@ class TestHDP22StackAdvisor(TestCase): } clusterData = { "mapMemory": 760, - "amMemory": 2048, + "amMemory": 2000, "reduceMemory": 760, "containers": 3, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "tez-site": { @@ -887,8 +884,7 @@ class TestHDP22StackAdvisor(TestCase): clusterData = { "cpu": 4, "containers" : 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "yarn-env": { @@ -927,8 +923,7 @@ class TestHDP22StackAdvisor(TestCase): clusterData = { "cpu": 4, "containers": 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "spark-defaults": { @@ -960,8 +955,7 @@ class TestHDP22StackAdvisor(TestCase): clusterData = { "cpu": 4, "containers" : 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "yarn-env": { @@ -1222,8 +1216,7 @@ class TestHDP22StackAdvisor(TestCase): clusterData = { "cpu": 4, "containers" : 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } services = { @@ -1270,8 +1263,7 @@ class TestHDP22StackAdvisor(TestCase): "amMemory": 2000, "reduceMemory": 2056, "containers": 3, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { @@ -1762,7 +1754,6 @@ class TestHDP22StackAdvisor(TestCase): "containers" : 7, "ramPerContainer": 256, "totalAvailableRam": 4096, - "minContainerRam": 256 } expected = { "cluster-env": { @@ -2030,8 +2021,7 @@ class TestHDP22StackAdvisor(TestCase): clusterData = { "cpu": 4, "containers" : 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "yarn-env": { @@ -3819,7 +3809,7 @@ class TestHDP22StackAdvisor(TestCase): "yarn.nodemanager.linux-container-executor.cgroups.mount-path": "/cgroup", "yarn.nodemanager.linux-container-executor.cgroups.mount": "true", "yarn.nodemanager.resource.memory-mb": "39424", - "yarn.scheduler.minimum-allocation-mb": "1024", + "yarn.scheduler.minimum-allocation-mb": "3584", "yarn.scheduler.maximum-allocation-vcores": "4", "yarn.scheduler.minimum-allocation-vcores": "1", "yarn.nodemanager.resource.cpu-vcores": "4", @@ -3878,7 +3868,7 @@ class TestHDP22StackAdvisor(TestCase): "yarn.nodemanager.linux-container-executor.cgroups.mount-path": "/cgroup", "yarn.nodemanager.linux-container-executor.cgroups.mount": "true", "yarn.nodemanager.resource.memory-mb": "39424", - "yarn.scheduler.minimum-allocation-mb": "1024", + "yarn.scheduler.minimum-allocation-mb": "3584", "yarn.scheduler.maximum-allocation-vcores": "4", "yarn.scheduler.minimum-allocation-vcores": "1", "yarn.nodemanager.resource.cpu-vcores": "4", http://git-wip-us.apache.org/repos/asf/ambari/blob/87fa076e/ambari-server/src/test/python/stacks/2.3/common/test_stack_advisor.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.3/common/test_stack_advisor.py b/ambari-server/src/test/python/stacks/2.3/common/test_stack_advisor.py index 88fa6d0..378522e 100644 --- a/ambari-server/src/test/python/stacks/2.3/common/test_stack_advisor.py +++ b/ambari-server/src/test/python/stacks/2.3/common/test_stack_advisor.py @@ -1165,7 +1165,7 @@ class TestHDP23StackAdvisor(TestCase): "tez.runtime.io.sort.mb": "202", "tez.session.am.dag.submit.timeout.secs": "600", "tez.runtime.unordered.output.buffer.size-mb": "57", - "tez.am.resource.memory.mb": "2000", + "tez.am.resource.memory.mb": "4000", "tez.queue.name": "queue2", } }, http://git-wip-us.apache.org/repos/asf/ambari/blob/87fa076e/ambari-server/src/test/python/stacks/2.5/common/test_stack_advisor.py ---------------------------------------------------------------------- diff --git a/ambari-server/src/test/python/stacks/2.5/common/test_stack_advisor.py b/ambari-server/src/test/python/stacks/2.5/common/test_stack_advisor.py index d4a301c..a7ccdfe 100644 --- a/ambari-server/src/test/python/stacks/2.5/common/test_stack_advisor.py +++ b/ambari-server/src/test/python/stacks/2.5/common/test_stack_advisor.py @@ -374,8 +374,7 @@ class TestHDP25StackAdvisor(TestCase): clusterData = { "cpu": 4, "containers": 5, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { "spark2-defaults": { @@ -802,8 +801,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 1024 - }, - "minContainerRam": 512 + } } @@ -971,8 +969,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 1024 - }, - "minContainerRam": 512 + } } configurations = { @@ -1152,8 +1149,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 1024 - }, - "minContainerRam": 512 + } } @@ -1352,8 +1348,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 2048 - }, - "minContainerRam": 512 + } } @@ -1547,8 +1542,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 1024, "referenceNodeManagerHost" : { "total_mem" : 51200 * 1024 - }, - "minContainerRam": 1024 + } } configurations = { @@ -1750,8 +1744,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 2048, "referenceNodeManagerHost" : { "total_mem" : 40960 * 1024 - }, - "minContainerRam": 1024 + } } configurations = { @@ -1947,8 +1940,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 341, "referenceNodeManagerHost" : { "total_mem" : 12288 * 1024 - }, - "minContainerRam": 341 + } } @@ -2146,8 +2138,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 2048, "referenceNodeManagerHost" : { "total_mem" : 204800 * 1024 - }, - "minContainerRam": 1024 + } } configurations = { @@ -2349,8 +2340,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 3072, "referenceNodeManagerHost" : { "total_mem" : 40960 * 1024 - }, - "minContainerRam": 1024 + } } @@ -2548,8 +2538,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 341, "referenceNodeManagerHost" : { "total_mem" : 204800 * 1024 - }, - "minContainerRam": 341 + } } @@ -2746,8 +2735,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 82240, "referenceNodeManagerHost" : { "total_mem" : 204800 * 1024 - }, - "minContainerRam": 1024 + } } configurations = { @@ -2939,8 +2927,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 82240, "referenceNodeManagerHost" : { "total_mem" : 328960 * 1024 - }, - "minContainerRam": 1024 + } } @@ -3161,8 +3148,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 1024 - }, - "minContainerRam": 512 + } } configurations = { @@ -3388,8 +3374,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 1024 - }, - "minContainerRam": 512 + } } configurations = { @@ -3581,8 +3566,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 82240, "referenceNodeManagerHost" : { "total_mem" : 328960 * 1024 - }, - "minContainerRam": 1024 + } } configurations = { @@ -3764,8 +3748,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 82240, "referenceNodeManagerHost" : { "total_mem" : 328960 * 1024 - }, - "minContainerRam": 1024 + } } configurations = { @@ -3897,8 +3880,7 @@ class TestHDP25StackAdvisor(TestCase): "ramPerContainer": 512, "referenceNodeManagerHost" : { "total_mem" : 10240 * 1024 - }, - "minContainerRam": 512 + } } configurations = { @@ -3943,8 +3925,7 @@ class TestHDP25StackAdvisor(TestCase): "amMemory": 2000, "reduceMemory": 2056, "containers": 3, - "ramPerContainer": 256, - "minContainerRam": 256 + "ramPerContainer": 256 } expected = { 'application-properties': { @@ -4679,8 +4660,7 @@ class TestHDP25StackAdvisor(TestCase): "cpu": 4, "referenceNodeManagerHost" : { "total_mem" : 328960 * 1024 - }, - "minContainerRam": 256 + } } hosts = { "items" : [
