This is an automated email from the ASF dual-hosted git repository.

adoroszlai pushed a commit to branch trunk
in repository https://gitbox.apache.org/repos/asf/ambari.git


The following commit(s) were added to refs/heads/trunk by this push:
     new 90b7c14  AMBARI-22726. Fix dependent tests in BlueprintImplTest (#133)
90b7c14 is described below

commit 90b7c1417484ad8c4792b3996603a326ae9618a4
Author: Reed Oei <[email protected]>
AuthorDate: Wed Jan 17 13:42:37 2018 -0600

    AMBARI-22726. Fix dependent tests in BlueprintImplTest (#133)
---
 .../ambari/server/topology/BlueprintImplTest.java  | 26 +++++++++++-----------
 1 file changed, 13 insertions(+), 13 deletions(-)

diff --git 
a/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
 
b/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
index dd0adcc..d34526a 100644
--- 
a/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
+++ 
b/ambari-server/src/test/java/org/apache/ambari/server/topology/BlueprintImplTest.java
@@ -61,9 +61,10 @@ public class BlueprintImplTest {
   Map<String, Map<String, String>> properties = new HashMap<>();
   Map<String, String> hdfsProps = new HashMap<>();
   Configuration configuration = new Configuration(properties, 
EMPTY_ATTRIBUTES, EMPTY_CONFIGURATION);
+  org.apache.ambari.server.configuration.Configuration serverConfig;
 
   @Before
-  public void setup() {
+  public void setup() throws NoSuchFieldException, IllegalAccessException {
     properties.put("hdfs-site", hdfsProps);
     hdfsProps.put("foo", "val");
     hdfsProps.put("bar", "val");
@@ -104,11 +105,12 @@ public class BlueprintImplTest {
     requiredService2Properties.add(new Stack.ConfigProperty("category2", 
"prop2", null));
     
expect(stack.getRequiredConfigurationProperties("HDFS")).andReturn(requiredHDFSProperties).anyTimes();
     
expect(stack.getRequiredConfigurationProperties("SERVICE2")).andReturn(requiredService2Properties).anyTimes();
+
+    serverConfig = setupConfigurationWithGPLLicense(true);
   }
 
   @Test
   public void testValidateConfigurations__basic_positive() throws Exception {
-    org.apache.ambari.server.configuration.Configuration serverConfig = 
setupConfigurationWithGPLLicense(true);
     expect(group1.getCardinality()).andReturn("1").atLeastOnce();
     expect(group1.getComponents()).andReturn(Arrays.asList(new 
Component("c1"), new Component("c2"))).atLeastOnce();
     expect(group2.getCardinality()).andReturn("1").atLeastOnce();
@@ -138,7 +140,6 @@ public class BlueprintImplTest {
     group2Props.put("category2", group2Category2Props);
     group2Category2Props.put("prop2", "val");
 
-    org.apache.ambari.server.configuration.Configuration serverConfig = 
setupConfigurationWithGPLLicense(true);
     // set config for group2 which contains a required property
     Configuration group2Configuration = new Configuration(group2Props, 
EMPTY_ATTRIBUTES, configuration);
     
expect(group2.getConfiguration()).andReturn(group2Configuration).atLeastOnce();
@@ -178,6 +179,7 @@ public class BlueprintImplTest {
     Configuration group2Configuration = new Configuration(group2Props, 
EMPTY_ATTRIBUTES, configuration);
     
expect(group2.getConfiguration()).andReturn(group2Configuration).atLeastOnce();
 
+
     expect(group1.getCardinality()).andReturn("1").atLeastOnce();
     expect(group1.getComponents()).andReturn(Arrays.asList(new 
Component("NAMENODE"),new Component("ZKFC"))).atLeastOnce();
     expect(group2.getCardinality()).andReturn("1").atLeastOnce();
@@ -198,13 +200,13 @@ public class BlueprintImplTest {
     properties.put("hadoop-env", hadoopProps);
     hadoopProps.put("dfs_ha_initial_namenode_active", "%HOSTGROUP::group1%");
     hadoopProps.put("dfs_ha_initial_namenode_standby", "%HOSTGROUP::group2%");
-    replay(stack, group1, group2);
+    replay(stack, group1, group2, serverConfig);
 
     Blueprint blueprint = new BlueprintImpl("test", hostGroups, stack, 
configuration, null);
     blueprint.validateRequiredProperties();
     BlueprintEntity entity = blueprint.toEntity();
 
-    verify(stack, group1, group2);
+    verify(stack, group1, group2, serverConfig);
     assertTrue(entity.getSecurityType() == SecurityType.NONE);
     assertTrue(entity.getSecurityDescriptorReference() == null);
   }
@@ -240,10 +242,10 @@ public class BlueprintImplTest {
     properties.put("hadoop-env", hadoopProps);
     hadoopProps.put("dfs_ha_initial_namenode_active", "%HOSTGROUP::group2%");
     hadoopProps.put("dfs_ha_initial_namenode_standby", "%HOSTGROUP::group3%");
-    replay(stack, group1, group2);
+    replay(stack, group1, group2, serverConfig);
     Blueprint blueprint = new BlueprintImpl("test", hostGroups, stack, 
configuration, null);
     blueprint.validateRequiredProperties();
-    verify(stack, group1, group2);
+    verify(stack, group1, group2, serverConfig);
   }
   @Test(expected= IllegalArgumentException.class)
   public void 
testValidateConfigurations__hostGroupConfigForNameNodeHAMappedSameHostGroup() 
throws Exception {
@@ -276,10 +278,10 @@ public class BlueprintImplTest {
     properties.put("hadoop-env", hadoopProps);
     hadoopProps.put("dfs_ha_initial_namenode_active", "%HOSTGROUP::group2%");
     hadoopProps.put("dfs_ha_initial_namenode_standby", "%HOSTGROUP::group2%");
-    replay(stack, group1, group2);
+    replay(stack, group1, group2, serverConfig);
     Blueprint blueprint = new BlueprintImpl("test", hostGroups, stack, 
configuration, null);
     blueprint.validateRequiredProperties();
-    verify(stack, group1, group2);
+    verify(stack, group1, group2, serverConfig);
   }
   @Test(expected = InvalidTopologyException.class)
   public void testValidateConfigurations__secretReference() throws 
InvalidTopologyException,
@@ -287,7 +289,6 @@ public class BlueprintImplTest {
     Map<String, Map<String, String>> group2Props = new HashMap<>();
     Map<String, String> group2Category2Props = new HashMap<>();
 
-    org.apache.ambari.server.configuration.Configuration serverConfig = 
setupConfigurationWithGPLLicense(true);
     group2Props.put("category2", group2Category2Props);
     group2Category2Props.put("prop2", "val");
     hdfsProps.put("secret", "SECRET:hdfs-site:1:test");
@@ -307,7 +308,7 @@ public class BlueprintImplTest {
     }});
     Configuration lzoUsageConfiguration = new Configuration(lzoProperties, 
EMPTY_ATTRIBUTES, EMPTY_CONFIGURATION);
 
-    org.apache.ambari.server.configuration.Configuration serverConfig = 
setupConfigurationWithGPLLicense(false);
+    serverConfig = setupConfigurationWithGPLLicense(false);
     replay(stack, group1, group2, serverConfig);
 
     Blueprint blueprint = new BlueprintImpl("test", hostGroups, stack, 
lzoUsageConfiguration, null);
@@ -324,7 +325,7 @@ public class BlueprintImplTest {
     }});
     Configuration lzoUsageConfiguration = new Configuration(lzoProperties, 
EMPTY_ATTRIBUTES, EMPTY_CONFIGURATION);
 
-    org.apache.ambari.server.configuration.Configuration serverConfig = 
setupConfigurationWithGPLLicense(false);
+    serverConfig = setupConfigurationWithGPLLicense(false);
     replay(stack, group1, group2, serverConfig);
 
     Blueprint blueprint = new BlueprintImpl("test", hostGroups, stack, 
lzoUsageConfiguration, null);
@@ -342,7 +343,6 @@ public class BlueprintImplTest {
     }});
     Configuration lzoUsageConfiguration = new Configuration(lzoProperties, 
EMPTY_ATTRIBUTES, EMPTY_CONFIGURATION);
 
-    org.apache.ambari.server.configuration.Configuration serverConfig = 
setupConfigurationWithGPLLicense(true);
     
expect(group2.getConfiguration()).andReturn(EMPTY_CONFIGURATION).atLeastOnce();
     replay(stack, group1, group2, serverConfig);
 

-- 
To stop receiving notification emails like this one, please contact
['"[email protected]" <[email protected]>'].

Reply via email to