http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ClientServerTransactionDUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ClientServerTransactionDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ClientServerTransactionDUnitTest.java index d49b116..8115701 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ClientServerTransactionDUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ClientServerTransactionDUnitTest.java @@ -16,6 +16,8 @@ */ package com.gemstone.gemfire.internal.cache; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; + import com.gemstone.gemfire.cache.*; import com.gemstone.gemfire.cache.Region.Entry; import com.gemstone.gemfire.cache.client.*; @@ -147,10 +149,10 @@ public class ClientServerTransactionDUnitTest extends RemoteTransactionDUnitTest ClientCacheFactory ccf = new ClientCacheFactory(); ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); // these settings were used to manually check that tx operation stats were being updated - //ccf.set(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - //ccf.set(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "clientStats.gfs"); + //ccf.set(STATISTIC_SAMPLING_ENABLED, "true"); + //ccf.set(STATISTIC_ARCHIVE_FILE, "clientStats.gfs"); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<Integer, String> crf = cCache .createClientRegionFactory(isEmpty ? ClientRegionShortcut.PROXY @@ -207,7 +209,7 @@ public class ClientServerTransactionDUnitTest extends RemoteTransactionDUnitTest ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port1); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); @@ -278,7 +280,7 @@ public class ClientServerTransactionDUnitTest extends RemoteTransactionDUnitTest ClientCacheFactory ccf = new ClientCacheFactory(); ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port1); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(cachingProxy ? ClientRegionShortcut.CACHING_PROXY : ClientRegionShortcut.PROXY); @@ -1333,7 +1335,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ClientCacheFactory ccf = new ClientCacheFactory(); ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port1); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(ClientRegionShortcut.PROXY); @@ -1429,7 +1431,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost", port2); ccf.setPoolLoadConditioningInterval(1); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(ClientRegionShortcut.PROXY); @@ -1546,7 +1548,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { if (port2 != 0) ccf.addPoolServer("localhost", port2); if (port3 != 0) ccf.addPoolServer("localhost", port3); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(cachingProxy ? ClientRegionShortcut.CACHING_PROXY : ClientRegionShortcut.PROXY); @@ -2058,7 +2060,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost", port2); ccf.setPoolSubscriptionEnabled(false); ccf.setPoolLoadConditioningInterval(1); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(ClientRegionShortcut.PROXY); @@ -2451,7 +2453,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.setPoolMinConnections(5); ccf.setPoolLoadConditioningInterval(-1); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); Region r1 = cCache.createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY).create("r1"); Region r2 = cCache.createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY).create("r2"); @@ -2684,10 +2686,10 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port); ccf.addPoolServer("localhost", port2); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); // these settings were used to manually check that tx operation stats were being updated - //ccf.set(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - //ccf.set(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "clientStats.gfs"); + //ccf.set(STATISTIC_SAMPLING_ENABLED, "true"); + //ccf.set(STATISTIC_ARCHIVE_FILE, "clientStats.gfs"); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<Integer, String> crf = cCache .createClientRegionFactory(ClientRegionShortcut.PROXY); @@ -2986,7 +2988,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port1); if (port2 != 0) ccf.addPoolServer("localhost", port2); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY); @@ -3127,7 +3129,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost", port2); ccf.setPoolMinConnections(0); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY); @@ -3215,7 +3217,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port1); ccf.setPoolMinConnections(0); ccf.setPoolSubscriptionEnabled(false); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<CustId, Customer> custrf = cCache .createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY); @@ -3305,7 +3307,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.setPoolMinConnections(0); ccf.setPoolSubscriptionEnabled(true); ccf.setPoolSubscriptionRedundancy(0); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); Region r = cCache.createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY).addCacheListener(new ClientListener()).create(regionName); r.registerInterestRegex(".*"); @@ -3321,7 +3323,7 @@ public void testClientCommitAndDataStoreGetsEvent() throws Exception { ccf.addPoolServer("localhost"/*getServerHostName(Host.getHost(0))*/, port1); ccf.setPoolMinConnections(0); ccf.setPoolSubscriptionEnabled(true); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, getDUnitLogLevel()); + ccf.set(LOG_LEVEL, getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); Region r = cCache.createClientRegionFactory(ClientRegionShortcut.CACHING_PROXY).create(regionName); getCache().getCacheTransactionManager().begin();
http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConcurrentMapOpsDUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConcurrentMapOpsDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConcurrentMapOpsDUnitTest.java index 8a1586c..197b028 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConcurrentMapOpsDUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConcurrentMapOpsDUnitTest.java @@ -31,7 +31,6 @@ import com.gemstone.gemfire.cache.server.CacheServer; import com.gemstone.gemfire.cache.util.CacheListenerAdapter; import com.gemstone.gemfire.cache30.CacheTestCase; import com.gemstone.gemfire.distributed.DistributedMember; -import com.gemstone.gemfire.distributed.internal.DistributionConfig; import com.gemstone.gemfire.distributed.internal.membership.MembershipManager; import com.gemstone.gemfire.distributed.internal.membership.gms.MembershipManagerHelper; import com.gemstone.gemfire.internal.AvailablePort; @@ -46,6 +45,8 @@ import java.util.HashSet; import java.util.Set; import java.util.concurrent.atomic.AtomicInteger; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; + /** * tests for the concurrentMapOperations. there are more tests in ClientServerMiscDUnitTest * @@ -135,7 +136,7 @@ public class ConcurrentMapOpsDUnitTest extends CacheTestCase { ccf.addPoolServer("localhost", port2); } ccf.setPoolSubscriptionEnabled(true); - ccf.set(DistributionConfig.LOG_LEVEL_NAME, LogWriterUtils.getDUnitLogLevel()); + ccf.set(LOG_LEVEL, LogWriterUtils.getDUnitLogLevel()); ClientCache cCache = getClientCache(ccf); ClientRegionFactory<Integer, String> crf = cCache .createClientRegionFactory(isEmpty ? ClientRegionShortcut.PROXY http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConnectDisconnectDUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConnectDisconnectDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConnectDisconnectDUnitTest.java index b6d5a12..34ad0fe 100755 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConnectDisconnectDUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/ConnectDisconnectDUnitTest.java @@ -22,8 +22,7 @@ import com.gemstone.gemfire.test.dunit.*; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.LOCATORS; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.START_LOCATOR; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; /** A test of 46438 - missing response to an update attributes message */ public class ConnectDisconnectDUnitTest extends CacheTestCase { @@ -151,8 +150,8 @@ public class ConnectDisconnectDUnitTest extends CacheTestCase { @Override public Properties getDistributedSystemProperties() { Properties props = super.getDistributedSystemProperties(); - props.setProperty(DistributionConfig.LOG_LEVEL_NAME, "info"); - props.setProperty(DistributionConfig.CONSERVE_SOCKETS_NAME, "false"); + props.setProperty(LOG_LEVEL, "info"); + props.setProperty(CONSERVE_SOCKETS, "false"); if (LOCATOR_PORT > 0) { props.setProperty(START_LOCATOR, "localhost[" + LOCATOR_PORT + "]"); props.setProperty(LOCATORS, LOCATORS_STRING); http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DeltaPropagationDUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DeltaPropagationDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DeltaPropagationDUnitTest.java index 7624d12..151a8e5 100755 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DeltaPropagationDUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DeltaPropagationDUnitTest.java @@ -47,8 +47,7 @@ import com.gemstone.gemfire.test.dunit.*; import java.io.File; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.LOCATORS; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; /** * @since GemFire 6.1 @@ -649,8 +648,8 @@ public class DeltaPropagationDUnitTest extends DistributedTestCase { Properties properties = new Properties(); properties.setProperty(MCAST_PORT, "0"); properties.setProperty(LOCATORS, ""); - properties.setProperty(DistributionConfig.DURABLE_CLIENT_ID_NAME, durableClientId); - properties.setProperty(DistributionConfig.DURABLE_CLIENT_TIMEOUT_NAME, String.valueOf(60)); + properties.setProperty(DURABLE_CLIENT_ID, durableClientId); + properties.setProperty(DURABLE_CLIENT_TIMEOUT, String.valueOf(60)); createDurableCacheClient(((PoolFactoryImpl)pf).getPoolAttributes(), regionName, properties, new Integer(DURABLE_CLIENT_LISTENER), Boolean.TRUE); @@ -722,8 +721,8 @@ public class DeltaPropagationDUnitTest extends DistributedTestCase { Properties properties = new Properties(); properties.setProperty(MCAST_PORT, "0"); properties.setProperty(LOCATORS, ""); - properties.setProperty(DistributionConfig.DURABLE_CLIENT_ID_NAME, durableClientId); - properties.setProperty(DistributionConfig.DURABLE_CLIENT_TIMEOUT_NAME, String.valueOf(60)); + properties.setProperty(DURABLE_CLIENT_ID, durableClientId); + properties.setProperty(DURABLE_CLIENT_TIMEOUT, String.valueOf(60)); createDurableCacheClient(((PoolFactoryImpl)pf).getPoolAttributes(), regionName, properties, new Integer(DURABLE_CLIENT_LISTENER), Boolean.FALSE); @@ -1314,7 +1313,7 @@ public class DeltaPropagationDUnitTest extends DistributedTestCase { Properties props = new Properties(); props.setProperty(MCAST_PORT, "0"); props.setProperty(LOCATORS, ""); - props.setProperty(DistributionConfig.CLIENT_CONFLATION_PROP_NAME, conflate); + props.setProperty(CONFLATE_EVENTS, conflate); new DeltaPropagationDUnitTest("temp").createCache(props); AttributesFactory factory = new AttributesFactory(); pool = ClientServerTestCase.configureConnectionPool(factory, "localhost", ports, http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOfflineCompactionJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOfflineCompactionJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOfflineCompactionJUnitTest.java index d57f706..60a0548 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOfflineCompactionJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOfflineCompactionJUnitTest.java @@ -33,8 +33,7 @@ import java.io.File; import java.io.IOException; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.LOCATORS; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; import static org.junit.Assert.assertEquals; /** @@ -64,10 +63,10 @@ public class DiskOfflineCompactionJUnitTest Properties props = new Properties(); props.setProperty(MCAST_PORT, "0"); props.setProperty(LOCATORS, ""); - props.setProperty(DistributionConfig.LOG_LEVEL_NAME, "config"); // to keep diskPerf logs smaller - props.setProperty(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - props.setProperty(DistributionConfig.ENABLE_TIME_STATISTICS_NAME, "true"); - props.setProperty(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "stats.gfs"); + props.setProperty(LOG_LEVEL, "config"); // to keep diskPerf logs smaller + props.setProperty(STATISTIC_SAMPLING_ENABLED, "true"); + props.setProperty(ENABLE_TIME_STATISTICS, "true"); + props.setProperty(STATISTIC_ARCHIVE_FILE, "stats.gfs"); ds = DistributedSystem.connect(props); cache = CacheFactory.create(ds); } http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOldAPIsJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOldAPIsJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOldAPIsJUnitTest.java index 316813d..0ac0d11 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOldAPIsJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskOldAPIsJUnitTest.java @@ -29,10 +29,8 @@ import java.io.File; import java.util.Properties; import java.util.Set; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.LOCATORS; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; -import static org.junit.Assert.assertEquals; -import static org.junit.Assert.assertTrue; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; +import static org.junit.Assert.*; /** * Tests the old disk apis to make sure they do the correct thing. @@ -52,10 +50,10 @@ public class DiskOldAPIsJUnitTest static { props.setProperty(MCAST_PORT, "0"); props.setProperty(LOCATORS, ""); - props.setProperty(DistributionConfig.LOG_LEVEL_NAME, "config"); // to keep diskPerf logs smaller - props.setProperty(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - props.setProperty(DistributionConfig.ENABLE_TIME_STATISTICS_NAME, "true"); - props.setProperty(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "stats.gfs"); + props.setProperty(LOG_LEVEL, "config"); // to keep diskPerf logs smaller + props.setProperty(STATISTIC_SAMPLING_ENABLED, "true"); + props.setProperty(ENABLE_TIME_STATISTICS, "true"); + props.setProperty(STATISTIC_ARCHIVE_FILE, "stats.gfs"); } @Before http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCacheXmlJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCacheXmlJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCacheXmlJUnitTest.java index b1a67e3..57bde39 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCacheXmlJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCacheXmlJUnitTest.java @@ -20,7 +20,6 @@ import com.gemstone.gemfire.SystemFailure; import com.gemstone.gemfire.cache.*; import com.gemstone.gemfire.distributed.DistributedSystem; import com.gemstone.gemfire.distributed.SystemConfigurationProperties; -import com.gemstone.gemfire.distributed.internal.DistributionConfig; import com.gemstone.gemfire.test.junit.categories.IntegrationTest; import com.gemstone.gemfire.util.test.TestUtil; import org.junit.After; @@ -32,7 +31,7 @@ import java.util.Arrays; import java.util.Iterator; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; import static org.junit.Assert.assertTrue; import static org.junit.Assert.fail; @@ -73,7 +72,7 @@ public class DiskRegCacheXmlJUnitTest props.setProperty(SystemConfigurationProperties.NAME, "test"); String path = TestUtil.getResourcePath(getClass(), "DiskRegCacheXmlJUnitTest.xml"); props.setProperty(MCAST_PORT, "0"); - props.setProperty(DistributionConfig.CACHE_XML_FILE_NAME, path); + props.setProperty(CACHE_XML_FILE, path); ds = DistributedSystem.connect(props); try { // Create the cache which causes the cache-xml-file to be parsed http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCachexmlGeneratorJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCachexmlGeneratorJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCachexmlGeneratorJUnitTest.java index a71be75..c7f1d18 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCachexmlGeneratorJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegCachexmlGeneratorJUnitTest.java @@ -21,7 +21,6 @@ import com.gemstone.gemfire.cache.Region; import com.gemstone.gemfire.cache.Scope; import com.gemstone.gemfire.distributed.DistributedSystem; import com.gemstone.gemfire.distributed.SystemConfigurationProperties; -import com.gemstone.gemfire.distributed.internal.DistributionConfig; import com.gemstone.gemfire.internal.cache.xmlcache.CacheXmlGenerator; import com.gemstone.gemfire.test.junit.categories.IntegrationTest; import org.junit.After; @@ -34,7 +33,7 @@ import java.io.FileWriter; import java.io.PrintWriter; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; import static org.junit.Assert.fail; /** @@ -50,53 +49,9 @@ public class DiskRegCachexmlGeneratorJUnitTest extends DiskRegionTestingBase DiskRegionProperties diskProps = new DiskRegionProperties(); - DiskRegionProperties diskProps1 = new DiskRegionProperties(); + DiskRegionProperties[] diskRegionProperties = new DiskRegionProperties[12]; - DiskRegionProperties diskProps2 = new DiskRegionProperties(); - - DiskRegionProperties diskProps3 = new DiskRegionProperties(); - - DiskRegionProperties diskProps4 = new DiskRegionProperties(); - - DiskRegionProperties diskProps5 = new DiskRegionProperties(); - - DiskRegionProperties diskProps6 = new DiskRegionProperties(); - - DiskRegionProperties diskProps7 = new DiskRegionProperties(); - - DiskRegionProperties diskProps8 = new DiskRegionProperties(); - - DiskRegionProperties diskProps9 = new DiskRegionProperties(); - - DiskRegionProperties diskProps10 = new DiskRegionProperties(); - - DiskRegionProperties diskProps11 = new DiskRegionProperties(); - - DiskRegionProperties diskProps12 = new DiskRegionProperties(); - - Region region1; - - Region region2; - - Region region3; - - Region region4; - - Region region5; - - Region region6; - - Region region7; - - Region region8; - - Region region9; - - Region region10; - - Region region11; - - Region region12; + Region[] regions = new Region[12]; @Before public void setUp() throws Exception @@ -107,18 +62,17 @@ public class DiskRegCachexmlGeneratorJUnitTest extends DiskRegionTestingBase diskDirSize[1] = Integer.MAX_VALUE; diskDirSize[2] = 1073741824; diskDirSize[3] = 2073741824; - diskProps1.setDiskDirsAndSizes(dirs, diskDirSize); - diskProps2.setDiskDirs(dirs); - diskProps3.setDiskDirs(dirs); - diskProps4.setDiskDirs(dirs); - diskProps5.setDiskDirs(dirs); - diskProps6.setDiskDirs(dirs); - diskProps7.setDiskDirs(dirs); - diskProps8.setDiskDirs(dirs); - diskProps9.setDiskDirs(dirs); - diskProps10.setDiskDirs(dirs); - diskProps11.setDiskDirs(dirs); - diskProps12.setDiskDirs(dirs); + + for (int i = 0; i < diskRegionProperties.length; i++) { + diskRegionProperties[i] = new DiskRegionProperties(); + if(i == 0) + { + diskRegionProperties[i].setDiskDirsAndSizes(dirs, diskDirSize); + } + else{ + diskRegionProperties[i].setDiskDirs(dirs); + } + } } @After @@ -130,97 +84,97 @@ public class DiskRegCachexmlGeneratorJUnitTest extends DiskRegionTestingBase public void createCacheXML() { - // create the region1 which is SyncPersistOnly and set DiskWriteAttibutes - diskProps1.setRolling(true); - diskProps1.setMaxOplogSize(1073741824L); - diskProps1.setRegionName("region1"); - region1 = DiskRegionHelperFactory.getSyncPersistOnlyRegion(cache, - diskProps1, Scope.LOCAL); - - // create the region2 which is SyncPersistOnly and set DiskWriteAttibutes - - diskProps2.setRolling(false); - diskProps2.setRegionName("region2"); - region2 = DiskRegionHelperFactory.getSyncPersistOnlyRegion(cache, - diskProps2, Scope.LOCAL); - - // create the region3 which AsyncPersistOnly, No buffer and Rolling oplog - diskProps3.setRolling(true); - diskProps3.setMaxOplogSize(1073741824L); - diskProps3.setRegionName("region3"); - region3 = DiskRegionHelperFactory.getAsyncPersistOnlyRegion(cache, - diskProps3); - - // create the region4 which is AsynchPersistonly, No buffer and fixed oplog - diskProps4.setRolling(false); - diskProps4.setRegionName("region4"); - region4 = DiskRegionHelperFactory.getAsyncPersistOnlyRegion(cache, - diskProps4); - - // create the region5 which is SynchOverflowOnly, Rolling oplog - diskProps5.setRolling(true); - diskProps5.setMaxOplogSize(1073741824L); - diskProps5.setRegionName("region5"); - region5 = DiskRegionHelperFactory.getSyncOverFlowOnlyRegion(cache, - diskProps5); - - // create the region6 which is SyncOverflowOnly, Fixed oplog - diskProps6.setRolling(false); - diskProps6.setRegionName("region6"); - region6 = DiskRegionHelperFactory.getSyncOverFlowOnlyRegion(cache, - diskProps6); - - // create the region7 which is AsyncOverflow, with Buffer and rolling oplog - diskProps7.setRolling(true); - diskProps7.setMaxOplogSize(1073741824L); - diskProps7.setBytesThreshold(10000l); - diskProps7.setTimeInterval(15l); - diskProps7.setRegionName("region7"); - region7 = DiskRegionHelperFactory.getAsyncOverFlowOnlyRegion(cache, - diskProps7); - - // create the region8 which is AsyncOverflow ,Time base buffer-zero byte + // create the regions[0] which is SyncPersistOnly and set DiskWriteAttibutes + diskRegionProperties[0].setRolling(true); + diskRegionProperties[0].setMaxOplogSize(1073741824L); + diskRegionProperties[0].setRegionName("regions1"); + regions[0] = DiskRegionHelperFactory.getSyncPersistOnlyRegion(cache, + diskRegionProperties[0], Scope.LOCAL); + + // create the regions[1] which is SyncPersistOnly and set DiskWriteAttibutes + + diskRegionProperties[1].setRolling(false); + diskRegionProperties[1].setRegionName("regions2"); + regions[1] = DiskRegionHelperFactory.getSyncPersistOnlyRegion(cache, + diskRegionProperties[1], Scope.LOCAL); + + // create the regions[2] which AsyncPersistOnly, No buffer and Rolling oplog + diskRegionProperties[2].setRolling(true); + diskRegionProperties[2].setMaxOplogSize(1073741824L); + diskRegionProperties[2].setRegionName("regions3"); + regions[2] = DiskRegionHelperFactory.getAsyncPersistOnlyRegion(cache, + diskRegionProperties[2]); + + // create the regions[3] which is AsynchPersistonly, No buffer and fixed oplog + diskRegionProperties[3].setRolling(false); + diskRegionProperties[3].setRegionName("regions4"); + regions[3] = DiskRegionHelperFactory.getAsyncPersistOnlyRegion(cache, + diskRegionProperties[3]); + + // create the regions[4] which is SynchOverflowOnly, Rolling oplog + diskRegionProperties[4].setRolling(true); + diskRegionProperties[4].setMaxOplogSize(1073741824L); + diskRegionProperties[4].setRegionName("regions5"); + regions[4] = DiskRegionHelperFactory.getSyncOverFlowOnlyRegion(cache, + diskRegionProperties[4]); + + // create the regions[5] which is SyncOverflowOnly, Fixed oplog + diskRegionProperties[5].setRolling(false); + diskRegionProperties[5].setRegionName("regions6"); + regions[5] = DiskRegionHelperFactory.getSyncOverFlowOnlyRegion(cache, + diskRegionProperties[5]); + + // create the regions[6] which is AsyncOverflow, with Buffer and rolling oplog + diskRegionProperties[6].setRolling(true); + diskRegionProperties[6].setMaxOplogSize(1073741824L); + diskRegionProperties[6].setBytesThreshold(10000l); + diskRegionProperties[6].setTimeInterval(15l); + diskRegionProperties[6].setRegionName("regions7"); + regions[6] = DiskRegionHelperFactory.getAsyncOverFlowOnlyRegion(cache, + diskRegionProperties[6]); + + // create the regions[7] which is AsyncOverflow ,Time base buffer-zero byte // buffer // and Fixed oplog - diskProps8.setRolling(false); - diskProps8.setTimeInterval(15l); - diskProps8.setBytesThreshold(0l); - diskProps8.setRegionName("region8"); - region8 = DiskRegionHelperFactory.getAsyncOverFlowOnlyRegion(cache, - diskProps8); - - // create the region9 which is SyncPersistOverflow, Rolling oplog - diskProps9.setRolling(true); - diskProps9.setMaxOplogSize(1073741824L); - diskProps9.setRegionName("region9"); - region9 = DiskRegionHelperFactory.getSyncOverFlowAndPersistRegion(cache, - diskProps9); - - // create the region10 which is Sync PersistOverflow, fixed oplog - diskProps10.setRolling(false); - diskProps10.setRegionName("region10"); - region10 = DiskRegionHelperFactory.getSyncOverFlowAndPersistRegion(cache, - diskProps10); - // create the region11 which is Async Overflow Persist ,with buffer and + diskRegionProperties[7].setRolling(false); + diskRegionProperties[7].setTimeInterval(15l); + diskRegionProperties[7].setBytesThreshold(0l); + diskRegionProperties[7].setRegionName("regions8"); + regions[7] = DiskRegionHelperFactory.getAsyncOverFlowOnlyRegion(cache, + diskRegionProperties[7]); + + // create the regions[8] which is SyncPersistOverflow, Rolling oplog + diskRegionProperties[8].setRolling(true); + diskRegionProperties[8].setMaxOplogSize(1073741824L); + diskRegionProperties[8].setRegionName("regions9"); + regions[8] = DiskRegionHelperFactory.getSyncOverFlowAndPersistRegion(cache, + diskRegionProperties[8]); + + // create the regions[9] which is Sync PersistOverflow, fixed oplog + diskRegionProperties[9].setRolling(false); + diskRegionProperties[9].setRegionName("regions10"); + regions[9] = DiskRegionHelperFactory.getSyncOverFlowAndPersistRegion(cache, + diskRegionProperties[9]); + // create the regions[10] which is Async Overflow Persist ,with buffer and // rollong // oplog - diskProps11.setRolling(true); - diskProps11.setMaxOplogSize(1073741824L); - diskProps11.setBytesThreshold(10000l); - diskProps11.setTimeInterval(15l); - diskProps11.setRegionName("region11"); - region11 = DiskRegionHelperFactory.getAsyncOverFlowAndPersistRegion(cache, - diskProps11); - - // create the region12 which is Async Persist Overflow with time based + diskRegionProperties[10].setRolling(true); + diskRegionProperties[10].setMaxOplogSize(1073741824L); + diskRegionProperties[10].setBytesThreshold(10000l); + diskRegionProperties[10].setTimeInterval(15l); + diskRegionProperties[10].setRegionName("regions11"); + regions[10] = DiskRegionHelperFactory.getAsyncOverFlowAndPersistRegion(cache, + diskRegionProperties[10]); + + // create the regions[11] which is Async Persist Overflow with time based // buffer // and Fixed oplog - diskProps12.setRolling(false); - diskProps12.setBytesThreshold(0l); - diskProps12.setTimeInterval(15l); - diskProps12.setRegionName("region12"); - region12 = DiskRegionHelperFactory.getAsyncOverFlowAndPersistRegion(cache, - diskProps12); + diskRegionProperties[11].setRolling(false); + diskRegionProperties[11].setBytesThreshold(0l); + diskRegionProperties[11].setTimeInterval(15l); + diskRegionProperties[11].setRegionName("regions12"); + regions[11] = DiskRegionHelperFactory.getAsyncOverFlowAndPersistRegion(cache, + diskRegionProperties[11]); //cacheXmlGenerator: generates cacheXml file @@ -246,58 +200,46 @@ public class DiskRegCachexmlGeneratorJUnitTest extends DiskRegionTestingBase props.setProperty(SystemConfigurationProperties.NAME, "DiskRegCachexmlGeneratorJUnitTest"); props.setProperty(MCAST_PORT, "0"); String path = "DiskRegCachexmlGeneratorJUnitTest.xml"; - props.setProperty(DistributionConfig.CACHE_XML_FILE_NAME, path); + props.setProperty(CACHE_XML_FILE, path); ds = DistributedSystem.connect(props); // Create the cache which causes the cache-xml-file to be parsed cache = CacheFactory.create(ds); - // Get the region1 - region1 = cache.getRegion("region1"); - verify((LocalRegion)region1, diskProps1); + // Get the regions[0] + verify((LocalRegion) cache.getRegion("regions1"), diskRegionProperties[0]); - // Get the region2 - Region region2 = cache.getRegion("region2"); - verify((LocalRegion)region2, diskProps2); + // Get the regions[1] + verify((LocalRegion) cache.getRegion("regions2"), diskRegionProperties[1]); - // Get the region3 - Region region3 = cache.getRegion("region3"); - verify((LocalRegion)region3, diskProps3); + // Get the regions[2] + verify((LocalRegion) cache.getRegion("regions3"), diskRegionProperties[2]); - // Get the region4 - Region region4 = cache.getRegion("region4"); - verify((LocalRegion)region4, diskProps4); + // Get the regions[3] + verify((LocalRegion) cache.getRegion("regions4"), diskRegionProperties[3]); - // Get the region5 - Region region5 = cache.getRegion("region5"); - verify((LocalRegion)region5, diskProps5); + // Get the regions[4] + verify((LocalRegion) cache.getRegion("regions5"), diskRegionProperties[4]); - // Get the region6 - Region region6 = cache.getRegion("region6"); - verify((LocalRegion)region6, diskProps6); + // Get the regions[5] + verify((LocalRegion) cache.getRegion("regions6"), diskRegionProperties[5]); - // Get the region7 - Region region7 = cache.getRegion("region7"); - verify((LocalRegion)region7, diskProps7); + // Get the regions[6] + verify((LocalRegion) cache.getRegion("regions7"), diskRegionProperties[6]); - // Get the region8 - Region region8 = cache.getRegion("region8"); - verify((LocalRegion)region8, diskProps8); + // Get the regions[7] + verify((LocalRegion) cache.getRegion("regions8"), diskRegionProperties[7]); - // Get the region9 - Region region9 = cache.getRegion("region9"); - verify((LocalRegion)region9, diskProps9); + // Get the regions[8] + verify((LocalRegion) cache.getRegion("regions9"), diskRegionProperties[8]); - // Get the region10 - Region region10 = cache.getRegion("region10"); - verify((LocalRegion)region10, diskProps10); + // Get the regions[9] + verify((LocalRegion) cache.getRegion("regions10"), diskRegionProperties[9]); - // Get the region11 - Region region11 = cache.getRegion("region11"); - verify((LocalRegion)region11, diskProps11); + // Get the regions[10] + verify((LocalRegion) cache.getRegion("regions11"), diskRegionProperties[10]); - // Get the region12 - Region region12 = cache.getRegion("region12"); - verify((LocalRegion)region12, diskProps12); + // Get the regions[11] + verify((LocalRegion) cache.getRegion("regions12"), diskRegionProperties[11]); } }// end of DiskRegCachexmlGeneratorJUnitTest http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalArguementsJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalArguementsJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalArguementsJUnitTest.java index 09f95f8..005555a 100755 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalArguementsJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalArguementsJUnitTest.java @@ -16,6 +16,8 @@ */ package com.gemstone.gemfire.internal.cache; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; + import com.gemstone.gemfire.cache.Cache; import com.gemstone.gemfire.cache.CacheFactory; import com.gemstone.gemfire.cache.DiskStoreFactory; @@ -54,10 +56,10 @@ public class DiskRegionIllegalArguementsJUnitTest static { props.setProperty(MCAST_PORT, "0"); props.setProperty(LOCATORS, ""); - props.setProperty(DistributionConfig.LOG_LEVEL_NAME, "config"); // to keep diskPerf logs smaller - props.setProperty(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - props.setProperty(DistributionConfig.ENABLE_TIME_STATISTICS_NAME, "true"); - props.setProperty(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "stats.gfs"); + props.setProperty(LOG_LEVEL, "config"); // to keep diskPerf logs smaller + props.setProperty(STATISTIC_SAMPLING_ENABLED, "true"); + props.setProperty(ENABLE_TIME_STATISTICS, "true"); + props.setProperty(STATISTIC_ARCHIVE_FILE, "stats.gfs"); } @Before http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalCacheXMLvaluesJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalCacheXMLvaluesJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalCacheXMLvaluesJUnitTest.java index cac5087..490aa8b 100755 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalCacheXMLvaluesJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionIllegalCacheXMLvaluesJUnitTest.java @@ -19,7 +19,6 @@ package com.gemstone.gemfire.internal.cache; import com.gemstone.gemfire.cache.CacheFactory; import com.gemstone.gemfire.cache.CacheXmlException; import com.gemstone.gemfire.distributed.DistributedSystem; -import com.gemstone.gemfire.distributed.internal.DistributionConfig; import com.gemstone.gemfire.test.junit.categories.IntegrationTest; import com.gemstone.gemfire.util.test.TestUtil; import org.junit.Test; @@ -28,7 +27,7 @@ import org.junit.experimental.categories.Category; import java.io.File; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; import static org.junit.Assert.fail; /** @@ -52,7 +51,7 @@ public class DiskRegionIllegalCacheXMLvaluesJUnitTest dir.deleteOnExit(); Properties props = new Properties(); props.setProperty(MCAST_PORT, "0"); - props.setProperty(DistributionConfig.CACHE_XML_FILE_NAME, TestUtil.getResourcePath(getClass(), path)); + props.setProperty(CACHE_XML_FILE, TestUtil.getResourcePath(getClass(), path)); ds = DistributedSystem.connect(props); try { http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionTestingBase.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionTestingBase.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionTestingBase.java index 08772a4..83a29ae 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionTestingBase.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskRegionTestingBase.java @@ -20,6 +20,8 @@ */ package com.gemstone.gemfire.internal.cache; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; + import com.gemstone.gemfire.LogWriter; import com.gemstone.gemfire.SystemFailure; import com.gemstone.gemfire.cache.*; @@ -77,10 +79,10 @@ public class DiskRegionTestingBase { props.setProperty(MCAST_PORT, "0"); props.setProperty(LOCATORS, ""); - props.setProperty(DistributionConfig.LOG_LEVEL_NAME, "config"); // to keep diskPerf logs smaller - props.setProperty(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - props.setProperty(DistributionConfig.ENABLE_TIME_STATISTICS_NAME, "true"); - props.setProperty(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "stats.gfs"); + props.setProperty(LOG_LEVEL, "config"); // to keep diskPerf logs smaller + props.setProperty(STATISTIC_SAMPLING_ENABLED, "true"); + props.setProperty(ENABLE_TIME_STATISTICS, "true"); + props.setProperty(STATISTIC_ARCHIVE_FILE, "stats.gfs"); File testingDirectory = new File("testingDirectory"); testingDirectory.mkdir(); @@ -173,8 +175,8 @@ public class DiskRegionTestingBase protected Cache createCache() { // useful for debugging: -// props.put(DistributionConfig.LOG_FILE_NAME, "diskRegionTestingBase_system.log"); -// props.put(DistributionConfig.LOG_LEVEL_NAME, getGemFireLogLevel()); +// props.put(LOG_FILE, "diskRegionTestingBase_system.log"); +// props.put(LOG_LEVEL, getGemFireLogLevel()); cache = new CacheFactory(props).create(); ds = cache.getDistributedSystem(); logWriter = cache.getLogger(); http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskStoreFactoryJUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskStoreFactoryJUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskStoreFactoryJUnitTest.java index f2685f6..84dcd1b 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskStoreFactoryJUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/DiskStoreFactoryJUnitTest.java @@ -30,8 +30,7 @@ import java.io.FilenameFilter; import java.util.Arrays; import java.util.Properties; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.LOCATORS; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; import static org.junit.Assert.*; /** @@ -51,10 +50,10 @@ public class DiskStoreFactoryJUnitTest static { props.setProperty(MCAST_PORT, "0"); props.setProperty(LOCATORS, ""); - props.setProperty(DistributionConfig.LOG_LEVEL_NAME, "config"); // to keep diskPerf logs smaller - props.setProperty(DistributionConfig.STATISTIC_SAMPLING_ENABLED_NAME, "true"); - props.setProperty(DistributionConfig.ENABLE_TIME_STATISTICS_NAME, "true"); - props.setProperty(DistributionConfig.STATISTIC_ARCHIVE_FILE_NAME, "stats.gfs"); + props.setProperty(LOG_LEVEL, "config"); // to keep diskPerf logs smaller + props.setProperty(STATISTIC_SAMPLING_ENABLED, "true"); + props.setProperty(ENABLE_TIME_STATISTICS, "true"); + props.setProperty(STATISTIC_ARCHIVE_FILE, "stats.gfs"); } @Before http://git-wip-us.apache.org/repos/asf/incubator-geode/blob/cb291539/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/FixedPRSinglehopDUnitTest.java ---------------------------------------------------------------------- diff --git a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/FixedPRSinglehopDUnitTest.java b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/FixedPRSinglehopDUnitTest.java index c97ac23..8ffcf07 100644 --- a/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/FixedPRSinglehopDUnitTest.java +++ b/geode-core/src/test/java/com/gemstone/gemfire/internal/cache/FixedPRSinglehopDUnitTest.java @@ -25,7 +25,6 @@ import com.gemstone.gemfire.cache.server.CacheServer; import com.gemstone.gemfire.cache30.CacheTestCase; import com.gemstone.gemfire.distributed.DistributedSystem; import com.gemstone.gemfire.distributed.Locator; -import com.gemstone.gemfire.distributed.internal.DistributionConfig; import com.gemstone.gemfire.internal.AvailablePort; import com.gemstone.gemfire.internal.cache.partitioned.fixed.QuarterPartitionResolver; import com.gemstone.gemfire.internal.cache.partitioned.fixed.SingleHopQuarterPartitionResolver; @@ -38,8 +37,7 @@ import java.io.File; import java.io.IOException; import java.util.*; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.LOCATORS; -import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.MCAST_PORT; +import static com.gemstone.gemfire.distributed.SystemConfigurationProperties.*; public class FixedPRSinglehopDUnitTest extends CacheTestCase { @@ -412,7 +410,7 @@ public class FixedPRSinglehopDUnitTest extends CacheTestCase { File logFile = new File("locator-" + locatorPort + ".log"); Properties props = new Properties(); - props.setProperty(DistributionConfig.ENABLE_CLUSTER_CONFIGURATION_NAME, "true"); + props.setProperty(ENABLE_CLUSTER_CONFIGURATION, "true"); try { locator = Locator.startLocatorAndDS(locatorPort, logFile, null, props); }
