http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineReaderImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineReaderImpl.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineReaderImpl.java index e864d61..b7804e7 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineReaderImpl.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineReaderImpl.java @@ -107,7 +107,6 @@ public class TestFileSystemTimelineReaderImpl { entity11.setId("id_1"); entity11.setType("app"); entity11.setCreatedTime(1425016502000L); - entity11.setModifiedTime(1425016502050L); Map<String, Object> info1 = new HashMap<String, Object>(); info1.put("info1", "val1"); entity11.addInfo(info1); @@ -136,7 +135,6 @@ public class TestFileSystemTimelineReaderImpl { TimelineEntity entity12 = new TimelineEntity(); entity12.setId("id_1"); entity12.setType("app"); - entity12.setModifiedTime(1425016503000L); configs.clear(); configs.put("config_2", "23"); configs.put("config_3", "abc"); @@ -166,7 +164,6 @@ public class TestFileSystemTimelineReaderImpl { entity2.setId("id_2"); entity2.setType("app"); entity2.setCreatedTime(1425016501050L); - entity2.setModifiedTime(1425016502010L); Map<String, Object> info2 = new HashMap<String, Object>(); info1.put("info2", 4); entity2.addInfo(info2); @@ -203,7 +200,6 @@ public class TestFileSystemTimelineReaderImpl { entity3.setId("id_3"); entity3.setType("app"); entity3.setCreatedTime(1425016501050L); - entity3.setModifiedTime(1425016502010L); Map<String, Object> info3 = new HashMap<String, Object>(); info3.put("info2", 3.5); entity3.addInfo(info3); @@ -239,7 +235,6 @@ public class TestFileSystemTimelineReaderImpl { entity4.setId("id_4"); entity4.setType("app"); entity4.setCreatedTime(1425016502050L); - entity4.setModifiedTime(1425016503010L); TimelineEvent event44 = new TimelineEvent(); event44.setId("event_4"); event44.setTimestamp(1425016502003L); @@ -252,7 +247,6 @@ public class TestFileSystemTimelineReaderImpl { entity5.setId("id_5"); entity5.setType("app"); entity5.setCreatedTime(1425016502050L); - entity5.setModifiedTime(1425016503010L); writeEntityFile(entity5, appDir2); } @@ -263,7 +257,7 @@ public class TestFileSystemTimelineReaderImpl { @Test public void testGetEntityDefaultView() throws Exception { // If no fields are specified, entity is returned with default view i.e. - // only the id, created and modified time + // only the id, type and created time. TimelineEntity result = reader.getEntity("user1", "cluster1", "flow1", 1L, "app1", "app", "id_1", null, null, null); @@ -271,7 +265,6 @@ public class TestFileSystemTimelineReaderImpl { (new TimelineEntity.Identifier("app", "id_1")).toString(), result.getIdentifier().toString()); Assert.assertEquals(1425016502000L, result.getCreatedTime()); - Assert.assertEquals(1425016503000L, result.getModifiedTime()); Assert.assertEquals(0, result.getConfigs().size()); Assert.assertEquals(0, result.getMetrics().size()); } @@ -286,7 +279,6 @@ public class TestFileSystemTimelineReaderImpl { (new TimelineEntity.Identifier("app", "id_1")).toString(), result.getIdentifier().toString()); Assert.assertEquals(1425016502000L, result.getCreatedTime()); - Assert.assertEquals(1425016503000L, result.getModifiedTime()); Assert.assertEquals(0, result.getConfigs().size()); Assert.assertEquals(0, result.getMetrics().size()); } @@ -303,7 +295,6 @@ public class TestFileSystemTimelineReaderImpl { (new TimelineEntity.Identifier("app", "id_5")).toString(), result.getIdentifier().toString()); Assert.assertEquals(1425016502050L, result.getCreatedTime()); - Assert.assertEquals(1425016503010L, result.getModifiedTime()); } @Test @@ -317,7 +308,6 @@ public class TestFileSystemTimelineReaderImpl { (new TimelineEntity.Identifier("app", "id_1")).toString(), result.getIdentifier().toString()); Assert.assertEquals(1425016502000L, result.getCreatedTime()); - Assert.assertEquals(1425016503000L, result.getModifiedTime()); Assert.assertEquals(3, result.getConfigs().size()); Assert.assertEquals(3, result.getMetrics().size()); Assert.assertEquals(1, result.getInfo().size()); @@ -335,7 +325,6 @@ public class TestFileSystemTimelineReaderImpl { (new TimelineEntity.Identifier("app", "id_1")).toString(), result.getIdentifier().toString()); Assert.assertEquals(1425016502000L, result.getCreatedTime()); - Assert.assertEquals(1425016503000L, result.getModifiedTime()); Assert.assertEquals(3, result.getConfigs().size()); Assert.assertEquals(3, result.getMetrics().size()); // All fields including events will be returned. @@ -346,8 +335,8 @@ public class TestFileSystemTimelineReaderImpl { public void testGetAllEntities() throws Exception { Set<TimelineEntity> result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, null, null, null, null, null, - null, null, null, null); + null, null, null, null, null, null, null, null, null, null, null, + null); // All 3 entities will be returned Assert.assertEquals(4, result.size()); } @@ -356,8 +345,8 @@ public class TestFileSystemTimelineReaderImpl { public void testGetEntitiesWithLimit() throws Exception { Set<TimelineEntity> result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - 2L, null, null, null, null, null, null, null, null, null, - null, null, null, null); + 2L, null, null, null, null, null, null, null, null, null, null, + null); Assert.assertEquals(2, result.size()); // Needs to be rewritten once hashcode and equals for // TimelineEntity is implemented @@ -370,8 +359,8 @@ public class TestFileSystemTimelineReaderImpl { } result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - 3L, null, null, null, null, null, null, null, null, null, - null, null, null, null); + 3L, null, null, null, null, null, null, null, null, null, null, + null); // Even though 2 entities out of 4 have same created time, one entity // is left out due to limit Assert.assertEquals(3, result.size()); @@ -383,7 +372,7 @@ public class TestFileSystemTimelineReaderImpl { Set<TimelineEntity> result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", null, 1425016502030L, 1425016502060L, null, null, null, null, null, - null, null, null, null, null, null); + null, null, null, null); Assert.assertEquals(1, result.size()); // Only one entity with ID id_4 should be returned. for (TimelineEntity entity : result) { @@ -396,7 +385,7 @@ public class TestFileSystemTimelineReaderImpl { result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", null, null, 1425016502010L, null, null, null, null, null, null, - null, null, null, null, null); + null, null, null); Assert.assertEquals(3, result.size()); for (TimelineEntity entity : result) { if (entity.getId().equals("id_4")) { @@ -408,50 +397,13 @@ public class TestFileSystemTimelineReaderImpl { result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", null, 1425016502010L, null, null, null, null, null, null, null, - null, null, null, null, null); + null, null, null); Assert.assertEquals(1, result.size()); for (TimelineEntity entity : result) { if (!entity.getId().equals("id_4")) { Assert.fail("Incorrect filtering based on created time range"); } } - - // Get entities based on modified time start and end time range. - result = - reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, 1425016502090L, 1425016503020L, null, null, null, - null, null, null, null, null, null); - Assert.assertEquals(2, result.size()); - // Two entities with IDs' id_1 and id_4 should be returned. - for (TimelineEntity entity : result) { - if (!entity.getId().equals("id_1") && !entity.getId().equals("id_4")) { - Assert.fail("Incorrect filtering based on modified time range"); - } - } - - // Get entities if only modified time end is specified. - result = - reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, 1425016502090L, null, null, null, null, - null, null, null, null, null); - Assert.assertEquals(2, result.size()); - for (TimelineEntity entity : result) { - if (entity.getId().equals("id_1") || entity.getId().equals("id_4")) { - Assert.fail("Incorrect filtering based on modified time range"); - } - } - - // Get entities if only modified time start is specified. - result = - reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, 1425016503005L, null, null, null, null, null, - null, null, null, null, null); - Assert.assertEquals(1, result.size()); - for (TimelineEntity entity : result) { - if (!entity.getId().equals("id_4")) { - Assert.fail("Incorrect filtering based on modified time range"); - } - } } @Test @@ -461,8 +413,8 @@ public class TestFileSystemTimelineReaderImpl { infoFilters.put("info2", 3.5); Set<TimelineEntity> result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, null, null, infoFilters, null, null, - null, null, null, null); + null, null, null, null, null, infoFilters, null, null, null, null, + null, null); Assert.assertEquals(1, result.size()); // Only one entity with ID id_3 should be returned. for (TimelineEntity entity : result) { @@ -477,8 +429,8 @@ public class TestFileSystemTimelineReaderImpl { configFilters.put("config_3", "abc"); result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, null, null, null, configFilters, null, - null, null, null, null); + null, null, null, null, null, null, configFilters, null, null, null, + null, null); Assert.assertEquals(2, result.size()); for (TimelineEntity entity : result) { if (!entity.getId().equals("id_1") && !entity.getId().equals("id_3")) { @@ -492,8 +444,8 @@ public class TestFileSystemTimelineReaderImpl { eventFilters.add("event_4"); result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, null, null, null, null, null, - eventFilters, null, null, null); + null, null, null, null, null, null, null, null, eventFilters, null, + null, null); Assert.assertEquals(1, result.size()); for (TimelineEntity entity : result) { if (!entity.getId().equals("id_3")) { @@ -506,8 +458,8 @@ public class TestFileSystemTimelineReaderImpl { metricFilters.add("metric3"); result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, null, null, null, null, metricFilters, - null, null, null, null); + null, null, null, null, null, null, null, metricFilters, null, null, + null, null); Assert.assertEquals(2, result.size()); // Two entities with IDs' id_1 and id_2 should be returned. for (TimelineEntity entity : result) { @@ -526,8 +478,8 @@ public class TestFileSystemTimelineReaderImpl { relatesTo.put("flow", relatesToIds); Set<TimelineEntity> result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, relatesTo, null, null, null, null, - null, null, null, null); + null, null, null, relatesTo, null, null, null, null, null, null, + null, null); Assert.assertEquals(1, result.size()); // Only one entity with ID id_1 should be returned. for (TimelineEntity entity : result) { @@ -543,8 +495,8 @@ public class TestFileSystemTimelineReaderImpl { isRelatedTo.put("type1", isRelatedToIds); result = reader.getEntities("user1", "cluster1", "flow1", 1L, "app1", "app", - null, null, null, null, null, null, isRelatedTo, null, null, null, - null, null, null, null); + null, null, null, null, isRelatedTo, null, null, null, null, null, + null, null); Assert.assertEquals(2, result.size()); // Two entities with IDs' id_1 and id_3 should be returned. for (TimelineEntity entity : result) {
http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineWriterImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineWriterImpl.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineWriterImpl.java index 50a9f60..5ce7d3b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineWriterImpl.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestFileSystemTimelineWriterImpl.java @@ -49,7 +49,6 @@ public class TestFileSystemTimelineWriterImpl { entity.setId(id); entity.setType(type); entity.setCreatedTime(1425016501000L); - entity.setModifiedTime(1425016502000L); te.addEntity(entity); FileSystemTimelineWriterImpl fsi = null; @@ -80,7 +79,7 @@ public class TestFileSystemTimelineWriterImpl { assertTrue(!(f.exists())); } finally { if (fsi != null) { - fsi.stop(); + fsi.close(); FileUtils.deleteDirectory(new File(fsi.getOutputRoot())); } } http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestHBaseTimelineStorage.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestHBaseTimelineStorage.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestHBaseTimelineStorage.java index 3198307..b74f5a1 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestHBaseTimelineStorage.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestHBaseTimelineStorage.java @@ -74,7 +74,6 @@ import org.junit.Test; import com.google.common.collect.ImmutableMap; import com.google.common.collect.ImmutableSet; -import com.google.common.collect.Maps; /** * Various tests to test writing entities to HBase and reading them back from @@ -112,9 +111,7 @@ public class TestHBaseTimelineStorage { entity.setId(id); entity.setType(TimelineEntityType.YARN_APPLICATION.toString()); Long cTime = 1425016501000L; - Long mTime = 1425026901000L; entity.setCreatedTime(cTime); - entity.setModifiedTime(mTime); // add the info map in Timeline Entity Map<String, Object> infoMap = new HashMap<String, Object>(); infoMap.put("infoMapKey1", "infoMapValue1"); @@ -177,7 +174,6 @@ public class TestHBaseTimelineStorage { entity1.setId(id1); entity1.setType(TimelineEntityType.YARN_APPLICATION.toString()); entity1.setCreatedTime(cTime); - entity1.setModifiedTime(mTime); // add the info map in Timeline Entity Map<String, Object> infoMap1 = new HashMap<String, Object>(); @@ -235,7 +231,6 @@ public class TestHBaseTimelineStorage { entity2.setId(id2); entity2.setType(TimelineEntityType.YARN_APPLICATION.toString()); entity2.setCreatedTime(cTime); - entity2.setModifiedTime(mTime); te2.addEntity(entity2); HBaseTimelineWriterImpl hbi = null; try { @@ -270,9 +265,7 @@ public class TestHBaseTimelineStorage { entity.setId(id); entity.setType(type); Long cTime = 1425016501000L; - Long mTime = 1425026901000L; entity.setCreatedTime(cTime); - entity.setModifiedTime(mTime); // add the info map in Timeline Entity Map<String, Object> infoMap = new HashMap<String, Object>(); infoMap.put("infoMapKey1", "infoMapValue1"); @@ -333,7 +326,6 @@ public class TestHBaseTimelineStorage { entity1.setId(id1); entity1.setType(type); entity1.setCreatedTime(cTime); - entity1.setModifiedTime(mTime); // add the info map in Timeline Entity Map<String, Object> infoMap1 = new HashMap<String, Object>(); @@ -390,7 +382,6 @@ public class TestHBaseTimelineStorage { entity2.setId(id2); entity2.setType(type); entity2.setCreatedTime(cTime); - entity2.setModifiedTime(mTime); te.addEntity(entity2); HBaseTimelineWriterImpl hbi = null; try { @@ -444,9 +435,7 @@ public class TestHBaseTimelineStorage { String appId = "application_1000178881110_2002"; entity.setId(appId); long cTime = 1425016501000L; - long mTime = 1425026901000L; entity.setCreatedTime(cTime); - entity.setModifiedTime(mTime); // add the info map in Timeline Entity Map<String, Object> infoMap = new HashMap<String, Object>(); @@ -522,7 +511,7 @@ public class TestHBaseTimelineStorage { Result result = new ApplicationTable().getResult(c1, conn, get); assertTrue(result != null); - assertEquals(16, result.size()); + assertEquals(15, result.size()); // check the row key byte[] row1 = result.getRow(); @@ -538,10 +527,6 @@ public class TestHBaseTimelineStorage { long cTime1 = val.longValue(); assertEquals(cTime1, cTime); - val = (Number) ApplicationColumn.MODIFIED_TIME.readResult(result); - long mTime1 = val.longValue(); - assertEquals(mTime1, mTime); - Map<String, Object> infoColumns = ApplicationColumnPrefix.INFO.readResults(result); assertEquals(infoMap, infoColumns); @@ -599,7 +584,6 @@ public class TestHBaseTimelineStorage { assertEquals(TimelineEntityType.YARN_APPLICATION.toString(), e1.getType()); assertEquals(cTime, e1.getCreatedTime()); - assertEquals(mTime, e1.getModifiedTime()); Map<String, Object> infoMap2 = e1.getInfo(); assertEquals(infoMap, infoMap2); @@ -635,9 +619,7 @@ public class TestHBaseTimelineStorage { entity.setId(id); entity.setType(type); long cTime = 1425016501000L; - long mTime = 1425026901000L; entity.setCreatedTime(cTime); - entity.setModifiedTime(mTime); // add the info map in Timeline Entity Map<String, Object> infoMap = new HashMap<String, Object>(); @@ -736,10 +718,6 @@ public class TestHBaseTimelineStorage { long cTime1 = val.longValue(); assertEquals(cTime1, cTime); - val = (Number) EntityColumn.MODIFIED_TIME.readResult(result); - long mTime1 = val.longValue(); - assertEquals(mTime1, mTime); - Map<String, Object> infoColumns = EntityColumnPrefix.INFO.readResults(result); assertEquals(infoMap, infoColumns); @@ -790,7 +768,7 @@ public class TestHBaseTimelineStorage { } } assertEquals(1, rowCount); - assertEquals(17, colCount); + assertEquals(16, colCount); // read the timeline entity using the reader this time TimelineEntity e1 = reader.getEntity(user, cluster, flow, runid, appName, @@ -798,8 +776,7 @@ public class TestHBaseTimelineStorage { EnumSet.of(TimelineReader.Field.ALL)); Set<TimelineEntity> es1 = reader.getEntities(user, cluster, flow, runid, appName, entity.getType(), null, null, null, null, null, null, null, - null, null, null, null, null, null, - EnumSet.of(TimelineReader.Field.ALL)); + null, null, null, null, EnumSet.of(TimelineReader.Field.ALL)); assertNotNull(e1); assertEquals(1, es1.size()); @@ -807,7 +784,6 @@ public class TestHBaseTimelineStorage { assertEquals(id, e1.getId()); assertEquals(type, e1.getType()); assertEquals(cTime, e1.getCreatedTime()); - assertEquals(mTime, e1.getModifiedTime()); Map<String, Object> infoMap2 = e1.getInfo(); assertEquals(infoMap, infoMap2); @@ -1041,8 +1017,7 @@ public class TestHBaseTimelineStorage { EnumSet.of(TimelineReader.Field.ALL)); Set<TimelineEntity> es1 = reader.getEntities(user, cluster, flow, runid, appName, entity.getType(), null, null, null, null, null, null, null, - null, null, null, null, null, null, - EnumSet.of(TimelineReader.Field.ALL)); + null, null, null, null, EnumSet.of(TimelineReader.Field.ALL)); assertNotNull(e1); assertEquals(1, es1.size()); @@ -1069,7 +1044,6 @@ public class TestHBaseTimelineStorage { String appId = "application_1000178881110_2002"; entityApp.setId(appId); entityApp.setCreatedTime(1425016501000L); - entityApp.setModifiedTime(1425026901000L); // add metrics with floating point values Set<TimelineMetric> metricsApp = new HashSet<>(); TimelineMetric mApp = new TimelineMetric(); @@ -1089,7 +1063,6 @@ public class TestHBaseTimelineStorage { entity.setId("hello"); entity.setType("world"); entity.setCreatedTime(1425016501000L); - entity.setModifiedTime(1425026901000L); // add metrics with floating point values Set<TimelineMetric> metricsEntity = new HashSet<>(); TimelineMetric mEntity = new TimelineMetric(); @@ -1135,8 +1108,8 @@ public class TestHBaseTimelineStorage { assertEquals(1, e1.getIsRelatedToEntities().size()); Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, null, null, null, null, - null, EnumSet.of(Field.ALL)); + null, null, null, null, null, null, null, null, null, null, null, + EnumSet.of(Field.ALL)); assertEquals(3, es1.size()); } @@ -1151,8 +1124,7 @@ public class TestHBaseTimelineStorage { e1.getRelatesToEntities().isEmpty()); Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, null, null, null, null, - null, null); + null, null, null, null, null, null, null, null, null, null, null, null); assertEquals(3, es1.size()); for (TimelineEntity e : es1) { assertTrue(e.getInfo().isEmpty() && e.getConfigs().isEmpty() && @@ -1172,8 +1144,8 @@ public class TestHBaseTimelineStorage { assertEquals(0, e1.getIsRelatedToEntities().size()); Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, null, null, null, null, - null, EnumSet.of(Field.IS_RELATED_TO, Field.METRICS)); + null, null, null, null, null, null, null, null, null, null, null, + EnumSet.of(Field.IS_RELATED_TO, Field.METRICS)); assertEquals(3, es1.size()); int metricsCnt = 0; int isRelatedToCnt = 0; @@ -1200,8 +1172,7 @@ public class TestHBaseTimelineStorage { assertEquals(1, e1.getConfigs().size()); Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, null, null, null, - list, null, null); + null, null, null, null, null, null, null, null, null, list, null, null); int cfgCnt = 0; for (TimelineEntity entity : es1) { cfgCnt += entity.getConfigs().size(); @@ -1217,8 +1188,8 @@ public class TestHBaseTimelineStorage { new TimelinePrefixFilter(TimelineCompareOp.EQUAL, "cfg_")); Set<TimelineEntity> entities = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, confFilters, null, null, - list, null, null); + null, null, null, null, null, null, confFilters, null, null, list, null, + null); assertEquals(1, entities.size()); int cfgCnt = 0; for (TimelineEntity entity : entities) { @@ -1239,8 +1210,7 @@ public class TestHBaseTimelineStorage { assertEquals(1, e1.getMetrics().size()); Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, null, null, null, null, - list, null); + null, null, null, null, null, null, null, null, null, null, list, null); int metricCnt = 0; for (TimelineEntity entity : es1) { metricCnt += entity.getMetrics().size(); @@ -1256,8 +1226,8 @@ public class TestHBaseTimelineStorage { new TimelinePrefixFilter(TimelineCompareOp.EQUAL, "MAP1_")); Set<TimelineEntity> entities = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, "application_1231111111_1111","world", - null, null, null, null, null, null, null, null, null, metricFilters, - null, null, list, null); + null, null, null, null, null, null, null, metricFilters, null, null, + list, null); assertEquals(1, entities.size()); int metricCnt = 0; for (TimelineEntity entity : entities) { @@ -1278,8 +1248,7 @@ public class TestHBaseTimelineStorage { Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, null, null, null, null, null, - EnumSet.of(Field.ALL)); + null, null, null, null, null, null, null, EnumSet.of(Field.ALL)); assertEquals(3, es1.size()); } @@ -1295,7 +1264,7 @@ public class TestHBaseTimelineStorage { Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, null, null, null, null, null, null); + null, null, null, null, null, null, null, null); assertEquals(3, es1.size()); for (TimelineEntity e : es1) { assertTrue(e.getInfo().isEmpty() && e.getConfigs().isEmpty() && @@ -1313,10 +1282,10 @@ public class TestHBaseTimelineStorage { assertNotNull(e1); assertEquals(3, e1.getConfigs().size()); assertEquals(0, e1.getIsRelatedToEntities().size()); - Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", - "some_flow_name", 1002345678919L, null, + Set<TimelineEntity> es1 = reader.getEntities( + "user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, null, null, null, null, null, + null, null, null, null, null, null, null, EnumSet.of(Field.IS_RELATED_TO, Field.METRICS)); assertEquals(3, es1.size()); int metricsCnt = 0; @@ -1342,10 +1311,10 @@ public class TestHBaseTimelineStorage { TimelineEntityType.YARN_APPLICATION.toString(), null, list, null, null); assertNotNull(e1); assertEquals(1, e1.getConfigs().size()); - Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", - "some_flow_name", 1002345678919L, null, + Set<TimelineEntity> es1 = reader.getEntities( + "user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, null, null, null, list, null, null); + null, null, null, null, null, list, null, null); int cfgCnt = 0; for (TimelineEntity entity : es1) { cfgCnt += entity.getConfigs().size(); @@ -1359,10 +1328,10 @@ public class TestHBaseTimelineStorage { TimelineFilterList list = new TimelineFilterList(Operator.OR, new TimelinePrefixFilter(TimelineCompareOp.EQUAL, "cfg_")); - Set<TimelineEntity> entities = reader.getEntities("user1", "cluster1", - "some_flow_name", 1002345678919L, null, + Set<TimelineEntity> entities = reader.getEntities( + "user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, confFilters, null, null, list, null, null); + null, null, confFilters, null, null, list, null, null); assertEquals(1, entities.size()); int cfgCnt = 0; for (TimelineEntity entity : entities) { @@ -1381,10 +1350,10 @@ public class TestHBaseTimelineStorage { TimelineEntityType.YARN_APPLICATION.toString(), null, null, list, null); assertNotNull(e1); assertEquals(1, e1.getMetrics().size()); - Set<TimelineEntity> es1 = reader.getEntities("user1", "cluster1", - "some_flow_name", 1002345678919L, null, + Set<TimelineEntity> es1 = reader.getEntities( + "user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, null, null, null, null, list, null); + null, null, null, null, null, null, list, null); int metricCnt = 0; for (TimelineEntity entity : es1) { metricCnt += entity.getMetrics().size(); @@ -1398,10 +1367,10 @@ public class TestHBaseTimelineStorage { new TimelineFilterList(Operator.OR, new TimelinePrefixFilter(TimelineCompareOp.EQUAL, "MAP1_")); Set<String> metricFilters = ImmutableSet.of("MAP1_SLOT_MILLIS"); - Set<TimelineEntity> entities = reader.getEntities("user1", "cluster1", - "some_flow_name", 1002345678919L, null, + Set<TimelineEntity> entities = reader.getEntities( + "user1", "cluster1", "some_flow_name", 1002345678919L, null, TimelineEntityType.YARN_APPLICATION.toString(), null, null, null, null, - null, null, null, null, null, metricFilters, null, null, list, null); + null, null, null, metricFilters, null, null, list, null); int metricCnt = 0; assertEquals(1, entities.size()); for (TimelineEntity entity : entities) { http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestPhoenixOfflineAggregationWriterImpl.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestPhoenixOfflineAggregationWriterImpl.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestPhoenixOfflineAggregationWriterImpl.java index de66a17..330adb3 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestPhoenixOfflineAggregationWriterImpl.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/TestPhoenixOfflineAggregationWriterImpl.java @@ -111,7 +111,6 @@ public class TestPhoenixOfflineAggregationWriterImpl extends BaseTest { entity.setId(id); entity.setType(type); entity.setCreatedTime(1425016501000L); - entity.setModifiedTime(1425016502000L); TimelineMetric metric = new TimelineMetric(); metric.setId("HDFS_BYTES_READ"); http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestFlowDataGenerator.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestFlowDataGenerator.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestFlowDataGenerator.java index d18613a..a4c06f2 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestFlowDataGenerator.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestFlowDataGenerator.java @@ -106,7 +106,6 @@ class TestFlowDataGenerator { long cTime = 20000000000000L; long mTime = 1425026901000L; entity.setCreatedTime(cTime); - entity.setModifiedTime(mTime); // add metrics Set<TimelineMetric> metrics = new HashSet<>(); TimelineMetric m1 = new TimelineMetric(); http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowActivity.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowActivity.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowActivity.java index eda1e21..cb3f7ee 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowActivity.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowActivity.java @@ -182,7 +182,7 @@ public class TestHBaseStorageFlowActivity { Set<TimelineEntity> entities = hbr.getEntities(null, cluster, null, null, null, TimelineEntityType.YARN_FLOW_ACTIVITY.toString(), 10L, null, null, - null, null, null, null, null, null, null, null, null, null, null); + null, null, null, null, null, null, null, null, null); assertEquals(1, entities.size()); for (TimelineEntity e : entities) { FlowActivityEntity flowActivity = (FlowActivityEntity)e; @@ -238,7 +238,7 @@ public class TestHBaseStorageFlowActivity { Set<TimelineEntity> entities = hbr.getEntities(user, cluster, flow, null, null, TimelineEntityType.YARN_FLOW_ACTIVITY.toString(), 10L, null, null, - null, null, null, null, null, null, null, null, null, null, null); + null, null, null, null, null, null, null, null, null); assertEquals(1, entities.size()); for (TimelineEntity e : entities) { FlowActivityEntity entity = (FlowActivityEntity)e; @@ -353,7 +353,7 @@ public class TestHBaseStorageFlowActivity { Set<TimelineEntity> entities = hbr.getEntities(null, cluster, null, null, null, TimelineEntityType.YARN_FLOW_ACTIVITY.toString(), 10L, null, null, - null, null, null, null, null, null, null, null, null, null, null); + null, null, null, null, null, null, null, null, null); assertEquals(1, entities.size()); for (TimelineEntity e : entities) { FlowActivityEntity flowActivity = (FlowActivityEntity)e; http://git-wip-us.apache.org/repos/asf/hadoop/blob/fd369a54/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowRun.java ---------------------------------------------------------------------- diff --git a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowRun.java b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowRun.java index 5da0192..57c326b 100644 --- a/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowRun.java +++ b/hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice/src/test/java/org/apache/hadoop/yarn/server/timelineservice/storage/flow/TestHBaseStorageFlowRun.java @@ -375,8 +375,7 @@ public class TestHBaseStorageFlowRun { Set<TimelineEntity> entities = hbr.getEntities(user, cluster, flow, runid, null, TimelineEntityType.YARN_FLOW_RUN.toString(), null, null, null, - null, null, null, null, null, null, null, null, null, - metricsToRetrieve, null); + null, null, null, null, null, null, null, metricsToRetrieve, null); assertEquals(1, entities.size()); for (TimelineEntity timelineEntity : entities) { Set<TimelineMetric> timelineMetrics = timelineEntity.getMetrics(); @@ -444,7 +443,7 @@ public class TestHBaseStorageFlowRun { hbr.start(); Set<TimelineEntity> entities = hbr.getEntities(user, cluster, flow, runid, null, TimelineEntityType.YARN_FLOW_RUN.toString(), null, null, null, - null, null, null, null, null, null, null, null, null, null, null); + null, null, null, null, null, null, null, null, null); assertEquals(1, entities.size()); for (TimelineEntity timelineEntity : entities) { assertEquals(0, timelineEntity.getMetrics().size()); @@ -452,8 +451,8 @@ public class TestHBaseStorageFlowRun { entities = hbr.getEntities(user, cluster, flow, runid, null, TimelineEntityType.YARN_FLOW_RUN.toString(), null, null, null, - null, null, null, null, null, null, null, null, null, - null, EnumSet.of(Field.METRICS)); + null, null, null, null, null, null, null, null, + EnumSet.of(Field.METRICS)); assertEquals(1, entities.size()); for (TimelineEntity timelineEntity : entities) { Set<TimelineMetric> timelineMetrics = timelineEntity.getMetrics(); --------------------------------------------------------------------- To unsubscribe, e-mail: common-commits-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-commits-h...@hadoop.apache.org