http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-bin.list ---------------------------------------------------------------------- diff --git a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-bin.list b/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-bin.list deleted file mode 100644 index d3861b9..0000000 --- a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-bin.list +++ /dev/null @@ -1,3 +0,0 @@ -mapred -yarn -container-executor
http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-client-2.7.3-api-report.json ---------------------------------------------------------------------- diff --git a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-client-2.7.3-api-report.json b/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-client-2.7.3-api-report.json deleted file mode 100644 index f62ee8e..0000000 --- a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-client-2.7.3-api-report.json +++ /dev/null @@ -1 +0,0 @@ -{"name":"hadoop-yarn-client","version":"2.7.3","classes":{"org.apache.hadoop.yarn.client.api.YarnClient":{"name":"org.apache.hadoop.yarn.client.api.YarnClient","methods":{"java.util.List getQueueAclsInfo() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getQueueAclsInfo","returnType":"java.util.List","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.client.api.YarnClient createYarnClient()":{"name":"createYarnClient","returnType":"org.apache.hadoop.yarn.client.api.YarnClient","args":[],"exceptions":[]},"java.util.List getApplications(java.util.Set) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplications","returnType":"java.util.List","args":["java.util.Set"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.Map getNodeToLabels() throws org.apache.hadoop.yarn.exceptions.YarnExceptio n, java.io.IOException":{"name":"getNodeToLabels","returnType":"java.util.Map","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.protocolrecords.ReservationUpdateResponse updateReservation(org.apache.hadoop.yarn.api.protocolrecords.ReservationUpdateRequest) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"updateReservation","returnType":"org.apache.hadoop.yarn.api.protocolrecords.ReservationUpdateResponse","args":["org.apache.hadoop.yarn.api.protocolrecords.ReservationUpdateRequest"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getAllQueues() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getAllQueues","returnType":"java.util.List","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getApplicationAttempts(org.apach e.hadoop.yarn.api.records.ApplicationId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplicationAttempts","returnType":"java.util.List","args":["org.apache.hadoop.yarn.api.records.ApplicationId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.client.api.YarnClientApplication createApplication() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"createApplication","returnType":"org.apache.hadoop.yarn.client.api.YarnClientApplication","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.Set getClusterNodeLabels() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getClusterNodeLabels","returnType":"java.util.Set","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"void moveApplicationAcrossQueues(org.apache .hadoop.yarn.api.records.ApplicationId, java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"moveApplicationAcrossQueues","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ApplicationId","java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.ApplicationAttemptReport getApplicationAttemptReport(org.apache.hadoop.yarn.api.records.ApplicationAttemptId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplicationAttemptReport","returnType":"org.apache.hadoop.yarn.api.records.ApplicationAttemptReport","args":["org.apache.hadoop.yarn.api.records.ApplicationAttemptId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.Token getRMDelegationToken(org.apache.hadoop.io.Text) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getRMDelegationToken","returnType":"org.apache.hadoop.yarn.api.records.Token","args":["org.apache.hadoop.io.Text"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.protocolrecords.ReservationSubmissionResponse submitReservation(org.apache.hadoop.yarn.api.protocolrecords.ReservationSubmissionRequest) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"submitReservation","returnType":"org.apache.hadoop.yarn.api.protocolrecords.ReservationSubmissionResponse","args":["org.apache.hadoop.yarn.api.protocolrecords.ReservationSubmissionRequest"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getContainers(org.apache.hadoop.yarn.api.records.ApplicationAttemptId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getContainers","returnType":"java.util.List","args":["org.apac he.hadoop.yarn.api.records.ApplicationAttemptId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.ApplicationReport getApplicationReport(org.apache.hadoop.yarn.api.records.ApplicationId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplicationReport","returnType":"org.apache.hadoop.yarn.api.records.ApplicationReport","args":["org.apache.hadoop.yarn.api.records.ApplicationId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.ApplicationId submitApplication(org.apache.hadoop.yarn.api.records.ApplicationSubmissionContext) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"submitApplication","returnType":"org.apache.hadoop.yarn.api.records.ApplicationId","args":["org.apache.hadoop.yarn.api.records.ApplicationSubmissionContext"],"exceptions":["org.apache.hadoop. yarn.exceptions.YarnException","java.io.IOException"]},"java.util.Map getLabelsToNodes() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getLabelsToNodes","returnType":"java.util.Map","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.security.token.Token getAMRMToken(org.apache.hadoop.yarn.api.records.ApplicationId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getAMRMToken","returnType":"org.apache.hadoop.security.token.Token","args":["org.apache.hadoop.yarn.api.records.ApplicationId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getApplications(java.util.EnumSet) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplications","returnType":"java.util.List","args":["java.util.EnumSet"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnExcept ion","java.io.IOException"]},"java.util.List getRootQueueInfos() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getRootQueueInfos","returnType":"java.util.List","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.QueueInfo getQueueInfo(java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getQueueInfo","returnType":"org.apache.hadoop.yarn.api.records.QueueInfo","args":["java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getChildQueueInfos(java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getChildQueueInfos","returnType":"java.util.List","args":["java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.Map getLabelsToNodes(java.util .Set) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getLabelsToNodes","returnType":"java.util.Map","args":["java.util.Set"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.protocolrecords.ReservationDeleteResponse deleteReservation(org.apache.hadoop.yarn.api.protocolrecords.ReservationDeleteRequest) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"deleteReservation","returnType":"org.apache.hadoop.yarn.api.protocolrecords.ReservationDeleteResponse","args":["org.apache.hadoop.yarn.api.protocolrecords.ReservationDeleteRequest"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.YarnClusterMetrics getYarnClusterMetrics() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getYarnClusterMetrics","returnType":"org.apache.hadoop.yarn.api.rec ords.YarnClusterMetrics","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getNodeReports([Lorg.apache.hadoop.yarn.api.records.NodeState;) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getNodeReports","returnType":"java.util.List","args":["[Lorg.apache.hadoop.yarn.api.records.NodeState;"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"void killApplication(org.apache.hadoop.yarn.api.records.ApplicationId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"killApplication","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ApplicationId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getApplications() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplications","returnType":"java.util.List","args ":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getApplications(java.util.Set, java.util.EnumSet) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplications","returnType":"java.util.List","args":["java.util.Set","java.util.EnumSet"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.ContainerReport getContainerReport(org.apache.hadoop.yarn.api.records.ContainerId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getContainerReport","returnType":"org.apache.hadoop.yarn.api.records.ContainerReport","args":["org.apache.hadoop.yarn.api.records.ContainerId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]}}},"org.apache.hadoop.yarn.client.api.async.NMClientAsync":{"name":"org.apache.hadoop.yarn.client.api.async.NMClientAsync","methods":{"v oid setClient(org.apache.hadoop.yarn.client.api.NMClient)":{"name":"setClient","returnType":"void","args":["org.apache.hadoop.yarn.client.api.NMClient"],"exceptions":[]},"void setCallbackHandler(org.apache.hadoop.yarn.client.api.async.NMClientAsync$CallbackHandler)":{"name":"setCallbackHandler","returnType":"void","args":["org.apache.hadoop.yarn.client.api.async.NMClientAsync$CallbackHandler"],"exceptions":[]},"void getContainerStatusAsync(org.apache.hadoop.yarn.api.records.ContainerId, org.apache.hadoop.yarn.api.records.NodeId)":{"name":"getContainerStatusAsync","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ContainerId","org.apache.hadoop.yarn.api.records.NodeId"],"exceptions":[]},"void startContainerAsync(org.apache.hadoop.yarn.api.records.Container, org.apache.hadoop.yarn.api.records.ContainerLaunchContext)":{"name":"startContainerAsync","returnType":"void","args":["org.apache.hadoop.yarn.api.records.Container","org.apache.hadoop.yarn.api.records.ContainerLaunch Context"],"exceptions":[]},"void stopContainerAsync(org.apache.hadoop.yarn.api.records.ContainerId, org.apache.hadoop.yarn.api.records.NodeId)":{"name":"stopContainerAsync","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ContainerId","org.apache.hadoop.yarn.api.records.NodeId"],"exceptions":[]},"org.apache.hadoop.yarn.client.api.NMClient getClient()":{"name":"getClient","returnType":"org.apache.hadoop.yarn.client.api.NMClient","args":[],"exceptions":[]},"org.apache.hadoop.yarn.client.api.async.NMClientAsync$CallbackHandler getCallbackHandler()":{"name":"getCallbackHandler","returnType":"org.apache.hadoop.yarn.client.api.async.NMClientAsync$CallbackHandler","args":[],"exceptions":[]},"org.apache.hadoop.yarn.client.api.async.NMClientAsync createNMClientAsync(org.apache.hadoop.yarn.client.api.async.NMClientAsync$CallbackHandler)":{"name":"createNMClientAsync","returnType":"org.apache.hadoop.yarn.client.api.async.NMClientAsync","args":["org.apache.hadoop.yarn.client.api. async.NMClientAsync$CallbackHandler"],"exceptions":[]}}},"org.apache.hadoop.yarn.client.api.AMRMClient":{"name":"org.apache.hadoop.yarn.client.api.AMRMClient","methods":{"org.apache.hadoop.yarn.client.api.NMTokenCache getNMTokenCache()":{"name":"getNMTokenCache","returnType":"org.apache.hadoop.yarn.client.api.NMTokenCache","args":[],"exceptions":[]},"void addContainerRequest(org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest)":{"name":"addContainerRequest","returnType":"void","args":["org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest"],"exceptions":[]},"void updateBlacklist(java.util.List, java.util.List)":{"name":"updateBlacklist","returnType":"void","args":["java.util.List","java.util.List"],"exceptions":[]},"java.util.List getMatchingRequests(org.apache.hadoop.yarn.api.records.Priority, java.lang.String, org.apache.hadoop.yarn.api.records.Resource)":{"name":"getMatchingRequests","returnType":"java.util.List","args":["org.apache.hadoop.yarn.api.records.Prior ity","java.lang.String","org.apache.hadoop.yarn.api.records.Resource"],"exceptions":[]},"void waitFor(com.google.common.base.Supplier) throws java.lang.InterruptedException":{"name":"waitFor","returnType":"void","args":["com.google.common.base.Supplier"],"exceptions":["java.lang.InterruptedException"]},"org.apache.hadoop.yarn.api.records.Resource getAvailableResources()":{"name":"getAvailableResources","returnType":"org.apache.hadoop.yarn.api.records.Resource","args":[],"exceptions":[]},"void waitFor(com.google.common.base.Supplier, int) throws java.lang.InterruptedException":{"name":"waitFor","returnType":"void","args":["com.google.common.base.Supplier","int"],"exceptions":["java.lang.InterruptedException"]},"org.apache.hadoop.yarn.api.protocolrecords.RegisterApplicationMasterResponse registerApplicationMaster(java.lang.String, int, java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"registerApplicationMaster","returnType":"org.ap ache.hadoop.yarn.api.protocolrecords.RegisterApplicationMasterResponse","args":["java.lang.String","int","java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"void unregisterApplicationMaster(org.apache.hadoop.yarn.api.records.FinalApplicationStatus, java.lang.String, java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"unregisterApplicationMaster","returnType":"void","args":["org.apache.hadoop.yarn.api.records.FinalApplicationStatus","java.lang.String","java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"void waitFor(com.google.common.base.Supplier, int, int) throws java.lang.InterruptedException":{"name":"waitFor","returnType":"void","args":["com.google.common.base.Supplier","int","int"],"exceptions":["java.lang.InterruptedException"]},"int getClusterNodeCount()":{"name":"getClusterNodeCount","returnType":"int","args":[], "exceptions":[]},"void releaseAssignedContainer(org.apache.hadoop.yarn.api.records.ContainerId)":{"name":"releaseAssignedContainer","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ContainerId"],"exceptions":[]},"void setNMTokenCache(org.apache.hadoop.yarn.client.api.NMTokenCache)":{"name":"setNMTokenCache","returnType":"void","args":["org.apache.hadoop.yarn.client.api.NMTokenCache"],"exceptions":[]},"void removeContainerRequest(org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest)":{"name":"removeContainerRequest","returnType":"void","args":["org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest"],"exceptions":[]},"org.apache.hadoop.yarn.client.api.AMRMClient createAMRMClient()":{"name":"createAMRMClient","returnType":"org.apache.hadoop.yarn.client.api.AMRMClient","args":[],"exceptions":[]},"org.apache.hadoop.yarn.api.protocolrecords.AllocateResponse allocate(float) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name": "allocate","returnType":"org.apache.hadoop.yarn.api.protocolrecords.AllocateResponse","args":["float"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]}}},"org.apache.hadoop.yarn.client.api.YarnClientApplication":{"name":"org.apache.hadoop.yarn.client.api.YarnClientApplication","methods":{"org.apache.hadoop.yarn.api.protocolrecords.GetNewApplicationResponse getNewApplicationResponse()":{"name":"getNewApplicationResponse","returnType":"org.apache.hadoop.yarn.api.protocolrecords.GetNewApplicationResponse","args":[],"exceptions":[]},"org.apache.hadoop.yarn.api.records.ApplicationSubmissionContext getApplicationSubmissionContext()":{"name":"getApplicationSubmissionContext","returnType":"org.apache.hadoop.yarn.api.records.ApplicationSubmissionContext","args":[],"exceptions":[]}}},"org.apache.hadoop.yarn.client.api.AHSClient":{"name":"org.apache.hadoop.yarn.client.api.AHSClient","methods":{"org.apache.hadoop.yarn.api.records.ApplicationAttemptReport g etApplicationAttemptReport(org.apache.hadoop.yarn.api.records.ApplicationAttemptId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplicationAttemptReport","returnType":"org.apache.hadoop.yarn.api.records.ApplicationAttemptReport","args":["org.apache.hadoop.yarn.api.records.ApplicationAttemptId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getApplications() throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplications","returnType":"java.util.List","args":[],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getContainers(org.apache.hadoop.yarn.api.records.ApplicationAttemptId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getContainers","returnType":"java.util.List","args":["org.apache.hadoop.yarn.api.records.ApplicationAttemptId"],"exceptions":["org .apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.ApplicationReport getApplicationReport(org.apache.hadoop.yarn.api.records.ApplicationId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplicationReport","returnType":"org.apache.hadoop.yarn.api.records.ApplicationReport","args":["org.apache.hadoop.yarn.api.records.ApplicationId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.client.api.AHSClient createAHSClient()":{"name":"createAHSClient","returnType":"org.apache.hadoop.yarn.client.api.AHSClient","args":[],"exceptions":[]},"org.apache.hadoop.yarn.api.records.ContainerReport getContainerReport(org.apache.hadoop.yarn.api.records.ContainerId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getContainerReport","returnType":"org.apache.hadoop.yarn.api.records.ContainerReport","args":["or g.apache.hadoop.yarn.api.records.ContainerId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"java.util.List getApplicationAttempts(org.apache.hadoop.yarn.api.records.ApplicationId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getApplicationAttempts","returnType":"java.util.List","args":["org.apache.hadoop.yarn.api.records.ApplicationId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]}}},"org.apache.hadoop.yarn.client.api.async.AMRMClientAsync":{"name":"org.apache.hadoop.yarn.client.api.async.AMRMClientAsync","methods":{"void addContainerRequest(org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest)":{"name":"addContainerRequest","returnType":"void","args":["org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest"],"exceptions":[]},"void updateBlacklist(java.util.List, java.util.List)":{"name":"updateBlacklist","returnType":"void","args":["java.uti l.List","java.util.List"],"exceptions":[]},"java.util.List getMatchingRequests(org.apache.hadoop.yarn.api.records.Priority, java.lang.String, org.apache.hadoop.yarn.api.records.Resource)":{"name":"getMatchingRequests","returnType":"java.util.List","args":["org.apache.hadoop.yarn.api.records.Priority","java.lang.String","org.apache.hadoop.yarn.api.records.Resource"],"exceptions":[]},"void waitFor(com.google.common.base.Supplier) throws java.lang.InterruptedException":{"name":"waitFor","returnType":"void","args":["com.google.common.base.Supplier"],"exceptions":["java.lang.InterruptedException"]},"org.apache.hadoop.yarn.client.api.async.AMRMClientAsync createAMRMClientAsync(int, org.apache.hadoop.yarn.client.api.async.AMRMClientAsync$CallbackHandler)":{"name":"createAMRMClientAsync","returnType":"org.apache.hadoop.yarn.client.api.async.AMRMClientAsync","args":["int","org.apache.hadoop.yarn.client.api.async.AMRMClientAsync$CallbackHandler"],"exceptions":[]},"org.apache.hadoop.yarn.api.r ecords.Resource getAvailableResources()":{"name":"getAvailableResources","returnType":"org.apache.hadoop.yarn.api.records.Resource","args":[],"exceptions":[]},"void waitFor(com.google.common.base.Supplier, int) throws java.lang.InterruptedException":{"name":"waitFor","returnType":"void","args":["com.google.common.base.Supplier","int"],"exceptions":["java.lang.InterruptedException"]},"org.apache.hadoop.yarn.client.api.async.AMRMClientAsync createAMRMClientAsync(org.apache.hadoop.yarn.client.api.AMRMClient, int, org.apache.hadoop.yarn.client.api.async.AMRMClientAsync$CallbackHandler)":{"name":"createAMRMClientAsync","returnType":"org.apache.hadoop.yarn.client.api.async.AMRMClientAsync","args":["org.apache.hadoop.yarn.client.api.AMRMClient","int","org.apache.hadoop.yarn.client.api.async.AMRMClientAsync$CallbackHandler"],"exceptions":[]},"org.apache.hadoop.yarn.api.protocolrecords.RegisterApplicationMasterResponse registerApplicationMaster(java.lang.String, int, java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"registerApplicationMaster","returnType":"org.apache.hadoop.yarn.api.protocolrecords.RegisterApplicationMasterResponse","args":["java.lang.String","int","java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"void unregisterApplicationMaster(org.apache.hadoop.yarn.api.records.FinalApplicationStatus, java.lang.String, java.lang.String) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"unregisterApplicationMaster","returnType":"void","args":["org.apache.hadoop.yarn.api.records.FinalApplicationStatus","java.lang.String","java.lang.String"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"void waitFor(com.google.common.base.Supplier, int, int) throws java.lang.InterruptedException":{"name":"waitFor","returnType":"void","args":["com.google.common.base.Supplier","int","int"],"exceptio ns":["java.lang.InterruptedException"]},"int getClusterNodeCount()":{"name":"getClusterNodeCount","returnType":"int","args":[],"exceptions":[]},"void releaseAssignedContainer(org.apache.hadoop.yarn.api.records.ContainerId)":{"name":"releaseAssignedContainer","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ContainerId"],"exceptions":[]},"void removeContainerRequest(org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest)":{"name":"removeContainerRequest","returnType":"void","args":["org.apache.hadoop.yarn.client.api.AMRMClient$ContainerRequest"],"exceptions":[]},"void setHeartbeatInterval(int)":{"name":"setHeartbeatInterval","returnType":"void","args":["int"],"exceptions":[]}}},"org.apache.hadoop.yarn.client.api.NMClient":{"name":"org.apache.hadoop.yarn.client.api.NMClient","methods":{"void stopContainer(org.apache.hadoop.yarn.api.records.ContainerId, org.apache.hadoop.yarn.api.records.NodeId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOExc eption":{"name":"stopContainer","returnType":"void","args":["org.apache.hadoop.yarn.api.records.ContainerId","org.apache.hadoop.yarn.api.records.NodeId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.api.records.ContainerStatus getContainerStatus(org.apache.hadoop.yarn.api.records.ContainerId, org.apache.hadoop.yarn.api.records.NodeId) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"getContainerStatus","returnType":"org.apache.hadoop.yarn.api.records.ContainerStatus","args":["org.apache.hadoop.yarn.api.records.ContainerId","org.apache.hadoop.yarn.api.records.NodeId"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.client.api.NMTokenCache getNMTokenCache()":{"name":"getNMTokenCache","returnType":"org.apache.hadoop.yarn.client.api.NMTokenCache","args":[],"exceptions":[]},"org.apache.hadoop.yarn.client.api.NMClient creat eNMClient()":{"name":"createNMClient","returnType":"org.apache.hadoop.yarn.client.api.NMClient","args":[],"exceptions":[]},"java.util.Map startContainer(org.apache.hadoop.yarn.api.records.Container, org.apache.hadoop.yarn.api.records.ContainerLaunchContext) throws org.apache.hadoop.yarn.exceptions.YarnException, java.io.IOException":{"name":"startContainer","returnType":"java.util.Map","args":["org.apache.hadoop.yarn.api.records.Container","org.apache.hadoop.yarn.api.records.ContainerLaunchContext"],"exceptions":["org.apache.hadoop.yarn.exceptions.YarnException","java.io.IOException"]},"org.apache.hadoop.yarn.client.api.NMClient createNMClient(java.lang.String)":{"name":"createNMClient","returnType":"org.apache.hadoop.yarn.client.api.NMClient","args":["java.lang.String"],"exceptions":[]},"void setNMTokenCache(org.apache.hadoop.yarn.client.api.NMTokenCache)":{"name":"setNMTokenCache","returnType":"void","args":["org.apache.hadoop.yarn.client.api.NMTokenCache"],"exceptions":[]},"void cleanupRunningContainersOnStop(boolean)":{"name":"cleanupRunningContainersOnStop","returnType":"void","args":["boolean"],"exceptions":[]}}}}} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-common-2.7.3-api-report.json ---------------------------------------------------------------------- diff --git a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-common-2.7.3-api-report.json b/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-common-2.7.3-api-report.json deleted file mode 100644 index b394bff..0000000 --- a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-common-2.7.3-api-report.json +++ /dev/null @@ -1 +0,0 @@ -{"name":"hadoop-yarn-common","version":"2.7.3","classes":{"org.apache.hadoop.yarn.security.ContainerTokenSelector":{"name":"org.apache.hadoop.yarn.security.ContainerTokenSelector","methods":{"org.apache.hadoop.security.token.Token selectToken(org.apache.hadoop.io.Text, java.util.Collection)":{"name":"selectToken","returnType":"org.apache.hadoop.security.token.Token","args":["org.apache.hadoop.io.Text","java.util.Collection"],"exceptions":[]}}},"org.apache.hadoop.yarn.security.ContainerManagerSecurityInfo":{"name":"org.apache.hadoop.yarn.security.ContainerManagerSecurityInfo","methods":{"org.apache.hadoop.security.KerberosInfo getKerberosInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getKerberosInfo","returnType":"org.apache.hadoop.security.KerberosInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]},"org.apache.hadoop.security.token.TokenInfo getTokenInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getTo kenInfo","returnType":"org.apache.hadoop.security.token.TokenInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]}}},"org.apache.hadoop.yarn.security.SchedulerSecurityInfo":{"name":"org.apache.hadoop.yarn.security.SchedulerSecurityInfo","methods":{"org.apache.hadoop.security.KerberosInfo getKerberosInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getKerberosInfo","returnType":"org.apache.hadoop.security.KerberosInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]},"org.apache.hadoop.security.token.TokenInfo getTokenInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getTokenInfo","returnType":"org.apache.hadoop.security.token.TokenInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]}}},"org.apache.hadoop.yarn.util.SystemClock":{"name":"org.apache.hadoop.yarn.util.SystemClock","methods":{"long getTime()":{"name":"getTime","returnType":"long", "args":[],"exceptions":[]}}},"org.apache.hadoop.yarn.security.client.RMDelegationTokenSelector":{"name":"org.apache.hadoop.yarn.security.client.RMDelegationTokenSelector","methods":{"org.apache.hadoop.security.token.Token selectToken(org.apache.hadoop.io.Text, java.util.Collection)":{"name":"selectToken","returnType":"org.apache.hadoop.security.token.Token","args":["org.apache.hadoop.io.Text","java.util.Collection"],"exceptions":[]}}},"org.apache.hadoop.yarn.security.client.ClientRMSecurityInfo":{"name":"org.apache.hadoop.yarn.security.client.ClientRMSecurityInfo","methods":{"org.apache.hadoop.security.KerberosInfo getKerberosInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getKerberosInfo","returnType":"org.apache.hadoop.security.KerberosInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]},"org.apache.hadoop.security.token.TokenInfo getTokenInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getTokenInfo","r eturnType":"org.apache.hadoop.security.token.TokenInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]}}},"org.apache.hadoop.yarn.security.admin.AdminSecurityInfo":{"name":"org.apache.hadoop.yarn.security.admin.AdminSecurityInfo","methods":{"org.apache.hadoop.security.KerberosInfo getKerberosInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getKerberosInfo","returnType":"org.apache.hadoop.security.KerberosInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]},"org.apache.hadoop.security.token.TokenInfo getTokenInfo(java.lang.Class, org.apache.hadoop.conf.Configuration)":{"name":"getTokenInfo","returnType":"org.apache.hadoop.security.token.TokenInfo","args":["java.lang.Class","org.apache.hadoop.conf.Configuration"],"exceptions":[]}}},"org.apache.hadoop.yarn.client.ClientRMProxy":{"name":"org.apache.hadoop.yarn.client.ClientRMProxy","methods":{"org.apache.hadoop.io.Text getRMDelegationTokenService (org.apache.hadoop.conf.Configuration)":{"name":"getRMDelegationTokenService","returnType":"org.apache.hadoop.io.Text","args":["org.apache.hadoop.conf.Configuration"],"exceptions":[]},"org.apache.hadoop.io.Text getAMRMTokenService(org.apache.hadoop.conf.Configuration)":{"name":"getAMRMTokenService","returnType":"org.apache.hadoop.io.Text","args":["org.apache.hadoop.conf.Configuration"],"exceptions":[]},"org.apache.hadoop.io.Text getTokenService(org.apache.hadoop.conf.Configuration, java.lang.String, java.lang.String, int)":{"name":"getTokenService","returnType":"org.apache.hadoop.io.Text","args":["org.apache.hadoop.conf.Configuration","java.lang.String","java.lang.String","int"],"exceptions":[]},"java.lang.Object createRMProxy(org.apache.hadoop.conf.Configuration, java.lang.Class) throws java.io.IOException":{"name":"createRMProxy","returnType":"java.lang.Object","args":["org.apache.hadoop.conf.Configuration","java.lang.Class"],"exceptions":["java.io.IOException"]}}},"org.apache.had oop.yarn.util.Clock":{"name":"org.apache.hadoop.yarn.util.Clock","methods":{"long getTime()":{"name":"getTime","returnType":"long","args":[],"exceptions":[]}}}}} \ No newline at end of file http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-jar.list ---------------------------------------------------------------------- diff --git a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-jar.list b/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-jar.list deleted file mode 100644 index 26613d4..0000000 --- a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn-jar.list +++ /dev/null @@ -1,38 +0,0 @@ -netty-3\.6\.2\.Final[\.\-_].*jar -leveldbjni-all-1\.8[\.\-_].*jar -jackson-core-asl-1\.9\.13[\.\-_].*jar -jackson-xc-1\.9\.13[\.\-_].*jar -jersey-server-1\.9[\.\-_].*jar -stax-api-1\.0-2[\.\-_].*jar -zookeeper-3\.4\.6[\.\-_].*jar -guice-3\.0[\.\-_].*jar -jaxb-impl-2\.2\.3-1[\.\-_].*jar -zookeeper-3\.4\.6.*-tests\.jar -jersey-client-1\.9[\.\-_].*jar -commons-cli-1\.2[\.\-_].*jar -log4j-1\.2\.17[\.\-_].*jar -jackson-mapper-asl-1\.9\.13[\.\-_].*jar -guava-11\.0\.2[\.\-_].*jar -jetty-6\.1\.26[\.\-_].*jar -commons-logging-1\.1\.3[\.\-_].*jar -jersey-core-1\.9[\.\-_].*jar -jersey-guice-1\.9[\.\-_].*jar -commons-compress-1\.4\.1[\.\-_].*jar -jettison-1\.1[\.\-_].*jar -commons-collections-3\.2\.[12][\.\-_].*jar -xz-1\.0[\.\-_].*jar -asm-3\.2[\.\-_].*jar -commons-codec-1\.4[\.\-_].*jar -aopalliance-1\.0[\.\-_].*jar -javax\.inject-1[\.\-_].*jar -commons-lang-2\.6[\.\-_].*jar -jetty-util-6\.1\.26[\.\-_].*jar -jsr305-3\.0\.0[\.\-_].*jar -protobuf-java-2\.5\.0[\.\-_].*jar -commons-io-2\.4[\.\-_].*jar -activation-1\.1[\.\-_].*jar -jersey-json-1\.9[\.\-_].*jar -jaxb-api-2\.2\.2[\.\-_].*jar -guice-servlet-3\.0[\.\-_].*jar -servlet-api-2\.5[\.\-_].*jar -jackson-jaxrs-1\.9\.13[\.\-_].*jar http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn.list ---------------------------------------------------------------------- diff --git a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn.list b/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn.list deleted file mode 100644 index bb88005..0000000 --- a/bigtop-tests/spec-tests/runtime/src/test/resources/hadoop-yarn.list +++ /dev/null @@ -1,74 +0,0 @@ -hadoop-yarn-server-sharedcachemanager.*\.jar -bin -bin/mapred -bin/container-executor -bin/yarn -sbin -sbin/yarn-daemon\.sh -sbin/yarn-daemons\.sh -hadoop-yarn-registry-2\.7\.[0-9].*\.jar -hadoop-yarn-applications-unmanaged-am-launcher-2\.7\.[0-9].*\.jar -hadoop-yarn-common-2\.7\.[0-9].*\.jar -hadoop-yarn-server-nodemanager.*\.jar -hadoop-yarn-server-applicationhistoryservice-2\.7\.[0-9].*\.jar -hadoop-yarn-server-common.*\.jar -etc -etc/hadoop -hadoop-yarn-server-common-2\.7\.[0-9].*\.jar -hadoop-yarn-server-tests.*\.jar -hadoop-yarn-server-resourcemanager.*\.jar -hadoop-yarn-server-web-proxy.*\.jar -hadoop-yarn-api-2\.7\.[0-9].*\.jar -hadoop-yarn-common.*\.jar -hadoop-yarn-server-web-proxy-2\.7\.[0-9].*\.jar -hadoop-yarn-applications-distributedshell-2\.7\.[0-9].*\.jar -hadoop-yarn-server-tests-2\.7\.[0-9].*\.jar -hadoop-yarn-server-resourcemanager-2\.7\.[0-9].*\.jar -hadoop-yarn-registry.*\.jar -hadoop-yarn-server-sharedcachemanager-2\.7\.[0-9].*\.jar -hadoop-yarn-client-2\.7\.[0-9].*\.jar -hadoop-yarn-applications-distributedshell.*\.jar -hadoop-yarn-server-nodemanager-2\.7\.[0-9].*\.jar -hadoop-yarn-api.*\.jar -hadoop-yarn-client.*\.jar -lib -lib/commons-cli-1\.2.*\.jar -lib/leveldbjni-all-1\.8.*\.jar -lib/jaxb-api-2\.2\.2.*\.jar -lib/jettison-1\.1.*\.jar -lib/commons-io-2\.4.*\.jar -lib/jetty-util-6\.1\.26.*\.jar -lib/jaxb-impl-2\.2\.3-1.*\.jar -lib/jersey-guice-1\.9.*\.jar -lib/netty-3\.6\.2\.Final.*\.jar -lib/jersey-core-1\.9.*\.jar -lib/jackson-mapper-asl-1\.9\.13.*\.jar -lib/asm-3\.2.*\.jar -lib/commons-compress-1\.4\.1.*\.jar -lib/aopalliance-1\.0.*\.jar -lib/jackson-xc-1\.9\.13.*\.jar -lib/jersey-json-1\.9.*\.jar -lib/commons-codec-1\.4.*\.jar -lib/jackson-core-asl-1\.9\.13.*\.jar -lib/servlet-api-2\.5.*\.jar -lib/jetty-6\.1\.26.*\.jar -lib/jersey-server-1\.9.*\.jar -lib/log4j-1\.2\.17.*\.jar -lib/zookeeper-3\.4\.6.*-tests\.jar -lib/stax-api-1\.0-2.*\.jar -lib/jersey-client-1\.9.*\.jar -lib/xz-1\.0.*\.jar -lib/zookeeper-3\.4\.6.*\.jar -lib/activation-1\.1.*\.jar -lib/javax\.inject-1.*\.jar -lib/protobuf-java-2\.5\.0.*\.jar -lib/guice-3\.0.*\.jar -lib/guava-11\.0\.2.*\.jar -lib/jsr305-3\.0\.0.*\.jar -lib/jackson-jaxrs-1\.9\.13.*\.jar -lib/commons-collections-3\.2\.[1-2].*\.jar -lib/commons-logging-1\.1\.3.*\.jar -lib/commons-lang-2\.6.*\.jar -lib/guice-servlet-3\.0.*\.jar -hadoop-yarn-server-applicationhistoryservice.*\.jar -hadoop-yarn-applications-unmanaged-am-launcher.*\.jar http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/bigtop-tests/spec-tests/runtime/src/test/resources/testRuntimeSpecConf.groovy ---------------------------------------------------------------------- diff --git a/bigtop-tests/spec-tests/runtime/src/test/resources/testRuntimeSpecConf.groovy b/bigtop-tests/spec-tests/runtime/src/test/resources/testRuntimeSpecConf.groovy deleted file mode 100644 index 339de4c..0000000 --- a/bigtop-tests/spec-tests/runtime/src/test/resources/testRuntimeSpecConf.groovy +++ /dev/null @@ -1,430 +0,0 @@ -/** - * Licensed to the Apache Software Foundation (ASF) under one - * or more contributor license agreements. See the NOTICE file - * distributed with this work for additional information - * regarding copyright ownership. The ASF licenses this file - * to you under the Apache License, Version 2.0 (the - * "License"); you may not use this file except in compliance - * with the License. You may obtain a copy of the License at - * <p/> - * http://www.apache.org/licenses/LICENSE-2.0 - * <p/> - * Unless required by applicable law or agreed to in writing, software - * distributed under the License is distributed on an "AS IS" BASIS, - * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. - * See the License for the specific language governing permissions and - * limitations under the License. - */ - -specs { - tests { - 'HADOOP_EJH1' { - name = 'HADOOP_EJH1' - type = 'envdir' - arguments { - envcmd = 'hadoop envvars' - variable = 'JAVA_HOME' - } - } - 'HADOOP_EC1' { - name = 'HADOOP_EC1' - type = 'envdir' - arguments { - envcmd = 'hadoop envvars' - variable = 'HADOOP_TOOLS_PATH' - donotcheckexistance = true - } - } - 'HADOOP_EC2' { - name = 'HADOOP_EC2' - type = 'envdir' - arguments { - envcmd = 'hadoop envvars' - variable = 'HADOOP_COMMON_HOME' - } - } - 'HADOOP_EC3' { - name = 'HADOOP_EC3' - type = 'envdir' - arguments { - envcmd = 'hadoop envvars' - variable = 'HADOOP_COMMON_DIR' - relative = true - } - } - 'HADOOP_EC4' { - name = 'HADOOP_EC4' - type = 'envdir' - arguments { - envcmd = 'hadoop envvars' - variable = 'HADOOP_COMMON_LIB_JARS_DIR' - relative = true - } - } - 'HADOOP_EC5' { - name = 'HADOOP_EC5' - type = 'envdir' - arguments { - envcmd = 'hadoop envvars' - variable = 'HADOOP_CONF_DIR' - } - } - 'HADOOP_EH1' { - name = 'HADOOP_EH1' - type = 'envdir' - arguments { - envcmd = 'hdfs envvars' - variable = 'HADOOP_HDFS_HOME' - } - } - 'HADOOP_EH2' { - name = 'HADOOP_EH2' - type = 'envdir' - arguments { - envcmd = 'hdfs envvars' - variable = 'HDFS_DIR' - relative = true - } - } - 'HADOOP_EH3' { - name = 'HADOOP_EH3' - type = 'envdir' - arguments { - envcmd = 'hdfs envvars' - variable = 'HDFS_LIB_JARS_DIR' - relative = true - } - } - 'HADOOP_EY1' { - name = 'HADOOP_EY1' - type = 'envdir' - arguments { - envcmd = 'yarn envvars' - variable = 'HADOOP_YARN_HOME' - } - } - 'HADOOP_EY2' { - name = 'HADOOP_EY2' - type = 'envdir' - arguments { - envcmd = 'yarn envvars' - variable = 'YARN_DIR' - relative = true - } - } - 'HADOOP_EY3' { - name = 'HADOOP_EY3' - type = 'envdir' - arguments { - envcmd = 'yarn envvars' - variable = 'YARN_LIB_JARS_DIR' - relative = true - } - } - 'HADOOP_EM1' { - name = 'HADOOP_EM1' - type = 'envdir' - arguments { - envcmd = 'mapred envvars' - variable = 'HADOOP_MAPRED_HOME' - } - } - 'HADOOP_EM2' { - name = 'HADOOP_EM2' - type = 'envdir' - arguments { - envcmd = 'mapred envvars' - variable = 'MAPRED_DIR' - relative = true - } - } - 'HADOOP_EM3' { - name = 'HADOOP_EM3' - type = 'envdir' - arguments { - envcmd = 'mapred envvars' - variable = 'MAPRED_LIB_JARS_DIR' - relative = true - } - } - 'HADOOP_EJH2_HADOOP' { - name = 'HADOOP_EJH2_HADOOP' - type = 'shell' - arguments { - command = '[ "${JAVA_HOME}xxx" != "xxx" ] || grep -E "^\\s*export\\s+JAVA_HOME=[\\w/]+" `hadoop envvars | grep HADOOP_CONF_DIR | sed "s|[^=]\\+=\'\\([^\']\\+\\)\'$|\\1|g"`/hadoop-env.sh' - message = 'JAVA_HOME is not set' - } - } - 'HADOOP_EJH2_YARN' { - name = 'HADOOP_EJH2_YARN' - type = 'shell' - arguments { - command = '[ "${JAVA_HOME}xxx" != "xxx" ] || grep -E "^\\s*export\\s+JAVA_HOME=[\\w/]+" `hadoop envvars | grep HADOOP_CONF_DIR | sed "s|[^=]\\+=\'\\([^\']\\+\\)\'$|\\1|g"`/yarn-env.sh' - message = 'JAVA_HOME is not set' - } - } - 'HADOOP_PLATVER_1' { - name = 'HADOOP_PLATVER' - type = 'shell' - arguments { - command = 'hadoop version | head -n 1 | grep -E \'Hadoop\\s+[0-9\\.]+[_\\-][A-Za-z_0-9]+\'' - message = 'Hadoop\'s version string is not correct' - } - } - 'HADOOP_DIRSTRUCT_COMMON' { - name = 'HADOOP_DIRSTRUCT_COMMON' - type = 'dirstruct' - arguments { - envcmd = 'hadoop envvars' - baseDirEnv = 'HADOOP_COMMON_HOME' - referenceList = 'hadoop-common.list' - } - } - 'HADOOP_DIRSTRUCT_HDFS' { - name = 'HADOOP_DIRSTRUCT_HDFS' - type = 'dirstruct' - arguments { - envcmd = 'hdfs envvars' - baseDirEnv = 'HADOOP_HDFS_HOME' - referenceList = 'hadoop-hdfs.list' - } - } - 'HADOOP_DIRSTRUCT_MAPRED' { - name = 'HADOOP_DIRSTRUCT_MAPRED' - type = 'dirstruct' - arguments { - envcmd = 'mapred envvars' - baseDirEnv = 'HADOOP_MAPRED_HOME' - referenceList = 'hadoop-mapreduce.list' - } - } - 'HADOOP_DIRSTRUCT_YARN' { - name = 'HADOOP_DIRSTRUCT_YARN' - type = 'dirstruct' - arguments { - envcmd = 'yarn envvars' - baseDirEnv = 'HADOOP_YARN_HOME' - referenceList = 'hadoop-yarn.list' - } - } - 'HADOOP_SUBPROJS' { - name = 'HADOOP_SUBPROJS' - type = 'dirstruct' - arguments { - envcmd = 'hadoop envvars' - baseDirEnv = 'HADOOP_COMMON_HOME' - referenceList = 'hadoop-subprojs.list' - } - } - 'HADOOP_BINCONTENT_COMMON' { - name = 'HADOOP_BINCONTENT_COMMON' - type = 'dirstruct' - arguments { - envcmd = 'hadoop envvars' - baseDirEnv = 'HADOOP_COMMON_HOME' - subDir = 'bin' - referenceList = 'hadoop-common-bin.list' - } - } - 'HADOOP_BINCONTENT_HDFS' { - name = 'HADOOP_BINCONTENT_HDFS' - type = 'dirstruct' - arguments { - envcmd = 'hdfs envvars' - baseDirEnv = 'HADOOP_HDFS_HOME' - subDir = 'bin' - referenceList = 'hadoop-hdfs-bin.list' - } - } - 'HADOOP_BINCONTENT_MAPRED' { - name = 'HADOOP_BINCONTENT_MAPRED' - type = 'dirstruct' - arguments { - envcmd = 'mapred envvars' - baseDirEnv = 'HADOOP_MAPRED_HOME' - subDir = 'bin' - referenceList = 'hadoop-mapreduce-bin.list' - } - } - 'HADOOP_BINCONTENT_YARN' { - name = 'HADOOP_BINCONTENT_YARN' - type = 'dirstruct' - arguments { - envcmd = 'yarn envvars' - baseDirEnv = 'HADOOP_YARN_HOME' - subDir = 'bin' - referenceList = 'hadoop-yarn-bin.list' - } - } - 'HADOOP_LIBJARSCONTENT_COMMON' { - name = 'HADOOP_JARCONTENT_COMMON' - type = 'dirstruct' - arguments { - envcmd = 'hadoop envvars' - baseDirEnv = 'HADOOP_COMMON_HOME' - subDirEnv = 'HADOOP_COMMON_LIB_JARS_DIR' - referenceList = 'hadoop-common-jar.list' - } - } - 'HADOOP_LIBJARSCONTENT_HDFS' { - name = 'HADOOP_JARCONTENT_HDFS' - type = 'dirstruct' - arguments { - envcmd = 'hdfs envvars' - baseDirEnv = 'HADOOP_HDFS_HOME' - subDirEnv = 'HDFS_LIB_JARS_DIR' - referenceList = 'hadoop-hdfs-jar.list' - } - } - 'HADOOP_LIBJARSCONTENT_MAPRED' { - name = 'HADOOP_JARCONTENT_MAPRED' - type = 'dirstruct' - arguments { - envcmd = 'mapred envvars' - baseDirEnv = 'HADOOP_MAPRED_HOME' - subDirEnv = 'MAPRED_LIB_JARS_DIR' - referenceList = 'hadoop-mapreduce-jar.list' - } - } - 'HADOOP_LIBJARSCONTENT_YARN' { - name = 'HADOOP_JARCONTENT_YARN' - type = 'dirstruct' - arguments { - envcmd = 'yarn envvars' - baseDirEnv = 'HADOOP_YARN_HOME' - subDirEnv = 'YARN_LIB_JARS_DIR' - referenceList = 'hadoop-yarn-jar.list' - } - } - 'HADOOP_GETCONF' { - name = 'HADOOP_GETCONF' - type = 'shell' - arguments { - command = '[ `hdfs getconf -confKey dfs.permissions.superusergroup >/dev/null 2>/dev/null; echo $?` == "0" ]' - message = 'It\' not possible to to determine key Hadoop configuration values by using ${HADOOP_HDFS_HOME}/bin/hdfs getconf' - } - } - 'HADOOP_CNATIVE1' { - name = 'HADOOP_CNATIVE1' - type = 'shell' - arguments { - command = 'hadoop checknative -a 2>/dev/null | grep hadoop | grep true' - message = 'hadoop-common-project must be build with -Pnative or -Pnative-win' - } - } - 'HADOOP_CNATIVE2' { - name = 'HADOOP_CNATIVE2' - type = 'shell' - arguments { - command = 'hadoop checknative -a 2>/dev/null | grep snappy | grep true' - message = 'hadoop-common-project must be build with -Prequire.snappy' - } - } - 'HADOOP_HNATIVE1' { - name = 'HADOOP_HNATIVE1' - type = 'shell' - arguments { - command = '[ ! -n ${HADOOP_COMMON_HOME} ] || HADOOP_COMMON_HOME=`hadoop envvars | grep HADOOP_COMMON_HOME | sed "s/.*=\'\\(.*\\)\'/\\1/"`; '+ - 'test -e $HADOOP_COMMON_HOME/lib/native/libhdfs.a' - message = 'hadoop-hdfs-project must be build with -Pnative or -Pnative-win' - } - } - 'HADOOP_YNATIVE1' { - name = 'HADOOP_YNATIVE1' - type = 'shell' - arguments { - command = '[ ! -n ${HADOOP_YARN_HOME} ] || HADOOP_YARN_HOME=`yarn envvars | grep HADOOP_YARN_HOME | sed "s/.*=\'\\(.*\\)\'/\\1/"`; '+ - 'echo $HADOOP_YARN_HOME; test -e $HADOOP_YARN_HOME/bin/container-executor' - message = 'hadoop-yarn-project must be build with -Pnative or -Pnative-win' - } - } - 'HADOOP_MNATIVE1' { - name = 'HADOOP_MNATIVE1' - type = 'shell' - arguments { - command = 'hadoop checknative -a 2>/dev/null | grep snappy | grep true' - message = 'hadoop-mapreduce-project must be build with -Prequire.snappy' - } - } - 'HADOOP_COMPRESSION' { - name = 'HADOOP_COMPRESSION' - type = 'shell' - arguments { - command = '[[ "$(hadoop checknative -a 2>/dev/null | egrep -e ^zlib -e ^snappy | sort -u | grep true | wc -l)" == 2 ]]' - message = 'hadoop must be built with -Dcompile.native=true' - } - } - 'HADOOP_TOOLS' { - name = 'HADOOP_TOOLS' - type = 'hadoop_tools' - arguments { - } - } - 'HADOOP_API1' { - name = "HADOOP_API1" - type = 'api_examination' - arguments { - baseDirEnv = 'HADOOP_COMMON_HOME' - libDir = 'HADOOP_COMMON_DIR' - envcmd = 'hadoop envvars' - jar = 'hadoop-common' - resourceFile = 'hadoop-common-2.7.3-api-report.json' - } - } - 'HADOOP_API2' { - name = "HADOOP_API2" - type = 'api_examination' - arguments { - baseDirEnv = 'HADOOP_HDFS_HOME' - libDir = 'HDFS_DIR' - envcmd = 'hdfs envvars' - jar = 'hadoop-hdfs' - resourceFile = 'hadoop-hdfs-2.7.3-api-report.json' - } - } - 'HADOOP_API3' { - name = "HADOOP_API3" - type = 'api_examination' - arguments { - baseDirEnv = 'HADOOP_YARN_HOME' - libDir = 'YARN_DIR' - envcmd = 'yarn envvars' - jar = 'hadoop-yarn-common' - resourceFile = 'hadoop-yarn-common-2.7.3-api-report.json' - } - } - 'HADOOP_API4' { - name = "HADOOP_API4" - type = 'api_examination' - arguments { - baseDirEnv = 'HADOOP_YARN_HOME' - libDir = 'YARN_DIR' - envcmd = 'yarn envvars' - jar = 'hadoop-yarn-client' - resourceFile = 'hadoop-yarn-client-2.7.3-api-report.json' - } - } - 'HADOOP_API5' { - name = "HADOOP_API5" - type = 'api_examination' - arguments { - baseDirEnv = 'HADOOP_YARN_HOME' - libDir = 'YARN_DIR' - envcmd = 'yarn envvars' - jar = 'hadoop-yarn-api' - resourceFile = 'hadoop-yarn-api-2.7.3-api-report.json' - } - } - 'HADOOP_API6' { - name = "HADOOP_API6" - type = 'api_examination' - arguments { - baseDirEnv = 'HADOOP_MAPRED_HOME' - libDir = 'MAPRED_DIR' - envcmd = 'mapred envvars' - jar = 'hadoop-mapreduce-client-core' - resourceFile = 'hadoop-mapreduce-client-core-2.7.3-api-report.json' - } - } - } -} http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/build.gradle ---------------------------------------------------------------------- diff --git a/build.gradle b/build.gradle index 101ecc7..b0a83b1 100644 --- a/build.gradle +++ b/build.gradle @@ -80,6 +80,8 @@ rat { "bigtop-tests/smoke-tests/phoenix/*.sql", "bigtop-tests/smoke-tests/ignite-hadoop/*.data", "bigtop-tests/smoke-tests/tajo/table1/*.csv", + "bigtop-tests/smoke-tests/odpi-runtime/src/test/resources/*.json", + "bigtop-tests/smoke-tests/odpi-runtime/src/test/resources/*.list", "**/target/**", "**/build/**", "**/.gradle/**", @@ -427,7 +429,6 @@ task "bigtop-slaves"(type:Exec, project.afterEvaluate { checkClusterTestProjects("smoke.tests") - checkClusterTestProjects("spec.tests") artifactToInstall(dependsOn: [installTopLevel, installCommon, installConf, installiTest]) } http://git-wip-us.apache.org/repos/asf/bigtop/blob/5e342c45/settings.gradle ---------------------------------------------------------------------- diff --git a/settings.gradle b/settings.gradle index 7ae4bf5..ac071f1 100644 --- a/settings.gradle +++ b/settings.gradle @@ -31,9 +31,4 @@ fTree.each() { buildFile -> def parent = buildFile.getParentFile().name include("bigtop-tests:smoke-tests:$parent") } -fTree = fileTree(dir: 'bigtop-tests/spec-tests', include: '*/build.gradle') -fTree.each() { buildFile -> - def parent = buildFile.getParentFile().name - include("bigtop-tests:spec-tests:$parent") -}
