[jira] [Created] (HADOOP-16147) Allow CopyListing sequence file keys and values to be more easily customized

2019-02-25 Thread Andrew Olson (JIRA)
Andrew Olson created HADOOP-16147:
-

 Summary: Allow CopyListing sequence file keys and values to be 
more easily customized
 Key: HADOOP-16147
 URL: https://issues.apache.org/jira/browse/HADOOP-16147
 Project: Hadoop Common
  Issue Type: Improvement
  Components: tools/distcp
Reporter: Andrew Olson


We have encountered a scenario where, when using the Crunch library to run a 
distributed copy (CRUNCH-660, CRUNCH-675) at the conclusion of a job we need to 
dynamically rename target paths to the preferred destination output part file 
names, rather than retaining the original source path names.

A custom CopyListing implementation appears to be the proper solution for this. 
However the place where the current SimpleCopyListing logic needs to be 
adjusted is in a private method (writeToFileListing), so a relatively large 
portion of the class would need to be cloned.

To minimize the amount of code duplication required for such a custom 
implementation, we propose adding two new protected methods to the CopyListing 
class, that can be used to change the actual keys and/or values written to the 
copy listing sequence file: 

{noformat}
protected Text getFileListingKey(Path sourcePathRoot, CopyListingFileStatus 
fileStatus);

protected CopyListingFileStatus getFileListingValue(CopyListingFileStatus 
fileStatus);
{noformat}

The SimpleCopyListing class would then be modified to consume these methods as 
follows,
{noformat}
fileListWriter.append(
   getFileListingKey(sourcePathRoot, fileStatus),
   getFileListingValue(fileStatus));
{noformat}

The default implementations would simply preserve the present behavior of the 
SimpleCopyListing class, and could reside in either CopyListing or 
SimpleCopyListing, whichever is preferable.

{noformat}
protected Text getFileListingKey(Path sourcePathRoot, CopyListingFileStatus 
fileStatus) {
   return new Text(DistCpUtils.getRelativePath(sourcePathRoot, 
fileStatus.getPath()));
}

protected CopyListingFileStatus getFileListingValue(CopyListingFileStatus 
fileStatus) {
   return fileStatus;
}
{noformat}

Please let me know if this proposal seems to be on the right track. If so I can 
provide a patch.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2019-02-25 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/

[Feb 24, 2019 4:31:39 AM] (aengineer) HDDS-1166. Fix checkstyle line length 
issues. Contributed by Nandakumar.
[Feb 24, 2019 4:45:01 AM] (aengineer) HDDS-1167. Error in 
hadoop-ozone/dev-support/checks/checkstyle.sh.
[Feb 24, 2019 4:56:24 AM] (aengineer) HDDS-1147. Fix failing unit tests in 
TestOzoneManager. Contributed by
[Feb 24, 2019 5:13:47 AM] (aengineer) HDDS-1155.Fix failing unit test methods 
of TestDeadNodeHandler.
[Feb 24, 2019 5:37:55 AM] (github) HDDS-1120. Add a config to disable checksum 
verification during read.
[Feb 24, 2019 9:07:06 AM] (sunilg) YARN-8821. [YARN-8851] GPU 
hierarchy/topology scheduling support based
[Feb 24, 2019 2:17:29 PM] (wwei) YARN-9300. Lazy preemption should trigger an 
update on queue preemption
[Feb 24, 2019 2:42:27 PM] (wwei) YARN-9316. 
TestPlacementConstraintsUtil#testInterAppConstraintsByAppID
[Feb 24, 2019 8:36:03 PM] (aengineer) HDDS-1168.Use random ports in 
TestBlockManager and TestDeletedBlockLog.
[Feb 24, 2019 9:13:37 PM] (aengineer) HDDS-1154. Add tracing to the client side 
of
[Feb 24, 2019 9:22:36 PM] (aengineer) HDDS-1149. Change the default 
ozone.client.checksum.type. Contributed by
[Feb 24, 2019 9:53:38 PM] (aengineer) HDDS-1089. Disable 
OzoneFSStorageStatistics for hadoop versions older
[Feb 24, 2019 10:02:30 PM] (aengineer) HDDS-1090. Add a recipe to the ozone 
docs about using Spark, OzoneFS,
[Feb 24, 2019 10:40:52 PM] (aengineer) HDDS-1115. Provide ozone specific 
top-level pom.xml. Contributed by
[Feb 24, 2019 11:04:25 PM] (aengineer) HDDS-1145. Add optional web server to 
the Ozone freon test tool.
[Feb 24, 2019 11:20:34 PM] (aengineer) HDDS-1152. Add trace information for the 
client side of the datanode
[Feb 24, 2019 11:39:41 PM] (aengineer) HDDS-1070. Adding Node and Pipeline 
related metrics in SCM. Contributed




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.util.TestDiskCheckerWithDiskIo 
   hadoop.hdfs.server.datanode.TestDataNodeHotSwapVolumes 
   hadoop.hdfs.server.namenode.ha.TestHASafeMode 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.hdfs.server.namenode.ha.TestBootstrapAliasmap 
   hadoop.hdfs.qjournal.server.TestJournalNodeSync 
   hadoop.fs.viewfs.TestViewFileSystemLinkMergeSlash 
   hadoop.hdfs.server.namenode.TestNamenodeCapacityReport 
   hadoop.fs.http.client.TestHttpFSFWithSWebhdfsFileSystem 
   hadoop.mapred.TestJobCounters 
   hadoop.yarn.sls.TestSLSGenericSynth 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-compile-javac-root.txt
  [336K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-checkstyle-root.txt
  [17M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-patch-hadolint.txt
  [8.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-patch-pylint.txt
  [144K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-patch-shellcheck.txt
  [20K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/diff-patch-shelldocs.txt
  [12K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/whitespace-eol.txt
  [9.6M]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/whitespace-tabs.txt
  [1.1M]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/branch-findbugs-hadoop-hdds_client.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/branch-findbugs-hadoop-hdds_container-service.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/branch-findbugs-hadoop-hdds_framework.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/branch-findbugs-hadoop-hdds_server-scm.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1058/artifact/out/branch-findbugs-hadoop-hdds_tools.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trun

Apache Hadoop qbt Report: branch2+JDK7 on Linux/x86

2019-02-25 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/

No changes




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

XML :

   Parsing Error(s): 
   
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml
 
   hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml
 

FindBugs :

   module:hadoop-common-project/hadoop-common 
   Class org.apache.hadoop.fs.GlobalStorageStatistics defines non-transient 
non-serializable instance field map In GlobalStorageStatistics.java:instance 
field map In GlobalStorageStatistics.java 

FindBugs :

   module:hadoop-hdfs-project/hadoop-hdfs 
   Dead store to state in 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(OutputStream,
 INodeSymlink) At 
FSImageFormatPBINode.java:org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(OutputStream,
 INodeSymlink) At FSImageFormatPBINode.java:[line 623] 

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client
 
   Boxed value is unboxed and then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:[line 335] 

Failed junit tests :

   hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys 
   hadoop.hdfs.TestRollingUpgrade 
   hadoop.hdfs.TestEncryptionZonesWithKMS 
   hadoop.registry.secure.TestSecureLogins 
   hadoop.yarn.server.nodemanager.amrmproxy.TestFederationInterceptor 
   hadoop.yarn.server.TestContainerManagerSecurity 
   hadoop.yarn.server.timelineservice.security.TestTimelineAuthFilterForV2 
   hadoop.yarn.client.api.impl.TestAMRMProxy 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-compile-cc-root-jdk1.7.0_95.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-compile-javac-root-jdk1.7.0_95.txt
  [328K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-compile-cc-root-jdk1.8.0_191.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-compile-javac-root-jdk1.8.0_191.txt
  [308K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-checkstyle-root.txt
  [16M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-patch-hadolint.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-patch-pylint.txt
  [24K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-patch-shellcheck.txt
  [72K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/diff-patch-shelldocs.txt
  [8.0K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/whitespace-eol.txt
  [12M]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/whitespace-tabs.txt
  [1.2M]

   xml:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/xml.txt
  [8.0K]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/branch-findbugs-hadoop-hdfs-project_hadoop-hdfs-warnings.html
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/243/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timelineservice-hbase_hadoop-yarn-server-timelineservice-hbase-client-warnings.html
  [8.0K]

   javadoc:

   

[jira] [Created] (HADOOP-16146) Make start-build-env.sh safe in case of misusage of DOCKER_INTERACTIVE_RUN

2019-02-25 Thread Elek, Marton (JIRA)
Elek, Marton created HADOOP-16146:
-

 Summary: Make start-build-env.sh safe in case of misusage of 
DOCKER_INTERACTIVE_RUN
 Key: HADOOP-16146
 URL: https://issues.apache.org/jira/browse/HADOOP-16146
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Elek, Marton
Assignee: Elek, Marton


[~aw] reported the problem in HDDS-891:
{quote}DOCKER_INTERACTIVE_RUN opens the door for users to set command line 
options to docker. Most notably, -c and -v and a few others that share one 
particular characteristic: they reference the file system. As soon as shell 
code hits the file system, it is no longer safe to assume space delimited 
options. In other words, -c /My Cool Filesystem/Docker Files/config.json or -v 
/c_drive/Program Files/Data:/data may be something a user wants to do, but the 
script now breaks because of the IFS assumptions.
{quote}
DOCKER_INTERACTIVE_RUN was used in jenkins to run normal build process in 
docker. In case of DOCKER_INTERACTIVE_RUN was set to empty the docker container 
is started without the "-i -t" flags.

It can be improved by checking the value of the environment variable and enable 
only fixed set of values.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org