[jira] [Commented] (HADOOP-15696) KMS performance regression due to too many open file descriptors after Jetty migration

2018-08-27 Thread Misha Dmitriev (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594364#comment-16594364
 ] 

Misha Dmitriev commented on HADOOP-15696:
-

Answering my own question: yes, it looks like indeed the KMS client opens a new 
HTTP connection for every request, and doesn't seem to close it explicitly. See 
methods such in {{KMSClientProvider}} , e.g. {{decryptEncryptedKey()}}, 
{{getCurrentKey()}}, {{deleteKey()}} etc. They all invoke 
{{createConnection()}}, then {{call()}} with this connection to send a request 
and get a response, and then I don't see any code that would close that 
connection.

It's hard for me to say whether it's by design or by neglect, but anyway, such 
a frequent creation of TCP connection and no explicit close/reuse look unusual. 
At a minimum, I think it's worth documenting this behavior briefly somewhere, 
maybe in the top-level javadoc of the {{KMSClientProvider}} class. I suggest to 
do it in the above patch so that this is not left behind.

 

> KMS performance regression due to too many open file descriptors after Jetty 
> migration
> --
>
> Key: HADOOP-15696
> URL: https://issues.apache.org/jira/browse/HADOOP-15696
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.0.0-alpha2
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Attachments: HADOOP-15696.001.patch, Screen Shot 2018-08-22 at 
> 11.36.16 AM.png, Screen Shot 2018-08-22 at 4.26.51 PM.png, Screen Shot 
> 2018-08-22 at 4.26.51 PM.png, Screen Shot 2018-08-22 at 4.27.02 PM.png, 
> Screen Shot 2018-08-22 at 4.30.32 PM.png, Screen Shot 2018-08-22 at 4.30.39 
> PM.png, Screen Shot 2018-08-24 at 7.08.16 PM.png
>
>
> We recently found KMS performance regressed in Hadoop 3.0, possibly linking 
> to the migration from Tomcat to Jetty in HADOOP-13597.
> Symptoms:
> # Hadoop 3.x KMS open file descriptors quickly rises to more than 10 thousand 
> under stress, sometimes even exceeds 32K, which is the system limit, causing 
> failures for any access to encryption zones. Our internal testing shows the 
> openfd number was in the range of a few hundred in Hadoop 2.x, and it 
> increases by almost 100x in Hadoop 3.
> # Hadoop 3.x KMS as much as twice the heap size than in Hadoop 2.x. The same 
> heap size can go OOM in Hadoop 3.x. Jxray analysis suggests most of them are 
> temporary byte arrays associated with open SSL connections.
> # Due to the heap usage, Hadoop 3.x KMS has more frequent GC activities, and 
> we observed up to 20% performance reduction due to GC.
> A possible solution is to reduce the idle timeout setting in HttpServer2. It 
> is currently hard-coded 10 seconds. By setting it to 1 second, open fds 
> dropped from 20 thousand down to 3 thousand in my experiment.
> File this jira to invite open discussion for a solution.
> Credit: [~mi...@cloudera.com] for the proposed Jetty idle timeout remedy; 
> [~xiaochen] for digging into this problem.
> Screenshots:
> CDH5 (Hadoop 2) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.39 PM.png! 
> CDH6 (Hadoop 3) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.32 PM.png! 
> CDH5 (Hadoop 2) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.26.51 PM.png! 
> CDH6 (Hadoop 3) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.27.02 PM.png! 
> JXray report
>  !Screen Shot 2018-08-22 at 11.36.16 AM.png! 
> open fd drops from 20 k down to 3k after the proposed change.
>  !Screen Shot 2018-08-24 at 7.08.16 PM.png! 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15696) KMS performance regression due to too many open file descriptors after Jetty migration

2018-08-27 Thread Wei-Chiu Chuang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15696?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang updated HADOOP-15696:
-
Attachment: HADOOP-15696.001.patch

> KMS performance regression due to too many open file descriptors after Jetty 
> migration
> --
>
> Key: HADOOP-15696
> URL: https://issues.apache.org/jira/browse/HADOOP-15696
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.0.0-alpha2
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Attachments: HADOOP-15696.001.patch, Screen Shot 2018-08-22 at 
> 11.36.16 AM.png, Screen Shot 2018-08-22 at 4.26.51 PM.png, Screen Shot 
> 2018-08-22 at 4.26.51 PM.png, Screen Shot 2018-08-22 at 4.27.02 PM.png, 
> Screen Shot 2018-08-22 at 4.30.32 PM.png, Screen Shot 2018-08-22 at 4.30.39 
> PM.png, Screen Shot 2018-08-24 at 7.08.16 PM.png
>
>
> We recently found KMS performance regressed in Hadoop 3.0, possibly linking 
> to the migration from Tomcat to Jetty in HADOOP-13597.
> Symptoms:
> # Hadoop 3.x KMS open file descriptors quickly rises to more than 10 thousand 
> under stress, sometimes even exceeds 32K, which is the system limit, causing 
> failures for any access to encryption zones. Our internal testing shows the 
> openfd number was in the range of a few hundred in Hadoop 2.x, and it 
> increases by almost 100x in Hadoop 3.
> # Hadoop 3.x KMS as much as twice the heap size than in Hadoop 2.x. The same 
> heap size can go OOM in Hadoop 3.x. Jxray analysis suggests most of them are 
> temporary byte arrays associated with open SSL connections.
> # Due to the heap usage, Hadoop 3.x KMS has more frequent GC activities, and 
> we observed up to 20% performance reduction due to GC.
> A possible solution is to reduce the idle timeout setting in HttpServer2. It 
> is currently hard-coded 10 seconds. By setting it to 1 second, open fds 
> dropped from 20 thousand down to 3 thousand in my experiment.
> File this jira to invite open discussion for a solution.
> Credit: [~mi...@cloudera.com] for the proposed Jetty idle timeout remedy; 
> [~xiaochen] for digging into this problem.
> Screenshots:
> CDH5 (Hadoop 2) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.39 PM.png! 
> CDH6 (Hadoop 3) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.32 PM.png! 
> CDH5 (Hadoop 2) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.26.51 PM.png! 
> CDH6 (Hadoop 3) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.27.02 PM.png! 
> JXray report
>  !Screen Shot 2018-08-22 at 11.36.16 AM.png! 
> open fd drops from 20 k down to 3k after the proposed change.
>  !Screen Shot 2018-08-24 at 7.08.16 PM.png! 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15696) KMS performance regression due to too many open file descriptors after Jetty migration

2018-08-27 Thread Misha Dmitriev (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594339#comment-16594339
 ] 

Misha Dmitriev commented on HADOOP-15696:
-

To be precise, the suggested measure, that had such a big effect, was to adjust 
the timeout in the {{HttpServer2.configureChannelConnector(ServerConnector c)}} 
method. Currently it has the {{c.setIdleTimeout(1);}} line. This timeout 
should be made configurable in the first place, and looks like we need to 
adjust it to a (much) smaller value when {{HttpServer2}} is used by KMS.

Here is a question that I have in this regard. If by closing HTTP connections 
on the server side, and thus recycling more quickly, we make KMS work better - 
does it mean that the KMS client doesn't reuse any such connections, and/or 
doesn't close a connection when it doesn't need it anymore? If so, it doesn't 
sound very optimal. I wonder how to prove or disprove this theory.

> KMS performance regression due to too many open file descriptors after Jetty 
> migration
> --
>
> Key: HADOOP-15696
> URL: https://issues.apache.org/jira/browse/HADOOP-15696
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.0.0-alpha2
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Attachments: Screen Shot 2018-08-22 at 11.36.16 AM.png, Screen Shot 
> 2018-08-22 at 4.26.51 PM.png, Screen Shot 2018-08-22 at 4.26.51 PM.png, 
> Screen Shot 2018-08-22 at 4.27.02 PM.png, Screen Shot 2018-08-22 at 4.30.32 
> PM.png, Screen Shot 2018-08-22 at 4.30.39 PM.png, Screen Shot 2018-08-24 at 
> 7.08.16 PM.png
>
>
> We recently found KMS performance regressed in Hadoop 3.0, possibly linking 
> to the migration from Tomcat to Jetty in HADOOP-13597.
> Symptoms:
> # Hadoop 3.x KMS open file descriptors quickly rises to more than 10 thousand 
> under stress, sometimes even exceeds 32K, which is the system limit, causing 
> failures for any access to encryption zones. Our internal testing shows the 
> openfd number was in the range of a few hundred in Hadoop 2.x, and it 
> increases by almost 100x in Hadoop 3.
> # Hadoop 3.x KMS as much as twice the heap size than in Hadoop 2.x. The same 
> heap size can go OOM in Hadoop 3.x. Jxray analysis suggests most of them are 
> temporary byte arrays associated with open SSL connections.
> # Due to the heap usage, Hadoop 3.x KMS has more frequent GC activities, and 
> we observed up to 20% performance reduction due to GC.
> A possible solution is to reduce the idle timeout setting in HttpServer2. It 
> is currently hard-coded 10 seconds. By setting it to 1 second, open fds 
> dropped from 20 thousand down to 3 thousand in my experiment.
> File this jira to invite open discussion for a solution.
> Credit: [~mi...@cloudera.com] for the proposed Jetty idle timeout remedy; 
> [~xiaochen] for digging into this problem.
> Screenshots:
> CDH5 (Hadoop 2) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.39 PM.png! 
> CDH6 (Hadoop 3) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.32 PM.png! 
> CDH5 (Hadoop 2) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.26.51 PM.png! 
> CDH6 (Hadoop 3) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.27.02 PM.png! 
> JXray report
>  !Screen Shot 2018-08-22 at 11.36.16 AM.png! 
> open fd drops from 20 k down to 3k after the proposed change.
>  !Screen Shot 2018-08-24 at 7.08.16 PM.png! 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15633) fs.TrashPolicyDefault: Can't create trash directory

2018-08-27 Thread John Zhuge (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15633?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594308#comment-16594308
 ] 

John Zhuge commented on HADOOP-15633:
-

Committed to branch-3.1 and branch-3.0.

> fs.TrashPolicyDefault: Can't create trash directory
> ---
>
> Key: HADOOP-15633
> URL: https://issues.apache.org/jira/browse/HADOOP-15633
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
>Affects Versions: 2.8.3, 3.1.0, 3.0.3, 2.7.7
>Reporter: Fei Hui
>Assignee: Fei Hui
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HADOOP-15633.001.patch, HADOOP-15633.002.patch, 
> HADOOP-15633.003.patch, HADOOP-15633.004.patch, HADOOP-15633.005.patch
>
>
> Reproduce it as follow
> {code:java}
> hadoop fs -mkdir /user/hadoop/aaa
> hadoop fs -touchz /user/hadoop/aaa/bbb
> hadoop fs -rm /user/hadoop/aaa/bbb
> hadoop fs -mkdir /user/hadoop/aaa/bbb
> hadoop fs -touchz /user/hadoop/aaa/bbb/ccc
> hadoop fs -rm /user/hadoop/aaa/bbb/ccc
> {code}
> Then we get errors 
> {code:java}
> 18/07/26 17:55:24 WARN fs.TrashPolicyDefault: Can't create trash directory: 
> hdfs://xxx/user/hadoop/.Trash/Current/user/hadoop/aaa/bbb
> org.apache.hadoop.fs.FileAlreadyExistsException: Path is not a directory: 
> /user/hadoop/.Trash/Current/user/hadoop/aaa/bbb
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:65)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3961)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2115)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2111)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1867)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2111)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> at 
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
> at 
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
> at 
> org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3002)
> at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2970)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1047)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1043)
> at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1061)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1036)
> at 
> org.apache.hadoop.fs.TrashPolicyDefault.moveToTrash(TrashPolicyDefault.java:136)
> at org.apache.hadoop.fs.Trash.moveToTrash(Trash.java:114)
> at org.apache.hadoop.fs.Trash.moveToAppropriateTrash(Trash.java:95)
> at org.apache.hadoop.fs.shell.Delete$Rm.moveToTrash(Delete.java:118)
> at org.apache.hadoop.fs.shell.Delete$Rm.processPath(Delete.java:105)
> at org.apache.hadoop.fs.shell.Command.processPaths(Command.java:317)
> at 
> org.apache.hadoop.fs.shell.Command.processPathArgument(Command.java:289)
> at 
> org.apache.hadoop.fs.shell.Command.processArgument(Command.java:271)
> at 
> org.apache.hadoop.fs.shell.Command.processArguments(Command.java:255)
> at 
> org.apache.hadoop.fs.shell.Command.processRawArguments(Command.java:201)
> at org.apache.hadoop.fs.shell.Command.run(Command.java:165)
> at org.apache.hadoop.fs.FsShell.run(FsShe

[jira] [Commented] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594182#comment-16594182
 ] 

Hudson commented on HADOOP-15699:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14841 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14841/])
HADOOP-15699. Fix some of testContainerManager failures in Windows. (gifuma: 
rev 602d13844a8d4c7b08ce185da01fde098ff8b9a6)
* (edit) 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/TestContainerManager.java


> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Fix For: 3.2.0, 2.9.2
>
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread genericqa (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594178#comment-16594178
 ] 

genericqa commented on HADOOP-15691:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
12s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:red}-1{color} | {color:red} test4tests {color} | {color:red}  0m  
0s{color} | {color:red} The patch doesn't appear to include any new or modified 
tests. Please justify why no new tests are needed for this patch. Also please 
list what manual steps were performed to verify this patch. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
20s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m 
48s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 14m 
53s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  3m 
19s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  4m 
42s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
19m 40s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  6m  
7s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  3m 
34s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
21s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  3m 
28s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 16m 
23s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 16m 
23s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
3m 20s{color} | {color:orange} root: The patch generated 2 new + 454 unchanged 
- 0 fixed = 456 total (was 454) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  4m 
36s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 1 line(s) that end in whitespace. Use git 
apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply 
{color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 25s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  6m 
48s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} javadoc {color} | {color:red}  0m 
58s{color} | {color:red} hadoop-common-project_hadoop-common generated 2 new + 
0 unchanged - 0 fixed = 2 total (was 0) {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:red}-1{color} | {color:red} unit {color} | {color:red}  8m  2s{color} 
| {color:red} hadoop-common in the patch failed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
44s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  3m 
54s{color} | {color:green} hadoop-hdfs-httpfs in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  4m 
40s{color} | {color:green} hadoop-aws in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  1m 
17s{color} | {color:green} hadoop-azure in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  0m 
54s{color} | {color:green} hadoop-azure-datalake in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
39s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {co

[jira] [Commented] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Botong Huang (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594153#comment-16594153
 ] 

Botong Huang commented on HADOOP-15699:
---

Thanks [~giovanni.fumarola] for the quick review and commit!

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Fix For: 3.2.0, 2.9.2
>
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Giovanni Matteo Fumarola (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594150#comment-16594150
 ] 

Giovanni Matteo Fumarola commented on HADOOP-15699:
---

Thanks [~botong] . Committed to trunk, branch-2, branch-2.9.

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Fix For: 3.2.0, 2.9.2
>
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Giovanni Matteo Fumarola (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni Matteo Fumarola updated HADOOP-15699:
--
Resolution: Fixed
Status: Resolved  (was: Patch Available)

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Fix For: 3.2.0, 2.9.2
>
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Giovanni Matteo Fumarola (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Giovanni Matteo Fumarola updated HADOOP-15699:
--
Fix Version/s: 2.9.2
   3.2.0

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Fix For: 3.2.0, 2.9.2
>
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Botong Huang (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594133#comment-16594133
 ] 

Botong Huang commented on HADOOP-15699:
---

The failures before the patch is here 
[https://builds.apache.org/job/hadoop-trunk-win/533/testReport/]

With the patch they are passing in my local Windows box. 

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15669) ABFS: Improve HTTPS Performance

2018-08-27 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15669?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594128#comment-16594128
 ] 

Steve Loughran commented on HADOOP-15669:
-

thanks, just caught up with this after being on vacation briefly.

FWIW, I Think the user agent should include the values from 
org.apache.hadoop.util.VersionInfo.getBuildVersion(), as that shows which 
hadoop build is at work here. Handy if you have a large log of HTTP requests 
and are trying to work out which client  appis causing all your problems

> ABFS: Improve HTTPS Performance
> ---
>
> Key: HADOOP-15669
> URL: https://issues.apache.org/jira/browse/HADOOP-15669
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Reporter: Thomas Marquardt
>Assignee: Vishwajeet Dusane
>Priority: Major
> Attachments: ABFS - Improve HTTPS Performance Over Java Based 
> Client.pdf, HADOOP-15669-HADOOP-15407-01.patch, 
> HADOOP-15669-HADOOP-15407-02.patch, HADOOP-15669-HADOOP-15407-03.patch, 
> HADOOP-15669-HADOOP-15407-04.patch
>
>
> We see approximately 50% worse throughput for ABFS over HTTPs vs HTTP.  Lets 
> perform a detailed measurement and see what can be done to improve throughput.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15692) ABFS: extensible support for custom oauth

2018-08-27 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15692?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594121#comment-16594121
 ] 

Steve Loughran commented on HADOOP-15692:
-

I'm busy learning about DTs myself (HADOOP-14556) so am now roughly in a 
position to review this —at the same time, I'm still learning, so it's good to 
see your design and try to pick up ideas. What I'm thinking of implementing as 
pluging mechanism is similar to yours, though you are ahead of me in 
implementation.

h2. General

Some general questions about overall design.

* Is the expectation that there will be one token provider for every client 
configuration? As I've been looking at for S3A having per-fs configs here.
* related: is there ever going to be the need to have >1 link to the same FS 
URI for the same client? Daryn's patch for S3A has support for that, which I've 
been adopting, even though I'm not sure how I'd imagine using.

I'd like to propose making CustomDelegationTokenManager class a subclass of 
AbstractService, using its initialize/start/stop. AbfsDelegationTokenManager 
can then subclass CompositeService or simply invoke the lifecycle events of its 
tokenManager field as appropriate, as can AzureBlobFileSystem. This is what I 
intend to do with my s3a plugin. Why do this? 

* Gives everything a unified lifecycle, *including an operation to stop all 
threads running*
* Designed for easy subclassing; stable and robust lifecycle implementation.

If someone wanted to implement a plugin which did  like the restricted roles of 
AWS, where when a DT was requested a token which was returned which offered 
only access to the specific container, would this plugin design support it in 
future? The use case I'm looking at is: shared Hive cluster with user requests 
including DTs granting access only to the given containers. If the Oauth tokens 
don't offer such restrictions, I'm not so worried about this, but do consider 
this use rather than just "shared Oauth pool service".

h2. Specific


h3. AbfsConfiguration

* can you keep imports in alphabetical order; the renamed 
CustomTokenProviderAdapter import needs to move
* You can do a static import of FileSystemConfigurations if that would help the 
code readability

h3. AzureBlobFileSystem


L126: add a log entry to log the status of security & DT config; will be useful 
for debugging.

L833. Is the javadoc consistent with the current config option/behaviour (i.e. 
you need two switches for DTs)?
L840: Feel free to use ? : here rather than the if/else clause



h3. CustomDelegationTokenManager

L40: Add the FS URI to the constructor call. Eventually you'll need it, at 
which point all existing implementations will stop linking

L65: maybe have it return false if cancelling isn't supported? Because it's not 
in the same category of error as other IOEs.


h3. CustomTokenProviderAdaptee

I know Adaptee is technically correct, but I keep reading it as a typo for the 
much more common Adapter. 

L74: Recommend `java.time.ZonedDateTime` over `Date`; the latter is too 
ambiguous on a global scale. If I in GMT+1 auth with an Azure service in GMT+2, 
I still need to know the expiry time relative to me.

h3, DelegationTokenIdentifier order

importing: should be 

{code}
java.*
javax.*
--
all non org.apache.
--
org.apache.*
--
statics
{code}

I know, it's rarely changed in existing code, but it's good to get right in the 
new stuff. thanks.

L54: use a ref to the constant, to guarantee consistency/spelling

h3. AbfsTokenRenewer

same comment about import ordering.




> ABFS: extensible support for custom oauth
> -
>
> Key: HADOOP-15692
> URL: https://issues.apache.org/jira/browse/HADOOP-15692
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Reporter: Thomas Marquardt
>Assignee: Da Zhou
>Priority: Major
> Attachments: HADOOP-15692-HADOOP-15407-001.patch
>
>
> ABFS supports oauth in various forms and needs to export interfaces for 
> customization of FileSystem.getDelegationToken and getAccessToken.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread genericqa (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594116#comment-16594116
 ] 

genericqa commented on HADOOP-15699:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
20s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 
17s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  1m  
3s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
27s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
38s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
12m  5s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m  
0s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
27s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  0m 
39s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green}  0m 
59s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  0m 
22s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  0m 
36s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
12m 23s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  1m  
3s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  0m 
23s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 18m 
55s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
28s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 71m  5s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:ba1ab08 |
| JIRA Issue | HADOOP-15699 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12937317/HADOOP-15699.v1.patch 
|
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  findbugs  checkstyle  |
| uname | Linux 95eac0fd8260 3.13.0-144-generic #193-Ubuntu SMP Thu Mar 15 
17:03:53 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | trunk / c9b6395 |
| maven | version: Apache Maven 3.3.9 |
| Default Java | 1.8.0_181 |
| findbugs | v3.1.0-RC1 |
|  Test Results | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15098/testReport/ |
| Max. process+thread count | 302 (vs. ulimit of 1) |
| modules | C: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager
 U: 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager
 |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15098/console |
| Powered by | Apache Yetus 0.9.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Fix some of test

[jira] [Updated] (HADOOP-15633) fs.TrashPolicyDefault: Can't create trash directory

2018-08-27 Thread John Zhuge (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15633?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

John Zhuge updated HADOOP-15633:

   Resolution: Fixed
Fix Version/s: 3.2.0
   Status: Resolved  (was: Patch Available)

Committed to trunk. Tranks [~ferhui] for discovering and fixing the issue!

> fs.TrashPolicyDefault: Can't create trash directory
> ---
>
> Key: HADOOP-15633
> URL: https://issues.apache.org/jira/browse/HADOOP-15633
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
>Affects Versions: 2.8.3, 3.1.0, 3.0.3, 2.7.7
>Reporter: Fei Hui
>Assignee: Fei Hui
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HADOOP-15633.001.patch, HADOOP-15633.002.patch, 
> HADOOP-15633.003.patch, HADOOP-15633.004.patch, HADOOP-15633.005.patch
>
>
> Reproduce it as follow
> {code:java}
> hadoop fs -mkdir /user/hadoop/aaa
> hadoop fs -touchz /user/hadoop/aaa/bbb
> hadoop fs -rm /user/hadoop/aaa/bbb
> hadoop fs -mkdir /user/hadoop/aaa/bbb
> hadoop fs -touchz /user/hadoop/aaa/bbb/ccc
> hadoop fs -rm /user/hadoop/aaa/bbb/ccc
> {code}
> Then we get errors 
> {code:java}
> 18/07/26 17:55:24 WARN fs.TrashPolicyDefault: Can't create trash directory: 
> hdfs://xxx/user/hadoop/.Trash/Current/user/hadoop/aaa/bbb
> org.apache.hadoop.fs.FileAlreadyExistsException: Path is not a directory: 
> /user/hadoop/.Trash/Current/user/hadoop/aaa/bbb
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:65)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3961)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2115)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2111)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1867)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2111)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> at 
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
> at 
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
> at 
> org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3002)
> at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2970)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1047)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1043)
> at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1061)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1036)
> at 
> org.apache.hadoop.fs.TrashPolicyDefault.moveToTrash(TrashPolicyDefault.java:136)
> at org.apache.hadoop.fs.Trash.moveToTrash(Trash.java:114)
> at org.apache.hadoop.fs.Trash.moveToAppropriateTrash(Trash.java:95)
> at org.apache.hadoop.fs.shell.Delete$Rm.moveToTrash(Delete.java:118)
> at org.apache.hadoop.fs.shell.Delete$Rm.processPath(Delete.java:105)
> at org.apache.hadoop.fs.shell.Command.processPaths(Command.java:317)
> at 
> org.apache.hadoop.fs.shell.Command.processPathArgument(Command.java:289)
> at 
> org.apache.hadoop.fs.shell.Command.processArgument(Command.java:271)
> at 
> org.apache.hadoop.fs.shell.Command.processArguments(Command.java:255)
> at 
> org.apache.hadoop.fs.shell.Command.processRawArguments(Command.java:201)
> at org.apache.hadoop.fs.shell.C

[jira] [Comment Edited] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Giovanni Matteo Fumarola (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594061#comment-16594061
 ] 

Giovanni Matteo Fumarola edited comment on HADOOP-15699 at 8/27/18 6:25 PM:


Thanks [~botong] . Can you share the windows execution with and without the 
patch?
e.g. 
https://issues.apache.org/jira/browse/HDFS-13509?focusedCommentId=16457054&page=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16457054
 


was (Author: giovanni.fumarola):
Thanks [~botong] . Can you share the windows execution with and without the 
patch?
e.g. https://issues.apache.org/jira/browse/HDFS-13509 

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Giovanni Matteo Fumarola (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594061#comment-16594061
 ] 

Giovanni Matteo Fumarola edited comment on HADOOP-15699 at 8/27/18 6:24 PM:


Thanks [~botong] . Can you share the windows execution with and without the 
patch?
e.g. https://issues.apache.org/jira/browse/HDFS-13509 


was (Author: giovanni.fumarola):
Thanks [~botong] . Can you share the windows execution with and without the 
patch?

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Giovanni Matteo Fumarola (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594061#comment-16594061
 ] 

Giovanni Matteo Fumarola commented on HADOOP-15699:
---

Thanks [~botong] . Can you share the windows execution with and without the 
patch?

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Status: Patch Available  (was: Open)

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Attachment: HADOOP-15691-002.patch

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch, HADOOP-15691-002.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Botong Huang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Botong Huang updated HADOOP-15699:
--
Status: Patch Available  (was: Open)

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Botong Huang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Botong Huang updated HADOOP-15699:
--
Attachment: HADOOP-15699.v1.patch

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
> Attachments: HADOOP-15699.v1.patch
>
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15693) Typo in Secure Mode documentation

2018-08-27 Thread genericqa (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594035#comment-16594035
 ] 

genericqa commented on HADOOP-15693:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue}  0m 
17s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 19m 
30s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 11m 
39s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
42m 19s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 18m 
38s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 11m 
50s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green}  0m 
 0s{color} | {color:green} The patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
11m 16s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
30s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 85m 30s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:ba1ab08 |
| JIRA Issue | HADOOP-15693 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12936982/HADOOP-15693.patch |
| Optional Tests |  dupname  asflicense  mvnsite  |
| uname | Linux a95ffcbc199e 3.13.0-153-generic #203-Ubuntu SMP Thu Jun 14 
08:52:28 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/patchprocess/precommit/personality/provided.sh |
| git revision | trunk / 6eecd25 |
| maven | version: Apache Maven 3.3.9 |
| Max. process+thread count | 302 (vs. ulimit of 1) |
| modules | C: . U: . |
| Console output | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/15097/console |
| Powered by | Apache Yetus 0.9.0-SNAPSHOT   http://yetus.apache.org |


This message was automatically generated.



> Typo in Secure Mode documentation
> -
>
> Key: HADOOP-15693
> URL: https://issues.apache.org/jira/browse/HADOOP-15693
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
> Environment: OpenBSD 6.3 GENERIC.MP#4 amd64
>Reporter: Thomas Levine
>Assignee: Thomas Levine
>Priority: Trivial
>  Labels: documentation
> Attachments: HADOOP-15693.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Status: Open  (was: Patch Available)

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15698) KMS startup logs don't show

2018-08-27 Thread Ajay Kumar (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16594031#comment-16594031
 ] 

Ajay Kumar commented on HADOOP-15698:
-

[~knanasi] thanks for submitting the patch. Any plans to add a unit test ? 
Seems adding unit test is not straight forward but we can test logging in 
KMSWebServer.

> KMS startup logs don't show
> ---
>
> Key: HADOOP-15698
> URL: https://issues.apache.org/jira/browse/HADOOP-15698
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.1.0
>Reporter: Kitti Nanasi
>Assignee: Kitti Nanasi
>Priority: Major
> Attachments: HADOOP-15698.001.patch
>
>
> During KMs startup, log4j logs don't show up resulting in important logs 
> getting omitted. This happens because log4 initialisation only happens in 
> KMSWebApp#contextInitialized and logs written before that don't show up.
> For example the following log never shows up:
> [https://github.com/apache/hadoop/blob/a55d6bba71c81c1c4e9d8cd11f55c78f10a548b0/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/util/ZKSignerSecretProvider.java#L197-L199]
> Another example is that the KMS startup message never shows up in the kms 
> logs.
> Note that this works in the unit tests, because MiniKMS sets the log4j system 
> property.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Status: Patch Available  (was: Open)

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Botong Huang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15699?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Botong Huang updated HADOOP-15699:
--
Issue Type: Sub-task  (was: Task)
Parent: HADOOP-15475

> Fix some of testContainerManager failures in Windows
> 
>
> Key: HADOOP-15699
> URL: https://issues.apache.org/jira/browse/HADOOP-15699
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Botong Huang
>Assignee: Botong Huang
>Priority: Major
>
> TestContainerManager.testAuxPathHandler: The filename, directory name, or 
> volume label syntax is incorrect
> TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't 
> exist!
> TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't 
> exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15699) Fix some of testContainerManager failures in Windows

2018-08-27 Thread Botong Huang (JIRA)
Botong Huang created HADOOP-15699:
-

 Summary: Fix some of testContainerManager failures in Windows
 Key: HADOOP-15699
 URL: https://issues.apache.org/jira/browse/HADOOP-15699
 Project: Hadoop Common
  Issue Type: Task
Reporter: Botong Huang
Assignee: Botong Huang


TestContainerManager.testAuxPathHandler: The filename, directory name, or 
volume label syntax is incorrect
TestContainerManager.testForcefulShutdownSignal: ProcessStartFile doesn't exist!
TestContainerManager.testGracefulShutdownSignal: ProcessStartFile doesn't exist!
TestContainerManager.testOutputThreadDumpSignal: ProcessStartFile doesn't exist!



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15633) fs.TrashPolicyDefault: Can't create trash directory

2018-08-27 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15633?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593940#comment-16593940
 ] 

Hudson commented on HADOOP-15633:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14836 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14836/])
HADOOP-15633. fs.TrashPolicyDefault: Can't create trash directory. (jzhuge: rev 
f9c6fd94711458b77ecf3fa425aad7fda5089376)
* (edit) 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/TestTrash.java
* (edit) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/TrashPolicyDefault.java


> fs.TrashPolicyDefault: Can't create trash directory
> ---
>
> Key: HADOOP-15633
> URL: https://issues.apache.org/jira/browse/HADOOP-15633
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
>Affects Versions: 2.8.3, 3.1.0, 3.0.3, 2.7.7
>Reporter: Fei Hui
>Assignee: Fei Hui
>Priority: Major
> Attachments: HADOOP-15633.001.patch, HADOOP-15633.002.patch, 
> HADOOP-15633.003.patch, HADOOP-15633.004.patch, HADOOP-15633.005.patch
>
>
> Reproduce it as follow
> {code:java}
> hadoop fs -mkdir /user/hadoop/aaa
> hadoop fs -touchz /user/hadoop/aaa/bbb
> hadoop fs -rm /user/hadoop/aaa/bbb
> hadoop fs -mkdir /user/hadoop/aaa/bbb
> hadoop fs -touchz /user/hadoop/aaa/bbb/ccc
> hadoop fs -rm /user/hadoop/aaa/bbb/ccc
> {code}
> Then we get errors 
> {code:java}
> 18/07/26 17:55:24 WARN fs.TrashPolicyDefault: Can't create trash directory: 
> hdfs://xxx/user/hadoop/.Trash/Current/user/hadoop/aaa/bbb
> org.apache.hadoop.fs.FileAlreadyExistsException: Path is not a directory: 
> /user/hadoop/.Trash/Current/user/hadoop/aaa/bbb
> at 
> org.apache.hadoop.hdfs.server.namenode.FSDirMkdirOp.mkdirs(FSDirMkdirOp.java:65)
> at 
> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.mkdirs(FSNamesystem.java:3961)
> at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.mkdirs(NameNodeRpcServer.java:984)
> at 
> org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.mkdirs(ClientNamenodeProtocolServerSideTranslatorPB.java:622)
> at 
> org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
> at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:616)
> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:982)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2115)
> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2111)
> at java.security.AccessController.doPrivileged(Native Method)
> at javax.security.auth.Subject.doAs(Subject.java:422)
> at 
> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1867)
> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2111)
> at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native 
> Method)
> at 
> sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
> at 
> sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
> at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
> at 
> org.apache.hadoop.ipc.RemoteException.instantiateException(RemoteException.java:106)
> at 
> org.apache.hadoop.ipc.RemoteException.unwrapRemoteException(RemoteException.java:73)
> at 
> org.apache.hadoop.hdfs.DFSClient.primitiveMkdir(DFSClient.java:3002)
> at org.apache.hadoop.hdfs.DFSClient.mkdirs(DFSClient.java:2970)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1047)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem$21.doCall(DistributedFileSystem.java:1043)
> at 
> org.apache.hadoop.fs.FileSystemLinkResolver.resolve(FileSystemLinkResolver.java:81)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirsInternal(DistributedFileSystem.java:1061)
> at 
> org.apache.hadoop.hdfs.DistributedFileSystem.mkdirs(DistributedFileSystem.java:1036)
> at 
> org.apache.hadoop.fs.TrashPolicyDefault.moveToTrash(TrashPolicyDefault.java:136)
> at org.apache.hadoop.fs.Trash.moveToTrash(Trash.java:114)
> at org.apache.hadoop.fs.Trash.moveToAppropriateTrash(Trash.java:95)
> at org.apache.hadoop.fs.shell.Delete$Rm.moveToTrash(Delete.java:118)
> at org.apache.hadoop.fs.shell.Delete$Rm.processPath(Delete.java:105)
> at org.apache.hadoop.fs.shell.Command.processPaths(Command.java:317)
> at 
> org.apache.hadoop.fs.shell.Command.processPathArgument(Command.java:289)
>

[jira] [Commented] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593934#comment-16593934
 ] 

Steve Loughran commented on HADOOP-15691:
-

Should be of interest to: [~sanjay.radia], [~tmarquardt], [~fabbri], [~ehiggs] 
[~thomasdemoor], [~rdblue], [~arpitagarwal] [~jzhuge] [~sseth]

The goal here initially is just to let callers see if things like concat/append 
work before trying, but I really want to line things up for being able to ask 
if a store can do a direct write to a dest path without it being visible until 
closed (so you don't need to write to a temp path & rename for that 
one-operation commit), when to know that rename() doesn't do what posix 
expects, 

I know this breaks the idea that all filesystems should work just like HDFS, 
but that is true today: we just need to be able to determine which APIs *and 
semantics* are offered by specific FS instances.

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593926#comment-16593926
 ] 

Steve Loughran commented on HADOOP-15691:
-

Patch 001: initial implementation. This defines a new interface, 
PathCapabiliities, implemented by FileSystem, AbstractFileSystem and 
FileContext, to probe for features. It has one method

{code}
boolean hasPathCapability(Path path, String capability) throws IOException;
{code}

* There's a core list of features added in StreamCapabilities (right place?); 
for the main optional FS operations & features: permissions, xattrs, storage 
policies, snapshots, append, truncate, concat, and wired up for the all built 
in filesystems
* No tests; they can go into contract tests (e.g the append test can verify 
that having the append capability -> append works, etc)
* Base implementation returns false for everything but the "fs:symlinks" probe, 
which is forwarded to {{supportsSymlinks()}}

To do in separate JIRA;

* Object store operations/semantics,  where I want to be able to declare when a 
store does a PUT at the end of a write (e.g: no rename needed), and to declare 
when renames are slow/non-atomic for dirs. We can use that in the mrv2 
committers to know when to warn, perhaps.
* Multipart upload availablity

h3. Semantics of return value

w.r/t this patch, the spec says "return true iff you know the feature is 
available". This means "false" can mean any of

* The capability is unknown.
* The capability is known, but not available on this instance.
* The capability is known but the connector does not know if it is supported by 
this specific instance.

We could think about having an enum of all four states, so that you can 
distinguish them, but I fear it gets complicated with forwarded filesystems. 

h3. FSLinkResolver

new class {{org.apache.hadoop.fs.FSLinkResolver}} for implementing FileContext 
operations as closures, here
{code}
FsLinkResolution.resolve(this,
fixRelativePart(path),
(fs, p) -> fs.hasPathCapability(p, capability));
{code}

It should be straightforward to move all FileContext operations to this & make 
the file a lot leaner.

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15691) Add PathCapabilities to FS and FC to complement StreamCapabilities

2018-08-27 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15691?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15691:

Attachment: HADOOP-15691-001.patch

> Add PathCapabilities to FS and FC to complement StreamCapabilities
> --
>
> Key: HADOOP-15691
> URL: https://issues.apache.org/jira/browse/HADOOP-15691
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-15691-001.patch
>
>
> Add a {{PathCapabilities}} interface to both FileSystem and FileContext to 
> declare the capabilities under the path of an FS
> This is needed for 
> * HADOOP-15407: declare that a dest FS supports permissions
> * object stores to declare that they offer PUT-in-place alongside 
> (slow-rename)
> * Anything else where the implementation semantics of an FS is so different 
> caller apps would benefit from probing for the underlying semantics
> I know, we want all filesystem to work *exactly* the same. But it doesn't 
> hold, especially for object stores —and to efficiently use them, callers need 
> to be able to ask for specific features.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15693) Typo in Secure Mode documentation

2018-08-27 Thread Arpit Agarwal (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15693?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593898#comment-16593898
 ] 

Arpit Agarwal commented on HADOOP-15693:


Thanks [~tlevine]. +1 pending Jenkins.

I've added you as a Hadoop contributor and assigned to you.

> Typo in Secure Mode documentation
> -
>
> Key: HADOOP-15693
> URL: https://issues.apache.org/jira/browse/HADOOP-15693
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
> Environment: OpenBSD 6.3 GENERIC.MP#4 amd64
>Reporter: Thomas Levine
>Assignee: Thomas Levine
>Priority: Trivial
>  Labels: documentation
> Attachments: HADOOP-15693.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-15693) Typo in Secure Mode documentation

2018-08-27 Thread Arpit Agarwal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal reassigned HADOOP-15693:
--

Assignee: Thomas Levine

> Typo in Secure Mode documentation
> -
>
> Key: HADOOP-15693
> URL: https://issues.apache.org/jira/browse/HADOOP-15693
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
> Environment: OpenBSD 6.3 GENERIC.MP#4 amd64
>Reporter: Thomas Levine
>Assignee: Thomas Levine
>Priority: Trivial
>  Labels: documentation
> Attachments: HADOOP-15693.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15693) Typo in Secure Mode documentation

2018-08-27 Thread Arpit Agarwal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15693?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Arpit Agarwal updated HADOOP-15693:
---
Status: Patch Available  (was: Open)

> Typo in Secure Mode documentation
> -
>
> Key: HADOOP-15693
> URL: https://issues.apache.org/jira/browse/HADOOP-15693
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
> Environment: OpenBSD 6.3 GENERIC.MP#4 amd64
>Reporter: Thomas Levine
>Assignee: Thomas Levine
>Priority: Trivial
>  Labels: documentation
> Attachments: HADOOP-15693.patch
>
>   Original Estimate: 0h
>  Remaining Estimate: 0h
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15697) memory leak in distcp run method

2018-08-27 Thread mahesh kumar behera (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15697?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

mahesh kumar behera updated HADOOP-15697:
-
Attachment: gc-root.png

> memory leak in distcp run method
> 
>
> Key: HADOOP-15697
> URL: https://issues.apache.org/jira/browse/HADOOP-15697
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Priority: Major
> Attachments: gc-root.png
>
>
> in distcp run method, execute is called but the job is not closed. This is 
> causing OOM error in Hive replication. I think in run method , shouldBlock 
> should be set to true and execute should be called within try-with-resources.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15697) memory leak in distcp run method

2018-08-27 Thread mahesh kumar behera (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593883#comment-16593883
 ] 

mahesh kumar behera commented on HADOOP-15697:
--

 !gc-root.png! 

> memory leak in distcp run method
> 
>
> Key: HADOOP-15697
> URL: https://issues.apache.org/jira/browse/HADOOP-15697
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Priority: Major
>
> in distcp run method, execute is called but the job is not closed. This is 
> causing OOM error in Hive replication. I think in run method , shouldBlock 
> should be set to true and execute should be called within try-with-resources.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15697) memory leak in distcp run method

2018-08-27 Thread mahesh kumar behera (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593882#comment-16593882
 ] 

mahesh kumar behera commented on HADOOP-15697:
--

https://docs.google.com/document/d/1RNcAcRbATM7RfYboGLSpTmbzHbHPsrU8tRo9O9pPOZM/edit?usp=sharing

[~jojochuang]
The hprof file is shared. 



> memory leak in distcp run method
> 
>
> Key: HADOOP-15697
> URL: https://issues.apache.org/jira/browse/HADOOP-15697
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Priority: Major
>
> in distcp run method, execute is called but the job is not closed. This is 
> causing OOM error in Hive replication. I think in run method , shouldBlock 
> should be set to true and execute should be called within try-with-resources.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15697) memory leak in distcp run method

2018-08-27 Thread Wei-Chiu Chuang (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15697?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593620#comment-16593620
 ] 

Wei-Chiu Chuang commented on HADOOP-15697:
--

Hi [~maheshk114] thanks for reporting the issue. Do you have a heap dump after 
the OOM? Without a heapdump or a stacktrace it's hard to know what went wrong 
exactly.

One source of OOM that we are aware of is the file lists (distcp has to get a 
list of every files) and it can grow to several gigabytes for a large 
directory. This known issue is addressed by HADOOP-12502. But there could be 
other issues as well so we'll need more data to understand better.

> memory leak in distcp run method
> 
>
> Key: HADOOP-15697
> URL: https://issues.apache.org/jira/browse/HADOOP-15697
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: mahesh kumar behera
>Priority: Major
>
> in distcp run method, execute is called but the job is not closed. This is 
> causing OOM error in Hive replication. I think in run method , shouldBlock 
> should be set to true and execute should be called within try-with-resources.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-08-27 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593567#comment-16593567
 ] 

Gabor Bota edited comment on HADOOP-15621 at 8/27/18 12:28 PM:
---

Some comments on the description: 
bq. I think we need a new column in the dynamo table "entry last written time". 
This is updated each time the entry is written to dynamo.
The current implementation uses {{mod_time}} field when using prune. It would 
be wise to use the same because this is an online version of prune. Thus, we 
don't need to add a new field to the item.

{quote}After that we can either
1. Have the client simply ignore / elide any entries that are older than the 
configured TTL.
2. Have the client delete entries older than the TTL.{quote}
There will be a switch for this - the user can set if she wants the item to be 
deleted automatically if the {{current_time - mod_time > ttl}}.

The implementation will behave like the following:
1. {{DynamoDBMetadataStore}} will ignore any entries that are older than the 
configured TTL. That means the entry will not be included in the directory 
listing, the returned {{DirListingMetadata}} won't be authoritative.
There's an issue with this setting: the listing won't be authoritative if the 
entry is older than TTL, but if the client does another full listing, and saves 
it to the metadata store, the next listing request will still be automatically 
{{isAuthoritative=false}} because of the leftover expired entry in the same 
directory. The solution could be that the {{DDBPathMetadata}}, so the dynamo 
entry could store if the file already caused a directory listing to be 
{{isAuthoritative=false}}, so it won't cause the listing to not be 
authoritative again. Of course, if there will be another entry with the same 
name, the flag should be removed (overwritten). This could be a flag named 
{{expired_entry}}.

2. {{DynamoDBMetadataStore}} deletes entries which are older than the TTL. With 
this setting we won't have the same issue as in 1. setting, but will cost more 
as we have to calculate with the deletion round trip time to dynamo.


was (Author: gabor.bota):
Some comments on the description: 
bq. I think we need a new column in the dynamo table "entry last written time". 
This is updated each time the entry is written to dynamo.
The current implementation uses {{mod_time}} field when using prune. It would 
be wise to use the same because this is an online version of prune. Thus, we 
don't need to add a new field to the item.

{quote}After that we can either
1. Have the client simply ignore / elide any entries that are older than the 
configured TTL.
2. Have the client delete entries older than the TTL.{quote}
There will be a switch for this - the user can set if she wants the item to be 
deleted automatically if the {{current_time - mod_time > ttl}}.

The implementation will behave like the following:
1. {{DynamoDBMetadataStore}} will ignore any entries that are older than the 
configured TTL. That means the entry will not be included in the directory 
listing, the returned {{DirListingMetadata}} won't be authoritative.
There's an issue with this setting: the listing won't be authoritative if the 
entry is older than TTL, but if the client does another full listing, and saves 
it to the metadata store, the next listing request will still be automatically 
{{isAuthoritative=false}} because of the leftover expired entry in the same 
dictionary. The solution could be that the {{DDBPathMetadata}}, so the dynamo 
entry could store if the file already caused a directory listing to be 
{{isAuthoritative=false}}, so it won't cause the listing to not be 
authoritative again. Of course, if there will be another entry with the same 
name, the flag should be removed (overwritten).
2. {{DynamoDBMetadataStore}} deletes entries which are older than the TTL. With 
this setting we won't have the same issue as in 1. setting, but will cost more 
as we have to calculate with the deletion round trip time to dynamo.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> Think of this as the "online algorithm" version of the CLI prune() function, 
> which is the "offline algorithm".
> Why: 
> 1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metada

[jira] [Updated] (HADOOP-15696) KMS performance regression due to too many open file descriptors after Jetty migration

2018-08-27 Thread Wei-Chiu Chuang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15696?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang updated HADOOP-15696:
-
Affects Version/s: (was: 3.0.0)
   3.0.0-alpha2

> KMS performance regression due to too many open file descriptors after Jetty 
> migration
> --
>
> Key: HADOOP-15696
> URL: https://issues.apache.org/jira/browse/HADOOP-15696
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.0.0-alpha2
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Attachments: Screen Shot 2018-08-22 at 11.36.16 AM.png, Screen Shot 
> 2018-08-22 at 4.26.51 PM.png, Screen Shot 2018-08-22 at 4.26.51 PM.png, 
> Screen Shot 2018-08-22 at 4.27.02 PM.png, Screen Shot 2018-08-22 at 4.30.32 
> PM.png, Screen Shot 2018-08-22 at 4.30.39 PM.png, Screen Shot 2018-08-24 at 
> 7.08.16 PM.png
>
>
> We recently found KMS performance regressed in Hadoop 3.0, possibly linking 
> to the migration from Tomcat to Jetty in HADOOP-13597.
> Symptoms:
> # Hadoop 3.x KMS open file descriptors quickly rises to more than 10 thousand 
> under stress, sometimes even exceeds 32K, which is the system limit, causing 
> failures for any access to encryption zones. Our internal testing shows the 
> openfd number was in the range of a few hundred in Hadoop 2.x, and it 
> increases by almost 100x in Hadoop 3.
> # Hadoop 3.x KMS as much as twice the heap size than in Hadoop 2.x. The same 
> heap size can go OOM in Hadoop 3.x. Jxray analysis suggests most of them are 
> temporary byte arrays associated with open SSL connections.
> # Due to the heap usage, Hadoop 3.x KMS has more frequent GC activities, and 
> we observed up to 20% performance reduction due to GC.
> A possible solution is to reduce the idle timeout setting in HttpServer2. It 
> is currently hard-coded 10 seconds. By setting it to 1 second, open fds 
> dropped from 20 thousand down to 3 thousand in my experiment.
> File this jira to invite open discussion for a solution.
> Credit: [~mi...@cloudera.com] for the proposed Jetty idle timeout remedy; 
> [~xiaochen] for digging into this problem.
> Screenshots:
> CDH5 (Hadoop 2) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.39 PM.png! 
> CDH6 (Hadoop 3) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.32 PM.png! 
> CDH5 (Hadoop 2) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.26.51 PM.png! 
> CDH6 (Hadoop 3) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.27.02 PM.png! 
> JXray report
>  !Screen Shot 2018-08-22 at 11.36.16 AM.png! 
> open fd drops from 20 k down to 3k after the proposed change.
>  !Screen Shot 2018-08-24 at 7.08.16 PM.png! 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15621) s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store

2018-08-27 Thread Gabor Bota (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15621?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593567#comment-16593567
 ] 

Gabor Bota commented on HADOOP-15621:
-

Some comments on the description: 
bq. I think we need a new column in the dynamo table "entry last written time". 
This is updated each time the entry is written to dynamo.
The current implementation uses {{mod_time}} field when using prune. It would 
be wise to use the same because this is an online version of prune. Thus, we 
don't need to add a new field to the item.

{quote}After that we can either
1. Have the client simply ignore / elide any entries that are older than the 
configured TTL.
2. Have the client delete entries older than the TTL.{quote}
There will be a switch for this - the user can set if she wants the item to be 
deleted automatically if the {{current_time - mod_time > ttl}}.

The implementation will behave like the following:
1. {{DynamoDBMetadataStore}} will ignore any entries that are older than the 
configured TTL. That means the entry will not be included in the directory 
listing, the returned {{DirListingMetadata}} won't be authoritative.
There's an issue with this setting: the listing won't be authoritative if the 
entry is older than TTL, but if the client does another full listing, and saves 
it to the metadata store, the next listing request will still be automatically 
{{isAuthoritative=false}} because of the leftover expired entry in the same 
dictionary. The solution could be that the {{DDBPathMetadata}}, so the dynamo 
entry could store if the file already caused a directory listing to be 
{{isAuthoritative=false}}, so it won't cause the listing to not be 
authoritative again. Of course, if there will be another entry with the same 
name, the flag should be removed (overwritten).
2. {{DynamoDBMetadataStore}} deletes entries which are older than the TTL. With 
this setting we won't have the same issue as in 1. setting, but will cost more 
as we have to calculate with the deletion round trip time to dynamo.

> s3guard: implement time-based (TTL) expiry for DynamoDB Metadata Store
> --
>
> Key: HADOOP-15621
> URL: https://issues.apache.org/jira/browse/HADOOP-15621
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Aaron Fabbri
>Assignee: Gabor Bota
>Priority: Minor
>
> Similar to HADOOP-13649, I think we should add a TTL (time to live) feature 
> to the Dynamo metadata store (MS) for S3Guard.
> Think of this as the "online algorithm" version of the CLI prune() function, 
> which is the "offline algorithm".
> Why: 
> 1. Self healing (soft state): since we do not implement transactions around 
> modification of the two systems (s3 and metadata store), certain failures can 
> lead to inconsistency between S3 and the metadata store (MS) state.  Having a 
> time to live (TTL) on each entry in S3Guard means that any inconsistencies 
> will be time bound.  Thus "wait and restart your job" becomes a valid, if 
> ugly, way to get around any issues with FS client failure leaving things in a 
> bad state.
> 2. We could make manual invocation of `hadoop s3guard prune ...` unnecessary, 
> depending on the implementation.
> 3. Makes it possible to fix the problem that dynamo MS prune() doesn't prune 
> directories due to the lack of true modification time.
> How:
> I think we need a new column in the dynamo table "entry last written time".  
> This is updated each time the entry is written to dynamo.
> After that we can either
> 1. Have the client simply ignore / elide any entries that are older than the 
> configured TTL.
> 2. Have the client delete entries older than the TTL.
> The issue with #2 is it will increase latency if done inline in the context 
> of an FS operation. We could mitigate this some by using an async helper 
> thread, or probabilistically doing it "some times" to amortize the expense of 
> deleting stale entries (allowing some batching as well).
> Caveats:
> - Clock synchronization as usual is a concern. Many clusters already keep 
> clocks close enough via NTP. We should at least document the requirement 
> along with the configuration knob that enables the feature.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15676) Cleanup TestSSLHttpServer

2018-08-27 Thread Zoltan Siegl (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15676?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593538#comment-16593538
 ] 

Zoltan Siegl commented on HADOOP-15676:
---

Hey [~snemeth]! 
Did some review on the patch, and I have a few comments might worth considering.
 * {{org/apache/hadoop/http/TestSSLHttpServer.java:70}}

{code:java}
public static final String SERVLET_NAME_LONGHEADER = "longheader";
public static final String SERVLET_PATH_LONGHEADER =
"/" + SERVLET_NAME_LONGHEADER;
public static final String SERVLET_NAME_ECHO = "echo";
public static final String SERVLET_PATH_ECHO = "/" + SERVLET_NAME_ECHO;{code}
 
 If there is not specific reason for these to be public, maybe we should 
consider making these private.
 * {{org/apache/hadoop/http/TestSSLHttpServer.java:248}}

It looks like there has been an additional empty line added in the patch. If 
you create a new patch anyways, should we consider removing this?

 

 

 

> Cleanup TestSSLHttpServer
> -
>
> Key: HADOOP-15676
> URL: https://issues.apache.org/jira/browse/HADOOP-15676
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: common
>Affects Versions: 2.6.0
>Reporter: Szilard Nemeth
>Assignee: Szilard Nemeth
>Priority: Minor
> Attachments: HADOOP-15676.001.patch
>
>
> This issue will fix: 
> * Several typos in this class
> * Code is not very well readable in some of the places.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15698) KMS startup logs don't show

2018-08-27 Thread Kitti Nanasi (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16593456#comment-16593456
 ] 

Kitti Nanasi commented on HADOOP-15698:
---

I tried my patch on a cluster and the logs mentioned in the description do show 
up in the kms.log file.

> KMS startup logs don't show
> ---
>
> Key: HADOOP-15698
> URL: https://issues.apache.org/jira/browse/HADOOP-15698
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.1.0
>Reporter: Kitti Nanasi
>Assignee: Kitti Nanasi
>Priority: Major
> Attachments: HADOOP-15698.001.patch
>
>
> During KMs startup, log4j logs don't show up resulting in important logs 
> getting omitted. This happens because log4 initialisation only happens in 
> KMSWebApp#contextInitialized and logs written before that don't show up.
> For example the following log never shows up:
> [https://github.com/apache/hadoop/blob/a55d6bba71c81c1c4e9d8cd11f55c78f10a548b0/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/util/ZKSignerSecretProvider.java#L197-L199]
> Another example is that the KMS startup message never shows up in the kms 
> logs.
> Note that this works in the unit tests, because MiniKMS sets the log4j system 
> property.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15698) KMS startup logs don't show

2018-08-27 Thread Kitti Nanasi (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15698?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kitti Nanasi updated HADOOP-15698:
--
Attachment: HADOOP-15698.001.patch

> KMS startup logs don't show
> ---
>
> Key: HADOOP-15698
> URL: https://issues.apache.org/jira/browse/HADOOP-15698
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.1.0
>Reporter: Kitti Nanasi
>Assignee: Kitti Nanasi
>Priority: Major
> Attachments: HADOOP-15698.001.patch
>
>
> During KMs startup, log4j logs don't show up resulting in important logs 
> getting omitted. This happens because log4 initialisation only happens in 
> KMSWebApp#contextInitialized and logs written before that don't show up.
> For example the following log never shows up:
> [https://github.com/apache/hadoop/blob/a55d6bba71c81c1c4e9d8cd11f55c78f10a548b0/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/util/ZKSignerSecretProvider.java#L197-L199]
> Another example is that the KMS startup message never shows up in the kms 
> logs.
> Note that this works in the unit tests, because MiniKMS sets the log4j system 
> property.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15698) KMS startup logs don't show

2018-08-27 Thread Kitti Nanasi (JIRA)
Kitti Nanasi created HADOOP-15698:
-

 Summary: KMS startup logs don't show
 Key: HADOOP-15698
 URL: https://issues.apache.org/jira/browse/HADOOP-15698
 Project: Hadoop Common
  Issue Type: Bug
  Components: kms
Affects Versions: 3.1.0
Reporter: Kitti Nanasi
Assignee: Kitti Nanasi


During KMs startup, log4j logs don't show up resulting in important logs 
getting omitted. This happens because log4 initialisation only happens in 
KMSWebApp#contextInitialized and logs written before that don't show up.

For example the following log never shows up:

[https://github.com/apache/hadoop/blob/a55d6bba71c81c1c4e9d8cd11f55c78f10a548b0/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/util/ZKSignerSecretProvider.java#L197-L199]

Another example is that the KMS startup message never shows up in the kms logs.

Note that this works in the unit tests, because MiniKMS sets the log4j system 
property.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15697) memory leak in distcp run method

2018-08-27 Thread mahesh kumar behera (JIRA)
mahesh kumar behera created HADOOP-15697:


 Summary: memory leak in distcp run method
 Key: HADOOP-15697
 URL: https://issues.apache.org/jira/browse/HADOOP-15697
 Project: Hadoop Common
  Issue Type: Bug
Reporter: mahesh kumar behera


in distcp run method, execute is called but the job is not closed. This is 
causing OOM error in Hive replication. I think in run method , shouldBlock 
should be set to true and execute should be called within try-with-resources.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org