[jira] [Commented] (HDFS-8135) Remove the deprecated FSConstants class

2015-05-20 Thread zhangduo (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8135?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14551916#comment-14551916
 ] 

zhangduo commented on HDFS-8135:


HBase uses DistributedFileSystem.setSafeMode to check whether an hdfs cluster 
is in safe mode. This method needs {{HdfsConstants.SafeModeAction}}.

I found that there is a {{HdfsUtils.isHealthy}}, we could try to make use of 
this method instead of depending on hdfs private classes.

Thanks.

 Remove the deprecated FSConstants class
 ---

 Key: HDFS-8135
 URL: https://issues.apache.org/jira/browse/HDFS-8135
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Li Lu
 Fix For: 2.8.0

 Attachments: HDFS-8135-041315.patch


 The {{FSConstants}} class has been marked as deprecated since 0.23. There is 
 no uses of this class in the current code base and it can be removed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread Akira AJISAKA (JIRA)
Akira AJISAKA created HDFS-8443:
---

 Summary: Document dfs.namenode.service.handler.count in 
hdfs-site.xml
 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA


When dfs.namenode.servicerpc-address is configured, NameNode launches an extra 
RPC server to handle requests from non-client nodes. 
dfs.namenode.service.handler.count specifies the number of threads for the 
server but the parameter is not documented anywhere.

I found a mail for asking about the parameter. 
http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8413) Directories are not listed recursively when fs.defaultFs is viewFs

2015-05-20 Thread Ajith S (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ajith S updated HDFS-8413:
--
Attachment: HDFS-8413.patch

 Directories are not listed recursively when fs.defaultFs is viewFs
 --

 Key: HDFS-8413
 URL: https://issues.apache.org/jira/browse/HDFS-8413
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Ajith S
Assignee: Ajith S
  Labels: viewfs
 Attachments: HDFS-8413.patch


 Mount a cluster on client throught viewFs mount table
 Example:
 {quote}
  property
 namefs.defaultFS/name
 valueviewfs:value
   /property
 property
 namefs.viewfs.mounttable.default.link./nn1/name
 valuehdfs://ns1//value  !-- HA nameservice --
 /property
 property
 namefs.viewfs.mounttable.default.link./user/name
 valuehdfs://host-72:8020//value
 /property
  property
 {quote}
 Try to list the files recursively *(hdfs dfs -ls -R / or hadoop fs -ls -R /)* 
 only the parent folders are listed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8442) Remove ServerLifecycleListener from kms/server.xml.

2015-05-20 Thread nijel (JIRA)
nijel created HDFS-8442:
---

 Summary: Remove ServerLifecycleListener from kms/server.xml.
 Key: HDFS-8442
 URL: https://issues.apache.org/jira/browse/HDFS-8442
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: nijel
Assignee: nijel


Remove ServerLifecycleListener from kms/server.xml.

From tomcat Tomcat 7.0.9 the support for ServerLifecycleListener is removed
ref : https://tomcat.apache.org/tomcat-7.0-doc/changelog.html

Remove ServerLifecycleListener. This was already removed from server.xml and 
with the Lifecycle re-factoring is no longer required. (markt)

So if the build env is with tomcat later than this, kms startup is failing
{code}
SEVERE: Begin event threw exception
java.lang.ClassNotFoundException: 
org.apache.catalina.mbeans.ServerLifecycleListener
{code}
can we remove this listener ? 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8413) Directories are not listed recursively when fs.defaultFs is viewFs

2015-05-20 Thread Ajith S (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ajith S updated HDFS-8413:
--
Status: Patch Available  (was: Open)

*Resolution* : decide if the mount point is a directory or not by querying the 
target filesystem
Submitting the patch for same. Please review

 Directories are not listed recursively when fs.defaultFs is viewFs
 --

 Key: HDFS-8413
 URL: https://issues.apache.org/jira/browse/HDFS-8413
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Ajith S
Assignee: Ajith S
  Labels: viewfs
 Attachments: HDFS-8413.patch


 Mount a cluster on client throught viewFs mount table
 Example:
 {quote}
  property
 namefs.defaultFS/name
 valueviewfs:value
   /property
 property
 namefs.viewfs.mounttable.default.link./nn1/name
 valuehdfs://ns1//value  !-- HA nameservice --
 /property
 property
 namefs.viewfs.mounttable.default.link./user/name
 valuehdfs://host-72:8020//value
 /property
  property
 {quote}
 Try to list the files recursively *(hdfs dfs -ls -R / or hadoop fs -ls -R /)* 
 only the parent folders are listed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8413) Directories are not listed recursively when fs.defaultFs is viewFs

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8413?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14551940#comment-14551940
 ] 

Hadoop QA commented on HDFS-8413:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 29s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 26s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 36s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 22s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   1m  7s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 36s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   1m 40s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:red}-1{color} | common tests |  23m  6s | Tests failed in 
hadoop-common. |
| | |  59m 58s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.fs.viewfs.TestViewFileSystemLocalFileSystem |
|   | hadoop.fs.viewfs.TestViewFileSystemWithAuthorityLocalFileSystem |
|   | hadoop.security.token.delegation.web.TestWebDelegationToken |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734076/HDFS-8413.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / ce53c8e |
| hadoop-common test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11057/artifact/patchprocess/testrun_hadoop-common.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11057/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf906.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11057/console |


This message was automatically generated.

 Directories are not listed recursively when fs.defaultFs is viewFs
 --

 Key: HDFS-8413
 URL: https://issues.apache.org/jira/browse/HDFS-8413
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.0
Reporter: Ajith S
Assignee: Ajith S
  Labels: viewfs
 Attachments: HDFS-8413.patch


 Mount a cluster on client throught viewFs mount table
 Example:
 {quote}
  property
 namefs.defaultFS/name
 valueviewfs:value
   /property
 property
 namefs.viewfs.mounttable.default.link./nn1/name
 valuehdfs://ns1//value  !-- HA nameservice --
 /property
 property
 namefs.viewfs.mounttable.default.link./user/name
 valuehdfs://host-72:8020//value
 /property
  property
 {quote}
 Try to list the files recursively *(hdfs dfs -ls -R / or hadoop fs -ls -R /)* 
 only the parent folders are listed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8427) Remove dataBlockNum and parityBlockNum from BlockInfoStriped

2015-05-20 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-8427:
-
Status: Patch Available  (was: Open)

 Remove dataBlockNum and parityBlockNum from BlockInfoStriped
 

 Key: HDFS-8427
 URL: https://issues.apache.org/jira/browse/HDFS-8427
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-7285
Reporter: Kai Sasaki
Assignee: Kai Sasaki
 Fix For: HDFS-7285

 Attachments: HDFS-8427-HDFS-7285-00.patch, 
 HDFS-8427-HDFS-7285-01.patch


 Remove unnecessary members such as {{dataBlockNum}} and {{parityBlockNum}} 
 from {{BlockInfoStriped}}. These are included in {{ECShema}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552064#comment-14552064
 ] 

Hadoop QA commented on HDFS-8441:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 39s | Pre-patch HDFS-7285 compilation 
is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 28s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 40s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 15s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 36s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 38s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   3m 14s | The patch appears to introduce 6 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 14s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 172m 54s | Tests failed in hadoop-hdfs. |
| | | 214m 15s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
|  |  Inconsistent synchronization of 
org.apache.hadoop.hdfs.DFSOutputStream.streamer; locked 89% of time  
Unsynchronized access at DFSOutputStream.java:89% of time  Unsynchronized 
access at DFSOutputStream.java:[line 146] |
|  |  Possible null pointer dereference of arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:[line 194] |
|  |  Unread field:field be static?  At ErasureCodingWorker.java:[line 254] |
|  |  Should 
org.apache.hadoop.hdfs.server.datanode.erasurecode.ErasureCodingWorker$StripedReader
 be a _static_ inner class?  At ErasureCodingWorker.java:inner class?  At 
ErasureCodingWorker.java:[lines 907-914] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:[line 108] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:[line 409] |
| Failed unit tests | hadoop.hdfs.TestEncryptedTransfer |
|   | hadoop.hdfs.tools.TestDFSZKFailoverController |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
|   | hadoop.hdfs.server.blockmanagement.TestBlockInfo |
|   | hadoop.hdfs.server.namenode.TestFileTruncate |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734073/HDFS-8441-HDFS-7285.001.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | HDFS-7285 / 4dd4aa5 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11056/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11056/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11056/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11056/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11056/console |


This message was automatically generated.

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task

[jira] [Updated] (HDFS-8427) Remove dataBlockNum and parityBlockNum from BlockInfoStriped

2015-05-20 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-8427:
-
Attachment: HDFS-8427-HDFS-7285-01.patch

 Remove dataBlockNum and parityBlockNum from BlockInfoStriped
 

 Key: HDFS-8427
 URL: https://issues.apache.org/jira/browse/HDFS-8427
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-7285
Reporter: Kai Sasaki
Assignee: Kai Sasaki
 Fix For: HDFS-7285

 Attachments: HDFS-8427-HDFS-7285-00.patch, 
 HDFS-8427-HDFS-7285-01.patch


 Remove unnecessary members such as {{dataBlockNum}} and {{parityBlockNum}} 
 from {{BlockInfoStriped}}. These are included in {{ECShema}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8382) Remove chunkSize parameter from initialize method of raw erasure coder

2015-05-20 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552079#comment-14552079
 ] 

Vinayakumar B commented on HDFS-8382:
-

bq. Updated the patch also removing initialize method per the suggestion.
I see {{initialize()}} and {{chunkSize}} in {{AbstractErasureCoder}}
currently this is the only place where chunkSize was passed from ECSchema to 
coder. So HDFS-8374 depends on this.

 Remove chunkSize parameter from initialize method of raw erasure coder
 --

 Key: HDFS-8382
 URL: https://issues.apache.org/jira/browse/HDFS-8382
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Zheng
 Attachments: HDFS-8382-HDFS-7285-v1.patch, 
 HDFS-8382-HDFS-7285-v2.patch, HDFS-8382-HDFS-7285-v3.patch


 Per discussion in HDFS-8347, we need to support encoding/decoding variable 
 width units data instead of predefined fixed width like {{chunkSize}}. Have 
 this issue to remove chunkSize in the general raw erasure coder API. Specific 
 coder will support fixed chunkSize using hard-coded or specific schema 
 customizing way if necessary, like HitchHiker coder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8382) Remove chunkSize parameter from initialize method of raw erasure coder

2015-05-20 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552170#comment-14552170
 ] 

Kai Zheng commented on HDFS-8382:
-

Thanks Vinay for the comments. Yes I was going to get all of this done here but 
looks like I missed some places. Good catch! Will address them in the following 
patch. Thanks.

 Remove chunkSize parameter from initialize method of raw erasure coder
 --

 Key: HDFS-8382
 URL: https://issues.apache.org/jira/browse/HDFS-8382
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Zheng
 Attachments: HDFS-8382-HDFS-7285-v1.patch, 
 HDFS-8382-HDFS-7285-v2.patch, HDFS-8382-HDFS-7285-v3.patch


 Per discussion in HDFS-8347, we need to support encoding/decoding variable 
 width units data instead of predefined fixed width like {{chunkSize}}. Have 
 this issue to remove chunkSize in the general raw erasure coder API. Specific 
 coder will support fixed chunkSize using hard-coded or specific schema 
 customizing way if necessary, like HitchHiker coder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8444) Erasure Coding: fix cannot rename a zone dir

2015-05-20 Thread Walter Su (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Walter Su updated HDFS-8444:

Status: Patch Available  (was: Open)

 Erasure Coding: fix cannot rename a zone dir
 

 Key: HDFS-8444
 URL: https://issues.apache.org/jira/browse/HDFS-8444
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8444-HDFS-7285.001.patch


 We create a EC zone {{/my_ec_zone}}.
 We want to rename it to {{/myZone}}.
 But it failed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8444) Erasure Coding: fix cannot rename a zone dir

2015-05-20 Thread Walter Su (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8444?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Walter Su updated HDFS-8444:

Attachment: HDFS-8444-HDFS-7285.001.patch

 Erasure Coding: fix cannot rename a zone dir
 

 Key: HDFS-8444
 URL: https://issues.apache.org/jira/browse/HDFS-8444
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8444-HDFS-7285.001.patch


 We create a EC zone {{/my_ec_zone}}.
 We want to rename it to {{/myZone}}.
 But it failed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8131) Implement a space balanced block placement policy

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552191#comment-14552191
 ] 

Hudson commented on HDFS-8131:
--

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #202 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/202/])
HDFS-8131. Implement a space balanced block placement policy. Contributed by 
Liu Shaohui. (kihwal: rev de30d66b2673d0344346fb985e786247ca682317)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestAvailableSpaceBlockPlacementPolicy.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockPlacementPolicyDefault.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/AvailableSpaceBlockPlacementPolicy.java


 Implement a space balanced block placement policy
 -

 Key: HDFS-8131
 URL: https://issues.apache.org/jira/browse/HDFS-8131
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.0.0
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
  Labels: BlockPlacementPolicy
 Fix For: 2.8.0

 Attachments: HDFS-8131-v1.diff, HDFS-8131-v2.diff, HDFS-8131-v3.diff, 
 HDFS-8131.004.patch, HDFS-8131.005.patch, HDFS-8131.006.patch, balanced.png


 The default block placement policy will choose datanodes for new blocks 
 randomly, which will result in unbalanced space used percent among datanodes 
 after an cluster expansion. The old datanodes always are in high used percent 
 of space and new added ones are in low percent.
 Through we can used the external balance tool to balance the space used rate, 
 it will cost extra network IO and it's not easy to control the balance speed.
 An easy solution is to implement an balanced block placement policy which 
 will choose low used percent datanodes for new blocks with a little high 
 possibility. In a not long term, the used percent of datanodes will trend to 
 be balanced.
 Suggestions and discussions are welcomed. Thanks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8404) Pending block replication can get stuck using older genstamp

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552189#comment-14552189
 ] 

Hudson commented on HDFS-8404:
--

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #202 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/202/])
HDFS-8404. Pending block replication can get stuck using older genstamp. 
Contributed by Nathan Roberts. (kihwal: rev 
8860e352c394372e4eb3ebdf82ea899567f34e4e)
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestPendingReplication.java


 Pending block replication can get stuck using older genstamp
 

 Key: HDFS-8404
 URL: https://issues.apache.org/jira/browse/HDFS-8404
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.6.0, 2.7.0
Reporter: Nathan Roberts
Assignee: Nathan Roberts
 Fix For: 2.7.1

 Attachments: HDFS-8404-v0.patch, HDFS-8404-v1.patch


 If an under-replicated block gets into the pending-replication list, but 
 later the  genstamp of that block ends up being newer than the one originally 
 submitted for replication, the block will fail replication until the NN is 
 restarted. 
 It will be safer if processPendingReplications()  gets up-to-date blockinfo 
 before resubmitting replication work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8404) Pending block replication can get stuck using older genstamp

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552211#comment-14552211
 ] 

Hudson commented on HDFS-8404:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #933 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/933/])
HDFS-8404. Pending block replication can get stuck using older genstamp. 
Contributed by Nathan Roberts. (kihwal: rev 
8860e352c394372e4eb3ebdf82ea899567f34e4e)
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestPendingReplication.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java


 Pending block replication can get stuck using older genstamp
 

 Key: HDFS-8404
 URL: https://issues.apache.org/jira/browse/HDFS-8404
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.6.0, 2.7.0
Reporter: Nathan Roberts
Assignee: Nathan Roberts
 Fix For: 2.7.1

 Attachments: HDFS-8404-v0.patch, HDFS-8404-v1.patch


 If an under-replicated block gets into the pending-replication list, but 
 later the  genstamp of that block ends up being newer than the one originally 
 submitted for replication, the block will fail replication until the NN is 
 restarted. 
 It will be safer if processPendingReplications()  gets up-to-date blockinfo 
 before resubmitting replication work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8131) Implement a space balanced block placement policy

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552213#comment-14552213
 ] 

Hudson commented on HDFS-8131:
--

SUCCESS: Integrated in Hadoop-Yarn-trunk #933 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/933/])
HDFS-8131. Implement a space balanced block placement policy. Contributed by 
Liu Shaohui. (kihwal: rev de30d66b2673d0344346fb985e786247ca682317)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockPlacementPolicyDefault.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/AvailableSpaceBlockPlacementPolicy.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestAvailableSpaceBlockPlacementPolicy.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java


 Implement a space balanced block placement policy
 -

 Key: HDFS-8131
 URL: https://issues.apache.org/jira/browse/HDFS-8131
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.0.0
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
  Labels: BlockPlacementPolicy
 Fix For: 2.8.0

 Attachments: HDFS-8131-v1.diff, HDFS-8131-v2.diff, HDFS-8131-v3.diff, 
 HDFS-8131.004.patch, HDFS-8131.005.patch, HDFS-8131.006.patch, balanced.png


 The default block placement policy will choose datanodes for new blocks 
 randomly, which will result in unbalanced space used percent among datanodes 
 after an cluster expansion. The old datanodes always are in high used percent 
 of space and new added ones are in low percent.
 Through we can used the external balance tool to balance the space used rate, 
 it will cost extra network IO and it's not easy to control the balance speed.
 An easy solution is to implement an balanced block placement policy which 
 will choose low used percent datanodes for new blocks with a little high 
 possibility. In a not long term, the used percent of datanodes will trend to 
 be balanced.
 Suggestions and discussions are welcomed. Thanks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8429) Death of watcherThread making other local read blocked

2015-05-20 Thread zhouyingchao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552095#comment-14552095
 ] 

zhouyingchao commented on HDFS-8429:


Colin, thank you for the great comments.  In this case, I think the bottom line 
is that the death of the watcher thread should not block other threads and the 
client side should be indicated to fall through to other ways as quick as 
possible.
I created a patch trying to resolve the blocking. Besides that, I also changed 
the native getAndClearReadableFds method to throw exception as Colin mentioned. 
 Please feel free to post your thoughts and comments. Thank you.

 Death of watcherThread making other local read blocked
 --

 Key: HDFS-8429
 URL: https://issues.apache.org/jira/browse/HDFS-8429
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: zhouyingchao
Assignee: zhouyingchao

 In our cluster, an application is hung when doing a short circuit read of 
 local hdfs block. By looking into the log, we found the DataNode's 
 DomainSocketWatcher.watcherThread has exited with following log:
 {code}
 ERROR org.apache.hadoop.net.unix.DomainSocketWatcher: 
 Thread[Thread-25,5,main] terminating on unexpected exception
 java.lang.NullPointerException
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher$2.run(DomainSocketWatcher.java:463)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 The line 463 is following code snippet:
 {code}
  try {
 for (int fd : fdSet.getAndClearReadableFds()) {
   sendCallbackAndRemove(getAndClearReadableFds, entries, fdSet,
 fd);
 }
 {code}
 getAndClearReadableFds is a native method which will malloc an int array. 
 Since our memory is very tight, it looks like the malloc failed and a NULL 
 pointer is returned.
 The bad thing is that other threads then blocked in stack like this:
 {code}
 DataXceiver for client 
 unix:/home/work/app/hdfs/c3prc-micloud/datanode/dn_socket [Waiting for 
 operation #1] daemon prio=10 tid=0x7f0c9c086d90 nid=0x8fc3 waiting on 
 condition [0x7f09b9856000]
java.lang.Thread.State: WAITING (parking)
 at sun.misc.Unsafe.park(Native Method)
 - parking to wait for  0x0007b0174808 (a 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
 at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
 at 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher.add(DomainSocketWatcher.java:323)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.createNewMemorySegment(ShortCircuitRegistry.java:322)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.requestShortCircuitShm(DataXceiver.java:403)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opRequestShortCircuitShm(Receiver.java:214)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:95)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 IMO, we should exit the DN so that the users can know that something go  
 wrong  and fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8444) Erasure Coding: fix cannot rename a zone dir

2015-05-20 Thread Walter Su (JIRA)
Walter Su created HDFS-8444:
---

 Summary: Erasure Coding: fix cannot rename a zone dir
 Key: HDFS-8444
 URL: https://issues.apache.org/jira/browse/HDFS-8444
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su


We create a EC zone {{/my_ec_zone}}.
We want to rename it to {{/myZone}}.
But it failed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8429) Death of watcherThread making other local read blocked

2015-05-20 Thread zhouyingchao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhouyingchao updated HDFS-8429:
---
Status: Patch Available  (was: Open)

 Death of watcherThread making other local read blocked
 --

 Key: HDFS-8429
 URL: https://issues.apache.org/jira/browse/HDFS-8429
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: zhouyingchao
Assignee: zhouyingchao
 Attachments: HDFS-8429-001.patch


 In our cluster, an application is hung when doing a short circuit read of 
 local hdfs block. By looking into the log, we found the DataNode's 
 DomainSocketWatcher.watcherThread has exited with following log:
 {code}
 ERROR org.apache.hadoop.net.unix.DomainSocketWatcher: 
 Thread[Thread-25,5,main] terminating on unexpected exception
 java.lang.NullPointerException
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher$2.run(DomainSocketWatcher.java:463)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 The line 463 is following code snippet:
 {code}
  try {
 for (int fd : fdSet.getAndClearReadableFds()) {
   sendCallbackAndRemove(getAndClearReadableFds, entries, fdSet,
 fd);
 }
 {code}
 getAndClearReadableFds is a native method which will malloc an int array. 
 Since our memory is very tight, it looks like the malloc failed and a NULL 
 pointer is returned.
 The bad thing is that other threads then blocked in stack like this:
 {code}
 DataXceiver for client 
 unix:/home/work/app/hdfs/c3prc-micloud/datanode/dn_socket [Waiting for 
 operation #1] daemon prio=10 tid=0x7f0c9c086d90 nid=0x8fc3 waiting on 
 condition [0x7f09b9856000]
java.lang.Thread.State: WAITING (parking)
 at sun.misc.Unsafe.park(Native Method)
 - parking to wait for  0x0007b0174808 (a 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
 at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
 at 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher.add(DomainSocketWatcher.java:323)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.createNewMemorySegment(ShortCircuitRegistry.java:322)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.requestShortCircuitShm(DataXceiver.java:403)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opRequestShortCircuitShm(Receiver.java:214)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:95)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 IMO, we should exit the DN so that the users can know that something go  
 wrong  and fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8429) Death of watcherThread making other local read blocked

2015-05-20 Thread zhouyingchao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

zhouyingchao updated HDFS-8429:
---
Attachment: HDFS-8429-001.patch

Test following cases  : 
TestDomainSocket,TestDomainSocketWatcher,TestParallelShortCircuitRead,TestFsDatasetCacheRevocation,TestFatasetCacheRevocation,TestScrLazyPersistFiles,TestParallelShortCircuitReadNoChecksum,TestDFSInputStream,TestBlockReaderFactory,TestParallelUnixDomainRead,TestParallelShortCircuitReadUnCached,TestBlockReaderLocalLegacy,TestPeerCache,TestShortCircuitCache,TestShortCircuitLocalRead,TestBlockReaderLocal,TestParallelShortCircuitLegacyRead,TestTracingShortCircuitLocalRead,TestEnhancedByteBufferAccess

 Death of watcherThread making other local read blocked
 --

 Key: HDFS-8429
 URL: https://issues.apache.org/jira/browse/HDFS-8429
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: zhouyingchao
Assignee: zhouyingchao
 Attachments: HDFS-8429-001.patch


 In our cluster, an application is hung when doing a short circuit read of 
 local hdfs block. By looking into the log, we found the DataNode's 
 DomainSocketWatcher.watcherThread has exited with following log:
 {code}
 ERROR org.apache.hadoop.net.unix.DomainSocketWatcher: 
 Thread[Thread-25,5,main] terminating on unexpected exception
 java.lang.NullPointerException
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher$2.run(DomainSocketWatcher.java:463)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 The line 463 is following code snippet:
 {code}
  try {
 for (int fd : fdSet.getAndClearReadableFds()) {
   sendCallbackAndRemove(getAndClearReadableFds, entries, fdSet,
 fd);
 }
 {code}
 getAndClearReadableFds is a native method which will malloc an int array. 
 Since our memory is very tight, it looks like the malloc failed and a NULL 
 pointer is returned.
 The bad thing is that other threads then blocked in stack like this:
 {code}
 DataXceiver for client 
 unix:/home/work/app/hdfs/c3prc-micloud/datanode/dn_socket [Waiting for 
 operation #1] daemon prio=10 tid=0x7f0c9c086d90 nid=0x8fc3 waiting on 
 condition [0x7f09b9856000]
java.lang.Thread.State: WAITING (parking)
 at sun.misc.Unsafe.park(Native Method)
 - parking to wait for  0x0007b0174808 (a 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
 at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
 at 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher.add(DomainSocketWatcher.java:323)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.createNewMemorySegment(ShortCircuitRegistry.java:322)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.requestShortCircuitShm(DataXceiver.java:403)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opRequestShortCircuitShm(Receiver.java:214)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:95)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 IMO, we should exit the DN so that the users can know that something go  
 wrong  and fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8429) Death of watcherThread making other local read blocked

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552147#comment-14552147
 ] 

Hadoop QA commented on HDFS-8429:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 39s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 31s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 42s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m  5s | The applied patch generated  2 
new checkstyle issues (total was 19, now 21). |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 34s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   1m 40s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | common tests |  22m 53s | Tests passed in 
hadoop-common. |
| | |  60m  2s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734105/HDFS-8429-001.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / ce53c8e |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/11061/artifact/patchprocess/diffcheckstylehadoop-common.txt
 |
| hadoop-common test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11061/artifact/patchprocess/testrun_hadoop-common.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11061/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf907.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11061/console |


This message was automatically generated.

 Death of watcherThread making other local read blocked
 --

 Key: HDFS-8429
 URL: https://issues.apache.org/jira/browse/HDFS-8429
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: zhouyingchao
Assignee: zhouyingchao
 Attachments: HDFS-8429-001.patch


 In our cluster, an application is hung when doing a short circuit read of 
 local hdfs block. By looking into the log, we found the DataNode's 
 DomainSocketWatcher.watcherThread has exited with following log:
 {code}
 ERROR org.apache.hadoop.net.unix.DomainSocketWatcher: 
 Thread[Thread-25,5,main] terminating on unexpected exception
 java.lang.NullPointerException
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher$2.run(DomainSocketWatcher.java:463)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 The line 463 is following code snippet:
 {code}
  try {
 for (int fd : fdSet.getAndClearReadableFds()) {
   sendCallbackAndRemove(getAndClearReadableFds, entries, fdSet,
 fd);
 }
 {code}
 getAndClearReadableFds is a native method which will malloc an int array. 
 Since our memory is very tight, it looks like the malloc failed and a NULL 
 pointer is returned.
 The bad thing is that other threads then blocked in stack like this:
 {code}
 DataXceiver for client 
 unix:/home/work/app/hdfs/c3prc-micloud/datanode/dn_socket [Waiting for 
 operation #1] daemon prio=10 tid=0x7f0c9c086d90 nid=0x8fc3 waiting on 
 condition [0x7f09b9856000]
java.lang.Thread.State: WAITING (parking)
 at sun.misc.Unsafe.park(Native Method)
 - parking to wait for  0x0007b0174808 (a 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
 at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
 at 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher.add(DomainSocketWatcher.java:323)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.createNewMemorySegment(ShortCircuitRegistry.java:322)
 at 
 

[jira] [Updated] (HDFS-8420) Erasure Coding: ECZoneManager#getECZoneInfo is not resolving the path properly if zone dir itself is the snapshottable dir

2015-05-20 Thread Rakesh R (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rakesh R updated HDFS-8420:
---
Attachment: HDFS-8320-HDFS-7285-01.patch

 Erasure Coding: ECZoneManager#getECZoneInfo is not resolving the path 
 properly if zone dir itself is the snapshottable dir
 --

 Key: HDFS-8420
 URL: https://issues.apache.org/jira/browse/HDFS-8420
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Rakesh R
Assignee: Rakesh R
 Attachments: HDFS-8320-HDFS-7285-00.patch, 
 HDFS-8320-HDFS-7285-01.patch


 Presently the resultant zone dir will come with {{.snapshot}} only when the 
 zone dir itself is snapshottable dir. It will return the path including the 
 snapshot name like, {{/zone/.snapshot/snap1}}. Instead could improve this by 
 returning only path {{/zone}}.
 Thanks [~vinayrpet] for the helpful 
 [discussion|https://issues.apache.org/jira/browse/HDFS-8266?focusedCommentId=14543821page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14543821]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8408) Revisit and refactor ErasureCodingInfo

2015-05-20 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8408?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552220#comment-14552220
 ] 

Tsz Wo Nicholas Sze commented on HDFS-8408:
---

{quote}
There is EncryptionZone.java which provides details about EncryptionZone, 
returned with listEncryptionZones().

So IMO, keeping ErasureCodingZoneInfo would be helpful atleast for admin API. 
If we can include zoneDir also in HdfsFileStatus, then this may be unnecessary.
{quote}
Thanks for clarifying it.  I agree that we should keep something like 
ErasureCodingZoneInfo.  However, how about renaming it to ErasureCodingZone?  
The ec zone related code should correspond to encryption zone code.  Otherwise, 
it will be quite confusing.

 Revisit and refactor ErasureCodingInfo
 --

 Key: HDFS-8408
 URL: https://issues.apache.org/jira/browse/HDFS-8408
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Vinayakumar B
Assignee: Vinayakumar B

 As mentioned in HDFS-8375 
 [here|https://issues.apache.org/jira/browse/HDFS-8375?focusedCommentId=14544618page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14544618]
  
 {{ErasureCodingInfo}} needs a revisit.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8408) Revisit and refactor ErasureCodingInfo

2015-05-20 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8408?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552059#comment-14552059
 ] 

Vinayakumar B commented on HDFS-8408:
-

There is a slight difference in directly using the HdfsFileStatus.
HdfsFileStatus as of now carries schema and cellSize, but it doesnot carry the 
zonedir. In Serverside {{zone dir}} is the one which stores the schema and 
cellsize as xattr. And these zonedirs are not tracked separately in any other 
data structure as in EncryptionZoneManager.

As of now, renames are not allowed from one zonedir to another zonedir as 
encoding/decoding may have issues, but we can move zone dir's parent dir itself.
So using ErasureCodingZoneInfo admin can get the zone dir and decide to move 
zonedir itself

bq. Note that we neither have StorageTypeInfo for StorageType nor 
EncryptionZoneInfo for EncryptionZone
There is {{EncryptionZone.java}} which provides details about EncryptionZone, 
returned with {{listEncryptionZones()}}.


So IMO, keeping ErasureCodingZoneInfo would be helpful atleast for admin API. 
If we can include zoneDir also in HdfsFileStatus, then this may be unnecessary.

 Revisit and refactor ErasureCodingInfo
 --

 Key: HDFS-8408
 URL: https://issues.apache.org/jira/browse/HDFS-8408
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Vinayakumar B
Assignee: Vinayakumar B

 As mentioned in HDFS-8375 
 [here|https://issues.apache.org/jira/browse/HDFS-8375?focusedCommentId=14544618page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14544618]
  
 {{ErasureCodingInfo}} needs a revisit.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina reassigned HDFS-8443:


Assignee: J.Andreina

 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina

 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8382) Remove chunkSize parameter from initialize method of raw erasure coder

2015-05-20 Thread Kai Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Zheng updated HDFS-8382:

Attachment: HDFS-8382-HDFS-7285-v3.patch

Updated the patch also removing {{initialize}} method per the suggestion.

 Remove chunkSize parameter from initialize method of raw erasure coder
 --

 Key: HDFS-8382
 URL: https://issues.apache.org/jira/browse/HDFS-8382
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Zheng
 Attachments: HDFS-8382-HDFS-7285-v1.patch, 
 HDFS-8382-HDFS-7285-v2.patch, HDFS-8382-HDFS-7285-v3.patch


 Per discussion in HDFS-8347, we need to support encoding/decoding variable 
 width units data instead of predefined fixed width like {{chunkSize}}. Have 
 this issue to remove chunkSize in the general raw erasure coder API. Specific 
 coder will support fixed chunkSize using hard-coded or specific schema 
 customizing way if necessary, like HitchHiker coder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8382) Remove chunkSize parameter from initialize method of raw erasure coder

2015-05-20 Thread Kai Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Zheng updated HDFS-8382:

Status: Open  (was: Patch Available)

The patch pending on HADOOP-11847.

 Remove chunkSize parameter from initialize method of raw erasure coder
 --

 Key: HDFS-8382
 URL: https://issues.apache.org/jira/browse/HDFS-8382
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Zheng
 Attachments: HDFS-8382-HDFS-7285-v1.patch, 
 HDFS-8382-HDFS-7285-v2.patch, HDFS-8382-HDFS-7285-v3.patch


 Per discussion in HDFS-8347, we need to support encoding/decoding variable 
 width units data instead of predefined fixed width like {{chunkSize}}. Have 
 this issue to remove chunkSize in the general raw erasure coder API. Specific 
 coder will support fixed chunkSize using hard-coded or specific schema 
 customizing way if necessary, like HitchHiker coder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Walter Su (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Walter Su updated HDFS-8441:

Description: 
Changes:
1. {{UnsupportedActionException}} is more user-firendly.
2. check condition before update quota count
3. support create EC file with replication=0

  was:
Changes:
1. {{UnsupportedActionException}} is more user-firendly.
2. check condition before update quota count



 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update quota count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Walter Su (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Walter Su updated HDFS-8441:

Description: 
Changes:
1. {{UnsupportedActionException}} is more user-firendly.
2. check condition before update storage count
3. support create EC file with replication=0

  was:
Changes:
1. {{UnsupportedActionException}} is more user-firendly.
2. check condition before update quota count
3. support create EC file with replication=0


 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update storage count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Walter Su (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Walter Su updated HDFS-8441:

Attachment: HDFS-8441-HDFS-7285.002.patch

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update quota count



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Walter Su (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552069#comment-14552069
 ] 

Walter Su commented on HDFS-8441:
-

002 patch add support create EC file with replication=0

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update quota count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread J.Andreina (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14551970#comment-14551970
 ] 

J.Andreina commented on HDFS-8443:
--

I would like update the patch.

 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina

 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Work started] (HDFS-8382) Remove chunkSize parameter from initialize method of raw erasure coder

2015-05-20 Thread Kai Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8382?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Work on HDFS-8382 started by Kai Zheng.
---
 Remove chunkSize parameter from initialize method of raw erasure coder
 --

 Key: HDFS-8382
 URL: https://issues.apache.org/jira/browse/HDFS-8382
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Zheng
 Attachments: HDFS-8382-HDFS-7285-v1.patch, 
 HDFS-8382-HDFS-7285-v2.patch, HDFS-8382-HDFS-7285-v3.patch


 Per discussion in HDFS-8347, we need to support encoding/decoding variable 
 width units data instead of predefined fixed width like {{chunkSize}}. Have 
 this issue to remove chunkSize in the general raw erasure coder API. Specific 
 coder will support fixed chunkSize using hard-coded or specific schema 
 customizing way if necessary, like HitchHiker coder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8375) Add cellSize as an XAttr to ECZone

2015-05-20 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8375?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552063#comment-14552063
 ] 

Vinayakumar B commented on HDFS-8375:
-

Thanks [~zhz].

 Add cellSize as an XAttr to ECZone
 --

 Key: HDFS-8375
 URL: https://issues.apache.org/jira/browse/HDFS-8375
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Fix For: HDFS-7285

 Attachments: HDFS-8375-HDFS-7285-01.patch, 
 HDFS-8375-HDFS-7285-02.patch, HDFS-8375-HDFS-7285-03.patch, 
 HDFS-8375-HDFS-7285-04.patch


 Add {{cellSize}} as an Xattr for ECZone. as discussed 
 [here|https://issues.apache.org/jira/browse/HDFS-8347?focusedCommentId=14539108page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14539108]



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-8443:
-
Attachment: HDFS-8443.1.patch

Attached an initial patch.
Please review.

 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina
 Attachments: HDFS-8443.1.patch


 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8382) Remove chunkSize parameter from initialize method of raw erasure coder

2015-05-20 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8382?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552089#comment-14552089
 ] 

Vinayakumar B commented on HDFS-8382:
-

I see that patch covers 'Raw' coders. But similar treatment to be done for 
'coder' classes also. Could it be done in this patch itself. ?

 Remove chunkSize parameter from initialize method of raw erasure coder
 --

 Key: HDFS-8382
 URL: https://issues.apache.org/jira/browse/HDFS-8382
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Zheng
 Attachments: HDFS-8382-HDFS-7285-v1.patch, 
 HDFS-8382-HDFS-7285-v2.patch, HDFS-8382-HDFS-7285-v3.patch


 Per discussion in HDFS-8347, we need to support encoding/decoding variable 
 width units data instead of predefined fixed width like {{chunkSize}}. Have 
 this issue to remove chunkSize in the general raw erasure coder API. Specific 
 coder will support fixed chunkSize using hard-coded or specific schema 
 customizing way if necessary, like HitchHiker coder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-8443:
-
Status: Patch Available  (was: Open)

 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina
 Attachments: HDFS-8443.1.patch


 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552227#comment-14552227
 ] 

Kai Zheng commented on HDFS-8441:
-

The patch looks good. One comment regarding the following change. Maybe we 
could add a method in {{FSDirectory#isInECZone(src)}} and use it to tell the 
condition?
{code}
+final ErasureCodingZoneInfo ecZoneInfo = dir.getECZoneInfo(iip);
+if (ecZoneInfo == null) {
+  blockManager.verifyReplication(src, replication, clientMachine);
+}
{code}

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update storage count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552252#comment-14552252
 ] 

Rakesh R commented on HDFS-8441:


Thanks [~walter.k.su] for taking care this. Just one comment:

1) Please add message to the {{fail()}} statement like,
{code}
+  fail(Shouldn't allow to set replication to a file with striped blocks);
{code}

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update storage count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-3854) Implement a fence method which should fence the BK shared storage.

2015-05-20 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-3854?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552283#comment-14552283
 ] 

Rakesh R commented on HDFS-3854:


[~umamaheswararao] like you mentioned, HDFS-3862 has a proposal to skip the 
fencer logic as BK has already a built-in fencing mechanism. Can I close this 
issue ?

 Implement a fence method which should fence the BK shared storage.
 --

 Key: HDFS-3854
 URL: https://issues.apache.org/jira/browse/HDFS-3854
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Uma Maheswara Rao G
Assignee: Rakesh R

 Currently when machine down or network down, SSHFence can not ensure that, 
 other node is completely down. So, fence will fail and switch will not happen.
 [ internally we did work around to return true when machine is not reachable, 
 as BKJM already has fencing]
 It may be good idea to implement a fence method, which should ensure shared 
 storage fenced propertly and return true.
 We can plug in this new method in ZKFC fence methods.
 only pain points what I can see is, we may have to put the BKJM jar in ZKFC 
 lib for running this fence method.
 thoughts?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8427) Remove dataBlockNum and parityBlockNum from BlockInfoStriped

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8427?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552292#comment-14552292
 ] 

Hadoop QA commented on HDFS-8427:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 46s | Pre-patch HDFS-7285 compilation 
is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 28s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 40s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 15s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   2m 14s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 35s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   3m 12s | The patch appears to introduce 6 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 14s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 171m 10s | Tests failed in hadoop-hdfs. |
| | | 214m 11s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
|  |  Inconsistent synchronization of 
org.apache.hadoop.hdfs.DFSOutputStream.streamer; locked 89% of time  
Unsynchronized access at DFSOutputStream.java:89% of time  Unsynchronized 
access at DFSOutputStream.java:[line 146] |
|  |  Possible null pointer dereference of arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:[line 194] |
|  |  Unread field:field be static?  At ErasureCodingWorker.java:[line 254] |
|  |  Should 
org.apache.hadoop.hdfs.server.datanode.erasurecode.ErasureCodingWorker$StripedReader
 be a _static_ inner class?  At ErasureCodingWorker.java:inner class?  At 
ErasureCodingWorker.java:[lines 907-914] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:[line 108] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:[line 409] |
| Failed unit tests | hadoop.hdfs.TestEncryptedTransfer |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
|   | hadoop.hdfs.server.blockmanagement.TestBlockInfo |
|   | hadoop.hdfs.server.namenode.TestFileTruncate |
|   | hadoop.hdfs.server.blockmanagement.TestReplicationPolicy |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734100/HDFS-8427-HDFS-7285-01.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | HDFS-7285 / 4dd4aa5 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11058/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11058/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11058/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11058/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11058/console |


This message was automatically generated.

 Remove dataBlockNum and parityBlockNum from BlockInfoStriped
 

 Key: 

[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552299#comment-14552299
 ] 

Hadoop QA commented on HDFS-8441:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  15m 20s | Pre-patch HDFS-7285 compilation 
is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 44s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 58s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 15s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 38s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 40s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   3m 17s | The patch appears to introduce 6 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 19s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 173m 12s | Tests failed in hadoop-hdfs. |
| | | 216m  4s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
|  |  Inconsistent synchronization of 
org.apache.hadoop.hdfs.DFSOutputStream.streamer; locked 89% of time  
Unsynchronized access at DFSOutputStream.java:89% of time  Unsynchronized 
access at DFSOutputStream.java:[line 146] |
|  |  Possible null pointer dereference of arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:[line 194] |
|  |  Unread field:field be static?  At ErasureCodingWorker.java:[line 254] |
|  |  Should 
org.apache.hadoop.hdfs.server.datanode.erasurecode.ErasureCodingWorker$StripedReader
 be a _static_ inner class?  At ErasureCodingWorker.java:inner class?  At 
ErasureCodingWorker.java:[lines 907-914] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:[line 108] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:[line 409] |
| Failed unit tests | hadoop.hdfs.server.namenode.TestFileTruncate |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS |
|   | hadoop.hdfs.TestEncryptedTransfer |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
|   | hadoop.hdfs.server.blockmanagement.TestBlockInfo |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734101/HDFS-8441-HDFS-7285.002.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | HDFS-7285 / 4dd4aa5 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11059/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11059/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11059/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11059/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf904.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11059/console |


This message was automatically generated.

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: 

[jira] [Commented] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552333#comment-14552333
 ] 

Hadoop QA commented on HDFS-8443:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 44s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 34s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 41s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 22s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 36s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | native |   3m 16s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 164m 13s | Tests failed in hadoop-hdfs. |
| | | 202m  4s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.server.namenode.TestFileTruncate |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734104/HDFS-8443.1.patch |
| Optional Tests | javadoc javac unit |
| git revision | trunk / ce53c8e |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11060/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11060/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf900.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11060/console |


This message was automatically generated.

 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina
 Attachments: HDFS-8443.1.patch


 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8404) Pending block replication can get stuck using older genstamp

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552600#comment-14552600
 ] 

Hudson commented on HDFS-8404:
--

SUCCESS: Integrated in Hadoop-Mapreduce-trunk #2149 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2149/])
HDFS-8404. Pending block replication can get stuck using older genstamp. 
Contributed by Nathan Roberts. (kihwal: rev 
8860e352c394372e4eb3ebdf82ea899567f34e4e)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestPendingReplication.java


 Pending block replication can get stuck using older genstamp
 

 Key: HDFS-8404
 URL: https://issues.apache.org/jira/browse/HDFS-8404
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.6.0, 2.7.0
Reporter: Nathan Roberts
Assignee: Nathan Roberts
 Fix For: 2.7.1

 Attachments: HDFS-8404-v0.patch, HDFS-8404-v1.patch


 If an under-replicated block gets into the pending-replication list, but 
 later the  genstamp of that block ends up being newer than the one originally 
 submitted for replication, the block will fail replication until the NN is 
 restarted. 
 It will be safer if processPendingReplications()  gets up-to-date blockinfo 
 before resubmitting replication work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8131) Implement a space balanced block placement policy

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552602#comment-14552602
 ] 

Hudson commented on HDFS-8131:
--

SUCCESS: Integrated in Hadoop-Mapreduce-trunk #2149 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2149/])
HDFS-8131. Implement a space balanced block placement policy. Contributed by 
Liu Shaohui. (kihwal: rev de30d66b2673d0344346fb985e786247ca682317)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/AvailableSpaceBlockPlacementPolicy.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestAvailableSpaceBlockPlacementPolicy.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockPlacementPolicyDefault.java


 Implement a space balanced block placement policy
 -

 Key: HDFS-8131
 URL: https://issues.apache.org/jira/browse/HDFS-8131
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.0.0
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
  Labels: BlockPlacementPolicy
 Fix For: 2.8.0

 Attachments: HDFS-8131-v1.diff, HDFS-8131-v2.diff, HDFS-8131-v3.diff, 
 HDFS-8131.004.patch, HDFS-8131.005.patch, HDFS-8131.006.patch, balanced.png


 The default block placement policy will choose datanodes for new blocks 
 randomly, which will result in unbalanced space used percent among datanodes 
 after an cluster expansion. The old datanodes always are in high used percent 
 of space and new added ones are in low percent.
 Through we can used the external balance tool to balance the space used rate, 
 it will cost extra network IO and it's not easy to control the balance speed.
 An easy solution is to implement an balanced block placement policy which 
 will choose low used percent datanodes for new blocks with a little high 
 possibility. In a not long term, the used percent of datanodes will trend to 
 be balanced.
 Suggestions and discussions are welcomed. Thanks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8404) Pending block replication can get stuck using older genstamp

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8404?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552556#comment-14552556
 ] 

Hudson commented on HDFS-8404:
--

SUCCESS: Integrated in Hadoop-Mapreduce-trunk-Java8 #201 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/201/])
HDFS-8404. Pending block replication can get stuck using older genstamp. 
Contributed by Nathan Roberts. (kihwal: rev 
8860e352c394372e4eb3ebdf82ea899567f34e4e)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockManager.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestPendingReplication.java


 Pending block replication can get stuck using older genstamp
 

 Key: HDFS-8404
 URL: https://issues.apache.org/jira/browse/HDFS-8404
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.6.0, 2.7.0
Reporter: Nathan Roberts
Assignee: Nathan Roberts
 Fix For: 2.7.1

 Attachments: HDFS-8404-v0.patch, HDFS-8404-v1.patch


 If an under-replicated block gets into the pending-replication list, but 
 later the  genstamp of that block ends up being newer than the one originally 
 submitted for replication, the block will fail replication until the NN is 
 restarted. 
 It will be safer if processPendingReplications()  gets up-to-date blockinfo 
 before resubmitting replication work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8131) Implement a space balanced block placement policy

2015-05-20 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552558#comment-14552558
 ] 

Hudson commented on HDFS-8131:
--

SUCCESS: Integrated in Hadoop-Mapreduce-trunk-Java8 #201 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/201/])
HDFS-8131. Implement a space balanced block placement policy. Contributed by 
Liu Shaohui. (kihwal: rev de30d66b2673d0344346fb985e786247ca682317)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/AvailableSpaceBlockPlacementPolicy.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockPlacementPolicyDefault.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/blockmanagement/TestAvailableSpaceBlockPlacementPolicy.java


 Implement a space balanced block placement policy
 -

 Key: HDFS-8131
 URL: https://issues.apache.org/jira/browse/HDFS-8131
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.0.0
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
  Labels: BlockPlacementPolicy
 Fix For: 2.8.0

 Attachments: HDFS-8131-v1.diff, HDFS-8131-v2.diff, HDFS-8131-v3.diff, 
 HDFS-8131.004.patch, HDFS-8131.005.patch, HDFS-8131.006.patch, balanced.png


 The default block placement policy will choose datanodes for new blocks 
 randomly, which will result in unbalanced space used percent among datanodes 
 after an cluster expansion. The old datanodes always are in high used percent 
 of space and new added ones are in low percent.
 Through we can used the external balance tool to balance the space used rate, 
 it will cost extra network IO and it's not easy to control the balance speed.
 An easy solution is to implement an balanced block placement policy which 
 will choose low used percent datanodes for new blocks with a little high 
 possibility. In a not long term, the used percent of datanodes will trend to 
 be balanced.
 Suggestions and discussions are welcomed. Thanks



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8444) Erasure Coding: fix cannot rename a zone dir

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552633#comment-14552633
 ] 

Hadoop QA commented on HDFS-8444:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  15m 11s | Pre-patch HDFS-7285 compilation 
is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 46s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 55s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 15s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 38s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 38s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   3m 16s | The patch appears to introduce 6 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 21s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 182m 35s | Tests failed in hadoop-hdfs. |
| | | 225m 12s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
|  |  Inconsistent synchronization of 
org.apache.hadoop.hdfs.DFSOutputStream.streamer; locked 89% of time  
Unsynchronized access at DFSOutputStream.java:89% of time  Unsynchronized 
access at DFSOutputStream.java:[line 146] |
|  |  Possible null pointer dereference of arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:[line 194] |
|  |  Unread field:field be static?  At ErasureCodingWorker.java:[line 254] |
|  |  Should 
org.apache.hadoop.hdfs.server.datanode.erasurecode.ErasureCodingWorker$StripedReader
 be a _static_ inner class?  At ErasureCodingWorker.java:inner class?  At 
ErasureCodingWorker.java:[lines 907-914] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:[line 108] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:[line 409] |
| Failed unit tests | hadoop.hdfs.TestEncryptedTransfer |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
|   | hadoop.hdfs.server.blockmanagement.TestBlockInfo |
|   | hadoop.hdfs.server.namenode.TestFileTruncate |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS |
|   | hadoop.hdfs.server.blockmanagement.TestReplicationPolicy |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734114/HDFS-8444-HDFS-7285.001.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | HDFS-7285 / 4dd4aa5 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11062/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11062/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11062/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11062/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf903.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11062/console |


This message was automatically generated.

 Erasure Coding: fix cannot rename a zone dir
 

 Key: HDFS-8444
 URL: https://issues.apache.org/jira/browse/HDFS-8444
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: 

[jira] [Commented] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552620#comment-14552620
 ] 

Arpit Agarwal commented on HDFS-8443:
-

Thanks for fixing this [~andreina]. I suggest a grammatical fix. +1 otherwise.

_thread that listens_ -- _threads that listen_.


 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina
 Attachments: HDFS-8443.1.patch


 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8420) Erasure Coding: ECZoneManager#getECZoneInfo is not resolving the path properly if zone dir itself is the snapshottable dir

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8420?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552622#comment-14552622
 ] 

Hadoop QA commented on HDFS-8420:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 42s | Pre-patch HDFS-7285 compilation 
is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 35s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 37s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 15s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 37s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 38s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   3m 14s | The patch appears to introduce 6 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 14s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 174m 24s | Tests failed in hadoop-hdfs. |
| | | 215m 54s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
|  |  Inconsistent synchronization of 
org.apache.hadoop.hdfs.DFSOutputStream.streamer; locked 89% of time  
Unsynchronized access at DFSOutputStream.java:89% of time  Unsynchronized 
access at DFSOutputStream.java:[line 146] |
|  |  Possible null pointer dereference of arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:[line 194] |
|  |  Unread field:field be static?  At ErasureCodingWorker.java:[line 254] |
|  |  Should 
org.apache.hadoop.hdfs.server.datanode.erasurecode.ErasureCodingWorker$StripedReader
 be a _static_ inner class?  At ErasureCodingWorker.java:inner class?  At 
ErasureCodingWorker.java:[lines 907-914] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:[line 108] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:[line 409] |
| Failed unit tests | hadoop.hdfs.server.blockmanagement.TestBlockInfo |
|   | hadoop.hdfs.server.namenode.TestFileTruncate |
|   | hadoop.hdfs.server.blockmanagement.TestBlockTokenWithDFS |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
|   | hadoop.hdfs.TestEncryptedTransfer |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734113/HDFS-8320-HDFS-7285-01.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | HDFS-7285 / 4dd4aa5 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11063/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11063/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11063/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11063/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf901.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11063/console |


This message was automatically generated.

 Erasure Coding: ECZoneManager#getECZoneInfo is not resolving the path 
 properly if zone dir itself is the snapshottable dir
 --

 Key: HDFS-8420
 URL: https://issues.apache.org/jira/browse/HDFS-8420

[jira] [Commented] (HDFS-8439) Adding more slow action log in critical read path

2015-05-20 Thread stack (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552970#comment-14552970
 ] 

stack commented on HDFS-8439:
-

s/cost/duration/g  I could also do on commit. I like how you log the culprit or 
pipeline. Are you running this in production [~liushaohui]?



 Adding more slow action log in critical read path
 -

 Key: HDFS-8439
 URL: https://issues.apache.org/jira/browse/HDFS-8439
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
 Attachments: HDFS-8439.001.patch


 To dig a HBase read spike issue, we'd better to add more slow pread/seek log 
 in read flow to get the abnormal datanodes.
 Patch will be uploaded soon.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8427) Remove dataBlockNum and parityBlockNum from BlockInfoStriped

2015-05-20 Thread Jing Zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8427?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jing Zhao updated HDFS-8427:

  Resolution: Fixed
Hadoop Flags: Reviewed
  Status: Resolved  (was: Patch Available)

+1. I've committed this to the feature branch. Thanks for the contribution, 
[~kaisasak]!

 Remove dataBlockNum and parityBlockNum from BlockInfoStriped
 

 Key: HDFS-8427
 URL: https://issues.apache.org/jira/browse/HDFS-8427
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-7285
Reporter: Kai Sasaki
Assignee: Kai Sasaki
 Fix For: HDFS-7285

 Attachments: HDFS-8427-HDFS-7285-00.patch, 
 HDFS-8427-HDFS-7285-01.patch


 Remove unnecessary members such as {{dataBlockNum}} and {{parityBlockNum}} 
 from {{BlockInfoStriped}}. These are included in {{ECShema}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8348) WebHDFS Concat - Support sources list passed a POST body

2015-05-20 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8348?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan updated HDFS-8348:
--
Summary: WebHDFS Concat - Support sources list passed a POST body  (was: 
Concat - Support sources list passed a POST body)

 WebHDFS Concat - Support sources list passed a POST body
 

 Key: HDFS-8348
 URL: https://issues.apache.org/jira/browse/HDFS-8348
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: webhdfs
Reporter: vishwajeet dusane

 Problem - Current webhdfs behavior for concat support sources list to be 
 passed as query parameter. This approach limits on the number of sources list 
 send as query parameter. 
 Proposed Solution - Add support to send sources list as part of the request 
 body instead of query parameter.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8446) Separate safemode related operations in GetBlockLocations()

2015-05-20 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-8446:
-
Attachment: HDFS-8446.000.patch

 Separate safemode related operations in GetBlockLocations()
 ---

 Key: HDFS-8446
 URL: https://issues.apache.org/jira/browse/HDFS-8446
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
Priority: Minor
 Attachments: HDFS-8446.000.patch


 Currently {{FSNamesystem#GetBlockLocations()}} has some special cases when 
 the NN is in SafeMode. This jira proposes to refactor the code to improve 
 readability.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a command to get the bandwidth of balancer

2015-05-20 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552774#comment-14552774
 ] 

Rakesh R commented on HDFS-7116:


Hi [~ajisakaa], its really an interesting feature and would like to take this 
ahead. To begin with, I have gone through the {{-setBalancerBandwidth}} 
feature. Here, Namenode is sending the {{DNA_BALANCERBANDWIDTHUPDATE}} command 
to the Datanodes as a heartbeat response and Datanode will update the new 
bandwidth value. In normal case, after setting the bandwidth all the datanodes 
will have the same value. But I could see few corner cases where all the 
Datanodes in a cluster not having unique bandwidth value.

Case-1) Newly added Datanodes will not aware about the new value and will have 
the default bandwidth value. 
Case-2) Assume we have 10 Datanodes, after sending heartbeat responses to 1 to 
5 Datanodes, unfortunately Namenode restarted. Since bandwidth value is not 
persisted restarted Namenode will not have the new bandwidth value. Now, the 
rest of the 6 to 10 Datanodes will be having old value.

Is this expected behavior?. If yes, we have to design the 
{{-getBalancerBandwidth}} command in such a way to list each Datanode with the 
corresponding bandwidth value like, {{each Datanode = bandwidth value}}. Any 
thoughts?

 Add a command to get the bandwidth of balancer
 --

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA

 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the parameter via CLI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8429) Death of watcherThread making other local read blocked

2015-05-20 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8429?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14552875#comment-14552875
 ] 

Colin Patrick McCabe commented on HDFS-8429:


Thanks.  We shouldn't need to modify {{DomainSocketWatcher#add}} and 
{{DomainSocketWatcher#remove}}, since the finally block clears {{toAdd}} and 
{{toRemove}} prior to calling {{signalAll}}.  The modifications to the finally 
block look reasonable.  I think we are going to need a unit test for this as 
well.  Perhaps simply send an interrupt to a thread in the unit test.

 Death of watcherThread making other local read blocked
 --

 Key: HDFS-8429
 URL: https://issues.apache.org/jira/browse/HDFS-8429
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.6.0
Reporter: zhouyingchao
Assignee: zhouyingchao
 Attachments: HDFS-8429-001.patch


 In our cluster, an application is hung when doing a short circuit read of 
 local hdfs block. By looking into the log, we found the DataNode's 
 DomainSocketWatcher.watcherThread has exited with following log:
 {code}
 ERROR org.apache.hadoop.net.unix.DomainSocketWatcher: 
 Thread[Thread-25,5,main] terminating on unexpected exception
 java.lang.NullPointerException
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher$2.run(DomainSocketWatcher.java:463)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 The line 463 is following code snippet:
 {code}
  try {
 for (int fd : fdSet.getAndClearReadableFds()) {
   sendCallbackAndRemove(getAndClearReadableFds, entries, fdSet,
 fd);
 }
 {code}
 getAndClearReadableFds is a native method which will malloc an int array. 
 Since our memory is very tight, it looks like the malloc failed and a NULL 
 pointer is returned.
 The bad thing is that other threads then blocked in stack like this:
 {code}
 DataXceiver for client 
 unix:/home/work/app/hdfs/c3prc-micloud/datanode/dn_socket [Waiting for 
 operation #1] daemon prio=10 tid=0x7f0c9c086d90 nid=0x8fc3 waiting on 
 condition [0x7f09b9856000]
java.lang.Thread.State: WAITING (parking)
 at sun.misc.Unsafe.park(Native Method)
 - parking to wait for  0x0007b0174808 (a 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject)
 at java.util.concurrent.locks.LockSupport.park(LockSupport.java:156)
 at 
 java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQueuedSynchronizer.java:1987)
 at 
 org.apache.hadoop.net.unix.DomainSocketWatcher.add(DomainSocketWatcher.java:323)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.createNewMemorySegment(ShortCircuitRegistry.java:322)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.requestShortCircuitShm(DataXceiver.java:403)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opRequestShortCircuitShm(Receiver.java:214)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:95)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:235)
 at java.lang.Thread.run(Thread.java:662)
 {code}
 IMO, we should exit the DN so that the users can know that something go  
 wrong  and fix it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8446) Separate safemode related operations in GetBlockLocations()

2015-05-20 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-8446:


 Summary: Separate safemode related operations in 
GetBlockLocations()
 Key: HDFS-8446
 URL: https://issues.apache.org/jira/browse/HDFS-8446
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
Priority: Minor


Currently {{FSNamesystem#GetBlockLocations()}} has some special cases when the 
NN is in SafeMode. This jira proposes to refactor the code to improve 
readability.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8446) Separate safemode related operations in GetBlockLocations()

2015-05-20 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-8446:
-
Status: Patch Available  (was: Open)

 Separate safemode related operations in GetBlockLocations()
 ---

 Key: HDFS-8446
 URL: https://issues.apache.org/jira/browse/HDFS-8446
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
Priority: Minor
 Attachments: HDFS-8446.000.patch


 Currently {{FSNamesystem#GetBlockLocations()}} has some special cases when 
 the NN is in SafeMode. This jira proposes to refactor the code to improve 
 readability.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8294) Erasure Coding: Fix Findbug warnings present in erasure coding

2015-05-20 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8294?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553247#comment-14553247
 ] 

Zhe Zhang commented on HDFS-8294:
-

Good point Rakesh. I agree. Could you update the patch just for the new 
{{StripedBlockUtil}} then?

 Erasure Coding: Fix Findbug warnings present in erasure coding
 --

 Key: HDFS-8294
 URL: https://issues.apache.org/jira/browse/HDFS-8294
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Rakesh R
Assignee: Rakesh R
  Labels: BB2015-05-RFC
 Attachments: FindBugs Report in EC feature.html, 
 HDFS-8294-HDFS-7285.00.patch, HDFS-8294-HDFS-7285.01.patch, 
 HDFS-8294-HDFS-7285.02.patch, HDFS-8294-HDFS-7285.03.patch, 
 HDFS-8294-HDFS-7285.04.patch, HDFS-8294-HDFS-7285.05.patch


 This jira is to address the findbug issues reported in erasure coding feature.
 Attached sheet which contains the details of the findbug issues reported in 
 the erasure coding feature. I've taken this report from the jenkins build : 
 https://builds.apache.org/job/PreCommit-HDFS-Build/10848/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8444) Erasure Coding: fix cannot rename a zone dir

2015-05-20 Thread Walter Su (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8444?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553474#comment-14553474
 ] 

Walter Su commented on HDFS-8444:
-

jenkins not related. Please review.

 Erasure Coding: fix cannot rename a zone dir
 

 Key: HDFS-8444
 URL: https://issues.apache.org/jira/browse/HDFS-8444
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8444-HDFS-7285.001.patch


 We create a EC zone {{/my_ec_zone}}.
 We want to rename it to {{/myZone}}.
 But it failed.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8421) Move startFile() and related operations into FSDirWriteFileOp

2015-05-20 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553416#comment-14553416
 ] 

Jing Zhao commented on HDFS-8421:
-

# We do not need to remove the following check from the beginning of the 
startFile call.
{code}
checkOperation(OperationCategory.WRITE);
{code}
# The safemode check should only be in the real file creation section
# The following log seems unnecessary to me: in case of Standby NN this change 
will cause a lot of warning msg in the log
{code}
+} catch (IOException e) {
+  NameNode.stateChangeLog.warn(DIR* NameSystem.startFile:  + src +   +
+  e.getMessage());
{code}

 Move startFile() and related operations into FSDirWriteFileOp
 -

 Key: HDFS-8421
 URL: https://issues.apache.org/jira/browse/HDFS-8421
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-8421.000.patch, HDFS-8421.001.patch


 This jira proposes to move startFile() and related functions into 
 FSDirWriteFileOp.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8421) Move startFile() and related operations into FSDirWriteFileOp

2015-05-20 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8421?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553421#comment-14553421
 ] 

Jing Zhao commented on HDFS-8421:
-

+1 after addressing the comments.

 Move startFile() and related operations into FSDirWriteFileOp
 -

 Key: HDFS-8421
 URL: https://issues.apache.org/jira/browse/HDFS-8421
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-8421.000.patch, HDFS-8421.001.patch


 This jira proposes to move startFile() and related functions into 
 FSDirWriteFileOp.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8448) Create REST Interface

2015-05-20 Thread Anu Engineer (JIRA)
Anu Engineer created HDFS-8448:
--

 Summary: Create REST Interface
 Key: HDFS-8448
 URL: https://issues.apache.org/jira/browse/HDFS-8448
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Anu Engineer
Assignee: Anu Engineer


Create REST interfaces as specified in the architecture document.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8344) NameNode doesn't recover lease for files with missing blocks

2015-05-20 Thread Ravi Prakash (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravi Prakash updated HDFS-8344:
---
Status: Patch Available  (was: Open)

 NameNode doesn't recover lease for files with missing blocks
 

 Key: HDFS-8344
 URL: https://issues.apache.org/jira/browse/HDFS-8344
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.0
Reporter: Ravi Prakash
Assignee: Ravi Prakash
 Attachments: HDFS-8344.01.patch, HDFS-8344.02.patch, 
 HDFS-8344.03.patch


 I found another\(?) instance in which the lease is not recovered. This is 
 reproducible easily on a pseudo-distributed single node cluster
 # Before you start it helps if you set. This is not necessary, but simply 
 reduces how long you have to wait
 {code}
   public static final long LEASE_SOFTLIMIT_PERIOD = 30 * 1000;
   public static final long LEASE_HARDLIMIT_PERIOD = 2 * 
 LEASE_SOFTLIMIT_PERIOD;
 {code}
 # Client starts to write a file. (could be less than 1 block, but it hflushed 
 so some of the data has landed on the datanodes) (I'm copying the client code 
 I am using. I generate a jar and run it using $ hadoop jar TestHadoop.jar)
 # Client crashes. (I simulate this by kill -9 the $(hadoop jar 
 TestHadoop.jar) process after it has printed Wrote to the bufferedWriter
 # Shoot the datanode. (Since I ran on a pseudo-distributed cluster, there was 
 only 1)
 I believe the lease should be recovered and the block should be marked 
 missing. However this is not happening. The lease is never recovered.
 The effect of this bug for us was that nodes could not be decommissioned 
 cleanly. Although we knew that the client had crashed, the Namenode never 
 released the leases (even after restarting the Namenode) (even months 
 afterwards). There are actually several other cases too where we don't 
 consider what happens if ALL the datanodes die while the file is being 
 written, but I am going to punt on that for another time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a command to get the bandwidth of balancer

2015-05-20 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553425#comment-14553425
 ] 

Allen Wittenauer commented on HDFS-7116:


Why not just add this to dfsadmin -report?

 Add a command to get the bandwidth of balancer
 --

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA

 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the parameter via CLI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8448) Create REST Interface for Volumes

2015-05-20 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-8448:
---
Description: 
Create REST interfaces as specified in the architecture document.
This Jira is for creating the Volume Interface

  was:Create REST interfaces as specified in the architecture document.

Summary: Create REST Interface for Volumes  (was: Create REST Interface)

 Create REST Interface for Volumes
 -

 Key: HDFS-8448
 URL: https://issues.apache.org/jira/browse/HDFS-8448
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Anu Engineer
Assignee: Anu Engineer

 Create REST interfaces as specified in the architecture document.
 This Jira is for creating the Volume Interface



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7609) startup used too much time to load edits

2015-05-20 Thread Jing Zhao (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7609?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553292#comment-14553292
 ] 

Jing Zhao commented on HDFS-7609:
-

Thanks for working on this, [~mingma]! Your analysis makes sense to me and I 
can also reproduce the same scenario. Your patch also looks good to me.

One extra issue, which exists before your patch, is that the retry cache may 
not be hit during the NameNode failover. For example, suppose the following 
event sequence:
# a delete request is served by the ANN but the client fails to receive the 
response 
# NN failover happens
# the first retry request gets StandbyException since the old ANN becomes 
standby
# the second retry request is sent to the other NN, which at this time has not 
started the transition yet
# no cached entry has been found in the retry cache
# before running {{FSNamesystem#delete}}, the transition starts, which blocks 
the {{FSNamesystem#delete}} call
# the {{FSNamesystem#delete}} is called and fails

If the above scenario stands, maybe we can use this chance to fix it? In your 
patch, if the NN is in standby state, and there is no retry cache entry, should 
we directly throw StandbyException instead of checking it again in 
FSNamesystem? 

 startup used too much time to load edits
 

 Key: HDFS-7609
 URL: https://issues.apache.org/jira/browse/HDFS-7609
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 2.2.0
Reporter: Carrey Zhan
Assignee: Ming Ma
  Labels: BB2015-05-RFC
 Attachments: HDFS-7609-CreateEditsLogWithRPCIDs.patch, 
 HDFS-7609.patch, recovery_do_not_use_retrycache.patch


 One day my namenode crashed because of two journal node timed out at the same 
 time under very high load, leaving behind about 100 million transactions in 
 edits log.(I still have no idea why they were not rolled into fsimage.)
 I tryed to restart namenode, but it showed that almost 20 hours would be 
 needed before finish, and it was loading fsedits most of the time. I also 
 tryed to restart namenode in recover mode, the loading speed had no different.
 I looked into the stack trace, judged that it is caused by the retry cache. 
 So I set dfs.namenode.enable.retrycache to false, the restart process 
 finished in half an hour.
 I think the retry cached is useless during startup, at least during recover 
 process.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8210) Ozone: Implement storage container manager

2015-05-20 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HDFS-8210:

Hadoop Flags: Reviewed

Thank you for the new patch, Jitendra.  The test failure is unrelated.  The 
checkstyle nits are mostly not actionable, but we could fix these two:

{code}
./hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/storagecontainer/protocol/StorageContainer.java:23:
 First sentence should end with a period.
./hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/storagecontainer/StorageContainerMap.java:86:
 First sentence should end with a period.
{code}

+1 for committing to the feature branch after those are addressed.  Thanks!

 Ozone: Implement storage container manager 
 ---

 Key: HDFS-8210
 URL: https://issues.apache.org/jira/browse/HDFS-8210
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HDFS-8210-HDFS-7240.1.patch, 
 HDFS-8210-HDFS-7240.2.patch, HDFS-8210-HDFS-7240.3.patch


 The storage container manager collects datanode heartbeats, manages 
 replication and exposes API to lookup containers. This jira implements 
 storage container manager by re-using the block manager implementation in 
 namenode. This jira provides initial implementation that works with 
 datanodes. The additional protocols will be added in subsequent jiras.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a command to get the bandwidth of balancer

2015-05-20 Thread Kengo Seki (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553399#comment-14553399
 ] 

Kengo Seki commented on HDFS-7116:
--

These are expected behavior, so displaying bandwidth values for each DataNodes 
seems appropriate. In addition, it may be useful that the command can take a 
list of DataNodes to be processed, in case the cluster is large.

 Add a command to get the bandwidth of balancer
 --

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA

 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the parameter via CLI.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8344) NameNode doesn't recover lease for files with missing blocks

2015-05-20 Thread Ravi Prakash (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8344?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Ravi Prakash updated HDFS-8344:
---
Attachment: HDFS-8344.03.patch

Hi Kihwal! Here's a patch which improves the test to ensure that data that was 
once hsynced is read back correctly. Could you please review it?

 NameNode doesn't recover lease for files with missing blocks
 

 Key: HDFS-8344
 URL: https://issues.apache.org/jira/browse/HDFS-8344
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.0
Reporter: Ravi Prakash
Assignee: Ravi Prakash
 Attachments: HDFS-8344.01.patch, HDFS-8344.02.patch, 
 HDFS-8344.03.patch


 I found another\(?) instance in which the lease is not recovered. This is 
 reproducible easily on a pseudo-distributed single node cluster
 # Before you start it helps if you set. This is not necessary, but simply 
 reduces how long you have to wait
 {code}
   public static final long LEASE_SOFTLIMIT_PERIOD = 30 * 1000;
   public static final long LEASE_HARDLIMIT_PERIOD = 2 * 
 LEASE_SOFTLIMIT_PERIOD;
 {code}
 # Client starts to write a file. (could be less than 1 block, but it hflushed 
 so some of the data has landed on the datanodes) (I'm copying the client code 
 I am using. I generate a jar and run it using $ hadoop jar TestHadoop.jar)
 # Client crashes. (I simulate this by kill -9 the $(hadoop jar 
 TestHadoop.jar) process after it has printed Wrote to the bufferedWriter
 # Shoot the datanode. (Since I ran on a pseudo-distributed cluster, there was 
 only 1)
 I believe the lease should be recovered and the block should be marked 
 missing. However this is not happening. The lease is never recovered.
 The effect of this bug for us was that nodes could not be decommissioned 
 cleanly. Although we knew that the client had crashed, the Namenode never 
 released the leases (even after restarting the Namenode) (even months 
 afterwards). There are actually several other cases too where we don't 
 consider what happens if ALL the datanodes die while the file is being 
 written, but I am going to punt on that for another time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HDFS-8435) createNonRecursive support needed in WebHdfsFileSystem to support HBase

2015-05-20 Thread Jakob Homan (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8435?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jakob Homan reassigned HDFS-8435:
-

Assignee: Jakob Homan

 createNonRecursive support needed in WebHdfsFileSystem to support HBase
 ---

 Key: HDFS-8435
 URL: https://issues.apache.org/jira/browse/HDFS-8435
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: webhdfs
Reporter: Vinoth Sathappan
Assignee: Jakob Homan

 The WebHdfsFileSystem implementation doesn't support createNonRecursive. 
 HBase extensively depends on that for proper functioning. Currently, when the 
 region servers are started over web hdfs, they crash due with -
 createNonRecursive unsupported for this filesystem class 
 org.apache.hadoop.hdfs.web.SWebHdfsFileSystem
 at 
 org.apache.hadoop.fs.FileSystem.createNonRecursive(FileSystem.java:1137)
 at 
 org.apache.hadoop.fs.FileSystem.createNonRecursive(FileSystem.java:1112)
 at 
 org.apache.hadoop.fs.FileSystem.createNonRecursive(FileSystem.java:1088)
 at 
 org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter.init(ProtobufLogWriter.java:85)
 at 
 org.apache.hadoop.hbase.regionserver.wal.HLogFactory.createWriter(HLogFactory.java:198)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8447) Decouple information of files in GetLocatedBlocks

2015-05-20 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-8447:
-
Issue Type: Improvement  (was: Bug)

 Decouple information of files in GetLocatedBlocks
 -

 Key: HDFS-8447
 URL: https://issues.apache.org/jira/browse/HDFS-8447
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-8447.000.patch


 The current implementation of {{BlockManager.getLocatedBlocks()}} requires 
 the information of files to be passed as parameters. These information does 
 not affect the results of getting the physical locations of blocks.
 This jira proposes to refactor the call so that 
 {{BlockManager.getLocatedBlocks()}} depends only on the block information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8447) Decouple information of files in GetLocatedBlocks

2015-05-20 Thread Haohui Mai (JIRA)
Haohui Mai created HDFS-8447:


 Summary: Decouple information of files in GetLocatedBlocks
 Key: HDFS-8447
 URL: https://issues.apache.org/jira/browse/HDFS-8447
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai


The current implementation of {{BlockManager.getLocatedBlocks()}} requires the 
information of files to be passed as parameters. These information does not 
affect the results of getting the physical locations of blocks.

This jira proposes to refactor the call so that 
{{BlockManager.getLocatedBlocks()}} depends only on the block information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8447) Decouple information of files in GetLocatedBlocks

2015-05-20 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-8447:
-
Attachment: HDFS-8447.000.patch

 Decouple information of files in GetLocatedBlocks
 -

 Key: HDFS-8447
 URL: https://issues.apache.org/jira/browse/HDFS-8447
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-8447.000.patch


 The current implementation of {{BlockManager.getLocatedBlocks()}} requires 
 the information of files to be passed as parameters. These information does 
 not affect the results of getting the physical locations of blocks.
 This jira proposes to refactor the call so that 
 {{BlockManager.getLocatedBlocks()}} depends only on the block information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8447) Decouple information of files in GetLocatedBlocks

2015-05-20 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-8447:
-
Status: Patch Available  (was: Open)

 Decouple information of files in GetLocatedBlocks
 -

 Key: HDFS-8447
 URL: https://issues.apache.org/jira/browse/HDFS-8447
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-8447.000.patch


 The current implementation of {{BlockManager.getLocatedBlocks()}} requires 
 the information of files to be passed as parameters. These information does 
 not affect the results of getting the physical locations of blocks.
 This jira proposes to refactor the call so that 
 {{BlockManager.getLocatedBlocks()}} depends only on the block information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8446) Separate safemode related operations in GetBlockLocations()

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8446?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553328#comment-14553328
 ] 

Hadoop QA commented on HDFS-8446:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 36s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 32s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 38s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 22s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 51s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 36s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   3m  6s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 16s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 164m 25s | Tests failed in hadoop-hdfs. |
| | | 205m 59s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.TestEncryptionZones |
|   | hadoop.hdfs.TestReservedRawPaths |
|   | hadoop.hdfs.TestEncryptionZonesWithKMS |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734210/HDFS-8446.000.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 4aa730c |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11064/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11064/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf901.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11064/console |


This message was automatically generated.

 Separate safemode related operations in GetBlockLocations()
 ---

 Key: HDFS-8446
 URL: https://issues.apache.org/jira/browse/HDFS-8446
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
Priority: Minor
 Attachments: HDFS-8446.000.patch


 Currently {{FSNamesystem#GetBlockLocations()}} has some special cases when 
 the NN is in SafeMode. This jira proposes to refactor the code to improve 
 readability.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8448) Create REST Interface for Volumes

2015-05-20 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-8448:
---
Attachment: hdfs-8448-hdfs-7240.001.patch

This patch creates a Volume Interface for the Hadoop Object Store.



 Create REST Interface for Volumes
 -

 Key: HDFS-8448
 URL: https://issues.apache.org/jira/browse/HDFS-8448
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Anu Engineer
Assignee: Anu Engineer
 Attachments: hdfs-8448-hdfs-7240.001.patch


 Create REST interfaces as specified in the architecture document.
 This Jira is for creating the Volume Interface



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8448) Create REST Interface for Volumes

2015-05-20 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8448?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer updated HDFS-8448:
---
Status: Patch Available  (was: Open)

 Create REST Interface for Volumes
 -

 Key: HDFS-8448
 URL: https://issues.apache.org/jira/browse/HDFS-8448
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Anu Engineer
Assignee: Anu Engineer
 Attachments: hdfs-8448-hdfs-7240.001.patch


 Create REST interfaces as specified in the architecture document.
 This Jira is for creating the Volume Interface



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Walter Su (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553511#comment-14553511
 ] 

Walter Su commented on HDFS-8441:
-

... Maybe we could add a method in {{FSDirectory#isInECZone(src)}} and use it 
to tell the condition?
Good idea. I found {{dir#isInECZone(iip)}} exists. So I add 
{{ns#isInEcZone(src)}}. I think usually {{ns}} takes {{src}} arg, and {{dir}} 
takes {{iip}}.
... Please add message to the fail() statement like...
Done.
Thank you both. Uploaded 003 patch.

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch, HDFS-8441-HDFS-7285.003.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update storage count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread J.Andreina (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553575#comment-14553575
 ] 

J.Andreina commented on HDFS-8443:
--

Thanks [~arpitagarwal] for the review comments.
Please review the updated patch.


 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina
 Attachments: HDFS-8443.1.patch


 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8344) NameNode doesn't recover lease for files with missing blocks

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8344?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553625#comment-14553625
 ] 

Hadoop QA commented on HDFS-8344:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 56s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 33s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 44s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   2m 22s | The applied patch generated  3 
new checkstyle issues (total was 488, now 489). |
| {color:green}+1{color} | whitespace |   0m  1s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 36s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 35s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   3m 10s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 17s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 164m 35s | Tests failed in hadoop-hdfs. |
| | | 208m 17s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.TestAppendSnapshotTruncate |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734291/HDFS-8344.03.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 6329bd0 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/11066/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11066/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11066/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf903.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11066/console |


This message was automatically generated.

 NameNode doesn't recover lease for files with missing blocks
 

 Key: HDFS-8344
 URL: https://issues.apache.org/jira/browse/HDFS-8344
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.0
Reporter: Ravi Prakash
Assignee: Ravi Prakash
 Attachments: HDFS-8344.01.patch, HDFS-8344.02.patch, 
 HDFS-8344.03.patch


 I found another\(?) instance in which the lease is not recovered. This is 
 reproducible easily on a pseudo-distributed single node cluster
 # Before you start it helps if you set. This is not necessary, but simply 
 reduces how long you have to wait
 {code}
   public static final long LEASE_SOFTLIMIT_PERIOD = 30 * 1000;
   public static final long LEASE_HARDLIMIT_PERIOD = 2 * 
 LEASE_SOFTLIMIT_PERIOD;
 {code}
 # Client starts to write a file. (could be less than 1 block, but it hflushed 
 so some of the data has landed on the datanodes) (I'm copying the client code 
 I am using. I generate a jar and run it using $ hadoop jar TestHadoop.jar)
 # Client crashes. (I simulate this by kill -9 the $(hadoop jar 
 TestHadoop.jar) process after it has printed Wrote to the bufferedWriter
 # Shoot the datanode. (Since I ran on a pseudo-distributed cluster, there was 
 only 1)
 I believe the lease should be recovered and the block should be marked 
 missing. However this is not happening. The lease is never recovered.
 The effect of this bug for us was that nodes could not be decommissioned 
 cleanly. Although we knew that the client had crashed, the Namenode never 
 released the leases (even after restarting the Namenode) (even months 
 afterwards). There are actually several other cases too where we don't 
 consider what happens if ALL the datanodes die while the file is being 
 written, but I am going to punt on that for another time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8210) Ozone: Implement storage container manager

2015-05-20 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HDFS-8210:
---
Status: Open  (was: Patch Available)

 Ozone: Implement storage container manager 
 ---

 Key: HDFS-8210
 URL: https://issues.apache.org/jira/browse/HDFS-8210
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HDFS-8210-HDFS-7240.1.patch, 
 HDFS-8210-HDFS-7240.2.patch, HDFS-8210-HDFS-7240.3.patch, 
 HDFS-8210-HDFS-7240.4.patch


 The storage container manager collects datanode heartbeats, manages 
 replication and exposes API to lookup containers. This jira implements 
 storage container manager by re-using the block manager implementation in 
 namenode. This jira provides initial implementation that works with 
 datanodes. The additional protocols will be added in subsequent jiras.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8210) Ozone: Implement storage container manager

2015-05-20 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HDFS-8210:
---
Attachment: HDFS-8210-HDFS-7240.4.patch

The attached patch addresses the two style issues.

 Ozone: Implement storage container manager 
 ---

 Key: HDFS-8210
 URL: https://issues.apache.org/jira/browse/HDFS-8210
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HDFS-8210-HDFS-7240.1.patch, 
 HDFS-8210-HDFS-7240.2.patch, HDFS-8210-HDFS-7240.3.patch, 
 HDFS-8210-HDFS-7240.4.patch


 The storage container manager collects datanode heartbeats, manages 
 replication and exposes API to lookup containers. This jira implements 
 storage container manager by re-using the block manager implementation in 
 namenode. This jira provides initial implementation that works with 
 datanodes. The additional protocols will be added in subsequent jiras.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8210) Ozone: Implement storage container manager

2015-05-20 Thread Jitendra Nath Pandey (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8210?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Jitendra Nath Pandey updated HDFS-8210:
---
Status: Patch Available  (was: Open)

 Ozone: Implement storage container manager 
 ---

 Key: HDFS-8210
 URL: https://issues.apache.org/jira/browse/HDFS-8210
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Jitendra Nath Pandey
Assignee: Jitendra Nath Pandey
 Attachments: HDFS-8210-HDFS-7240.1.patch, 
 HDFS-8210-HDFS-7240.2.patch, HDFS-8210-HDFS-7240.3.patch, 
 HDFS-8210-HDFS-7240.4.patch


 The storage container manager collects datanode heartbeats, manages 
 replication and exposes API to lookup containers. This jira implements 
 storage container manager by re-using the block manager implementation in 
 namenode. This jira provides initial implementation that works with 
 datanodes. The additional protocols will be added in subsequent jiras.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Walter Su (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Walter Su updated HDFS-8441:

Attachment: HDFS-8441-HDFS-7285.003.patch

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch, HDFS-8441-HDFS-7285.003.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update storage count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553526#comment-14553526
 ] 

Kai Zheng commented on HDFS-8441:
-

Thanks for the update. Could we skip the lock and operation check stuffs since 
it's private and could be expected to be only used internally?
{code}
+  private boolean isInECZone(String src) throws IOException {
+byte[][] pathComponents = 
FSDirectory.getPathComponentsForReservedPath(src);
+readLock();
+try {
+  checkOperation(OperationCategory.READ);
+  src = FSDirectory.resolvePath(src, pathComponents, dir);
+  final INodesInPath iip = dir.getINodesInPath(src, true);
+  return dir.isInECZone(iip);
+} finally {
+  readUnlock();
+}
+  }
+
{code}

 Erasure Coding: make condition check earlier for setReplication
 ---

 Key: HDFS-8441
 URL: https://issues.apache.org/jira/browse/HDFS-8441
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Walter Su
Assignee: Walter Su
 Attachments: HDFS-8441-HDFS-7285.001.patch, 
 HDFS-8441-HDFS-7285.002.patch, HDFS-8441-HDFS-7285.003.patch


 Changes:
 1. {{UnsupportedActionException}} is more user-firendly.
 2. check condition before update storage count
 3. support create EC file with replication=0



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8294) Erasure Coding: Fix Findbug warnings present in erasure coding

2015-05-20 Thread Rakesh R (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8294?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rakesh R updated HDFS-8294:
---
Attachment: HDFS-8294-HDFS-7285.06.patch

 Erasure Coding: Fix Findbug warnings present in erasure coding
 --

 Key: HDFS-8294
 URL: https://issues.apache.org/jira/browse/HDFS-8294
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Rakesh R
Assignee: Rakesh R
  Labels: BB2015-05-RFC
 Attachments: FindBugs Report in EC feature.html, 
 HDFS-8294-HDFS-7285.00.patch, HDFS-8294-HDFS-7285.01.patch, 
 HDFS-8294-HDFS-7285.02.patch, HDFS-8294-HDFS-7285.03.patch, 
 HDFS-8294-HDFS-7285.04.patch, HDFS-8294-HDFS-7285.05.patch, 
 HDFS-8294-HDFS-7285.06.patch


 This jira is to address the findbug issues reported in erasure coding feature.
 Attached sheet which contains the details of the findbug issues reported in 
 the erasure coding feature. I've taken this report from the jenkins build : 
 https://builds.apache.org/job/PreCommit-HDFS-Build/10848/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8020) Erasure Coding: restore BlockGroup and schema info from stripping coding command

2015-05-20 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8020?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553568#comment-14553568
 ] 

Uma Maheswara Rao G commented on HDFS-8020:
---

Do we need this JIRA still? we are already sending EC Command to DN and DN 
already getting required information from it. So, can we we resolve this?

 Erasure Coding: restore BlockGroup and schema info from stripping coding 
 command
 

 Key: HDFS-8020
 URL: https://issues.apache.org/jira/browse/HDFS-8020
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Sasaki

 As a task of HDFS-7344, to process *stripping* coding commands from NameNode 
 or other scheduler services/tools, we need to first be able to restore 
 BlockGroup and schema information in DataNode, which will be used to 
 construct coding work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8020) Erasure Coding: restore BlockGroup and schema info from stripping coding command

2015-05-20 Thread Kai Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8020?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Zheng updated HDFS-8020:

Parent Issue: HDFS-8031  (was: HDFS-7285)

 Erasure Coding: restore BlockGroup and schema info from stripping coding 
 command
 

 Key: HDFS-8020
 URL: https://issues.apache.org/jira/browse/HDFS-8020
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Sasaki

 As a task of HDFS-7344, to process *stripping* coding commands from NameNode 
 or other scheduler services/tools, we need to first be able to restore 
 BlockGroup and schema information in DataNode, which will be used to 
 construct coding work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8442) Remove ServerLifecycleListener from kms/server.xml.

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8442?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553619#comment-14553619
 ] 

Hadoop QA commented on HDFS-8442:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  14m 40s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 36s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 40s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 33s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | common tests |   1m 38s | Tests passed in 
hadoop-kms. |
| | |  36m  7s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12734324/HDFS-8442-1.patch |
| Optional Tests | javadoc javac unit |
| git revision | trunk / fb6b38d |
| hadoop-kms test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11073/artifact/patchprocess/testrun_hadoop-kms.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11073/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf900.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/11073/console |


This message was automatically generated.

 Remove ServerLifecycleListener from kms/server.xml.
 ---

 Key: HDFS-8442
 URL: https://issues.apache.org/jira/browse/HDFS-8442
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: nijel
Assignee: nijel
 Attachments: HDFS-8442-1.patch


 Remove ServerLifecycleListener from kms/server.xml.
 From tomcat Tomcat 7.0.9 the support for ServerLifecycleListener is removed
 ref : https://tomcat.apache.org/tomcat-7.0-doc/changelog.html
 Remove ServerLifecycleListener. This was already removed from server.xml and 
 with the Lifecycle re-factoring is no longer required. (markt)
 So if the build env is with tomcat later than this, kms startup is failing
 {code}
 SEVERE: Begin event threw exception
 java.lang.ClassNotFoundException: 
 org.apache.catalina.mbeans.ServerLifecycleListener
 {code}
 can we remove this listener ? 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8439) Adding more slow action log in critical read path

2015-05-20 Thread Liu Shaohui (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8439?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553577#comment-14553577
 ] 

Liu Shaohui commented on HDFS-8439:
---

[~stack]
{quote}
Are you running this in production?
{quote}
Just in test cluster currently. We are trying to push this to production.


 Adding more slow action log in critical read path
 -

 Key: HDFS-8439
 URL: https://issues.apache.org/jira/browse/HDFS-8439
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
 Attachments: HDFS-8439.001.patch, HDFS-8439.002.patch


 To dig a HBase read spike issue, we'd better to add more slow pread/seek log 
 in read flow to get the abnormal datanodes.
 Patch will be uploaded soon.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8439) Adding more slow action log in critical read path

2015-05-20 Thread Liu Shaohui (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8439?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Liu Shaohui updated HDFS-8439:
--
Attachment: HDFS-8439.002.patch

Update for @stack's review
- replace cost with duration
- Using Time.monotonicNow() instead of System.currentTimeMillis()

 Adding more slow action log in critical read path
 -

 Key: HDFS-8439
 URL: https://issues.apache.org/jira/browse/HDFS-8439
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Liu Shaohui
Assignee: Liu Shaohui
Priority: Minor
 Attachments: HDFS-8439.001.patch, HDFS-8439.002.patch


 To dig a HBase read spike issue, we'd better to add more slow pread/seek log 
 in read flow to get the abnormal datanodes.
 Patch will be uploaded soon.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8443) Document dfs.namenode.service.handler.count in hdfs-site.xml

2015-05-20 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8443?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-8443:
-
Attachment: HDFS-8443.2.patch

 Document dfs.namenode.service.handler.count in hdfs-site.xml
 

 Key: HDFS-8443
 URL: https://issues.apache.org/jira/browse/HDFS-8443
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: documentation
Reporter: Akira AJISAKA
Assignee: J.Andreina
 Attachments: HDFS-8443.1.patch, HDFS-8443.2.patch


 When dfs.namenode.servicerpc-address is configured, NameNode launches an 
 extra RPC server to handle requests from non-client nodes. 
 dfs.namenode.service.handler.count specifies the number of threads for the 
 server but the parameter is not documented anywhere.
 I found a mail for asking about the parameter. 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201505.mbox/%3CE0D5A619-BDEA-44D2-81EB-C32B8464133D%40gmail.com%3E



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8442) Remove ServerLifecycleListener from kms/server.xml.

2015-05-20 Thread nijel (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8442?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

nijel updated HDFS-8442:

Status: Patch Available  (was: Open)

 Remove ServerLifecycleListener from kms/server.xml.
 ---

 Key: HDFS-8442
 URL: https://issues.apache.org/jira/browse/HDFS-8442
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: nijel
Assignee: nijel
 Attachments: HDFS-8442-1.patch


 Remove ServerLifecycleListener from kms/server.xml.
 From tomcat Tomcat 7.0.9 the support for ServerLifecycleListener is removed
 ref : https://tomcat.apache.org/tomcat-7.0-doc/changelog.html
 Remove ServerLifecycleListener. This was already removed from server.xml and 
 with the Lifecycle re-factoring is no longer required. (markt)
 So if the build env is with tomcat later than this, kms startup is failing
 {code}
 SEVERE: Begin event threw exception
 java.lang.ClassNotFoundException: 
 org.apache.catalina.mbeans.ServerLifecycleListener
 {code}
 can we remove this listener ? 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8442) Remove ServerLifecycleListener from kms/server.xml.

2015-05-20 Thread nijel (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8442?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

nijel updated HDFS-8442:

Attachment: HDFS-8442-1.patch

please find the patch

 Remove ServerLifecycleListener from kms/server.xml.
 ---

 Key: HDFS-8442
 URL: https://issues.apache.org/jira/browse/HDFS-8442
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: nijel
Assignee: nijel
 Attachments: HDFS-8442-1.patch


 Remove ServerLifecycleListener from kms/server.xml.
 From tomcat Tomcat 7.0.9 the support for ServerLifecycleListener is removed
 ref : https://tomcat.apache.org/tomcat-7.0-doc/changelog.html
 Remove ServerLifecycleListener. This was already removed from server.xml and 
 with the Lifecycle re-factoring is no longer required. (markt)
 So if the build env is with tomcat later than this, kms startup is failing
 {code}
 SEVERE: Begin event threw exception
 java.lang.ClassNotFoundException: 
 org.apache.catalina.mbeans.ServerLifecycleListener
 {code}
 can we remove this listener ? 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8441) Erasure Coding: make condition check earlier for setReplication

2015-05-20 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8441?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553586#comment-14553586
 ] 

Hadoop QA commented on HDFS-8441:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  15m 23s | Pre-patch HDFS-7285 compilation 
is healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 46s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 54s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 14s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 38s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 38s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 31s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   3m 17s | The patch appears to introduce 6 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 18s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests |  48m 33s | Tests failed in hadoop-hdfs. |
| | |  91m 16s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
|  |  Inconsistent synchronization of 
org.apache.hadoop.hdfs.DFSOutputStream.streamer; locked 89% of time  
Unsynchronized access at DFSOutputStream.java:89% of time  Unsynchronized 
access at DFSOutputStream.java:[line 146] |
|  |  Possible null pointer dereference of arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:arr$ in 
org.apache.hadoop.hdfs.server.blockmanagement.BlockInfoStripedUnderConstruction.initializeBlockRecovery(long)
  Dereferenced at BlockInfoStripedUnderConstruction.java:[line 194] |
|  |  Unread field:field be static?  At ErasureCodingWorker.java:[line 254] |
|  |  Should 
org.apache.hadoop.hdfs.server.datanode.erasurecode.ErasureCodingWorker$StripedReader
 be a _static_ inner class?  At ErasureCodingWorker.java:inner class?  At 
ErasureCodingWorker.java:[lines 907-914] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.constructInternalBlock(LocatedStripedBlock,
 int, int, int, int)  At StripedBlockUtil.java:[line 108] |
|  |  Result of integer multiplication cast to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:to long in 
org.apache.hadoop.hdfs.util.StripedBlockUtil.getStartOffsetsForInternalBlocks(ECSchema,
 int, LocatedStripedBlock, long)  At StripedBlockUtil.java:[line 409] |
| Failed unit tests | hadoop.hdfs.server.blockmanagement.TestDatanodeManager |
|   | hadoop.hdfs.server.namenode.snapshot.TestUpdatePipelineWithSnapshots |
|   | hadoop.hdfs.TestModTime |
|   | hadoop.fs.TestUrlStreamHandler |
|   | hadoop.hdfs.security.TestDelegationToken |
|   | hadoop.hdfs.server.namenode.TestBlockPlacementPolicyRackFaultTolarent |
|   | hadoop.hdfs.server.namenode.TestFileLimit |
|   | hadoop.hdfs.TestParallelShortCircuitRead |
|   | hadoop.hdfs.server.namenode.snapshot.TestFileContextSnapshot |
|   | hadoop.hdfs.server.blockmanagement.TestBlockInfoStriped |
|   | hadoop.hdfs.server.namenode.TestEditLogAutoroll |
|   | hadoop.TestRefreshCallQueue |
|   | hadoop.hdfs.protocolPB.TestPBHelper |
|   | hadoop.hdfs.server.namenode.ha.TestStandbyCheckpoints |
|   | hadoop.cli.TestCryptoAdminCLI |
|   | hadoop.hdfs.TestSetrepDecreasing |
|   | hadoop.hdfs.server.datanode.TestDiskError |
|   | hadoop.fs.viewfs.TestViewFsWithAcls |
|   | hadoop.hdfs.server.datanode.TestDataNodeHotSwapVolumes |
|   | hadoop.hdfs.server.namenode.TestAddStripedBlocks |
|   | hadoop.hdfs.server.namenode.TestFSEditLogLoader |
|   | hadoop.hdfs.server.namenode.TestHostsFiles |
|   | hadoop.hdfs.server.datanode.TestTransferRbw |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistPolicy |
|   | hadoop.fs.contract.hdfs.TestHDFSContractDelete |
|   | hadoop.hdfs.server.namenode.TestFileContextAcl |
|   | hadoop.fs.TestFcHdfsSetUMask |
|   | hadoop.fs.TestUnbuffer |
|   | hadoop.hdfs.server.namenode.TestClusterId |
|   | hadoop.hdfs.server.namenode.TestDeleteRace |
|   | hadoop.hdfs.TestPread |
|   | hadoop.hdfs.server.namenode.TestFSDirectory |
|   

[jira] [Commented] (HDFS-8407) libhdfs hdfsListDirectory() API has different behavior than documentation

2015-05-20 Thread Juan Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553640#comment-14553640
 ] 

Juan Yu commented on HDFS-8407:
---

I think move errno = ret; before the if check will cover both cases, right?

 libhdfs hdfsListDirectory() API has different behavior than documentation
 -

 Key: HDFS-8407
 URL: https://issues.apache.org/jira/browse/HDFS-8407
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Reporter: Juan Yu
Assignee: Masatake Iwasaki
 Attachments: HDFS-8407.001.patch, HDFS-8407.002.patch, 
 HDFS-8407.003.patch


 The documentation says it returns NULL on error, but it could also return 
 NULL when the directory is empty.
 /** 
  * hdfsListDirectory - Get list of files/directories for a given
  * directory-path. hdfsFreeFileInfo should be called to deallocate 
 memory. 
  * @param fs The configured filesystem handle.
  * @param path The path of the directory. 
  * @param numEntries Set to the number of files/directories in path.
  * @return Returns a dynamically-allocated array of hdfsFileInfo
  * objects; NULL on error.
  */
 {code}
 hdfsFileInfo *pathList = NULL; 
 ...
 //Figure out the number of entries in that directory
 jPathListSize = (*env)-GetArrayLength(env, jPathList);
 if (jPathListSize == 0) {
 ret = 0;
 goto done;
 }
 ...
 if (ret) {
 hdfsFreeFileInfo(pathList, jPathListSize);
 errno = ret;
 return NULL;
 }
 *numEntries = jPathListSize;
 return pathList;
 {code}
 Either change the implementation to match the doc, or fix the doc to match 
 the implementation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8446) Separate safemode related operations in GetBlockLocations()

2015-05-20 Thread Haohui Mai (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8446?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Haohui Mai updated HDFS-8446:
-
Attachment: HDFS-8446.001.patch

 Separate safemode related operations in GetBlockLocations()
 ---

 Key: HDFS-8446
 URL: https://issues.apache.org/jira/browse/HDFS-8446
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Haohui Mai
Assignee: Haohui Mai
Priority: Minor
 Attachments: HDFS-8446.000.patch, HDFS-8446.001.patch


 Currently {{FSNamesystem#GetBlockLocations()}} has some special cases when 
 the NN is in SafeMode. This jira proposes to refactor the code to improve 
 readability.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8020) Erasure Coding: restore BlockGroup and schema info from stripping coding command

2015-05-20 Thread Kai Zheng (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8020?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14553574#comment-14553574
 ] 

Kai Zheng commented on HDFS-8020:
-

Hi Uma, I think we might still need this to make use of the {{ErasureCoder}} 
API in next phase. Yes currently it works because we're using the 
{{RawErasureCoder}} API directly.

 Erasure Coding: restore BlockGroup and schema info from stripping coding 
 command
 

 Key: HDFS-8020
 URL: https://issues.apache.org/jira/browse/HDFS-8020
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Kai Zheng
Assignee: Kai Sasaki

 As a task of HDFS-7344, to process *stripping* coding commands from NameNode 
 or other scheduler services/tools, we need to first be able to restore 
 BlockGroup and schema information in DataNode, which will be used to 
 construct coding work.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >