[jira] [Commented] (HDFS-8373) Ec files can't be deleted into Trash because of that Trash isn't EC zone.

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711665#comment-14711665
 ] 

Zhe Zhang commented on HDFS-8373:
-

Good catch [~brahmareddy]. Once HDFS-8833 is finalized let's close this one as 
duplicate.

 Ec files can't be deleted into Trash because of that Trash isn't EC zone.
 -

 Key: HDFS-8373
 URL: https://issues.apache.org/jira/browse/HDFS-8373
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: HDFS-7285
Reporter: GAO Rui
Assignee: Brahma Reddy Battula
  Labels: EC

 When EC files were deleted, they would be moved into {{Trash}} directory. 
 But, EC files can only be placed under EC zone. So, EC files which have been 
 deleted can not be moved to {{Trash}} directory.
 Problem could be solved by creating a EC zone(floder) inside {{Trash}} to 
 contain deleted EC files.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8943) Read apis in ByteRangeInputStream does not read all the bytes specified when chunked transfer-encoding is used in the server

2015-08-25 Thread Shradha Revankar (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711664#comment-14711664
 ] 

Shradha Revankar commented on HDFS-8943:


Thanks, I've created this https://issues.apache.org/jira/browse/HBASE-14307

 Read apis in ByteRangeInputStream does not read all the bytes specified when 
 chunked transfer-encoding is used in the server
 

 Key: HDFS-8943
 URL: https://issues.apache.org/jira/browse/HDFS-8943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar
 Attachments: HDFS-8943.000.patch


 With the default Webhdfs server implementation the read apis in 
 ByteRangeInputStream work as expected reading the correct number of bytes for 
 these apis :
 {{public int read(byte b[], int off, int len)}}
 {{public int read(long position, byte[] buffer, int offset, int length)}}
 But when a custom Webhdfs server implementation is plugged in which uses 
 chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
 would be to loop and read till bytes specified similar to {{readfully()}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8915) TestFSNamesystem.testFSLockGetWaiterCount fails intermittently in jenkins

2015-08-25 Thread Anu Engineer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8915?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711680#comment-14711680
 ] 

Anu Engineer commented on HDFS-8915:


This was a tracking bug to see if this failure happens in the apache tree.  
Since the only failure was reported when we had a VM outage, resolving this 
issue as won't fix. This test has been running without any issues for a week.



 TestFSNamesystem.testFSLockGetWaiterCount fails intermittently in jenkins
 -

 Key: HDFS-8915
 URL: https://issues.apache.org/jira/browse/HDFS-8915
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Affects Versions: 2.8.0
Reporter: Anu Engineer
Assignee: Anu Engineer

 This test was added as part of HDFS-8883, There is a race condition in the 
 test and it has failed *once* in the Apache Jenkins run.
 Here is the stack
 FAILED:  
 org.apache.hadoop.hdfs.server.namenode.TestFSNamesystem.testFSLockGetWaiterCount
 Error Message:
 Expected number of blocked thread not found expected:3 but was:1
 Stack Trace:
 java.lang.AssertionError: Expected number of blocked thread not found 
 expected:3 but was:1
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.hadoop.hdfs.server.namenode.TestFSNamesystem.testFSLockGetWaiterCount(TestFSNamesystem.java:261)
 From cursory code reading , even though we call into readlock.lock() there is 
 no guarantee that our thread is put in the wait queue. A proposed fix could 
 be to check for any thread in the lock queue instead of all 3, or disable the 
 test.
 It could also indicate an issue with the test infra-structure but any test 
 open to variations in result due to infra-structure issues creates noise in 
 tests so we are better off fixing it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-1148) Convert FSDataset to ReadWriteLock

2015-08-25 Thread Yong Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1148?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yong Zhang updated HDFS-1148:
-
Attachment: HDFS-1148.003.patch

 Convert FSDataset to ReadWriteLock
 --

 Key: HDFS-1148
 URL: https://issues.apache.org/jira/browse/HDFS-1148
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Reporter: Todd Lipcon
Assignee: Yong Zhang
 Attachments: HDFS-1148.001.patch, HDFS-1148.002.patch, 
 HDFS-1148.003.patch, hdfs-1148-old.txt, hdfs-1148-trunk.txt, 
 patch-HDFS-1148-rel0.20.2.txt


 In benchmarking HDFS-941 I noticed that for the random read workload, the 
 FSDataset lock is highly contended. After converting it to a 
 ReentrantReadWriteLock, I saw a ~25% improvement on both latency and 
 ops/second.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8894) Set SO_KEEPALIVE on DN server sockets

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8894:
--
Attachment: HDFS-8894-01.patch

 Set SO_KEEPALIVE on DN server sockets
 -

 Key: HDFS-8894
 URL: https://issues.apache.org/jira/browse/HDFS-8894
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Nathan Roberts
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8894-01.patch, HDFS-8894-01.patch


 SO_KEEPALIVE is not set on things like datastreamer sockets which can cause 
 lingering ESTABLISHED sockets when there is a network glitch.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Create edit log files with old layout version for upgrade testing

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711658#comment-14711658
 ] 

Zhe Zhang commented on HDFS-8846:
-

The change is test-only and is not related to {{TestEditLog}}. It also passes 
locally.

 Create edit log files with old layout version for upgrade testing
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-8915) TestFSNamesystem.testFSLockGetWaiterCount fails intermittently in jenkins

2015-08-25 Thread Anu Engineer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8915?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Anu Engineer resolved HDFS-8915.

Resolution: Invalid

 TestFSNamesystem.testFSLockGetWaiterCount fails intermittently in jenkins
 -

 Key: HDFS-8915
 URL: https://issues.apache.org/jira/browse/HDFS-8915
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Affects Versions: 2.8.0
Reporter: Anu Engineer
Assignee: Anu Engineer

 This test was added as part of HDFS-8883, There is a race condition in the 
 test and it has failed *once* in the Apache Jenkins run.
 Here is the stack
 FAILED:  
 org.apache.hadoop.hdfs.server.namenode.TestFSNamesystem.testFSLockGetWaiterCount
 Error Message:
 Expected number of blocked thread not found expected:3 but was:1
 Stack Trace:
 java.lang.AssertionError: Expected number of blocked thread not found 
 expected:3 but was:1
   at org.junit.Assert.fail(Assert.java:88)
   at org.junit.Assert.failNotEquals(Assert.java:743)
   at org.junit.Assert.assertEquals(Assert.java:118)
   at org.junit.Assert.assertEquals(Assert.java:555)
   at 
 org.apache.hadoop.hdfs.server.namenode.TestFSNamesystem.testFSLockGetWaiterCount(TestFSNamesystem.java:261)
 From cursory code reading , even though we call into readlock.lock() there is 
 no guarantee that our thread is put in the wait queue. A proposed fix could 
 be to check for any thread in the lock queue instead of all 3, or disable the 
 test.
 It could also indicate an issue with the test infra-structure but any test 
 open to variations in result due to infra-structure issues creates noise in 
 tests so we are better off fixing it.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8938) Refactor BlockManager in blockmanagement

2015-08-25 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu updated HDFS-8938:

Attachment: HDFS-8938.005.patch

 Refactor BlockManager in blockmanagement
 

 Key: HDFS-8938
 URL: https://issues.apache.org/jira/browse/HDFS-8938
 Project: Hadoop HDFS
  Issue Type: Task
  Components: build
Reporter: Mingliang Liu
Assignee: Mingliang Liu
 Attachments: HDFS-8938.000.patch, HDFS-8938.001.patch, 
 HDFS-8938.002.patch, HDFS-8938.003.patch, HDFS-8938.004.patch, 
 HDFS-8938.005.patch


 This lira tracks the effort of refactoring the {{BlockManager}} in 
 {{hdfs.server.blockmanagement}} package.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-6515) testPageRounder (org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache)

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-6515?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711289#comment-14711289
 ] 

Hadoop QA commented on HDFS-6515:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  19m 34s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 56s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  0s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   2m 36s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 29s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   4m 28s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | common tests |  23m  1s | Tests passed in 
hadoop-common. |
| {color:red}-1{color} | hdfs tests | 161m 42s | Tests failed in hadoop-hdfs. |
| | | 231m 44s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | 
hadoop.hdfs.server.blockmanagement.TestPendingInvalidateBlock |
|   | hadoop.hdfs.server.namenode.TestDeleteRace |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12677586/HDFS-6515-2.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12112/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| hadoop-common test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12112/artifact/patchprocess/testrun_hadoop-common.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12112/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12112/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf901.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12112/console |


This message was automatically generated.

 testPageRounder   (org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache)
 -

 Key: HDFS-6515
 URL: https://issues.apache.org/jira/browse/HDFS-6515
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 3.0.0, 2.4.0, 2.4.1
 Environment: Linux on PPC64
 Tested with Hadoop 3.0.0 SNAPSHOT, on RHEL 6.5, on Ubuntu 14.04, on Fedora 
 19, using mvn -Dtest=TestFsDatasetCache#testPageRounder -X test
Reporter: Tony Reix
Priority: Blocker
  Labels: BB2015-05-TBR, hadoop, test
 Attachments: HDFS-6515-1.patch, HDFS-6515-2.patch


 I have an issue with test :
testPageRounder
   (org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache)
 on Linux/PowerPC.
 On Linux/Intel, test runs fine.
 On Linux/PowerPC, I have:
 testPageRounder(org.apache.hadoop.hdfs.server.datanode.TestFsDatasetCache)  
 Time elapsed: 64.037 sec   ERROR!
 java.lang.Exception: test timed out after 6 milliseconds
 Looking at details, I see that some Failed to cache  messages appear in the 
 traces. Only 10 on Intel, but 186 on PPC64.
 On PPC64, it looks like some thread is waiting for something that never 
 happens, generating a TimeOut.
 I'm now using IBM JVM, however I've just checked that the issue also appears 
 with OpenJDK.
 I'm now using Hadoop latest, however, the issue appeared within Hadoop 2.4.0 .
 I need help for understanding what the test is doing, what traces are 
 expected, in order to understand what/where is the root cause.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711326#comment-14711326
 ] 

Hudson commented on HDFS-8900:
--

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #296 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/296/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8581) count cmd calculate wrong when huge files exist in one folder

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711399#comment-14711399
 ] 

Hadoop QA commented on HDFS-8581:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 43s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 52s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  0s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   1m 24s | There were no new checkstyle 
issues. |
| {color:red}-1{color} | whitespace |   0m  0s | The patch has 1  line(s) that 
end in whitespace. Use git apply --whitespace=fix. |
| {color:green}+1{color} | install |   1m 30s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 31s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 14s | Pre-build of native portion |
| {color:green}+1{color} | hdfs tests | 162m  9s | Tests passed in hadoop-hdfs. 
|
| | | 207m 24s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752211/HDFS-8581.2.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12115/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12115/artifact/patchprocess/whitespace.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12115/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12115/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf906.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12115/console |


This message was automatically generated.

 count cmd calculate wrong when huge files exist in one folder
 -

 Key: HDFS-8581
 URL: https://issues.apache.org/jira/browse/HDFS-8581
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Reporter: tongshiquan
Assignee: J.Andreina
Priority: Minor
 Attachments: HDFS-8581.1.patch, HDFS-8581.2.patch


 If one directory such as /result exists about 20 files, then when 
 execute hdfs dfs -count /, the result will go wrong. For all directories 
 whose name after /result, file num will not be included.
 My cluster see as below, /result_1433858936 is the directory exist huge 
 files, and files in /sparkJobHistory, /tmp, /user are not included
 vm-221:/export1/BigData/current # hdfs dfs -ls /
 15/06/11 11:00:17 INFO hdfs.PeerCache: SocketCache disabled.
 Found 9 items
 -rw-r--r--   3 hdfs   supergroup  0 2015-06-08 12:10 
 /PRE_CREATE_DIR.SUCCESS
 drwxr-x---   - flume  hadoop  0 2015-06-08 12:08 /flume
 drwx--   - hbase  hadoop  0 2015-06-10 15:25 /hbase
 drwxr-xr-x   - hdfs   supergroup  0 2015-06-10 17:19 /hyt
 drwxrwxrwx   - mapred hadoop  0 2015-06-08 12:08 /mr-history
 drwxr-xr-x   - hdfs   supergroup  0 2015-06-09 22:10 
 /result_1433858936
 drwxrwxrwx   - spark  supergroup  0 2015-06-10 19:15 /sparkJobHistory
 drwxrwxrwx   - hdfs   hadoop  0 2015-06-08 12:14 /tmp
 drwxrwxrwx   - hdfs   hadoop  0 2015-06-09 21:57 /user
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /
 15/06/11 11:00:24 INFO hdfs.PeerCache: SocketCache disabled.
 1043   171536 1756375688 /
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /PRE_CREATE_DIR.SUCCESS
 15/06/11 11:00:30 INFO hdfs.PeerCache: SocketCache disabled.
01  0 /PRE_CREATE_DIR.SUCCESS
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /flume
 15/06/11 11:00:41 INFO hdfs.PeerCache: SocketCache disabled.
10  0 /flume
 vm-221:/export1/BigData/current # 
 

[jira] [Updated] (HDFS-7765) FSOutputSummer throwing ArrayIndexOutOfBoundsException

2015-08-25 Thread Surendra Singh Lilhore (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7765?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Surendra Singh Lilhore updated HDFS-7765:
-
Assignee: Janmejay Singh  (was: Surendra Singh Lilhore)

 FSOutputSummer throwing ArrayIndexOutOfBoundsException
 --

 Key: HDFS-7765
 URL: https://issues.apache.org/jira/browse/HDFS-7765
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 2.6.0
 Environment: Centos 6, Open JDK 7, Amazon EC2, Accumulo 1.6.2RC4
Reporter: Keith Turner
Assignee: Janmejay Singh
 Attachments: 
 0001-PATCH-HDFS-7765-FSOutputSummer-throwing-ArrayIndexOu.patch, 
 HDFS-7765.patch


 While running an Accumulo test, saw exceptions like the following while 
 trying to write to write ahead log in HDFS. 
 The exception occurrs at 
 [FSOutputSummer.java:76|https://github.com/apache/hadoop/blob/release-2.6.0/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FSOutputSummer.java#L76]
  which is attempting to update a byte array.
 {noformat}
 2015-02-06 19:46:49,769 [log.DfsLogger] WARN : Exception syncing 
 java.lang.reflect.InvocationTargetException
 java.lang.ArrayIndexOutOfBoundsException: 4608
 at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:76)
 at 
 org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:50)
 at java.io.DataOutputStream.write(DataOutputStream.java:88)
 at java.io.DataOutputStream.writeByte(DataOutputStream.java:153)
 at 
 org.apache.accumulo.tserver.logger.LogFileKey.write(LogFileKey.java:87)
 at org.apache.accumulo.tserver.log.DfsLogger.write(DfsLogger.java:526)
 at 
 org.apache.accumulo.tserver.log.DfsLogger.logFileData(DfsLogger.java:540)
 at 
 org.apache.accumulo.tserver.log.DfsLogger.logManyTablets(DfsLogger.java:573)
 at 
 org.apache.accumulo.tserver.log.TabletServerLogger$6.write(TabletServerLogger.java:373)
 at 
 org.apache.accumulo.tserver.log.TabletServerLogger.write(TabletServerLogger.java:274)
 at 
 org.apache.accumulo.tserver.log.TabletServerLogger.logManyTablets(TabletServerLogger.java:365)
 at 
 org.apache.accumulo.tserver.TabletServer$ThriftClientHandler.flush(TabletServer.java:1667)
 at 
 org.apache.accumulo.tserver.TabletServer$ThriftClientHandler.closeUpdate(TabletServer.java:1754)
 at sun.reflect.GeneratedMethodAccessor8.invoke(Unknown Source)
 at 
 sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
 at java.lang.reflect.Method.invoke(Method.java:606)
 at 
 org.apache.accumulo.trace.instrument.thrift.RpcServerInvocationHandler.invoke(RpcServerInvocationHandler.java:46)
 at 
 org.apache.accumulo.server.util.RpcWrapper$1.invoke(RpcWrapper.java:47)
 at com.sun.proxy.$Proxy22.closeUpdate(Unknown Source)
 at 
 org.apache.accumulo.core.tabletserver.thrift.TabletClientService$Processor$closeUpdate.getResult(TabletClientService.java:2370)
 at 
 org.apache.accumulo.core.tabletserver.thrift.TabletClientService$Processor$closeUpdate.getResult(TabletClientService.java:2354)
 at org.apache.thrift.ProcessFunction.process(ProcessFunction.java:39)
 at org.apache.thrift.TBaseProcessor.process(TBaseProcessor.java:39)
 at 
 org.apache.accumulo.server.util.TServerUtils$TimedProcessor.process(TServerUtils.java:168)
 at 
 org.apache.thrift.server.AbstractNonblockingServer$FrameBuffer.invoke(AbstractNonblockingServer.java:516)
 at 
 org.apache.accumulo.server.util.CustomNonBlockingServer$1.run(CustomNonBlockingServer.java:77)
 at 
 java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
 at 
 java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
 at 
 org.apache.accumulo.trace.instrument.TraceRunnable.run(TraceRunnable.java:47)
 at 
 org.apache.accumulo.core.util.LoggingRunnable.run(LoggingRunnable.java:34)
 at java.lang.Thread.run(Thread.java:744)
 2015-02-06 19:46:49,769 [log.DfsLogger] WARN : Exception syncing 
 java.lang.reflect.InvocationTargetException
 2015-02-06 19:46:49,772 [log.DfsLogger] ERROR: 
 java.lang.ArrayIndexOutOfBoundsException: 4609
 java.lang.ArrayIndexOutOfBoundsException: 4609
 at org.apache.hadoop.fs.FSOutputSummer.write(FSOutputSummer.java:76)
 at 
 org.apache.hadoop.fs.FSDataOutputStream$PositionCache.write(FSDataOutputStream.java:50)
 at java.io.DataOutputStream.write(DataOutputStream.java:88)
 at java.io.DataOutputStream.writeByte(DataOutputStream.java:153)
 at 
 org.apache.accumulo.tserver.logger.LogFileKey.write(LogFileKey.java:87)
 at 

[jira] [Commented] (HDFS-8771) If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not send another RPC calls to Journalnodes

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711437#comment-14711437
 ] 

Hadoop QA commented on HDFS-8771:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 47s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 2 new or modified test files. |
| {color:green}+1{color} | javac |   7m 58s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  4s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 21s | The applied patch generated  1 
new checkstyle issues (total was 3, now 4). |
| {color:red}-1{color} | whitespace |   0m  0s | The patch has 1  line(s) that 
end in whitespace. Use git apply --whitespace=fix. |
| {color:green}+1{color} | install |   1m 31s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 32s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m  9s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 162m 26s | Tests failed in hadoop-hdfs. |
| | | 207m 47s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.server.namenode.TestFSNamesystem |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752216/HDFS-8771-01.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12116/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12116/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12116/artifact/patchprocess/whitespace.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12116/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12116/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf901.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12116/console |


This message was automatically generated.

 If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not 
 send another RPC calls to Journalnodes
 

 Key: HDFS-8771
 URL: https://issues.apache.org/jira/browse/HDFS-8771
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Takuya Fukudome
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8771-01.patch


 In our cluster, edits has became huge(about 50GB) accidentally and our 
 Jounalnodes' disks were busy, therefore {{purgeLogsOlderThan}} took more than 
 30secs. If {{IPCLoggerChannel#purgeLogsOlderThan}} takes too much time, 
 Namenode couldn't send other RPC calls to Journalnodes because 
 {{o.a.h.hdfs.qjournal.client.IPCLoggerChannel}}'s executor is single thread. 
 It will cause namenode shutting down.
 I think IPCLoggerChannel#purgeLogsOlderThan should not block other RPC calls 
 like sendEdits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8951) Move shortcircuit to hdfs-client

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712445#comment-14712445
 ] 

Hadoop QA commented on HDFS-8951:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  19m 28s | Pre-patch trunk has 6 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 55s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  3s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   2m 31s | The applied patch generated  
63 new checkstyle issues (total was 16, now 79). |
| {color:green}+1{color} | whitespace |   0m  1s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 41s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   4m 31s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 27s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 192m 47s | Tests failed in hadoop-hdfs. |
| {color:green}+1{color} | hdfs tests |   0m 29s | Tests passed in 
hadoop-hdfs-client. |
| | | 243m 54s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.server.namenode.TestFileTruncate |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752353/HDFS-8951.002.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / a4d9acc |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs-client.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/artifact/patchprocess/diffcheckstylehadoop-hdfs-client.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| hadoop-hdfs-client test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/artifact/patchprocess/testrun_hadoop-hdfs-client.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12127/console |


This message was automatically generated.

 Move shortcircuit to hdfs-client
 

 Key: HDFS-8951
 URL: https://issues.apache.org/jira/browse/HDFS-8951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: build
Reporter: Mingliang Liu
Assignee: Mingliang Liu
 Attachments: HDFS-8951.000.patch, HDFS-8951.001.patch, 
 HDFS-8951.002.patch


 This jira tracks the effort of moving the {{shortcircuit}} package into the 
 hdfs-client module.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8950) NameNode refresh doesn't remove DataNodes

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712448#comment-14712448
 ] 

Hadoop QA commented on HDFS-8950:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 43s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 51s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 13s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   1m 21s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 28s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 31s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m  8s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 166m 42s | Tests failed in hadoop-hdfs. |
| | | 211m 55s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.shortcircuit.TestShortCircuitCache |
|   | hadoop.hdfs.TestDecommission |
|   | hadoop.hdfs.server.namenode.TestAuditLogs |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752359/HDFS-8950.002.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / a4d9acc |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12128/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12128/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12128/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf900.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12128/console |


This message was automatically generated.

 NameNode refresh doesn't remove DataNodes
 -

 Key: HDFS-8950
 URL: https://issues.apache.org/jira/browse/HDFS-8950
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Daniel Templeton
Assignee: Daniel Templeton
 Fix For: 2.8.0

 Attachments: HDFS-8950.001.patch, HDFS-8950.002.patch


 If you remove a DN from NN's allowed host list (HDFS was HA) and then do NN 
 refresh, it doesn't remove it actually and the NN UI keeps showing that node. 
 It may try to allocate some blocks to that DN as well during an MR job.  This 
 issue is independent from DN decommission.
 To reproduce:
 1. Add a DN to dfs_hosts_allow
 2. Refresh NN
 3. Start DN. Now NN starts seeing DN.
 4. Stop DN
 5. Remove DN from dfs_hosts_allow
 6. Refresh NN - NN is still reporting DN as being used by HDFS.
 This is different from decom because there DN is added to exclude list in 
 addition to being removed from allowed list, and in that case everything 
 works correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8954) Not all the one block in RAM_DISK use lazy_persist policy

2015-08-25 Thread Xu Chen (JIRA)
Xu Chen created HDFS-8954:
-

 Summary: Not all the one block in RAM_DISK use lazy_persist policy
 Key: HDFS-8954
 URL: https://issues.apache.org/jira/browse/HDFS-8954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer  mover
Affects Versions: 2.7.1
Reporter: Xu Chen
Priority: Critical


I was try to use Heterogeneous Storage feature and I config RAM_DISK volume for 
each DN size of 20g , and set one directory policy of lazy_persist at HDFS , 
then command copyFromLocal to load a 79g file to the directory HDFS ,  it is 
have 593 block , and only 141 block in RAM_DISK , is that a bug ? or my config 
issue?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8955) Support 'hedged' write in DFSClient

2015-08-25 Thread bijaya (JIRA)
bijaya created HDFS-8955:


 Summary: Support 'hedged' write in DFSClient
 Key: HDFS-8955
 URL: https://issues.apache.org/jira/browse/HDFS-8955
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 3.0.0
Reporter: bijaya
Assignee: Liang Xie
 Fix For: 2.4.0


This is a placeholder of hdfs related stuff backport from 
https://issues.apache.org/jira/browse/HBASE-7509

The quorum read ability should be helpful especially to optimize read outliers

we can utilize dfs.dfsclient.quorum.read.threshold.millis  
dfs.dfsclient.quorum.read.threadpool.size to enable/disable the hedged read 
ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
could export the interested metric valus into client system(e.g. HBase's 
regionserver metric).

The core logic is in pread code path, we decide to goto the original 
fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
the above config items.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8955) Support 'hedged' write in DFSClient

2015-08-25 Thread bijaya (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

bijaya updated HDFS-8955:
-
Release Note: 
If a write from a block is slow, start up another parallel, 'hedged' write 
against a different set of replica. We need to get different set of 
replica/data pipeline from NN.  We then take the result of which ever write 
returns first (the outstanding write is cancelled).  This 'hedged' write 
feature will help rein in the outliers, the odd write that takes a long time 
because it hit a bad patch on the disc, etc.

This feature is off by default.  To enable this feature, set 
codedfs.client.hedged.write.threadpool.size/code to a positive number.  The 
threadpool size is how many threads to dedicate to the running of these 
'hedged', concurrent writes in your client.

Then set codedfs.client.hedged.write.threshold.millis/code to the number of 
milliseconds to wait before starting up a 'hedged' write.  For example, if you 
set this property to 10, then if a write has not returned within 10 
milliseconds, we will start up a new read against a different block replica.

This feature emits new metrics:

+ hedgedWriteOps
+ hedgeWriteOpsWin -- how many times the hedged write 'beat' the original write
+ hedgedWriteOpsInCurThread -- how many times we went to do a hedged write but 
we had to run it in the current thread because 
dfs.client.hedged.write.threadpool.size was at a maximum.

  was:
If a read from a block is slow, start up another parallel, 'hedged' read 
against a different block replica.  We then take the result of which ever read 
returns first (the outstanding read is cancelled).  This 'hedged' read feature 
will help rein in the outliers, the odd read that takes a long time because it 
hit a bad patch on the disc, etc.

This feature is off by default.  To enable this feature, set 
codedfs.client.hedged.read.threadpool.size/code to a positive number.  The 
threadpool size is how many threads to dedicate to the running of these 
'hedged', concurrent reads in your client.

Then set codedfs.client.hedged.read.threshold.millis/code to the number of 
milliseconds to wait before starting up a 'hedged' read.  For example, if you 
set this property to 10, then if a read has not returned within 10 
milliseconds, we will start up a new read against a different block replica.

This feature emits new metrics:

+ hedgedReadOps
+ hedgeReadOpsWin -- how many times the hedged read 'beat' the original read
+ hedgedReadOpsInCurThread -- how many times we went to do a hedged read but we 
had to run it in the current thread because 
dfs.client.hedged.read.threadpool.size was at a maximum.


 Support 'hedged' write in DFSClient
 ---

 Key: HDFS-8955
 URL: https://issues.apache.org/jira/browse/HDFS-8955
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 3.0.0
Reporter: bijaya
Assignee: bijaya
 Fix For: 2.4.0


 This is a placeholder of hdfs related stuff backport from 
 https://issues.apache.org/jira/browse/HBASE-7509
 The quorum read ability should be helpful especially to optimize read outliers
 we can utilize dfs.dfsclient.quorum.read.threshold.millis  
 dfs.dfsclient.quorum.read.threadpool.size to enable/disable the hedged read 
 ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
 could export the interested metric valus into client system(e.g. HBase's 
 regionserver metric).
 The core logic is in pread code path, we decide to goto the original 
 fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
 the above config items.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8956) Not able to start Datanode

2015-08-25 Thread sreelakshmi (JIRA)
sreelakshmi created HDFS-8956:
-

 Summary: Not able to start Datanode
 Key: HDFS-8956
 URL: https://issues.apache.org/jira/browse/HDFS-8956
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.7.0
 Environment: Centos
Reporter: sreelakshmi


Data node service is not started on one of the data nodes, java.net.bind 
exception is thrown. 
Verified that ports 50010,50070 and 50075 are not in use by any other 
application.

15/08/26 01:50:15 INFO http.HttpServer2: HttpServer.start() threw a non Bind 
IOException
java.net.BindException: Port in use: localhost:0
at 
org.apache.hadoop.http.HttpServer2.openListeners(HttpServer2.java:919)
at org.apache.hadoop.http.HttpServer2.start(HttpServer2.java:856)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.startInfoServer(DataNode.java:779)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:1134)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:434)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2404)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2291)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:2338)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2515)
at 
org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2539)
Caused by: java.net.BindException: Cannot assign requested address




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8897) Loadbalancer always exits with : java.io.IOException: Another Balancer is running.. Exiting ...

2015-08-25 Thread Xu Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8897?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712382#comment-14712382
 ] 

Xu Chen commented on HDFS-8897:
---

[~Alexandre LINTE] I was encounter this issue too , when I use hdfs balancer 
command , did you to do this patch ?

 Loadbalancer always exits with : java.io.IOException: Another Balancer is 
 running..  Exiting ...
 

 Key: HDFS-8897
 URL: https://issues.apache.org/jira/browse/HDFS-8897
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer  mover
Affects Versions: 2.7.1
 Environment: Centos 6.6
Reporter: LINTE

 When balancer is launched, it should test if there is already a 
 /system/balancer.id file in HDFS.
 When the file doesn't exist, the balancer don't want to run : 
 15/08/14 16:35:12 INFO balancer.Balancer: namenodes  = [hdfs://sandbox/, 
 hdfs://sandbox]
 15/08/14 16:35:12 INFO balancer.Balancer: parameters = 
 Balancer.Parameters[BalancingPolicy.Node, threshold=10.0, max idle iteration 
 = 5, number of nodes to be excluded = 0, number of nodes to be included = 0]
 Time Stamp   Iteration#  Bytes Already Moved  Bytes Left To Move  
 Bytes Being Moved
 15/08/14 16:35:14 INFO balancer.KeyManager: Block token params received from 
 NN: update interval=10hrs, 0sec, token lifetime=10hrs, 0sec
 15/08/14 16:35:14 INFO block.BlockTokenSecretManager: Setting block keys
 15/08/14 16:35:14 INFO balancer.KeyManager: Update block keys every 2hrs, 
 30mins, 0sec
 15/08/14 16:35:14 INFO block.BlockTokenSecretManager: Setting block keys
 15/08/14 16:35:14 INFO balancer.KeyManager: Block token params received from 
 NN: update interval=10hrs, 0sec, token lifetime=10hrs, 0sec
 15/08/14 16:35:14 INFO block.BlockTokenSecretManager: Setting block keys
 15/08/14 16:35:14 INFO balancer.KeyManager: Update block keys every 2hrs, 
 30mins, 0sec
 java.io.IOException: Another Balancer is running..  Exiting ...
 Aug 14, 2015 4:35:14 PM  Balancing took 2.408 seconds
 Looking at the audit log file when trying to run the balancer, the balancer 
 create the /system/balancer.id and then delete it on exiting ... 
 2015-08-14 16:37:45,844 INFO FSNamesystem.audit: allowed=true   
 ugi=hdfs@SANDBOX.HADOOP (auth:KERBEROS) ip=/x.x.x.x   cmd=getfileinfo 
 src=/system/balancer.id dst=nullperm=null   proto=rpc
 2015-08-14 16:37:45,900 INFO FSNamesystem.audit: allowed=true   
 ugi=hdfs@SANDBOX.HADOOP (auth:KERBEROS) ip=/x.x.x.x   cmd=create  
 src=/system/balancer.id dst=nullperm=hdfs:hadoop:rw-r-  
 proto=rpc
 2015-08-14 16:37:45,919 INFO FSNamesystem.audit: allowed=true   
 ugi=hdfs@SANDBOX.HADOOP (auth:KERBEROS) ip=/x.x.x.x   cmd=getfileinfo 
 src=/system/balancer.id dst=nullperm=null   proto=rpc
 2015-08-14 16:37:46,090 INFO FSNamesystem.audit: allowed=true   
 ugi=hdfs@SANDBOX.HADOOP (auth:KERBEROS) ip=/x.x.x.x   cmd=getfileinfo 
 src=/system/balancer.id dst=nullperm=null   proto=rpc
 2015-08-14 16:37:46,112 INFO FSNamesystem.audit: allowed=true   
 ugi=hdfs@SANDBOX.HADOOP (auth:KERBEROS) ip=/x.x.x.x   cmd=getfileinfo 
 src=/system/balancer.id dst=nullperm=null   proto=rpc
 2015-08-14 16:37:46,117 INFO FSNamesystem.audit: allowed=true   
 ugi=hdfs@SANDBOX.HADOOP (auth:KERBEROS) ip=/x.x.x.x   cmd=delete  
 src=/system/balancer.id dst=nullperm=null   proto=rpc
 The error seems to be located in 
 org/apache/hadoop/hdfs/server/balancer/NameNodeConnector.java 
 The function checkAndMarkRunning return null even if the /system/balancer.id 
 doesn't exist before entering this function; if it exists, then it is deleted 
 and the balancer exit with the same error.
 
   private OutputStream checkAndMarkRunning() throws IOException {
 try {
   if (fs.exists(idPath)) {
 // try appending to it so that it will fail fast if another balancer 
 is
 // running.
 IOUtils.closeStream(fs.append(idPath));
 fs.delete(idPath, true);
   }
   final FSDataOutputStream fsout = fs.create(idPath, false);
   // mark balancer idPath to be deleted during filesystem closure
   fs.deleteOnExit(idPath);
   if (write2IdFile) {
 fsout.writeBytes(InetAddress.getLocalHost().getHostName());
 fsout.hflush();
   }
   return fsout;
 } catch(RemoteException e) {
   
 if(AlreadyBeingCreatedException.class.getName().equals(e.getClassName())){
 return null;
   } else {
 throw e;
   }
 }
   }
 
 Regards



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8954) Not all the one block in RAM_DISK use lazy_persist policy

2015-08-25 Thread Xu Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Chen updated HDFS-8954:
--
Description: 
I was try to use Heterogeneous Storage feature and I config RAM_DISK volume for 
each DN size of 20g , and set one directory policy of lazy_persist at HDFS , 
then command copyFromLocal to load a 79g file to the directory HDFS ,  it is 
have 593 block , and only 141 block in RAM_DISK , I use command df to look 
tmpfs usage it only my execute command DN fill enough the tmpfs , other DN 
tmpfs is empty !   
that a bug ? or my config issue?

Regard

  was:I was try to use Heterogeneous Storage feature and I config RAM_DISK 
volume for each DN size of 20g , and set one directory policy of lazy_persist 
at HDFS , then command copyFromLocal to load a 79g file to the directory HDFS , 
 it is have 593 block , and only 141 block in RAM_DISK , is that a bug ? or my 
config issue?


 Not all the one block in RAM_DISK use lazy_persist policy
 -

 Key: HDFS-8954
 URL: https://issues.apache.org/jira/browse/HDFS-8954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer  mover
Affects Versions: 2.7.1
Reporter: Xu Chen
Priority: Critical

 I was try to use Heterogeneous Storage feature and I config RAM_DISK volume 
 for each DN size of 20g , and set one directory policy of lazy_persist at 
 HDFS , then command copyFromLocal to load a 79g file to the directory HDFS ,  
 it is have 593 block , and only 141 block in RAM_DISK , I use command df to 
 look tmpfs usage it only my execute command DN fill enough the tmpfs , other 
 DN tmpfs is empty !   
 that a bug ? or my config issue?
 Regard



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Assigned] (HDFS-8955) Support 'hedged' write in DFSClient

2015-08-25 Thread bijaya (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

bijaya reassigned HDFS-8955:


Assignee: bijaya  (was: Liang Xie)

 Support 'hedged' write in DFSClient
 ---

 Key: HDFS-8955
 URL: https://issues.apache.org/jira/browse/HDFS-8955
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 3.0.0
Reporter: bijaya
Assignee: bijaya
 Fix For: 2.4.0


 This is a placeholder of hdfs related stuff backport from 
 https://issues.apache.org/jira/browse/HBASE-7509
 The quorum read ability should be helpful especially to optimize read outliers
 we can utilize dfs.dfsclient.quorum.read.threshold.millis  
 dfs.dfsclient.quorum.read.threadpool.size to enable/disable the hedged read 
 ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
 could export the interested metric valus into client system(e.g. HBase's 
 regionserver metric).
 The core logic is in pread code path, we decide to goto the original 
 fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
 the above config items.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8954) Not balance block in RAM_DISK use lazy_persist policy on datanode local copy

2015-08-25 Thread Xu Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Chen updated HDFS-8954:
--
Description: 
I was try to use Heterogeneous Storage feature and I config RAM_DISK volume for 
each DN size of 20g , and set one directory policy of lazy_persist at HDFS , 
then command copyFromLocal to load a 79g file to the directory HDFS ,  it is 
have 593 block , and only 141 block in RAM_DISK , I use command df to look 
tmpfs usage it only my execute command DN fill enough the tmpfs , other DN 
tmpfs is empty !   

when I use other machine without datanode installed , it is worked fine all DN 
tmpfs is fill enough

Regards

  was:
I was try to use Heterogeneous Storage feature and I config RAM_DISK volume for 
each DN size of 20g , and set one directory policy of lazy_persist at HDFS , 
then command copyFromLocal to load a 79g file to the directory HDFS ,  it is 
have 593 block , and only 141 block in RAM_DISK , I use command df to look 
tmpfs usage it only my execute command DN fill enough the tmpfs , other DN 
tmpfs is empty !   
that a bug ? or my config issue?

Regards


 Not balance block in RAM_DISK use lazy_persist policy  on datanode local copy
 -

 Key: HDFS-8954
 URL: https://issues.apache.org/jira/browse/HDFS-8954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Xu Chen
Priority: Critical

 I was try to use Heterogeneous Storage feature and I config RAM_DISK volume 
 for each DN size of 20g , and set one directory policy of lazy_persist at 
 HDFS , then command copyFromLocal to load a 79g file to the directory HDFS ,  
 it is have 593 block , and only 141 block in RAM_DISK , I use command df to 
 look tmpfs usage it only my execute command DN fill enough the tmpfs , other 
 DN tmpfs is empty !   
 when I use other machine without datanode installed , it is worked fine all 
 DN tmpfs is fill enough
 Regards



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8954) Not balance block in RAM_DISK use lazy_persist policy on datanode local copy

2015-08-25 Thread Xu Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Chen updated HDFS-8954:
--
Summary: Not balance block in RAM_DISK use lazy_persist policy  on datanode 
local copy  (was: Not all the one block in RAM_DISK use lazy_persist policy)

 Not balance block in RAM_DISK use lazy_persist policy  on datanode local copy
 -

 Key: HDFS-8954
 URL: https://issues.apache.org/jira/browse/HDFS-8954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Xu Chen
Priority: Critical

 I was try to use Heterogeneous Storage feature and I config RAM_DISK volume 
 for each DN size of 20g , and set one directory policy of lazy_persist at 
 HDFS , then command copyFromLocal to load a 79g file to the directory HDFS ,  
 it is have 593 block , and only 141 block in RAM_DISK , I use command df to 
 look tmpfs usage it only my execute command DN fill enough the tmpfs , other 
 DN tmpfs is empty !   
 that a bug ? or my config issue?
 Regards



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8954) Not all the one block in RAM_DISK use lazy_persist policy

2015-08-25 Thread Xu Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Chen updated HDFS-8954:
--
Description: 
I was try to use Heterogeneous Storage feature and I config RAM_DISK volume for 
each DN size of 20g , and set one directory policy of lazy_persist at HDFS , 
then command copyFromLocal to load a 79g file to the directory HDFS ,  it is 
have 593 block , and only 141 block in RAM_DISK , I use command df to look 
tmpfs usage it only my execute command DN fill enough the tmpfs , other DN 
tmpfs is empty !   
that a bug ? or my config issue?

Regards

  was:
I was try to use Heterogeneous Storage feature and I config RAM_DISK volume for 
each DN size of 20g , and set one directory policy of lazy_persist at HDFS , 
then command copyFromLocal to load a 79g file to the directory HDFS ,  it is 
have 593 block , and only 141 block in RAM_DISK , I use command df to look 
tmpfs usage it only my execute command DN fill enough the tmpfs , other DN 
tmpfs is empty !   
that a bug ? or my config issue?

Regard


 Not all the one block in RAM_DISK use lazy_persist policy
 -

 Key: HDFS-8954
 URL: https://issues.apache.org/jira/browse/HDFS-8954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer  mover
Affects Versions: 2.7.1
Reporter: Xu Chen
Priority: Critical

 I was try to use Heterogeneous Storage feature and I config RAM_DISK volume 
 for each DN size of 20g , and set one directory policy of lazy_persist at 
 HDFS , then command copyFromLocal to load a 79g file to the directory HDFS ,  
 it is have 593 block , and only 141 block in RAM_DISK , I use command df to 
 look tmpfs usage it only my execute command DN fill enough the tmpfs , other 
 DN tmpfs is empty !   
 that a bug ? or my config issue?
 Regards



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8955) Support 'hedged' write in DFSClient

2015-08-25 Thread bijaya (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

bijaya updated HDFS-8955:
-
Affects Version/s: (was: 3.0.0)
   2.6.0

 Support 'hedged' write in DFSClient
 ---

 Key: HDFS-8955
 URL: https://issues.apache.org/jira/browse/HDFS-8955
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0
Reporter: bijaya
Assignee: bijaya

 We do have hedged read which serves redundancy on read failures due to bad 
 sector/patch in disk. We need to have similar feature for hdfs write. This 
 feature may come with cost but its something to must have for use case which 
 needs to guarantee write success regardless of degraded disk health. 
 Defination of degraded disk is highly debatable but this is what I would 
 define. Degraded disk is the disk which fails to read and write 
 intermittently



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8955) Support 'hedged' write in DFSClient

2015-08-25 Thread bijaya (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

bijaya updated HDFS-8955:
-
Description: We do have hedged read which serves redundancy on read 
failures due to bad sector/patch in disk. We need to have similar feature for 
hdfs write. This feature may come with cost but its something to must have for 
use case which needs to guarantee write success regardless of degraded disk 
health. Defination of degraded disk is highly debatable but this is what I 
would define. Degraded disk is the disk which fails to read and write 
intermittently  (was: This is a placeholder of hdfs related stuff backport 
from https://issues.apache.org/jira/browse/HBASE-7509

The quorum read ability should be helpful especially to optimize read outliers

we can utilize dfs.dfsclient.quorum.read.threshold.millis  
dfs.dfsclient.quorum.read.threadpool.size to enable/disable the hedged read 
ability from client side(e.g. HBase), and by using DFSQuorumReadMetrics, we 
could export the interested metric valus into client system(e.g. HBase's 
regionserver metric).

The core logic is in pread code path, we decide to goto the original 
fetchBlockByteRange or the new introduced fetchBlockByteRangeSpeculative per 
the above config items.)

 Support 'hedged' write in DFSClient
 ---

 Key: HDFS-8955
 URL: https://issues.apache.org/jira/browse/HDFS-8955
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0
Reporter: bijaya
Assignee: bijaya

 We do have hedged read which serves redundancy on read failures due to bad 
 sector/patch in disk. We need to have similar feature for hdfs write. This 
 feature may come with cost but its something to must have for use case which 
 needs to guarantee write success regardless of degraded disk health. 
 Defination of degraded disk is highly debatable but this is what I would 
 define. Degraded disk is the disk which fails to read and write 
 intermittently



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8955) Support 'hedged' write in DFSClient

2015-08-25 Thread bijaya (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

bijaya updated HDFS-8955:
-
Fix Version/s: (was: 2.4.0)

 Support 'hedged' write in DFSClient
 ---

 Key: HDFS-8955
 URL: https://issues.apache.org/jira/browse/HDFS-8955
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0
Reporter: bijaya
Assignee: bijaya

 We do have hedged read which serves redundancy on read failures due to bad 
 sector/patch in disk. We need to have similar feature for hdfs write. This 
 feature may come with cost but its something to must have for use case which 
 needs to guarantee write success regardless of degraded disk health. 
 Defination of degraded disk is highly debatable but this is what I would 
 define. Degraded disk is the disk which fails to read and write 
 intermittently



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-4750) Support NFSv3 interface to HDFS

2015-08-25 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-4750?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712494#comment-14712494
 ] 

Brandon Li commented on HDFS-4750:
--

Overlapped writes happen usually for the first block.
For example, a file has 100 bytes. User write this file to NFS mount point. 
After a while (the 100 bytes still in OS buffer cache), the user appends 
another 100 bytes. In this case, the NFS client will rewrite the first block 
(0-199th byte), and thus we see an overlapped write. These kind of case should 
be well handled by current code (but, still might have bugs there :-(   )

Based on your description, the problem seems not the above case. There is one 
case, we currently have not way to control it on the server side:

1. client sends write  0-99, we cache this write and do the write asynchronously
2. client sends another write  180-299, we cache it too, can't write since 
there is a hole
3. client sends another write 100-199, we will do the write since it's a 
sequential write
4. after we finish writing (0-99) and (100-199), we see an overlapped write 
(180-299) in the cache. This is where you see the error message since we are 
expecting another sequential write (200-xxx)

This kind of overlapped write happens very rarely. In case it happens, we have 
multiple copies of the same range (180-198 in the above example). The data 
could be different. When they are different, it could be hard to know which one 
is really expected by the client. 

 Support NFSv3 interface to HDFS
 ---

 Key: HDFS-4750
 URL: https://issues.apache.org/jira/browse/HDFS-4750
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: nfs
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Attachments: HADOOP-NFS-Proposal.pdf, HDFS-4750.patch, nfs-trunk.patch


 Access HDFS is usually done through HDFS Client or webHDFS. Lack of seamless 
 integration with client’s file system makes it difficult for users and 
 impossible for some applications to access HDFS. NFS interface support is one 
 way for HDFS to have such easy integration.
 This JIRA is to track the NFS protocol support for accessing HDFS. With HDFS 
 client, webHDFS and the NFS interface, HDFS will be easier to access and be 
 able support more applications and use cases. 
 We will upload the design document and the initial implementation. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8954) Not all the one block in RAM_DISK use lazy_persist policy

2015-08-25 Thread Xu Chen (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8954?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xu Chen updated HDFS-8954:
--
Component/s: (was: balancer  mover)
 datanode

 Not all the one block in RAM_DISK use lazy_persist policy
 -

 Key: HDFS-8954
 URL: https://issues.apache.org/jira/browse/HDFS-8954
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Xu Chen
Priority: Critical

 I was try to use Heterogeneous Storage feature and I config RAM_DISK volume 
 for each DN size of 20g , and set one directory policy of lazy_persist at 
 HDFS , then command copyFromLocal to load a 79g file to the directory HDFS ,  
 it is have 593 block , and only 141 block in RAM_DISK , I use command df to 
 look tmpfs usage it only my execute command DN fill enough the tmpfs , other 
 DN tmpfs is empty !   
 that a bug ? or my config issue?
 Regards



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8853) Erasure Coding: Provide ECSchema validation when creating ECZone

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712541#comment-14712541
 ] 

Zhe Zhang commented on HDFS-8853:
-

Thanks for the update [~andreina]. Since there are many failures (likely to be 
unrelated though), I just triggered Jenkins again.

 Erasure Coding: Provide ECSchema validation when creating ECZone
 

 Key: HDFS-8853
 URL: https://issues.apache.org/jira/browse/HDFS-8853
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Rakesh R
Assignee: J.Andreina
 Attachments: HDFS-8853-HDFS-7285-01.patch, 
 HDFS-8853-HDFS-7285-05.patch, HDFS-8853-HDFS-7285-merge-02.patch, 
 HDFS-8853-HDFS-7285-merge-03.patch, HDFS-8853-HDFS-7285-merge-04.patch


 Presently the {{DFS#createErasureCodingZone(path, ecSchema, cellSize)}} 
 doesn't have any validation that the given {{ecSchema}} is available in 
 {{ErasureCodingSchemaManager#activeSchemas}} list. Now, if it doesn't exists 
 then will create the ECZone with {{null}} schema. IMHO we could improve this 
 by doing necessary basic sanity checks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8581) count cmd calculate wrong when huge files exist in one folder

2015-08-25 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-8581:
-
Attachment: HDFS-8581.3.patch

Updated the patch.
Please review.

 count cmd calculate wrong when huge files exist in one folder
 -

 Key: HDFS-8581
 URL: https://issues.apache.org/jira/browse/HDFS-8581
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Reporter: tongshiquan
Assignee: J.Andreina
Priority: Minor
 Attachments: HDFS-8581.1.patch, HDFS-8581.2.patch, HDFS-8581.3.patch


 If one directory such as /result exists about 20 files, then when 
 execute hdfs dfs -count /, the result will go wrong. For all directories 
 whose name after /result, file num will not be included.
 My cluster see as below, /result_1433858936 is the directory exist huge 
 files, and files in /sparkJobHistory, /tmp, /user are not included
 vm-221:/export1/BigData/current # hdfs dfs -ls /
 15/06/11 11:00:17 INFO hdfs.PeerCache: SocketCache disabled.
 Found 9 items
 -rw-r--r--   3 hdfs   supergroup  0 2015-06-08 12:10 
 /PRE_CREATE_DIR.SUCCESS
 drwxr-x---   - flume  hadoop  0 2015-06-08 12:08 /flume
 drwx--   - hbase  hadoop  0 2015-06-10 15:25 /hbase
 drwxr-xr-x   - hdfs   supergroup  0 2015-06-10 17:19 /hyt
 drwxrwxrwx   - mapred hadoop  0 2015-06-08 12:08 /mr-history
 drwxr-xr-x   - hdfs   supergroup  0 2015-06-09 22:10 
 /result_1433858936
 drwxrwxrwx   - spark  supergroup  0 2015-06-10 19:15 /sparkJobHistory
 drwxrwxrwx   - hdfs   hadoop  0 2015-06-08 12:14 /tmp
 drwxrwxrwx   - hdfs   hadoop  0 2015-06-09 21:57 /user
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /
 15/06/11 11:00:24 INFO hdfs.PeerCache: SocketCache disabled.
 1043   171536 1756375688 /
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /PRE_CREATE_DIR.SUCCESS
 15/06/11 11:00:30 INFO hdfs.PeerCache: SocketCache disabled.
01  0 /PRE_CREATE_DIR.SUCCESS
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /flume
 15/06/11 11:00:41 INFO hdfs.PeerCache: SocketCache disabled.
10  0 /flume
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /hbase
 15/06/11 11:00:49 INFO hdfs.PeerCache: SocketCache disabled.
   36   18  14807 /hbase
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /hyt
 15/06/11 11:01:09 INFO hdfs.PeerCache: SocketCache disabled.
10  0 /hyt
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /mr-history
 15/06/11 11:01:18 INFO hdfs.PeerCache: SocketCache disabled.
30  0 /mr-history
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /result_1433858936
 15/06/11 11:01:29 INFO hdfs.PeerCache: SocketCache disabled.
 1001   171517 1756360881 /result_1433858936
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /sparkJobHistory
 15/06/11 11:01:41 INFO hdfs.PeerCache: SocketCache disabled.
13  21785 /sparkJobHistory
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /tmp
 15/06/11 11:01:48 INFO hdfs.PeerCache: SocketCache disabled.
   176  35958 /tmp
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /user
 15/06/11 11:01:55 INFO hdfs.PeerCache: SocketCache disabled.
   121  19077 /user



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8853) Erasure Coding: Provide ECSchema validation when creating ECZone

2015-08-25 Thread J.Andreina (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8853?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712520#comment-14712520
 ] 

J.Andreina commented on HDFS-8853:
--

Testcase failures are not related to this patch. Please review.

 Erasure Coding: Provide ECSchema validation when creating ECZone
 

 Key: HDFS-8853
 URL: https://issues.apache.org/jira/browse/HDFS-8853
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Rakesh R
Assignee: J.Andreina
 Attachments: HDFS-8853-HDFS-7285-01.patch, 
 HDFS-8853-HDFS-7285-05.patch, HDFS-8853-HDFS-7285-merge-02.patch, 
 HDFS-8853-HDFS-7285-merge-03.patch, HDFS-8853-HDFS-7285-merge-04.patch


 Presently the {{DFS#createErasureCodingZone(path, ecSchema, cellSize)}} 
 doesn't have any validation that the given {{ecSchema}} is available in 
 {{ErasureCodingSchemaManager#activeSchemas}} list. Now, if it doesn't exists 
 then will create the ECZone with {{null}} schema. IMHO we could improve this 
 by doing necessary basic sanity checks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8909) Erasure coding: update BlockInfoContiguousUC and BlockInfoStripedUC to use BlockUnderConstructionFeature

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710763#comment-14710763
 ] 

Hadoop QA commented on HDFS-8909:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  15m 35s | Findbugs (version ) appears to 
be broken on HDFS-7285. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 12 new or modified test files. |
| {color:green}+1{color} | javac |   7m 56s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 46s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 15s | The applied patch generated 
1 release audit warnings. |
| {color:green}+1{color} | checkstyle |   0m 32s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m 13s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 32s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   2m 41s | The patch appears to introduce 4 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m  4s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests |  20m 10s | Tests failed in hadoop-hdfs. |
| | |  62m 22s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
| Failed unit tests | hadoop.hdfs.TestErasureCodingZones |
|   | hadoop.hdfs.server.namenode.ha.TestXAttrsWithHA |
|   | hadoop.hdfs.server.datanode.TestDataNodeMXBean |
|   | hadoop.hdfs.server.namenode.ha.TestFailureOfSharedDir |
|   | hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes |
|   | hadoop.hdfs.server.namenode.TestEditLogRace |
|   | hadoop.hdfs.server.namenode.ha.TestStateTransitionFailure |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestScrLazyPersistFiles |
|   | hadoop.hdfs.TestDFSRename |
|   | hadoop.hdfs.server.datanode.TestBlockRecovery |
|   | hadoop.hdfs.server.blockmanagement.TestReplicationPolicyWithNodeGroup |
|   | hadoop.hdfs.TestDatanodeConfig |
|   | hadoop.hdfs.server.namenode.TestQuotaByStorageType |
|   | hadoop.hdfs.server.namenode.snapshot.TestSnapshotFileLength |
|   | hadoop.hdfs.server.namenode.TestBlockPlacementPolicyRackFaultTolerant |
|   | hadoop.hdfs.server.namenode.ha.TestStandbyCheckpoints |
|   | hadoop.hdfs.server.namenode.TestXAttrConfigFlag |
|   | hadoop.hdfs.TestBlockReaderFactory |
|   | hadoop.hdfs.server.namenode.ha.TestDFSUpgradeWithHA |
|   | hadoop.hdfs.server.namenode.TestFSNamesystem |
|   | hadoop.hdfs.TestDFSClientRetries |
|   | hadoop.hdfs.server.namenode.TestFSImage |
|   | hadoop.hdfs.TestFSOutputSummer |
|   | hadoop.hdfs.TestFileCreationDelete |
|   | hadoop.hdfs.server.datanode.TestDnRespectsBlockReportSplitThreshold |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestInterDatanodeProtocol |
|   | hadoop.hdfs.TestSnapshotCommands |
|   | hadoop.hdfs.TestFileConcurrentReader |
|   | hadoop.tools.TestJMXGet |
|   | hadoop.hdfs.server.namenode.TestCheckpoint |
|   | hadoop.hdfs.server.datanode.TestDiskError |
|   | hadoop.hdfs.server.datanode.TestRefreshNamenodes |
|   | hadoop.hdfs.TestFileAppend2 |
|   | hadoop.hdfs.tools.TestDFSAdmin |
|   | hadoop.fs.contract.hdfs.TestHDFSContractRootDirectory |
|   | hadoop.hdfs.TestBlockStoragePolicy |
|   | hadoop.hdfs.web.TestWebHdfsWithAuthenticationFilter |
|   | hadoop.hdfs.server.namenode.TestNNStorageRetentionFunctional |
|   | hadoop.fs.TestUrlStreamHandler |
|   | hadoop.hdfs.server.namenode.TestMetaSave |
|   | hadoop.hdfs.TestDFSRemove |
|   | hadoop.hdfs.TestLeaseRecovery |
|   | hadoop.hdfs.server.namenode.ha.TestBootstrapStandbyWithQJM |
|   | hadoop.hdfs.TestClientProtocolForPipelineRecovery |
|   | hadoop.hdfs.server.namenode.TestStartupProgressServlet |
|   | hadoop.hdfs.server.namenode.ha.TestEditLogTailer |
|   | hadoop.fs.contract.hdfs.TestHDFSContractSeek |
|   | hadoop.hdfs.server.datanode.TestHSync |
|   | hadoop.hdfs.tools.TestDelegationTokenFetcher |
|   | hadoop.tracing.TestTraceAdmin |
|   | hadoop.hdfs.server.namenode.TestGenericJournalConf |
|   | hadoop.hdfs.TestWriteConfigurationToDFS |
|   | hadoop.hdfs.server.blockmanagement.TestDatanodeManager |
|   | hadoop.hdfs.tools.TestStoragePolicyCommands |
|   | hadoop.hdfs.web.TestJsonUtil |
|   | hadoop.hdfs.server.namenode.TestNNThroughputBenchmark |
|   | hadoop.hdfs.TestReplication |
|   | hadoop.hdfs.server.namenode.snapshot.TestNestedSnapshots |
|   | hadoop.hdfs.TestWriteStripedFileWithFailure |
|   | hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork |
|   | hadoop.hdfs.server.namenode.TestAclConfigFlag |
|   | 

[jira] [Commented] (HDFS-8951) Move shortcircuit to hdfs-client

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710844#comment-14710844
 ] 

Hadoop QA commented on HDFS-8951:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  19m 18s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 55s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 56s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 16s | The applied patch generated 
1 release audit warnings. |
| {color:red}-1{color} | checkstyle |   2m 26s | The applied patch generated  
61 new checkstyle issues (total was 16, now 77). |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 39s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   4m 29s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 10s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests |  68m 23s | Tests failed in hadoop-hdfs. |
| {color:green}+1{color} | hdfs tests |   0m 28s | Tests passed in 
hadoop-hdfs-client. |
| | | 118m 38s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.server.namenode.TestSecondaryNameNodeUpgrade |
|   | hadoop.hdfs.server.namenode.web.resources.TestWebHdfsDataLocality |
|   | hadoop.hdfs.server.datanode.TestDataNodeMetrics |
|   | hadoop.hdfs.server.datanode.TestStorageReport |
|   | hadoop.hdfs.web.TestWebHDFSForHA |
|   | hadoop.hdfs.server.namenode.TestGenericJournalConf |
|   | hadoop.cli.TestCryptoAdminCLI |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistFiles |
|   | hadoop.hdfs.server.datanode.TestDataNodeHotSwapVolumes |
|   | hadoop.hdfs.TestSetrepIncreasing |
|   | hadoop.hdfs.server.namenode.TestSecondaryWebUi |
|   | hadoop.hdfs.TestFetchImage |
|   | hadoop.hdfs.server.namenode.ha.TestHAStateTransitions |
|   | hadoop.hdfs.server.datanode.TestReadOnlySharedStorage |
|   | hadoop.hdfs.server.namenode.snapshot.TestUpdatePipelineWithSnapshots |
|   | hadoop.hdfs.web.TestHttpsFileSystem |
|   | hadoop.hdfs.server.namenode.TestMetadataVersionOutput |
|   | 
hadoop.hdfs.server.namenode.snapshot.TestSnapshotNameWithInvalidCharacters |
|   | hadoop.hdfs.server.namenode.TestNameNodeResourceChecker |
|   | hadoop.hdfs.TestMiniDFSCluster |
|   | hadoop.hdfs.TestDFSClientRetries |
|   | hadoop.hdfs.TestEncryptionZonesWithKMS |
|   | hadoop.hdfs.TestReplication |
|   | hadoop.hdfs.server.namenode.TestCheckpoint |
|   | hadoop.hdfs.TestDFSOutputStream |
|   | hadoop.hdfs.server.datanode.fsdataset.impl.TestWriteToReplica |
|   | hadoop.hdfs.TestClientBlockVerification |
|   | hadoop.hdfs.TestDFSRollback |
|   | hadoop.hdfs.server.namenode.ha.TestPendingCorruptDnMessages |
|   | hadoop.hdfs.server.namenode.snapshot.TestXAttrWithSnapshot |
|   | hadoop.hdfs.server.namenode.TestCreateEditsLog |
|   | hadoop.tools.TestJMXGet |
|   | hadoop.hdfs.server.namenode.ha.TestGetGroupsWithHA |
|   | hadoop.hdfs.TestParallelRead |
|   | hadoop.hdfs.TestHDFSTrash |
|   | hadoop.hdfs.tools.TestDelegationTokenFetcher |
|   | hadoop.hdfs.server.blockmanagement.TestBlockReportRateLimiting |
|   | hadoop.hdfs.server.namenode.TestQuotaByStorageType |
|   | hadoop.hdfs.server.namenode.TestBlockUnderConstruction |
|   | hadoop.hdfs.server.datanode.TestIncrementalBlockReports |
|   | hadoop.hdfs.TestWriteRead |
|   | hadoop.hdfs.TestFileAppend4 |
|   | hadoop.hdfs.web.TestFSMainOperationsWebHdfs |
|   | hadoop.hdfs.TestSnapshotCommands |
|   | hadoop.hdfs.TestBlocksScheduledCounter |
|   | hadoop.hdfs.web.TestWebHdfsWithMultipleNameNodes |
|   | hadoop.hdfs.server.datanode.TestNNHandlesBlockReportPerStorage |
|   | hadoop.hdfs.TestClientProtocolForPipelineRecovery |
|   | hadoop.hdfs.crypto.TestHdfsCryptoStreams |
|   | hadoop.hdfs.TestLeaseRecovery2 |
|   | hadoop.hdfs.TestFSOutputSummer |
|   | hadoop.hdfs.server.namenode.TestDeleteRace |
|   | hadoop.hdfs.server.namenode.TestAddBlockRetry |
|   | hadoop.hdfs.server.namenode.TestCheckPointForSecurityTokens |
|   | hadoop.hdfs.server.namenode.snapshot.TestSnapshotFileLength |
|   | hadoop.hdfs.server.namenode.TestListCorruptFileBlocks |
|   | 

[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710857#comment-14710857
 ] 

Yi Liu commented on HDFS-8900:
--

Thanks Andrew, Colin and Brahma.

The findbugs doesn't exist. (The existing findbugs were in pre-patch).

Committed to trunk and branch-2.

 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Liu updated HDFS-8900:
-
   Resolution: Fixed
 Hadoop Flags: Incompatible change,Reviewed  (was: Incompatible change)
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7285) Erasure Coding Support inside HDFS

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7285?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710770#comment-14710770
 ] 

Zhe Zhang commented on HDFS-7285:
-

[~vinayrpet] Could you take a look at the results comparing the two 
rebase/merge efforts, and see if we should delete the {{HDFS-7285-REBASE}} 
branch? Thanks!

 Erasure Coding Support inside HDFS
 --

 Key: HDFS-7285
 URL: https://issues.apache.org/jira/browse/HDFS-7285
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Weihua Jiang
Assignee: Zhe Zhang
 Attachments: Compare-consolidated-20150824.diff, 
 Consolidated-20150707.patch, Consolidated-20150806.patch, 
 Consolidated-20150810.patch, ECAnalyzer.py, ECParser.py, 
 HDFS-7285-initial-PoC.patch, HDFS-7285-merge-consolidated-01.patch, 
 HDFS-7285-merge-consolidated-trunk-01.patch, 
 HDFS-7285-merge-consolidated.trunk.03.patch, 
 HDFS-7285-merge-consolidated.trunk.04.patch, 
 HDFS-EC-Merge-PoC-20150624.patch, HDFS-EC-merge-consolidated-01.patch, 
 HDFS-bistriped.patch, HDFSErasureCodingDesign-20141028.pdf, 
 HDFSErasureCodingDesign-20141217.pdf, HDFSErasureCodingDesign-20150204.pdf, 
 HDFSErasureCodingDesign-20150206.pdf, HDFSErasureCodingPhaseITestPlan.pdf, 
 HDFSErasureCodingSystemTestPlan-20150824.pdf, fsimage-analysis-20150105.pdf


 Erasure Coding (EC) can greatly reduce the storage overhead without sacrifice 
 of data reliability, comparing to the existing HDFS 3-replica approach. For 
 example, if we use a 10+4 Reed Solomon coding, we can allow loss of 4 blocks, 
 with storage overhead only being 40%. This makes EC a quite attractive 
 alternative for big data storage, particularly for cold data. 
 Facebook had a related open source project called HDFS-RAID. It used to be 
 one of the contribute packages in HDFS but had been removed since Hadoop 2.0 
 for maintain reason. The drawbacks are: 1) it is on top of HDFS and depends 
 on MapReduce to do encoding and decoding tasks; 2) it can only be used for 
 cold files that are intended not to be appended anymore; 3) the pure Java EC 
 coding implementation is extremely slow in practical use. Due to these, it 
 might not be a good idea to just bring HDFS-RAID back.
 We (Intel and Cloudera) are working on a design to build EC into HDFS that 
 gets rid of any external dependencies, makes it self-contained and 
 independently maintained. This design lays the EC feature on the storage type 
 support and considers compatible with existing HDFS features like caching, 
 snapshot, encryption, high availability and etc. This design will also 
 support different EC coding schemes, implementations and policies for 
 different deployment scenarios. By utilizing advanced libraries (e.g. Intel 
 ISA-L library), an implementation can greatly improve the performance of EC 
 encoding/decoding and makes the EC solution even more attractive. We will 
 post the design document soon. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8838) Tolerate datanode failures in DFSStripedOutputStream when the data length is small

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710775#comment-14710775
 ] 

Zhe Zhang commented on HDFS-8838:
-

Thanks Walter for the discussion. If you are still +1 on the latest patch, it's 
OK to commit to the HDFS-7285 branch now.

 Tolerate datanode failures in DFSStripedOutputStream when the data length is 
 small
 --

 Key: HDFS-8838
 URL: https://issues.apache.org/jira/browse/HDFS-8838
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client
Reporter: Tsz Wo Nicholas Sze
Assignee: Tsz Wo Nicholas Sze
 Attachments: HDFS-8838-HDFS-7285-000.patch, 
 HDFS-8838-HDFS-7285-20150809-test.patch, HDFS-8838-HDFS-7285-20150809.patch, 
 HDFS-8838-HDFS-7285-20150821.patch, h8838_20150729.patch, 
 h8838_20150731-HDFS-7285.patch, h8838_20150731.log, h8838_20150731.patch, 
 h8838_20150804-HDFS-7285.patch, h8838_20150809.patch


 Currently, DFSStripedOutputStream cannot tolerate datanode failures when the 
 data length is small.  We fix the bugs here and add more tests.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a command to get the balancer bandwidth

2015-08-25 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710820#comment-14710820
 ] 

Rakesh R commented on HDFS-7116:


It looks like test case failures are not related to the proposed patch, please 
ignore it !

 Add a command to get the balancer bandwidth
 ---

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA
Assignee: Rakesh R
 Attachments: HDFS-7116-00.patch, HDFS-7116-01.patch, 
 HDFS-7116-02.patch, HDFS-7116-03.patch, HDFS-7116-04.patch, 
 HDFS-7116-05.patch, HDFS-7116-06.patch


 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the value of the same. 
 This jira to discuss  implement the way to access the balancer bandwidth 
 value of the datanode.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8155) Support OAuth2 in WebHDFS

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710910#comment-14710910
 ] 

Hadoop QA commented on HDFS-8155:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  22m 16s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 4 new or modified test files. |
| {color:red}-1{color} | javac |   7m 50s | The applied patch generated  2  
additional warning messages. |
| {color:green}+1{color} | javadoc |   9m 59s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | site |   3m  3s | Site still builds. |
| {color:red}-1{color} | checkstyle |   2m 40s | The applied patch generated  
36 new checkstyle issues (total was 151, now 187). |
| {color:red}-1{color} | whitespace |   0m  1s | The patch has 7  line(s) that 
end in whitespace. Use git apply --whitespace=fix. |
| {color:green}+1{color} | install |   1m 30s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   4m 21s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 10s | Pre-build of native portion |
| {color:green}+1{color} | hdfs tests | 163m 36s | Tests passed in hadoop-hdfs. 
|
| {color:green}+1{color} | hdfs tests |   0m 27s | Tests passed in 
hadoop-hdfs-client. |
| | | 219m 54s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752142/HDFS-8155.002.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle site |
| git revision | trunk / 57c7ae1 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs-client.html
 |
| javac | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/diffJavacWarnings.txt
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/diffcheckstylehadoop-hdfs-client.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/whitespace.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| hadoop-hdfs-client test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/artifact/patchprocess/testrun_hadoop-hdfs-client.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf908.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12106/console |


This message was automatically generated.

 Support OAuth2 in WebHDFS
 -

 Key: HDFS-8155
 URL: https://issues.apache.org/jira/browse/HDFS-8155
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: webhdfs
Reporter: Jakob Homan
Assignee: Jakob Homan
 Attachments: HDFS-8155-1.patch, HDFS-8155.002.patch


 WebHDFS should be able to accept OAuth2 credentials.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710920#comment-14710920
 ] 

Hudson commented on HDFS-8900:
--

FAILURE: Integrated in Hadoop-trunk-Commit #8348 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8348/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8838) Tolerate datanode failures in DFSStripedOutputStream when the data length is small

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710773#comment-14710773
 ] 

Zhe Zhang commented on HDFS-8838:
-

{{HDFS-7285-merge}} and {{HDFS-7285-REBASE}} are intermediate branches. I have 
just updated the main {{HDFS-7285}} branch, so let's use it from now on. Sorry 
about the confusion.

 Tolerate datanode failures in DFSStripedOutputStream when the data length is 
 small
 --

 Key: HDFS-8838
 URL: https://issues.apache.org/jira/browse/HDFS-8838
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client
Reporter: Tsz Wo Nicholas Sze
Assignee: Tsz Wo Nicholas Sze
 Attachments: HDFS-8838-HDFS-7285-000.patch, 
 HDFS-8838-HDFS-7285-20150809-test.patch, HDFS-8838-HDFS-7285-20150809.patch, 
 HDFS-8838-HDFS-7285-20150821.patch, h8838_20150729.patch, 
 h8838_20150731-HDFS-7285.patch, h8838_20150731.log, h8838_20150731.patch, 
 h8838_20150804-HDFS-7285.patch, h8838_20150809.patch


 Currently, DFSStripedOutputStream cannot tolerate datanode failures when the 
 data length is small.  We fix the bugs here and add more tests.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8941) DistributedFileSystem listCorruptFileBlocks API should resolve relative path

2015-08-25 Thread Rakesh R (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8941?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rakesh R updated HDFS-8941:
---
Status: Patch Available  (was: Open)

 DistributedFileSystem listCorruptFileBlocks API should resolve relative path
 

 Key: HDFS-8941
 URL: https://issues.apache.org/jira/browse/HDFS-8941
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Rakesh R
Assignee: Rakesh R
 Attachments: HDFS-8941-00.patch


 Presently {{DFS#listCorruptFileBlocks(path)}} API is not resolving the given 
 path relative to the workingDir. This jira is to discuss and provide the 
 implementation of the same.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8941) DistributedFileSystem listCorruptFileBlocks API should resolve relative path

2015-08-25 Thread Rakesh R (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8941?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Rakesh R updated HDFS-8941:
---
Status: Open  (was: Patch Available)

 DistributedFileSystem listCorruptFileBlocks API should resolve relative path
 

 Key: HDFS-8941
 URL: https://issues.apache.org/jira/browse/HDFS-8941
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Rakesh R
Assignee: Rakesh R
 Attachments: HDFS-8941-00.patch


 Presently {{DFS#listCorruptFileBlocks(path)}} API is not resolving the given 
 path relative to the workingDir. This jira is to discuss and provide the 
 implementation of the same.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710718#comment-14710718
 ] 

Hadoop QA commented on HDFS-8900:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 50s | Pre-patch trunk has 2 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 2 new or modified test files. |
| {color:green}+1{color} | javac |   7m 58s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  7s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 22s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 25s | The applied patch generated  1 
new checkstyle issues (total was 516, now 511). |
| {color:green}+1{color} | whitespace |   0m  3s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 31s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 35s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   2m 36s | The patch appears to introduce 2 
new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 11s | Pre-build of native portion |
| {color:green}+1{color} | hdfs tests | 163m 27s | Tests passed in hadoop-hdfs. 
|
| | | 209m  9s | |
\\
\\
|| Reason || Tests ||
| FindBugs | module:hadoop-hdfs |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752124/HDFS-8900.005.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / 3b00eae |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12103/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12103/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12103/artifact/patchprocess/newPatchFindbugsWarningshadoop-hdfs.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12103/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12103/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf906.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12103/console |


This message was automatically generated.

 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8939) Test(S)WebHdfsFileContextMainOperations failing on branch-2

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8939?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710671#comment-14710671
 ] 

Hadoop QA commented on HDFS-8939:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  23m 15s | Pre-patch branch-2 has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   6m  8s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 46s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | checkstyle |   2m 55s | There were no new checkstyle 
issues. |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 16s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   4m 51s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   1m 52s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 164m 42s | Tests failed in hadoop-hdfs. |
| {color:green}+1{color} | hdfs tests |   0m 28s | Tests passed in 
hadoop-hdfs-client. |
| | | 216m 16s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.TestDistributedFileSystem |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752106/HDFS-8939-branch-2.002.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | branch-2 / 137bde0 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12101/artifact/patchprocess/branch-2FindbugsWarningshadoop-hdfs.html
 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12101/artifact/patchprocess/branch-2FindbugsWarningshadoop-hdfs-client.html
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12101/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| hadoop-hdfs-client test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12101/artifact/patchprocess/testrun_hadoop-hdfs-client.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12101/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf902.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12101/console |


This message was automatically generated.

 Test(S)WebHdfsFileContextMainOperations failing on branch-2
 ---

 Key: HDFS-8939
 URL: https://issues.apache.org/jira/browse/HDFS-8939
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.8.0
Reporter: Jakob Homan
Assignee: Jakob Homan
 Fix For: 2.8.0

 Attachments: HDFS-8939-branch-2.001.patch, 
 HDFS-8939-branch-2.002.patch


 After HDFS-8180, TestWebHdfsFileContextMainOperations and 
 TestSWebHdfsFileContextMainOperations are failing with runtime NPEs while 
 instantiating the wrapped WebHDFSFileSystems because {{getDefaultPort}} is 
 trying to access a conf that was never provided.  In the constructor both 
 both WebHdfs and SWebhdfs the underlying (S)WebHdfsFileSystems are 
 instantiated in the constructor and never have a chance to have their 
 {{setConf}} methods called:
 {code}  SWebHdfs(URI theUri, Configuration conf)
   throws IOException, URISyntaxException {
 super(theUri, new SWebHdfsFileSystem(), conf, SCHEME, false);
   }r{code}
 The test passes on trunk because HDFS-5321 removed the call to the 
 Configuration instance as part of {{getDefaultPort}}.  HDFS-5321 was applied 
 to branch-2 but reverted in HDFS-6632, so there's a bit of a difference in 
 how branch-2 versus trunk handles default values (branch-2 pulls them from 
 configs if specified, trunk just returns the hard-coded value from the 
 constants file).
 I've fixed this behave like trunk and return just the hard-coded value, which 
 causes the test to pass.
   There is no WebHdfsFileSystem that takes a Config, which would be another 
 way to fix this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8838) Tolerate datanode failures in DFSStripedOutputStream when the data length is small

2015-08-25 Thread Li Bo (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710695#comment-14710695
 ] 

Li Bo commented on HDFS-8838:
-

hi, Walter Su
我看了下branch, 跟7285相关的有
  remotes/origin/HDFS-7285
  remotes/origin/HDFS-7285-REBASE
  remotes/origin/HDFS-7285-merge

这几者目前是什么关系?
Thanks



 Tolerate datanode failures in DFSStripedOutputStream when the data length is 
 small
 --

 Key: HDFS-8838
 URL: https://issues.apache.org/jira/browse/HDFS-8838
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client
Reporter: Tsz Wo Nicholas Sze
Assignee: Tsz Wo Nicholas Sze
 Attachments: HDFS-8838-HDFS-7285-000.patch, 
 HDFS-8838-HDFS-7285-20150809-test.patch, HDFS-8838-HDFS-7285-20150809.patch, 
 HDFS-8838-HDFS-7285-20150821.patch, h8838_20150729.patch, 
 h8838_20150731-HDFS-7285.patch, h8838_20150731.log, h8838_20150731.patch, 
 h8838_20150804-HDFS-7285.patch, h8838_20150809.patch


 Currently, DFSStripedOutputStream cannot tolerate datanode failures when the 
 data length is small.  We fix the bugs here and add more tests.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8248) Store INodeId instead of the INodeFile object in BlockInfoContiguous

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710735#comment-14710735
 ] 

Hadoop QA commented on HDFS-8248:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 48s | Pre-patch trunk has 2 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 7 new or modified test files. |
| {color:green}+1{color} | javac |   7m 39s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 51s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 21s | The applied patch generated  3 
new checkstyle issues (total was 512, now 510). |
| {color:green}+1{color} | whitespace |   0m  2s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 27s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 33s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 29s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 12s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests |  66m 41s | Tests failed in hadoop-hdfs. |
| | | 111m 33s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.server.namenode.ha.TestXAttrsWithHA |
|   | hadoop.hdfs.server.namenode.ha.TestFailureOfSharedDir |
|   | hadoop.hdfs.server.balancer.TestBalancerWithMultipleNameNodes |
|   | hadoop.hdfs.server.namenode.TestEditLogRace |
|   | hadoop.hdfs.TestDFSRename |
|   | hadoop.hdfs.server.blockmanagement.TestReplicationPolicyWithNodeGroup |
|   | hadoop.hdfs.TestDatanodeConfig |
|   | hadoop.hdfs.server.namenode.TestQuotaByStorageType |
|   | hadoop.hdfs.server.namenode.snapshot.TestSnapshotFileLength |
|   | hadoop.hdfs.server.namenode.TestBlockPlacementPolicyRackFaultTolerant |
|   | hadoop.hdfs.server.namenode.TestXAttrConfigFlag |
|   | hadoop.hdfs.TestBlockReaderFactory |
|   | hadoop.hdfs.server.namenode.TestFSNamesystem |
|   | hadoop.hdfs.TestDFSClientRetries |
|   | hadoop.hdfs.server.namenode.TestFSImage |
|   | hadoop.hdfs.TestFSOutputSummer |
|   | hadoop.hdfs.TestFileCreationDelete |
|   | hadoop.hdfs.TestSnapshotCommands |
|   | hadoop.hdfs.TestFileConcurrentReader |
|   | hadoop.hdfs.server.namenode.TestCheckpoint |
|   | hadoop.hdfs.TestFileAppend2 |
|   | hadoop.hdfs.tools.TestDFSAdmin |
|   | hadoop.hdfs.TestBlockStoragePolicy |
|   | hadoop.hdfs.web.TestWebHdfsWithAuthenticationFilter |
|   | hadoop.hdfs.server.namenode.TestNNStorageRetentionFunctional |
|   | hadoop.hdfs.server.namenode.TestMetaSave |
|   | hadoop.hdfs.TestDFSRemove |
|   | hadoop.hdfs.TestLeaseRecovery |
|   | hadoop.hdfs.TestClientProtocolForPipelineRecovery |
|   | hadoop.hdfs.tools.TestDelegationTokenFetcher |
|   | hadoop.tracing.TestTraceAdmin |
|   | hadoop.hdfs.server.namenode.TestGenericJournalConf |
|   | hadoop.hdfs.TestWriteConfigurationToDFS |
|   | hadoop.hdfs.tools.TestStoragePolicyCommands |
|   | hadoop.hdfs.web.TestJsonUtil |
|   | hadoop.hdfs.server.namenode.TestNNThroughputBenchmark |
|   | hadoop.hdfs.TestReplication |
|   | hadoop.hdfs.server.namenode.snapshot.TestNestedSnapshots |
|   | hadoop.hdfs.server.blockmanagement.TestComputeInvalidateWork |
|   | hadoop.hdfs.server.namenode.TestAclConfigFlag |
|   | hadoop.hdfs.TestPipelines |
|   | hadoop.hdfs.tools.TestGetGroups |
|   | hadoop.hdfs.TestFileCreationClient |
|   | hadoop.hdfs.TestParallelShortCircuitRead |
|   | hadoop.hdfs.server.namenode.ha.TestHAFsck |
|   | hadoop.hdfs.server.blockmanagement.TestHeartbeatHandling |
|   | hadoop.hdfs.TestRenameWhileOpen |
|   | hadoop.hdfs.server.blockmanagement.TestSequentialBlockId |
|   | hadoop.hdfs.TestBlocksScheduledCounter |
|   | hadoop.hdfs.TestRollingUpgradeRollback |
|   | hadoop.hdfs.server.namenode.snapshot.TestSnapshotRename |
|   | hadoop.hdfs.server.namenode.snapshot.TestDisallowModifyROSnapshot |
|   | hadoop.hdfs.server.namenode.ha.TestGetGroupsWithHA |
|   | hadoop.hdfs.protocolPB.TestPBHelper |
|   | hadoop.hdfs.server.namenode.TestMalformedURLs |
|   | hadoop.hdfs.server.namenode.TestFSImageWithXAttr |
|   | hadoop.hdfs.server.namenode.TestFavoredNodesEndToEnd |
|   | hadoop.hdfs.server.balancer.TestBalancerWithEncryptedTransfer |
|   | hadoop.hdfs.server.namenode.ha.TestDNFencingWithReplication |
|   | hadoop.hdfs.server.namenode.TestDecommissioningStatus |
|   | hadoop.hdfs.server.balancer.TestBalancerWithSaslDataTransfer |
|   | 

[jira] [Commented] (HDFS-5851) Support memory as a storage medium

2015-08-25 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5851?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710744#comment-14710744
 ] 

Vinayakumar B commented on HDFS-5851:
-

This Jira should be marked as duplicate of HDFS-6581 ?

 Support memory as a storage medium
 --

 Key: HDFS-5851
 URL: https://issues.apache.org/jira/browse/HDFS-5851
 Project: Hadoop HDFS
  Issue Type: New Feature
Affects Versions: 3.0.0
Reporter: Arpit Agarwal
Assignee: Arpit Agarwal
 Attachments: 
 SupportingMemoryStorageinHDFSPersistentandDiscardableMemory.pdf, 
 SupportingMemoryStorageinHDFSPersistentandDiscardableMemory.pdf, 
 SupportingMemoryStorageinHDFSPersistentandDiscardableMemory.pdf


 Memory can be used as a storage medium for smaller/transient files for fast 
 write throughput.
 More information/design will be added later.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8894) Set SO_KEEPALIVE on DN server sockets

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8894?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711714#comment-14711714
 ] 

Hadoop QA commented on HDFS-8894:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  18m 28s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 58s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  4s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 23s | The applied patch generated  1 
new checkstyle issues (total was 150, now 150). |
| {color:red}-1{color} | whitespace |   0m  0s | The patch has 1  line(s) that 
end in whitespace. Use git apply --whitespace=fix. |
| {color:green}+1{color} | install |   1m 35s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 35s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   2m 17s | Post-patch findbugs 
hadoop-hdfs-project/hadoop-hdfs compilation is broken. |
| {color:green}+1{color} | findbugs |   2m 17s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   0m 24s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests |   0m 23s | Tests failed in hadoop-hdfs. |
| | |  43m 34s | |
\\
\\
|| Reason || Tests ||
| Failed build | hadoop-hdfs |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752268/HDFS-8894-01.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12118/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12118/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12118/artifact/patchprocess/whitespace.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12118/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12118/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12118/console |


This message was automatically generated.

 Set SO_KEEPALIVE on DN server sockets
 -

 Key: HDFS-8894
 URL: https://issues.apache.org/jira/browse/HDFS-8894
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Nathan Roberts
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8894-01.patch, HDFS-8894-01.patch


 SO_KEEPALIVE is not set on things like datastreamer sockets which can cause 
 lingering ESTABLISHED sockets when there is a network glitch.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8951) Move shortcircuit to hdfs-client

2015-08-25 Thread Haohui Mai (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711775#comment-14711775
 ] 

Haohui Mai commented on HDFS-8951:
--

Thanks for the work.

{code}
+package org.apache.hadoop.hdfs.util;
+
+import org.slf4j.Logger;
+
+import java.io.IOException;
+
+public class IOUtils {
+
+  /**
+   * Close the Closeable objects and bignore/b any {@link IOException} or
+   * null pointers. Must only be used for cleanup in exception handlers.
+   *
+   * @param log the log to record problems to at debug level. Can be null.
+   * @param closeables the objects to close
+   */
+  public static void cleanup(Logger log, java.io.Closeable... closeables) {
+for (java.io.Closeable c : closeables) {
+  if (c != null) {
+try {
+  c.close();
+} catch(Throwable e) {
+  if (log != null  log.isDebugEnabled()) {
+log.debug(Exception in closing  + c, e);
+  }
+}
+  }
+}
+  }
+
+}
{code}

The change is not required. It's better to change the callers to use 
try-with-resource statement in Java 7.

{code}
-final DataChecksum checksum = 
BlockMetadataHeader.readDataChecksum(srcMeta);
+final DataChecksum checksum = BlockMetadataHeader.readDataChecksum(srcMeta,
+DFSUtil.getIoFileBufferSize(new HdfsConfiguration()));
{code}

Creating a new configuration object will parse the XML configuration every 
single time. The caller needs to use the {{Configuration}} object in the 
{{FSDatasetImpl}} class. Therefore the changes need to be propagated all the 
way along the call chain.

 Move shortcircuit to hdfs-client
 

 Key: HDFS-8951
 URL: https://issues.apache.org/jira/browse/HDFS-8951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: build
Reporter: Mingliang Liu
Assignee: Mingliang Liu
 Attachments: HDFS-8951.000.patch, HDFS-8951.001.patch


 This jira tracks the effort of moving the {{shortcircuit}} package into the 
 hdfs-client module.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8894) Set SO_KEEPALIVE on DN server sockets

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8894:
--
Attachment: (was: HDFS-8894-01.patch)

 Set SO_KEEPALIVE on DN server sockets
 -

 Key: HDFS-8894
 URL: https://issues.apache.org/jira/browse/HDFS-8894
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Nathan Roberts
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8894-01.patch


 SO_KEEPALIVE is not set on things like datastreamer sockets which can cause 
 lingering ESTABLISHED sockets when there is a network glitch.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711003#comment-14711003
 ] 

Hudson commented on HDFS-8900:
--

FAILURE: Integrated in Hadoop-Yarn-trunk #1037 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/1037/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711045#comment-14711045
 ] 

Hudson commented on HDFS-8900:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk #2253 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2253/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8836) Skip newline on empty files with getMerge -nl

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8836:
--
Status: Patch Available  (was: Open)

Attached a patch to add a new flag {{-skip-empty-file}} and check the file 
length before copy content.

 Skip newline on empty files with getMerge -nl
 -

 Key: HDFS-8836
 URL: https://issues.apache.org/jira/browse/HDFS-8836
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.7.1, 2.6.0
Reporter: Jan Filipiak
Assignee: kanaka kumar avvaru
Priority: Trivial
 Attachments: HDFS-8836-01.patch


 Hello everyone,
 I recently was in the need of using the new line option -nl with getMerge 
 because the files I needed to merge simply didn't had one. I was merging all 
 the files from one directory and unfortunately this directory also included 
 empty files, which effectively led to multiple newlines append after some 
 files. I needed to remove them manually afterwards.
 In this situation it is maybe good to have another argument that allows 
 skipping empty files.
 Thing one could try to implement this feature:
 The call for IOUtils.copyBytes(in, out, getConf(), false); doesn't
 return the number of bytes copied which would be convenient as one could
 skip append the new line when 0 bytes where copied or one would check the 
 file size before.
 I posted this Idea on the mailing list 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201507.mbox/%3C55B25140.3060005%40trivago.com%3E
  but I didn't really get many responses, so I thought I my try this way.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8892) ShortCircuitCache.CacheCleaner can add Slot.isInvalid() check too

2015-08-25 Thread Ravikumar (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8892?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710959#comment-14710959
 ] 

Ravikumar commented on HDFS-8892:
-

I am fine if this runs in a separate thread as you said…

My app opens hadoop files only once. The block input-streams, once opened need 
to stay that way forever until app instructs to close it. The close call 
happens only for deleting the file or transferring the ownership of file to 
another server {Like Hbase regions}

With current ShortCircuitCache, I am not able to find any knobs to tune for 
such kind of behaviour.

I am planning to set all params of LRU cache high so that ShortCircuitCache 
never closes any open input-stream by itself. All the stream-close calls are 
handled only through invalid-slot-checker 

 ShortCircuitCache.CacheCleaner can add Slot.isInvalid() check too
 -

 Key: HDFS-8892
 URL: https://issues.apache.org/jira/browse/HDFS-8892
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.7.1
Reporter: Ravikumar
Assignee: kanaka kumar avvaru
Priority: Minor

 Currently CacheCleaner thread checks only for cache-expiry times. It would be 
 nice if it handles an invalid-slot too in an extra-pass of evictable map…
 for(ShortCircuitReplica replica:evictable.values()) {
  if(!scr.getSlot().isValid()) {
 purge(replica);
  }
 }
 //Existing code...
 int numDemoted = demoteOldEvictableMmaped(curMs);
 int numPurged = 0;
 Long evictionTimeNs = Long.valueOf(0);
 ….
 …..
 Apps like HBase can tweak the expiry/staleness/cache-size params in 
 DFS-Client, so that ShortCircuitReplica will never be closed except when Slot 
 is declared invalid. 
 I assume slot-invalidation will happen during block-invalidation/deletes 
 {Primarily triggered by compaction/shard-takeover etc..}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8853) Erasure Coding: Provide ECSchema validation when creating ECZone

2015-08-25 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8853?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-8853:
-
Attachment: HDFS-8853-HDFS-7285-05.patch

Thanks [~vinayrpet].
Have updated the patch fixing the comments.
Please review.

 Erasure Coding: Provide ECSchema validation when creating ECZone
 

 Key: HDFS-8853
 URL: https://issues.apache.org/jira/browse/HDFS-8853
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Rakesh R
Assignee: J.Andreina
 Attachments: HDFS-8853-HDFS-7285-01.patch, 
 HDFS-8853-HDFS-7285-05.patch, HDFS-8853-HDFS-7285-merge-02.patch, 
 HDFS-8853-HDFS-7285-merge-03.patch, HDFS-8853-HDFS-7285-merge-04.patch


 Presently the {{DFS#createErasureCodingZone(path, ecSchema, cellSize)}} 
 doesn't have any validation that the given {{ecSchema}} is available in 
 {{ErasureCodingSchemaManager#activeSchemas}} list. Now, if it doesn't exists 
 then will create the ECZone with {{null}} schema. IMHO we could improve this 
 by doing necessary basic sanity checks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8894) Set SO_KEEPALIVE on DN server sockets

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8894:
--
Status: Patch Available  (was: Open)

 Set SO_KEEPALIVE on DN server sockets
 -

 Key: HDFS-8894
 URL: https://issues.apache.org/jira/browse/HDFS-8894
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Nathan Roberts
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8894-01.patch


 SO_KEEPALIVE is not set on things like datastreamer sockets which can cause 
 lingering ESTABLISHED sockets when there is a network glitch.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8938) Refactor BlockManager in blockmanagement

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8938?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711041#comment-14711041
 ] 

Hadoop QA commented on HDFS-8938:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 38s | Pre-patch trunk has 2 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 50s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 59s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 23s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 21s | The applied patch generated  3 
new checkstyle issues (total was 311, now 295). |
| {color:green}+1{color} | whitespace |   0m  1s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 30s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 31s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings, and fixes 1 pre-existing warnings. |
| {color:green}+1{color} | native |   3m 13s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 162m 50s | Tests failed in hadoop-hdfs. |
| | | 207m 54s | |
\\
\\
|| Reason || Tests ||
| Timed out tests | 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeHotSwapVolumes |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752145/HDFS-8938.004.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / af78767 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12109/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12109/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12109/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12109/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf903.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12109/console |


This message was automatically generated.

 Refactor BlockManager in blockmanagement
 

 Key: HDFS-8938
 URL: https://issues.apache.org/jira/browse/HDFS-8938
 Project: Hadoop HDFS
  Issue Type: Task
  Components: build
Reporter: Mingliang Liu
Assignee: Mingliang Liu
 Attachments: HDFS-8938.000.patch, HDFS-8938.001.patch, 
 HDFS-8938.002.patch, HDFS-8938.003.patch, HDFS-8938.004.patch


 This lira tracks the effort of refactoring the {{BlockManager}} in 
 {{hdfs.server.blockmanagement}} package.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8763) Incremental blockreport order may replicate unnecessary block

2015-08-25 Thread Brahma Reddy Battula (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8763?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711052#comment-14711052
 ] 

Brahma Reddy Battula commented on HDFS-8763:


Change of {{dfs.namenode.replication.min}} to 3 (not 2) might solve this 
scenario.

Changing the min replication can lead to new problems. Like safemode on 
restart, write failures if not enough datanodes, etc.

So better to increase the replication check interval. Thats enough,I feel.

 Incremental blockreport order may replicate unnecessary block
 -

 Key: HDFS-8763
 URL: https://issues.apache.org/jira/browse/HDFS-8763
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Affects Versions: 2.4.0
Reporter: jiangyu
Assignee: Walter Su
Priority: Minor

 For our cluster, the NameNode is always very busy, so for every incremental 
 block report , the contention of lock is heavy.
 The logic of incremental block report is as follow, client send block to dn1 
 and dn1 mirror to dn2, dn2 mirror to dn3. After finish this block, all 
 datanode will report the newly received block to namenode. In NameNode side, 
 all will go to the method processIncrementalBlockReport in BlockManager 
 class. But the status of the block reported from dn2,dn3 is RECEIVING_BLOCK, 
 for dn1 is RECEIED_BLOCK. It is okay if dn2, dn3 report before dn1(that is 
 common), but in some busy environment, it is easy to find dn1 report before 
 dn2 or dn3, let’s assume dn2 report first, dn1 report second, and dn3 report 
 third.
 So dn1 will addStoredBlock and find the replica of this block is not reach 
 the the original number(which is 3), and the block will add to 
 neededReplications construction and soon ask some node in pipeline (dn1 or 
 dn2)to replica it dn4 . After sometime, dn4 and dn3 all report this block, 
 then choose one node to invalidate.
 Here is one log i found in our cluster:
 2015-07-08 01:05:34,675 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
 allocateBlock: 
 /logs/***_bigdata_spam/logs/application_1435099124107_470749/xx.xx.4.62_45454.tmp.
  BP-1386326728-xx.xx.2.131-1382089338395 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-a7c0f8f6-2399-4980-9479-efa08487b7b3:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-c75145a0-ed63-4180-87ee-d48ccaa647c5:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW]]}
 2015-07-08 01:05:34,689 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.7.75:50010 is added to 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-74ed264f-da43-4cc3-9fa9-164ba99f752a:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-56121ce1-8991-45b3-95bc-2a5357991512:NORMAL|RBW]]}
  size 0
 2015-07-08 01:05:34,689 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.4.62:50010 is added to 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-74ed264f-da43-4cc3-9fa9-164ba99f752a:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-56121ce1-8991-45b3-95bc-2a5357991512:NORMAL|RBW]]}
  size 0
 2015-07-08 01:05:35,003 INFO BlockStateChange: BLOCK* ask xx.xx.4.62:50010 to 
 replicate blk_3194502674_2121080184 to datanode(s) xx.xx.4.65:50010
 2015-07-08 01:05:35,403 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.7.73:50010 is added to blk_3194502674_2121080184 size 
 67750
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.4.65:50010 is added to blk_3194502674_2121080184 size 
 67750
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* InvalidateBlocks: add 
 blk_3194502674_2121080184 to xx.xx.7.75:50010
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* chooseExcessReplicates: 
 (xx.xx.7.75:50010, blk_3194502674_2121080184) is added to invalidated blocks 
 set
 2015-07-08 01:05:35,852 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
 InvalidateBlocks: ask xx.xx.7.75:50010 to delete [blk_3194502674_2121080184, 
 blk_3194497594_2121075104]
 Some day, the number of this situation can be 40, that is not good for 
 the performance and waste network band.
 Our base version is hadoop 2.4 and i have checked hadoop 2.7.1 didn’t find 
 any difference.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8520) Patch for PPC64 block size

2015-08-25 Thread Tony Reix (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8520?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710968#comment-14710968
 ] 

Tony Reix commented on HDFS-8520:
-

Hi Andrew
Sorry of answering so late. I was busy with other staff.

My understanding of this defect is:
- I had analyzed the code and had found hardcoded 4096 values for PageSize in 
many places, which does cover mainly basic x86 case but not other architectures
- on Linux on Power, PageSize is different
- I had proposed to use the getOperatingSystemPageSize() routine instead of 
hardcoded 4096 value in order to make the test generic (work on all targets 
with no change)
- You proposed instead to hardcode 4096 in the existing test and to add a new 
test with 64KB as hardcoded value (which would cover x86 Linux Large Page case)
 I guess that this would mean to add some code to check, before running 
these 2 tests, which ones are applicable on the targeted architecture.
What about other architectures that would not use 4096 nor 64KB ?

What is the status now for Hadoop 2.7.2 ?
Are you planning to make the code more homogeneous and to enable it to handle 
cases (like Linux on Power) where basic PageSize is NOT 4096 ?
It is very difficult for me to propose a patch since:
1) there are a lot of places where 4096 and PageSize appear, with no 
clear clue if that deals with my issue
2) your approach is very different from mine: hardcode values instead 
of making code generic
Moreover, the way the tests fail is a pain, since they end with TimeOut.

In my opinion, a huge work of rewriting test code so that it deals by default 
with very different architectures (including x86 4KB and 64KB cases) would be 
the best approach.
I also would suggest to add a Linux/PPC64 machine in the development/test 
environment in order to handle different architectures at soon as possible.

 Patch for PPC64 block size
 --

 Key: HDFS-8520
 URL: https://issues.apache.org/jira/browse/HDFS-8520
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.7.1
 Environment: RHEL 7.1 /PPC64
Reporter: Tony Reix
Assignee: Tony Reix
 Attachments: HDFS-8520-2.patch, HDFS-8520.patch


 The attached patch enables Hadoop to work on PPC64.
 That deals with SystemPageSize and BloclSize , which are not 4096 on PPC64.
 There are changes in 3 files:
 - 
 hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/nativeio/NativeIO.java
 - 
 hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/datanode/TestFsDatasetCache.java
 - 
 hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestCacheDirectives.java
 where 4096 is replaced by getOperatingSystemPageSize() or by using PAGE_SIZE
 The patch has been built on branch-2.7 .



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8836) Skip newline on empty files with getMerge -nl

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8836:
--
Attachment: HDFS-8836-01.patch

 Skip newline on empty files with getMerge -nl
 -

 Key: HDFS-8836
 URL: https://issues.apache.org/jira/browse/HDFS-8836
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0, 2.7.1
Reporter: Jan Filipiak
Assignee: kanaka kumar avvaru
Priority: Trivial
 Attachments: HDFS-8836-01.patch


 Hello everyone,
 I recently was in the need of using the new line option -nl with getMerge 
 because the files I needed to merge simply didn't had one. I was merging all 
 the files from one directory and unfortunately this directory also included 
 empty files, which effectively led to multiple newlines append after some 
 files. I needed to remove them manually afterwards.
 In this situation it is maybe good to have another argument that allows 
 skipping empty files.
 Thing one could try to implement this feature:
 The call for IOUtils.copyBytes(in, out, getConf(), false); doesn't
 return the number of bytes copied which would be convenient as one could
 skip append the new line when 0 bytes where copied or one would check the 
 file size before.
 I posted this Idea on the mailing list 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201507.mbox/%3C55B25140.3060005%40trivago.com%3E
  but I didn't really get many responses, so I thought I my try this way.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8894) Set SO_KEEPALIVE on DN server sockets

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8894:
--
Status: Open  (was: Patch Available)

 Set SO_KEEPALIVE on DN server sockets
 -

 Key: HDFS-8894
 URL: https://issues.apache.org/jira/browse/HDFS-8894
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Nathan Roberts
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8894-01.patch


 SO_KEEPALIVE is not set on things like datastreamer sockets which can cause 
 lingering ESTABLISHED sockets when there is a network glitch.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14710947#comment-14710947
 ] 

Hudson commented on HDFS-8900:
--

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #309 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/309/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8248) Store INodeId instead of the INodeFile object in BlockInfoContiguous

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8248?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712068#comment-14712068
 ] 

Hadoop QA commented on HDFS-8248:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  17m 49s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 7 new or modified test files. |
| {color:green}+1{color} | javac |   7m 56s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  7s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 22s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 26s | The applied patch generated  3 
new checkstyle issues (total was 512, now 510). |
| {color:green}+1{color} | whitespace |   0m  3s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 32s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   2m  8s | Post-patch findbugs 
hadoop-hdfs-project/hadoop-hdfs compilation is broken. |
| {color:green}+1{color} | findbugs |   2m  8s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   0m 22s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 159m 41s | Tests failed in hadoop-hdfs. |
| | | 202m  7s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.server.blockmanagement.TestNodeCount |
|   | hadoop.fs.TestHdfsNativeCodeLoader |
|   | hadoop.hdfs.server.blockmanagement.TestSequentialBlockId |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752141/HDFS-8248.007.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12122/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12122/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12122/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12122/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf900.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12122/console |


This message was automatically generated.

 Store INodeId instead of the INodeFile object in BlockInfoContiguous
 

 Key: HDFS-8248
 URL: https://issues.apache.org/jira/browse/HDFS-8248
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Haohui Mai
Assignee: Haohui Mai
 Attachments: HDFS-8248.000.patch, HDFS-8248.001.patch, 
 HDFS-8248.002.patch, HDFS-8248.003.patch, HDFS-8248.004.patch, 
 HDFS-8248.005.patch, HDFS-8248.006.patch, HDFS-8248.007.patch


 Currently the namespace and the block manager are tightly coupled together. 
 There are two couplings in terms of implementation:
 1. The {{BlockInfoContiguous}} stores a reference of the {{INodeFile}} that 
 owns the block, so that the block manager can look up the corresponding file 
 when replicating blocks, recovering from pipeline failures, etc.
 1. The {{INodeFile}} stores {{BlockInfoContiguous}} objects that the file 
 owns.
 Decoupling the namespace and the block manager allows the BM to be separated 
 out from the Java heap or even as a standalone process. This jira proposes to 
 remove the first coupling by storing the id of the inode instead of the 
 object reference of {{INodeFile}} in the {{BlockInfoContiguous}} class.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-1148) Convert FSDataset to ReadWriteLock

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1148?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711727#comment-14711727
 ] 

Hadoop QA commented on HDFS-1148:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  18m 25s | Pre-patch trunk has 4 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   8m  0s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m  5s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 24s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 25s | The applied patch generated  1 
new checkstyle issues (total was 121, now 111). |
| {color:green}+1{color} | whitespace |   0m  3s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 31s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:red}-1{color} | findbugs |   2m  1s | Post-patch findbugs 
hadoop-hdfs-project/hadoop-hdfs compilation is broken. |
| {color:green}+1{color} | findbugs |   2m  1s | The patch does not introduce 
any new Findbugs (version ) warnings. |
| {color:green}+1{color} | native |   0m 37s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests |   6m 39s | Tests failed in hadoop-hdfs. |
| | |  49m 49s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.TestFileStatus |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752270/HDFS-1148.003.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12117/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12117/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12117/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12117/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12117/console |


This message was automatically generated.

 Convert FSDataset to ReadWriteLock
 --

 Key: HDFS-1148
 URL: https://issues.apache.org/jira/browse/HDFS-1148
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode, performance
Reporter: Todd Lipcon
Assignee: Yong Zhang
 Attachments: HDFS-1148.001.patch, HDFS-1148.002.patch, 
 HDFS-1148.003.patch, hdfs-1148-old.txt, hdfs-1148-trunk.txt, 
 patch-HDFS-1148-rel0.20.2.txt


 In benchmarking HDFS-941 I noticed that for the random read workload, the 
 FSDataset lock is highly contended. After converting it to a 
 ReentrantReadWriteLock, I saw a ~25% improvement on both latency and 
 ops/second.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8950) NameNode refresh doesn't remove DataNodes

2015-08-25 Thread Aaron T. Myers (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Aaron T. Myers updated HDFS-8950:
-
Target Version/s: 2.8.0

 NameNode refresh doesn't remove DataNodes
 -

 Key: HDFS-8950
 URL: https://issues.apache.org/jira/browse/HDFS-8950
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Daniel Templeton
Assignee: Daniel Templeton

 If you remove a DN from NN's allowed host list (HDFS was HA) and then do NN 
 refresh, it doesn't remove it actually and the NN UI keeps showing that node. 
 It may try to allocate some blocks to that DN as well during an MR job.  This 
 issue is independent from DN decommission.
 To reproduce:
 1. Add a DN to dfs_hosts_allow
 2. Refresh NN
 3. Start DN. Now NN starts seeing DN.
 4. Stop DN
 5. Remove DN from dfs_hosts_allow
 6. Refresh NN - NN is still reporting DN as being used by HDFS.
 This is different from decom because there DN is added to exclude list in 
 addition to being removed from allowed list, and in that case everything 
 works correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8951) Move shortcircuit to hdfs-client

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712075#comment-14712075
 ] 

Hadoop QA commented on HDFS-8951:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  19m 59s | Pre-patch trunk has 6 extant 
Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:red}-1{color} | tests included |   0m  0s | The patch doesn't appear 
to include any new or modified tests.  Please justify why no new tests are 
needed for this patch. Also please list what manual steps were performed to 
verify this patch. |
| {color:green}+1{color} | javac |   7m 51s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 10s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 16s | The applied patch generated 
1 release audit warnings. |
| {color:red}-1{color} | checkstyle |   2m 31s | The applied patch generated  
62 new checkstyle issues (total was 16, now 78). |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 34s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   4m 30s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m 11s | Pre-build of native portion |
| {color:green}+1{color} | hdfs tests | 162m 22s | Tests passed in hadoop-hdfs. 
|
| {color:green}+1{color} | hdfs tests |   0m 29s | Tests passed in 
hadoop-hdfs-client. |
| | | 213m 29s | |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752146/HDFS-8951.001.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | trunk / eee0d45 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs.html
 |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/artifact/patchprocess/trunkFindbugsWarningshadoop-hdfs-client.html
 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/artifact/patchprocess/diffcheckstylehadoop-hdfs-client.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| hadoop-hdfs-client test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/artifact/patchprocess/testrun_hadoop-hdfs-client.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf900.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12121/console |


This message was automatically generated.

 Move shortcircuit to hdfs-client
 

 Key: HDFS-8951
 URL: https://issues.apache.org/jira/browse/HDFS-8951
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: build
Reporter: Mingliang Liu
Assignee: Mingliang Liu
 Attachments: HDFS-8951.000.patch, HDFS-8951.001.patch


 This jira tracks the effort of moving the {{shortcircuit}} package into the 
 hdfs-client module.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8909) Erasure coding: update BlockInfoContiguousUC and BlockInfoStripedUC to use BlockUnderConstructionFeature

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8909?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712083#comment-14712083
 ] 

Hadoop QA commented on HDFS-8909:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:red}-1{color} | pre-patch |  18m 16s | Pre-patch HDFS-7285 has 4 
extant Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 12 new or modified test files. |
| {color:green}+1{color} | javac |   7m 53s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 52s | There were no new javadoc 
warning messages. |
| {color:red}-1{color} | release audit |   0m 17s | The applied patch generated 
1 release audit warnings. |
| {color:red}-1{color} | checkstyle |   1m 28s | The applied patch generated  
14 new checkstyle issues (total was 1116, now 1086). |
| {color:green}+1{color} | whitespace |   0m 12s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 22s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 34s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m 36s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:green}+1{color} | native |   3m  5s | Pre-build of native portion |
| {color:red}-1{color} | hdfs tests | 208m 14s | Tests failed in hadoop-hdfs. |
| | | 253m 56s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.hdfs.TestFileAppend4 |
|   | hadoop.hdfs.server.namenode.TestFileTruncate |
|   | hadoop.hdfs.TestWriteStripedFileWithFailure |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752122/HDFS-8909-HDFS-7285.003.patch
 |
| Optional Tests | javadoc javac unit findbugs checkstyle |
| git revision | HDFS-7285 / 6b6a63b |
| Pre-patch Findbugs warnings | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12119/artifact/patchprocess/HDFS-7285FindbugsWarningshadoop-hdfs.html
 |
| Release Audit | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12119/artifact/patchprocess/patchReleaseAuditProblems.txt
 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12119/artifact/patchprocess/diffcheckstylehadoop-hdfs.txt
 |
| hadoop-hdfs test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12119/artifact/patchprocess/testrun_hadoop-hdfs.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12119/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf909.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12119/console |


This message was automatically generated.

 Erasure coding: update BlockInfoContiguousUC and BlockInfoStripedUC to use 
 BlockUnderConstructionFeature
 

 Key: HDFS-8909
 URL: https://issues.apache.org/jira/browse/HDFS-8909
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: HDFS-7285
Reporter: Zhe Zhang
Assignee: Jing Zhao
 Attachments: HDFS-8909-HDFS-7285.003.patch, HDFS-8909.000.patch, 
 HDFS-8909.001.patch, HDFS-8909.002.patch, HDFS-8909.003.patch


 HDFS-8801 converts {{BlockInfoUC}} as a feature. We should consolidate 
 {{BlockInfoContiguousUC}} and {{BlockInfoStripedUC}} logics to use this 
 feature.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8950) NameNode refresh doesn't remove DataNodes

2015-08-25 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-8950:
---
Attachment: 28531.patch

Here's a patch for a unit test and fix.  The HostFileManager now excludes any 
nodes not present in a non-empty includes file from the node report.  All tests 
in the HDFS project pass.

 NameNode refresh doesn't remove DataNodes
 -

 Key: HDFS-8950
 URL: https://issues.apache.org/jira/browse/HDFS-8950
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Daniel Templeton
Assignee: Daniel Templeton
 Attachments: 28531.patch


 If you remove a DN from NN's allowed host list (HDFS was HA) and then do NN 
 refresh, it doesn't remove it actually and the NN UI keeps showing that node. 
 It may try to allocate some blocks to that DN as well during an MR job.  This 
 issue is independent from DN decommission.
 To reproduce:
 1. Add a DN to dfs_hosts_allow
 2. Refresh NN
 3. Start DN. Now NN starts seeing DN.
 4. Stop DN
 5. Remove DN from dfs_hosts_allow
 6. Refresh NN - NN is still reporting DN as being used by HDFS.
 This is different from decom because there DN is added to exclude list in 
 addition to being removed from allowed list, and in that case everything 
 works correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Add a unit test for INotify functionality across a layout version upgrade

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712097#comment-14712097
 ] 

Hudson commented on HDFS-8846:
--

FAILURE: Integrated in Hadoop-Yarn-trunk #1038 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/1038/])
HDFS-8846. Add a unit test for INotify functionality across a layout version 
upgrade (Zhe Zhang via Colin P. McCabe) (cmccabe: rev 
a4d9acc51d1a977bc333da17780c00c72e8546f1)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgrade.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* hadoop-hdfs-project/hadoop-hdfs/src/test/resources/hadoop-252-dfs-dir.tgz
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSInotifyEventInputStream.java


 Add a unit test for INotify functionality across a layout version upgrade
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: 2.8.0

 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8885) ByteRangeInputStream used in webhdfs does not override available()

2015-08-25 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8885?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711935#comment-14711935
 ] 

Chris Nauroth commented on HDFS-8885:
-

Hello [~srevanka].  Thank you for posting the patch.  Here are a few notes.

# A {{ByteRangeInputStream}} starts out in {{SEEK}} status, which indicates the 
underlying HTTP stream needs to be opened.  If a caller calls {{available}} 
before any {{read}} calls, then the result won't be correct, because it has not 
yet determined the file length.  I recommend calling {{getInputStream}} first, 
which lazily opens the underlying HTTP stream if necessary.
# A similar situation occurs if a caller calls {{seek}} immediately before 
{{available}}.  Again, calling {{getInputStream}} first would handle this case 
correctly.
# It's typical for {{available}} to throw an {{IOException}} if called after 
the stream has been closed.  Once again, a call to {{getInputStream}} would 
cover this case, because that method is already implemented to throw an 
{{IOException}} if the stream is in {{CLOSED}} state.
# Can you please look into adding unit tests in {{TestByteRangeInputStream}}?


 ByteRangeInputStream used in webhdfs does not override available()
 --

 Key: HDFS-8885
 URL: https://issues.apache.org/jira/browse/HDFS-8885
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar
Priority: Minor
 Attachments: HDFS-8885.000.patch, HDFS-8885.001.patch


 ByteRangeInputStream has to override the method {{available()}} , the super 
 class implementation returns 0. Clients using the method {{available()}} to 
 rely on available bytes information will end up with errors during reads if 
 WebhdfsFileSystem is used. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8836) Skip newline on empty files with getMerge -nl

2015-08-25 Thread kanaka kumar avvaru (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711827#comment-14711827
 ] 

kanaka kumar avvaru commented on HDFS-8836:
---

Updated patch for jenkin reported errors

 Skip newline on empty files with getMerge -nl
 -

 Key: HDFS-8836
 URL: https://issues.apache.org/jira/browse/HDFS-8836
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0, 2.7.1
Reporter: Jan Filipiak
Assignee: kanaka kumar avvaru
Priority: Trivial
 Attachments: HDFS-8836-01.patch, HDFS-8836-02.patch


 Hello everyone,
 I recently was in the need of using the new line option -nl with getMerge 
 because the files I needed to merge simply didn't had one. I was merging all 
 the files from one directory and unfortunately this directory also included 
 empty files, which effectively led to multiple newlines append after some 
 files. I needed to remove them manually afterwards.
 In this situation it is maybe good to have another argument that allows 
 skipping empty files.
 Thing one could try to implement this feature:
 The call for IOUtils.copyBytes(in, out, getConf(), false); doesn't
 return the number of bytes copied which would be convenient as one could
 skip append the new line when 0 bytes where copied or one would check the 
 file size before.
 I posted this Idea on the mailing list 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201507.mbox/%3C55B25140.3060005%40trivago.com%3E
  but I didn't really get many responses, so I thought I my try this way.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8833) Erasure coding: store EC schema and cell size in INodeFile and eliminate notion of EC zones

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711882#comment-14711882
 ] 

Zhe Zhang commented on HDFS-8833:
-

{{TestFSNamesystem}} passes consistently on local node. 
{{testWriteStripedFileWithDNFailure}} is flaky even without the change. I'm 
debugging.

 Erasure coding: store EC schema and cell size in INodeFile and eliminate 
 notion of EC zones
 ---

 Key: HDFS-8833
 URL: https://issues.apache.org/jira/browse/HDFS-8833
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: HDFS-7285
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Attachments: HDFS-8833-HDFS-7285-merge.00.patch, 
 HDFS-8833-HDFS-7285-merge.01.patch, HDFS-8833-HDFS-7285.02.patch


 We have [discussed | 
 https://issues.apache.org/jira/browse/HDFS-7285?focusedCommentId=14357754page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14357754]
  storing EC schema with files instead of EC zones and recently revisited the 
 discussion under HDFS-8059.
 As a recap, the _zone_ concept has severe limitations including renaming and 
 nested configuration. Those limitations are valid in encryption for security 
 reasons and it doesn't make sense to carry them over in EC.
 This JIRA aims to store EC schema and cell size on {{INodeFile}} level. For 
 simplicity, we should first implement it as an xattr and consider memory 
 optimizations (such as moving it to file header) as a follow-on. We should 
 also disable changing EC policy on a non-empty file / dir in the first phase.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8894) Set SO_KEEPALIVE on DN server sockets

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8894?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8894:
--
Attachment: HDFS-8894-01.patch

Some problems in earlier build. Uploading same patch again for jenkins new 
build.

 Set SO_KEEPALIVE on DN server sockets
 -

 Key: HDFS-8894
 URL: https://issues.apache.org/jira/browse/HDFS-8894
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.7.1
Reporter: Nathan Roberts
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8894-01.patch, HDFS-8894-01.patch


 SO_KEEPALIVE is not set on things like datastreamer sockets which can cause 
 lingering ESTABLISHED sockets when there is a network glitch.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-6955) DN should reserve disk space for a full block when creating tmp files

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-6955:
--
Attachment: HDFS-6955-03.patch

updated patch for required check style errors. Test failure doesn't seem to be 
related to this patch

 DN should reserve disk space for a full block when creating tmp files
 -

 Key: HDFS-6955
 URL: https://issues.apache.org/jira/browse/HDFS-6955
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
Affects Versions: 2.5.0
Reporter: Arpit Agarwal
Assignee: kanaka kumar avvaru
 Attachments: HDFS-6955-01.patch, HDFS-6955-02.patch, 
 HDFS-6955-03.patch


 HDFS-6898 is introducing disk space reservation for RBW files to avoid 
 running out of disk space midway through block creation.
 This Jira is to introduce similar reservation for tmp files.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8155) Support OAuth2 in WebHDFS

2015-08-25 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8155?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711874#comment-14711874
 ] 

Chris Nauroth commented on HDFS-8155:
-

Hi [~jghoman].  Nice work!  I have a few comments.

# {{ConfRefreshTokenBasedAccessTokenProvider}}, 
{{ConfRefreshTokenBasedAccessTokenProvider}}: There are no timeouts specified 
in the calls to the refresh URL.  Timeouts can be controlled by calling 
{{client.setConnectTimeout}} and {{client.setReadTimeout}}.
# {{AccessTokenProvider}}: *Optional* - consider extending {{Configured}} so 
that it inherits the implementations of {{getConf}} and {{setConf}} for free.
# WebHDFS.md: Typo: toekns instead of tokens
# Please address the javac and checkstyle warnings.

[~chris.douglas], it looks like the most recent version of the patch has 
addressed your feedback.  Do you agree?

 Support OAuth2 in WebHDFS
 -

 Key: HDFS-8155
 URL: https://issues.apache.org/jira/browse/HDFS-8155
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: webhdfs
Reporter: Jakob Homan
Assignee: Jakob Homan
 Attachments: HDFS-8155-1.patch, HDFS-8155.002.patch


 WebHDFS should be able to accept OAuth2 credentials.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8784) BlockInfo#numNodes should be numStorages

2015-08-25 Thread Arpit Agarwal (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8784?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711936#comment-14711936
 ] 

Arpit Agarwal commented on HDFS-8784:
-

The NameNode does not support more than one replica of a given block per 
DataNode, so node/storage are interchangeable and node is probably better.

 BlockInfo#numNodes should be numStorages
 

 Key: HDFS-8784
 URL: https://issues.apache.org/jira/browse/HDFS-8784
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Jagadesh Kiran N
 Attachments: HDFS-8784-00.patch, HDFS-8784-01.patch


 The method actually returns the number of storages holding a block.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8833) Erasure coding: store EC schema and cell size in INodeFile and eliminate notion of EC zones

2015-08-25 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711937#comment-14711937
 ] 

Zhe Zhang commented on HDFS-8833:
-

The failure will be addressed by HDFS-8704.

 Erasure coding: store EC schema and cell size in INodeFile and eliminate 
 notion of EC zones
 ---

 Key: HDFS-8833
 URL: https://issues.apache.org/jira/browse/HDFS-8833
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: HDFS-7285
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Attachments: HDFS-8833-HDFS-7285-merge.00.patch, 
 HDFS-8833-HDFS-7285-merge.01.patch, HDFS-8833-HDFS-7285.02.patch


 We have [discussed | 
 https://issues.apache.org/jira/browse/HDFS-7285?focusedCommentId=14357754page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14357754]
  storing EC schema with files instead of EC zones and recently revisited the 
 discussion under HDFS-8059.
 As a recap, the _zone_ concept has severe limitations including renaming and 
 nested configuration. Those limitations are valid in encryption for security 
 reasons and it doesn't make sense to carry them over in EC.
 This JIRA aims to store EC schema and cell size on {{INodeFile}} level. For 
 simplicity, we should first implement it as an xattr and consider memory 
 optimizations (such as moving it to file header) as a follow-on. We should 
 also disable changing EC policy on a non-empty file / dir in the first phase.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8771) If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not send another RPC calls to Journalnodes

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8771:
--
Status: Patch Available  (was: Open)

Attached a patch to handle the {{purgeLogsOlderThan}} calls at {{JournalNode}} 
in a separate single threaded executor service so that edits are not blocked 
for the purge log completion.

 If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not 
 send another RPC calls to Journalnodes
 

 Key: HDFS-8771
 URL: https://issues.apache.org/jira/browse/HDFS-8771
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Takuya Fukudome
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8771-01.patch


 In our cluster, edits has became huge(about 50GB) accidentally and our 
 Jounalnodes' disks were busy, therefore {{purgeLogsOlderThan}} took more than 
 30secs. If {{IPCLoggerChannel#purgeLogsOlderThan}} takes too much time, 
 Namenode couldn't send other RPC calls to Journalnodes because 
 {{o.a.h.hdfs.qjournal.client.IPCLoggerChannel}}'s executor is single thread. 
 It will cause namenode shutting down.
 I think IPCLoggerChannel#purgeLogsOlderThan should not block other RPC calls 
 like sendEdits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-5696) Examples for httpfs REST API incorrect on apache.org

2015-08-25 Thread Harsh J (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-5696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1479#comment-1479
 ] 

Harsh J commented on HDFS-5696:
---

Thanks for reporting this. Care to submit a fix changing the op values to the 
right one? I suspect these may be leftovers from the Hoop days.

bq. Not sure what the convention should be for specifying the user.name. Use 
hdfs? or a name that is obviously an example?

Since these are curl based examples that also likely assume no kerberos setup, 
why not $USER or $(whoami) instead of a hardcoded value?

 Examples for httpfs REST API incorrect on apache.org
 

 Key: HDFS-5696
 URL: https://issues.apache.org/jira/browse/HDFS-5696
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Affects Versions: 2.2.0
 Environment: NA
Reporter: Casey Brotherton
Priority: Trivial

 The examples provided for the httpfs REST API are incorrect.
 http://hadoop.apache.org/docs/r2.2.0/hadoop-hdfs-httpfs/index.html
 http://hadoop.apache.org/docs/r2.0.5-alpha/hadoop-hdfs-httpfs/index.html
 From the documentation:
 *
 HttpFS is a separate service from Hadoop NameNode.
 HttpFS itself is Java web-application and it runs using a preconfigured 
 Tomcat bundled with HttpFS binary distribution.
 HttpFS HTTP web-service API calls are HTTP REST calls that map to a HDFS file 
 system operation. For example, using the curl Unix command:
 $ curl http://httpfs-host:14000/webhdfs/v1/user/foo/README.txt returns the 
 contents of the HDFS /user/foo/README.txt file.
 $ curl http://httpfs-host:14000/webhdfs/v1/user/foo?op=list returns the 
 contents of the HDFS /user/foo directory in JSON format.
 $ curl -X POST http://httpfs-host:14000/webhdfs/v1/user/foo/bar?op=mkdirs 
 creates the HDFS /user/foo.bar directory.
 ***
 The commands have incorrect operations. ( Verified through source code in 
 HttpFSFileSystem.java )
 In addition, although the webhdfs documentation specifies user.name as 
 optional, on my cluster, each action required a user.name
 It should be included in the short examples to allow for the greatest chance 
 of success.
 Three examples rewritten:
 curl -i -L 
 http://httpfs-host:14000/webhdfs/v1/user/foo/README.txt?op=openuser.name=hdfsuser;
 curl -i 
 http://httpfs-host:14000/webhdfs/v1/user/foo/?op=liststatususer.name=hdfsuser;
 curl -i -X PUT 
 http://httpfs-host:14000/webhdfs/v1/user/foo/bar?op=mkdirsuser.name=hdfsuser;
 Not sure what the convention should be for specifying the user.name. Use 
 hdfs? or a name that is obviously an example?
 It would also be beneficial if the HTTPfs page linked to the webhdfs 
 documentation page in the text instead of just on the menu sidebar.
 http://hadoop.apache.org/docs/r2.2.0/hadoop-project-dist/hadoop-hdfs/WebHDFS.html



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8836) Skip newline on empty files with getMerge -nl

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711131#comment-14711131
 ] 

Hadoop QA commented on HDFS-8836:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  20m  3s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 1 new or modified test files. |
| {color:green}+1{color} | javac |   7m 50s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |   9m 58s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 21s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | site |   3m  0s | Site still builds. |
| {color:red}-1{color} | checkstyle |   1m  8s | The applied patch generated  2 
new checkstyle issues (total was 45, now 47). |
| {color:red}-1{color} | whitespace |   0m  0s | The patch has 1  line(s) that 
end in whitespace. Use git apply --whitespace=fix. |
| {color:green}+1{color} | install |   1m 26s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 32s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   1m 53s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:red}-1{color} | common tests |  22m 16s | Tests failed in 
hadoop-common. |
| | |  68m 30s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.cli.TestCLI |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752203/HDFS-8836-01.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle site |
| git revision | trunk / eee0d45 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12114/artifact/patchprocess/diffcheckstylehadoop-common.txt
 |
| whitespace | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12114/artifact/patchprocess/whitespace.txt
 |
| hadoop-common test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12114/artifact/patchprocess/testrun_hadoop-common.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12114/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf903.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12114/console |


This message was automatically generated.

 Skip newline on empty files with getMerge -nl
 -

 Key: HDFS-8836
 URL: https://issues.apache.org/jira/browse/HDFS-8836
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0, 2.7.1
Reporter: Jan Filipiak
Assignee: kanaka kumar avvaru
Priority: Trivial
 Attachments: HDFS-8836-01.patch


 Hello everyone,
 I recently was in the need of using the new line option -nl with getMerge 
 because the files I needed to merge simply didn't had one. I was merging all 
 the files from one directory and unfortunately this directory also included 
 empty files, which effectively led to multiple newlines append after some 
 files. I needed to remove them manually afterwards.
 In this situation it is maybe good to have another argument that allows 
 skipping empty files.
 Thing one could try to implement this feature:
 The call for IOUtils.copyBytes(in, out, getConf(), false); doesn't
 return the number of bytes copied which would be convenient as one could
 skip append the new line when 0 bytes where copied or one would check the 
 file size before.
 I posted this Idea on the mailing list 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201507.mbox/%3C55B25140.3060005%40trivago.com%3E
  but I didn't really get many responses, so I thought I my try this way.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8771) If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not send another RPC calls to Journalnodes

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8771:
--
Attachment: HDFS-8771-01.patch

 If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not 
 send another RPC calls to Journalnodes
 

 Key: HDFS-8771
 URL: https://issues.apache.org/jira/browse/HDFS-8771
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Takuya Fukudome
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8771-01.patch


 In our cluster, edits has became huge(about 50GB) accidentally and our 
 Jounalnodes' disks were busy, therefore {{purgeLogsOlderThan}} took more than 
 30secs. If {{IPCLoggerChannel#purgeLogsOlderThan}} takes too much time, 
 Namenode couldn't send other RPC calls to Journalnodes because 
 {{o.a.h.hdfs.qjournal.client.IPCLoggerChannel}}'s executor is single thread. 
 It will cause namenode shutting down.
 I think IPCLoggerChannel#purgeLogsOlderThan should not block other RPC calls 
 like sendEdits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8581) count cmd calculate wrong when huge files exist in one folder

2015-08-25 Thread J.Andreina (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

J.Andreina updated HDFS-8581:
-
Attachment: HDFS-8581.2.patch

Patch is  updated. Please review.

 count cmd calculate wrong when huge files exist in one folder
 -

 Key: HDFS-8581
 URL: https://issues.apache.org/jira/browse/HDFS-8581
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Reporter: tongshiquan
Assignee: J.Andreina
Priority: Minor
 Attachments: HDFS-8581.1.patch, HDFS-8581.2.patch


 If one directory such as /result exists about 20 files, then when 
 execute hdfs dfs -count /, the result will go wrong. For all directories 
 whose name after /result, file num will not be included.
 My cluster see as below, /result_1433858936 is the directory exist huge 
 files, and files in /sparkJobHistory, /tmp, /user are not included
 vm-221:/export1/BigData/current # hdfs dfs -ls /
 15/06/11 11:00:17 INFO hdfs.PeerCache: SocketCache disabled.
 Found 9 items
 -rw-r--r--   3 hdfs   supergroup  0 2015-06-08 12:10 
 /PRE_CREATE_DIR.SUCCESS
 drwxr-x---   - flume  hadoop  0 2015-06-08 12:08 /flume
 drwx--   - hbase  hadoop  0 2015-06-10 15:25 /hbase
 drwxr-xr-x   - hdfs   supergroup  0 2015-06-10 17:19 /hyt
 drwxrwxrwx   - mapred hadoop  0 2015-06-08 12:08 /mr-history
 drwxr-xr-x   - hdfs   supergroup  0 2015-06-09 22:10 
 /result_1433858936
 drwxrwxrwx   - spark  supergroup  0 2015-06-10 19:15 /sparkJobHistory
 drwxrwxrwx   - hdfs   hadoop  0 2015-06-08 12:14 /tmp
 drwxrwxrwx   - hdfs   hadoop  0 2015-06-09 21:57 /user
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /
 15/06/11 11:00:24 INFO hdfs.PeerCache: SocketCache disabled.
 1043   171536 1756375688 /
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /PRE_CREATE_DIR.SUCCESS
 15/06/11 11:00:30 INFO hdfs.PeerCache: SocketCache disabled.
01  0 /PRE_CREATE_DIR.SUCCESS
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /flume
 15/06/11 11:00:41 INFO hdfs.PeerCache: SocketCache disabled.
10  0 /flume
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /hbase
 15/06/11 11:00:49 INFO hdfs.PeerCache: SocketCache disabled.
   36   18  14807 /hbase
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /hyt
 15/06/11 11:01:09 INFO hdfs.PeerCache: SocketCache disabled.
10  0 /hyt
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /mr-history
 15/06/11 11:01:18 INFO hdfs.PeerCache: SocketCache disabled.
30  0 /mr-history
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /result_1433858936
 15/06/11 11:01:29 INFO hdfs.PeerCache: SocketCache disabled.
 1001   171517 1756360881 /result_1433858936
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /sparkJobHistory
 15/06/11 11:01:41 INFO hdfs.PeerCache: SocketCache disabled.
13  21785 /sparkJobHistory
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /tmp
 15/06/11 11:01:48 INFO hdfs.PeerCache: SocketCache disabled.
   176  35958 /tmp
 vm-221:/export1/BigData/current # 
 vm-221:/export1/BigData/current # hdfs dfs -count /user
 15/06/11 11:01:55 INFO hdfs.PeerCache: SocketCache disabled.
   121  19077 /user



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1474#comment-1474
 ] 

Hudson commented on HDFS-8900:
--

SUCCESS: Integrated in Hadoop-Mapreduce-trunk-Java8 #304 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/304/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=1475#comment-1475
 ] 

Hudson commented on HDFS-8900:
--

FAILURE: Integrated in Hadoop-Hdfs-trunk #2234 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2234/])
HDFS-8900. Compact XAttrs to optimize memory footprint. (yliu) (yliu: rev 
eee0d4563c62647cfaaed6605ee713aaf69add78)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestStartup.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFormat.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/INodeDirectory.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirectory.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberMap.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSImageFormatPBINode.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestXAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/blockmanagement/BlockStoragePolicySuite.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrStorage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/SerialNumberManager.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/FSDirXAttrOp.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrPermissionFilter.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/server/namenode/XAttrFeature.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/XAttrHelper.java
* hadoop-hdfs-project/hadoop-hdfs/src/main/resources/hdfs-default.xml
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/web/JsonUtil.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/main/java/org/apache/hadoop/hdfs/DFSConfigKeys.java


 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: 2.8.0

 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch, HDFS-8900.004.patch, HDFS-8900.005.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8836) Skip newline on empty files with getMerge -nl

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8836:
--
Attachment: HDFS-8836-02.patch

 Skip newline on empty files with getMerge -nl
 -

 Key: HDFS-8836
 URL: https://issues.apache.org/jira/browse/HDFS-8836
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0, 2.7.1
Reporter: Jan Filipiak
Assignee: kanaka kumar avvaru
Priority: Trivial
 Attachments: HDFS-8836-01.patch, HDFS-8836-02.patch


 Hello everyone,
 I recently was in the need of using the new line option -nl with getMerge 
 because the files I needed to merge simply didn't had one. I was merging all 
 the files from one directory and unfortunately this directory also included 
 empty files, which effectively led to multiple newlines append after some 
 files. I needed to remove them manually afterwards.
 In this situation it is maybe good to have another argument that allows 
 skipping empty files.
 Thing one could try to implement this feature:
 The call for IOUtils.copyBytes(in, out, getConf(), false); doesn't
 return the number of bytes copied which would be convenient as one could
 skip append the new line when 0 bytes where copied or one would check the 
 file size before.
 I posted this Idea on the mailing list 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201507.mbox/%3C55B25140.3060005%40trivago.com%3E
  but I didn't really get many responses, so I thought I my try this way.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8771) If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not send another RPC calls to Journalnodes

2015-08-25 Thread kanaka kumar avvaru (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8771?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

kanaka kumar avvaru updated HDFS-8771:
--
Attachment: HDFS-8771-02.patch

 If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not 
 send another RPC calls to Journalnodes
 

 Key: HDFS-8771
 URL: https://issues.apache.org/jira/browse/HDFS-8771
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Takuya Fukudome
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8771-01.patch, HDFS-8771-02.patch


 In our cluster, edits has became huge(about 50GB) accidentally and our 
 Jounalnodes' disks were busy, therefore {{purgeLogsOlderThan}} took more than 
 30secs. If {{IPCLoggerChannel#purgeLogsOlderThan}} takes too much time, 
 Namenode couldn't send other RPC calls to Journalnodes because 
 {{o.a.h.hdfs.qjournal.client.IPCLoggerChannel}}'s executor is single thread. 
 It will cause namenode shutting down.
 I think IPCLoggerChannel#purgeLogsOlderThan should not block other RPC calls 
 like sendEdits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8771) If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not send another RPC calls to Journalnodes

2015-08-25 Thread kanaka kumar avvaru (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8771?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711839#comment-14711839
 ] 

kanaka kumar avvaru commented on HDFS-8771:
---

updated patch for checkstyle and white space erros reported in jenkins

 If IPCLoggerChannel#purgeLogsOlderThan takes too long, Namenode could not 
 send another RPC calls to Journalnodes
 

 Key: HDFS-8771
 URL: https://issues.apache.org/jira/browse/HDFS-8771
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Takuya Fukudome
Assignee: kanaka kumar avvaru
 Attachments: HDFS-8771-01.patch, HDFS-8771-02.patch


 In our cluster, edits has became huge(about 50GB) accidentally and our 
 Jounalnodes' disks were busy, therefore {{purgeLogsOlderThan}} took more than 
 30secs. If {{IPCLoggerChannel#purgeLogsOlderThan}} takes too much time, 
 Namenode couldn't send other RPC calls to Journalnodes because 
 {{o.a.h.hdfs.qjournal.client.IPCLoggerChannel}}'s executor is single thread. 
 It will cause namenode shutting down.
 I think IPCLoggerChannel#purgeLogsOlderThan should not block other RPC calls 
 like sendEdits.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8953) DataNode Metrics logging

2015-08-25 Thread kanaka kumar avvaru (JIRA)
kanaka kumar avvaru created HDFS-8953:
-

 Summary: DataNode Metrics logging
 Key: HDFS-8953
 URL: https://issues.apache.org/jira/browse/HDFS-8953
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: kanaka kumar avvaru
Assignee: kanaka kumar avvaru


HDFS-8880 added metrics logging at NameNode. Similarly, this JIRA is to  add a 
separate logger for metrics at DN



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8836) Skip newline on empty files with getMerge -nl

2015-08-25 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8836?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14711950#comment-14711950
 ] 

Hadoop QA commented on HDFS-8836:
-

\\
\\
| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | pre-patch |  22m 10s | Pre-patch trunk compilation is 
healthy. |
| {color:green}+1{color} | @author |   0m  0s | The patch does not contain any 
@author tags. |
| {color:green}+1{color} | tests included |   0m  0s | The patch appears to 
include 2 new or modified test files. |
| {color:green}+1{color} | javac |   8m 23s | There were no new javac warning 
messages. |
| {color:green}+1{color} | javadoc |  10m 52s | There were no new javadoc 
warning messages. |
| {color:green}+1{color} | release audit |   0m 26s | The applied patch does 
not increase the total number of release audit warnings. |
| {color:green}+1{color} | site |   3m  6s | Site still builds. |
| {color:red}-1{color} | checkstyle |   1m 10s | The applied patch generated  2 
new checkstyle issues (total was 45, now 46). |
| {color:green}+1{color} | whitespace |   0m  0s | The patch has no lines that 
end in whitespace. |
| {color:green}+1{color} | install |   1m 39s | mvn install still works. |
| {color:green}+1{color} | eclipse:eclipse |   0m 37s | The patch built with 
eclipse:eclipse. |
| {color:green}+1{color} | findbugs |   2m  6s | The patch does not introduce 
any new Findbugs (version 3.0.0) warnings. |
| {color:red}-1{color} | common tests |  23m  8s | Tests failed in 
hadoop-common. |
| | |  73m 40s | |
\\
\\
|| Reason || Tests ||
| Failed unit tests | hadoop.fs.sftp.TestSFTPFileSystem |
\\
\\
|| Subsystem || Report/Notes ||
| Patch URL | 
http://issues.apache.org/jira/secure/attachment/12752298/HDFS-8836-02.patch |
| Optional Tests | javadoc javac unit findbugs checkstyle site |
| git revision | trunk / eee0d45 |
| checkstyle |  
https://builds.apache.org/job/PreCommit-HDFS-Build/12123/artifact/patchprocess/diffcheckstylehadoop-common.txt
 |
| hadoop-common test log | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12123/artifact/patchprocess/testrun_hadoop-common.txt
 |
| Test Results | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12123/testReport/ |
| Java | 1.7.0_55 |
| uname | Linux asf904.gq1.ygridcore.net 3.13.0-36-lowlatency #63-Ubuntu SMP 
PREEMPT Wed Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Console output | 
https://builds.apache.org/job/PreCommit-HDFS-Build/12123/console |


This message was automatically generated.

 Skip newline on empty files with getMerge -nl
 -

 Key: HDFS-8836
 URL: https://issues.apache.org/jira/browse/HDFS-8836
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Affects Versions: 2.6.0, 2.7.1
Reporter: Jan Filipiak
Assignee: kanaka kumar avvaru
Priority: Trivial
 Attachments: HDFS-8836-01.patch, HDFS-8836-02.patch


 Hello everyone,
 I recently was in the need of using the new line option -nl with getMerge 
 because the files I needed to merge simply didn't had one. I was merging all 
 the files from one directory and unfortunately this directory also included 
 empty files, which effectively led to multiple newlines append after some 
 files. I needed to remove them manually afterwards.
 In this situation it is maybe good to have another argument that allows 
 skipping empty files.
 Thing one could try to implement this feature:
 The call for IOUtils.copyBytes(in, out, getConf(), false); doesn't
 return the number of bytes copied which would be convenient as one could
 skip append the new line when 0 bytes where copied or one would check the 
 file size before.
 I posted this Idea on the mailing list 
 http://mail-archives.apache.org/mod_mbox/hadoop-user/201507.mbox/%3C55B25140.3060005%40trivago.com%3E
  but I didn't really get many responses, so I thought I my try this way.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8950) NameNode refresh doesn't remove DataNodes

2015-08-25 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-8950:
---
Fix Version/s: 2.8.0
   Status: Patch Available  (was: In Progress)

Ready for review

 NameNode refresh doesn't remove DataNodes
 -

 Key: HDFS-8950
 URL: https://issues.apache.org/jira/browse/HDFS-8950
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Daniel Templeton
Assignee: Daniel Templeton
 Fix For: 2.8.0

 Attachments: HDFS-8950.001.patch


 If you remove a DN from NN's allowed host list (HDFS was HA) and then do NN 
 refresh, it doesn't remove it actually and the NN UI keeps showing that node. 
 It may try to allocate some blocks to that DN as well during an MR job.  This 
 issue is independent from DN decommission.
 To reproduce:
 1. Add a DN to dfs_hosts_allow
 2. Refresh NN
 3. Start DN. Now NN starts seeing DN.
 4. Stop DN
 5. Remove DN from dfs_hosts_allow
 6. Refresh NN - NN is still reporting DN as being used by HDFS.
 This is different from decom because there DN is added to exclude list in 
 addition to being removed from allowed list, and in that case everything 
 works correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8950) NameNode refresh doesn't remove DataNodes

2015-08-25 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-8950:
---
Attachment: HDFS-8950.002.patch

Just noticed an extraneous import in the patch. (Swing?!?  WTF?)  Removed and 
reposting patch.

 NameNode refresh doesn't remove DataNodes
 -

 Key: HDFS-8950
 URL: https://issues.apache.org/jira/browse/HDFS-8950
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Daniel Templeton
Assignee: Daniel Templeton
 Fix For: 2.8.0

 Attachments: HDFS-8950.001.patch, HDFS-8950.002.patch


 If you remove a DN from NN's allowed host list (HDFS was HA) and then do NN 
 refresh, it doesn't remove it actually and the NN UI keeps showing that node. 
 It may try to allocate some blocks to that DN as well during an MR job.  This 
 issue is independent from DN decommission.
 To reproduce:
 1. Add a DN to dfs_hosts_allow
 2. Refresh NN
 3. Start DN. Now NN starts seeing DN.
 4. Stop DN
 5. Remove DN from dfs_hosts_allow
 6. Refresh NN - NN is still reporting DN as being used by HDFS.
 This is different from decom because there DN is added to exclude list in 
 addition to being removed from allowed list, and in that case everything 
 works correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Add a unit test for INotify functionality across a layout version upgrade

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712233#comment-14712233
 ] 

Hudson commented on HDFS-8846:
--

SUCCESS: Integrated in Hadoop-Mapreduce-trunk #2254 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/2254/])
HDFS-8846. Add a unit test for INotify functionality across a layout version 
upgrade (Zhe Zhang via Colin P. McCabe) (cmccabe: rev 
a4d9acc51d1a977bc333da17780c00c72e8546f1)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgrade.java
* hadoop-hdfs-project/hadoop-hdfs/src/test/resources/hadoop-252-dfs-dir.tgz
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSInotifyEventInputStream.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java


 Add a unit test for INotify functionality across a layout version upgrade
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: 2.8.0

 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Add a unit test for INotify functionality across a layout version upgrade

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712265#comment-14712265
 ] 

Hudson commented on HDFS-8846:
--

FAILURE: Integrated in Hadoop-Hdfs-trunk-Java8 #297 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Java8/297/])
HDFS-8846. Add a unit test for INotify functionality across a layout version 
upgrade (Zhe Zhang via Colin P. McCabe) (cmccabe: rev 
a4d9acc51d1a977bc333da17780c00c72e8546f1)
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* hadoop-hdfs-project/hadoop-hdfs/src/test/resources/hadoop-252-dfs-dir.tgz
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgrade.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSInotifyEventInputStream.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java


 Add a unit test for INotify functionality across a layout version upgrade
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: 2.8.0

 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Add a unit test for INotify functionality across a layout version upgrade

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712277#comment-14712277
 ] 

Hudson commented on HDFS-8846:
--

FAILURE: Integrated in Hadoop-Hdfs-trunk #2235 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/2235/])
HDFS-8846. Add a unit test for INotify functionality across a layout version 
upgrade (Zhe Zhang via Colin P. McCabe) (cmccabe: rev 
a4d9acc51d1a977bc333da17780c00c72e8546f1)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSInotifyEventInputStream.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgrade.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java
* hadoop-hdfs-project/hadoop-hdfs/src/test/resources/hadoop-252-dfs-dir.tgz


 Add a unit test for INotify functionality across a layout version upgrade
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: 2.8.0

 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8950) NameNode refresh doesn't remove DataNodes

2015-08-25 Thread Daniel Templeton (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8950?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HDFS-8950:
---
Attachment: (was: 28531.patch)

 NameNode refresh doesn't remove DataNodes
 -

 Key: HDFS-8950
 URL: https://issues.apache.org/jira/browse/HDFS-8950
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.6.0
Reporter: Daniel Templeton
Assignee: Daniel Templeton

 If you remove a DN from NN's allowed host list (HDFS was HA) and then do NN 
 refresh, it doesn't remove it actually and the NN UI keeps showing that node. 
 It may try to allocate some blocks to that DN as well during an MR job.  This 
 issue is independent from DN decommission.
 To reproduce:
 1. Add a DN to dfs_hosts_allow
 2. Refresh NN
 3. Start DN. Now NN starts seeing DN.
 4. Stop DN
 5. Remove DN from dfs_hosts_allow
 6. Refresh NN - NN is still reporting DN as being used by HDFS.
 This is different from decom because there DN is added to exclude list in 
 addition to being removed from allowed list, and in that case everything 
 works correctly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Add a unit test for INotify functionality across a layout version upgrade

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712126#comment-14712126
 ] 

Hudson commented on HDFS-8846:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #305 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/305/])
HDFS-8846. Add a unit test for INotify functionality across a layout version 
upgrade (Zhe Zhang via Colin P. McCabe) (cmccabe: rev 
a4d9acc51d1a977bc333da17780c00c72e8546f1)
* hadoop-hdfs-project/hadoop-hdfs/src/test/resources/hadoop-252-dfs-dir.tgz
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSInotifyEventInputStream.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgrade.java


 Add a unit test for INotify functionality across a layout version upgrade
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: 2.8.0

 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8846) Add a unit test for INotify functionality across a layout version upgrade

2015-08-25 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8846?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14712135#comment-14712135
 ] 

Hudson commented on HDFS-8846:
--

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #310 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/310/])
HDFS-8846. Add a unit test for INotify functionality across a layout version 
upgrade (Zhe Zhang via Colin P. McCabe) (cmccabe: rev 
a4d9acc51d1a977bc333da17780c00c72e8546f1)
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgradeFromImage.java
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSUpgrade.java
* 
hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/TestDFSInotifyEventInputStream.java
* hadoop-hdfs-project/hadoop-hdfs/src/test/resources/hadoop-252-dfs-dir.tgz


 Add a unit test for INotify functionality across a layout version upgrade
 -

 Key: HDFS-8846
 URL: https://issues.apache.org/jira/browse/HDFS-8846
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.7.1
Reporter: Zhe Zhang
Assignee: Zhe Zhang
 Fix For: 2.8.0

 Attachments: HDFS-8846.00.patch, HDFS-8846.01.patch, 
 HDFS-8846.02.patch, HDFS-8846.03.patch


 Per discussion under HDFS-8480, we should create some edit log files with old 
 layout version, to test whether they can be correctly handled in upgrades.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >