[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851494#comment-16851494 ] Wei-Chiu Chuang commented on HDFS-14514: I am not sure if this is a bug prior to branch-2.9. HDFS-11402 was resolved in 2.9.0 and later. Prior to 2.9.0, we don't even support immutable snapshots and fixing this bug doesn't make much sense. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Fix For: 2.10.0, 2.9.3 > > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch, > HDFS-14514.branch-2.004.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851433#comment-16851433 ] Hadoop QA commented on HDFS-14514: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 10s{color} | {color:red} HDFS-14514 does not apply to branch-2. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | JIRA Issue | HDFS-14514 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12970255/HDFS-14514.branch-2.004.patch | | Console output | https://builds.apache.org/job/PreCommit-HDFS-Build/26865/console | | Powered by | Apache Yetus 0.8.0 http://yetus.apache.org | This message was automatically generated. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch, > HDFS-14514.branch-2.004.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851432#comment-16851432 ] Wei-Chiu Chuang commented on HDFS-14514: Pushed v003 to branch-3. There's a conflict in branch-2.8 so I'll wait for an update. I used the following commit message so it will be recognized as the contribution by both [~smeng] and [~sodonnell] on github. {noformat} Co-authored-by: Stephen O'Donnell {noformat} > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch, > HDFS-14514.branch-2.004.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851423#comment-16851423 ] Wei-Chiu Chuang commented on HDFS-14514: +1 looks great. I'll update the patch to convert into else-if upon commit and upload an updated patch for future reference. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851419#comment-16851419 ] Daniel Templeton commented on HDFS-14514: - LGTM. I'd like to see the last two {{if}}s in {{DFSInputStream}} be an {{if/else-if}}, but I can fix that on commit. If there are no complaints, I'll commit this later this evening. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851267#comment-16851267 ] Stephen O'Donnell commented on HDFS-14514: -- For the v003 patch the failed tests seem unrelated to this change. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851219#comment-16851219 ] Hadoop QA commented on HDFS-14514: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 13m 2s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} branch-2 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 17s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 8m 59s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 44s{color} | {color:green} branch-2 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 27s{color} | {color:green} branch-2 passed with JDK v1.8.0_212 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 35s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 27s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 18s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 37s{color} | {color:green} branch-2 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 11s{color} | {color:green} branch-2 passed with JDK v1.8.0_212 {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 10s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 44s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 44s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 27s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 27s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 32s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 33s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 27s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 24s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 54m 25s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 32s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}105m 6s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.server.namenode.ha.TestSeveralNameNodes | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:da67579 | | JIRA Issue | HDFS-14514 | | JIRA Patch URL |
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16851087#comment-16851087 ] Stephen O'Donnell commented on HDFS-14514: -- When a ByteBuffer is used to read data into, it is not valid to modify the limit of the ByteBuffer directly, as the client may be using that. However the buffer limit is also used to enforce how much data should be read in. In the short circuit read tests that failed with the v002 patch, the client was setting the buffer limit to a relatively high value (eg 16k). Then on each read from the file system, less than that (eg 4K) was read into the buffer. The code change in patch v002 set the limit of the buffer to 4K and then that made it back to the client which then believed the buffer was full. This caused it to stop reading and hence not read all the expected data. What you need to do instead is: 1. Clone the ByteBuffer, which does not clone the data, just the position and limit pointers. 2. Set the limit of the cloned buffer as required in the read method. 3. Then perform a read, and the number of bytes read will be returned. 4. Finally update the position of the non-tmp buffer to be origional_position + read_bytes This is largely how the code was changed to operate in the refactor of these methods (HDFS-8905). I have updated a v003 patch in which I got the failing tests to pass locally with the change described above. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch, HDFS-14514.branch-2.003.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16850720#comment-16850720 ] Stephen O'Donnell commented on HDFS-14514: -- The v002 patch has cleared most of the test failures, but there are still short Circuit read tests failing, probably due to this change. To answer the question on why is only a read from an Encrypted Zone affected: For a non-ez read, the code path ends up in DFSInputStream in the following method: {code} public synchronized int read(final byte buf[], int off, int len) throws IOException { validatePositionedReadArgs(pos, buf, off, len); if (len == 0) { return 0; } ReaderStrategy byteArrayReader = new ByteArrayStrategy(buf); ... {code} Note it is creating a byteArrayStrategy object to read the underlying stream. For the EZ read, the code goes to CrytpoInputStream which then calls a different read method in DFSInputStream: {code} public synchronized int read(final ByteBuffer buf) throws IOException { ReaderStrategy byteBufferReader = new ByteBufferStrategy(buf); ... {code} Note here it is creating a ByteBufferStrategy object rather than a ByteArrayStrategy object. It is within the ByteBufferStrategy object where this bug appears to be, and that explains why a non-EZ read is not impacted. The code does not use that object. I will look further to see if I can figure out why the ShortCircuit read tests are impact by this change. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch, > HDFS-14514.branch-2.002.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16850220#comment-16850220 ] Hadoop QA commented on HDFS-14514: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 26m 24s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} branch-2 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 42s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 12m 53s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 46s{color} | {color:green} branch-2 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 32s{color} | {color:green} branch-2 passed with JDK v1.8.0_212 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 38s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 29s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 27s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 34s{color} | {color:green} branch-2 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 13s{color} | {color:green} branch-2 passed with JDK v1.8.0_212 {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 10s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 19s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 46s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 29s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 29s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 30s{color} | {color:orange} hadoop-hdfs-project: The patch generated 1 new + 119 unchanged - 0 fixed = 120 total (was 119) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 17s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 29s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 22s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 1s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 20s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 74m 46s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 23s{color} | {color:red} The patch generated 12 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}143m 16s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.shortcircuit.TestShortCircuitLocalRead | | | hadoop.hdfs.server.namenode.TestNameNodeHttpServerXFrame | | | hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys | | | hadoop.hdfs.web.TestWebHdfsTimeouts | | | hadoop.hdfs.TestRollingUpgrade | | | hadoop.hdfs.server.balancer.TestBalancerRPCDelay | \\
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16850065#comment-16850065 ] Stephen O'Donnell commented on HDFS-14514: -- I can reproduce those test failures locally. Looking at the refactored code in HDFS-8905 (HADOOP 3.0 and later) where this issue does not occur, the code becomes: {code} public int readFromBlock(BlockReader blockReader, int length) throws IOException { ByteBuffer tmpBuf = readBuf.duplicate(); tmpBuf.limit(tmpBuf.position() + length); int nRead = blockReader.read(tmpBuf); // Only when data are read, update the position if (nRead > 0) { readBuf.position(readBuf.position() + nRead); } return nRead; } {code} Compared to the same code in the 2.x branch: {code} public int doRead(BlockReader blockReader, int off, int len) throws ChecksumException, IOException { int oldpos = buf.position(); int oldlimit = buf.limit(); boolean success = false; try { int ret = blockReader.read(buf); success = true; updateReadStatistics(readStatistics, ret, blockReader); return ret; } finally { if (!success) { // Reset to original state so that retries work correctly. buf.position(oldpos); buf.limit(oldlimit); } } } {code} Note that: 1. In the refactored code, the 'off' parameter has been dropped completely, indicating it is not needed but 'len' is now used. 2. In the 2.x code, off and len are parameters to the method but never used, and I believe "len" needs to be used or it causes this bug. In the v001 patch, if I remove the line setting the offset, it fixes the test failures I checked locally. I suspect with a ByteBufferReader, the offset is controlled elsewhere. To start with, I will push up a new patch with that change to see if all tests pass. Then we can look further into why non-encrypted reads are not affected. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16850047#comment-16850047 ] Stephen O'Donnell commented on HDFS-14514: -- I believe it only happens in snapshots due to, as that is the only place where the file length recorded in the namenode matters. In the snapshot, the length should be frozen, but outside of a snapshot, the reader will just read all available data based on the visible length returned from the datanode. I have not yet traced why this only impacts CryptoInputStream - the bug appears to be in ByteBufferStrategy, which is a static class within DFSInputStream, so it is probably related to the ReaderStrategy used by the calling class. I need to look a bit further to see why non-encrypted data is not affected. I agree that these test failures seem related. I will have a further look at them too. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16849963#comment-16849963 ] Wei-Chiu Chuang commented on HDFS-14514: Thanks [~smeng] [~sodonnell] this is a nasty bug. Haven't verified myself, but some of the failed tests seem related to the change. Any ideas why the length is only incorrect in this special case of encryption zone + snapshot + unclosed file? The change in the patch is in DFSInputStream and I would have expected this is also the case for no snapshots or outside of encryption zones. > Actual read size of open file in encryption zone still larger than listing > size even after enabling HDFS-11402 in Hadoop 2 > -- > > Key: HDFS-14514 > URL: https://issues.apache.org/jira/browse/HDFS-14514 > Project: Hadoop HDFS > Issue Type: Bug > Components: encryption, hdfs, snapshots >Affects Versions: 2.6.5, 2.9.2, 2.8.5, 2.7.7 >Reporter: Siyao Meng >Assignee: Siyao Meng >Priority: Major > Attachments: HDFS-14514.branch-2.001.patch > > > In Hadoop 2, when a file is opened for write in *encryption zone*, taken a > snapshot and appended, the read out file size in the snapshot is larger than > the listing size. This happens even when immutable snapshot HDFS-11402 is > enabled. > Note: The refactor HDFS-8905 happened in Hadoop 3.0 and later fixed the bug > silently (probably incidentally). Hadoop 2.x are still suffering from this > issue. > Thanks [~sodonnell] for locating the root cause in the codebase. > Repro: > 1. Set dfs.namenode.snapshot.capture.openfiles to true in hdfs-site.xml, > start HDFS cluster > 2. Create an empty directory /dataenc, create encryption zone and allow > snapshot on it > {code:bash} > hadoop key create reprokey > sudo -u hdfs hdfs dfs -mkdir /dataenc > sudo -u hdfs hdfs crypto -createZone -keyName reprokey -path /dataenc > sudo -u hdfs hdfs dfsadmin -allowSnapshot /dataenc > {code} > 3. Use a client that keeps a file open for write under /dataenc. For example, > I'm using Flume HDFS sink to tail a local file. > 4. Append the file several times using the client, keep the file open. > 5. Create a snapshot > {code:bash} > sudo -u hdfs hdfs dfs -createSnapshot /dataenc snap1 > {code} > 6. Append the file one or more times, but don't let the file size exceed the > block size limit. Wait for several seconds for the append to be flushed to DN. > 7. Do a -ls on the file inside the snapshot, then try to read the file using > -get, you should see the actual file size read is larger than the listing > size from -ls. > The patch and an updated unit test will be uploaded later. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: hdfs-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-issues-h...@hadoop.apache.org
[jira] [Commented] (HDFS-14514) Actual read size of open file in encryption zone still larger than listing size even after enabling HDFS-11402 in Hadoop 2
[ https://issues.apache.org/jira/browse/HDFS-14514?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16849520#comment-16849520 ] Hadoop QA commented on HDFS-14514: -- | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 22s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} branch-2 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 43s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 13m 34s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 2m 7s{color} | {color:green} branch-2 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 49s{color} | {color:green} branch-2 passed with JDK v1.8.0_212 {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 39s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 41s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 35s{color} | {color:green} branch-2 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 47s{color} | {color:green} branch-2 passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 6s{color} | {color:green} branch-2 passed with JDK v1.8.0_212 {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 10s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 1m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 42s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 1m 31s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 1m 31s{color} | {color:green} the patch passed {color} | | {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange} 0m 32s{color} | {color:orange} hadoop-hdfs-project: The patch generated 1 new + 119 unchanged - 0 fixed = 120 total (was 119) {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 24s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 3m 29s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 29s{color} | {color:green} the patch passed with JDK v1.7.0_95 {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 1m 0s{color} | {color:green} the patch passed with JDK v1.8.0_212 {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 1m 17s{color} | {color:green} hadoop-hdfs-client in the patch passed. {color} | | {color:red}-1{color} | {color:red} unit {color} | {color:red} 87m 19s{color} | {color:red} hadoop-hdfs in the patch failed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 34s{color} | {color:red} The patch generated 12 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}131m 52s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.hdfs.crypto.TestHdfsCryptoStreams | | | hadoop.hdfs.TestParallelRead | | | hadoop.hdfs.TestParallelShortCircuitLegacyRead | | | hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys | | | hadoop.hdfs.shortcircuit.TestShortCircuitLocalRead | | | hadoop.hdfs.server.namenode.snapshot.TestSnapshotFileLength