[jira] Updated: (HDFS-1105) Balancer improvement

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1105:


 Component/s: balancer
Hadoop Flags: [Reviewed]

> Balancer improvement
> 
>
> Key: HDFS-1105
> URL: https://issues.apache.org/jira/browse/HDFS-1105
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
> Fix For: 0.23.0
>
> Attachments: HDFS-1105.2.patch, HDFS-1105.3.patch, HDFS-1105.4.patch, 
> HDFS-1105.patch
>
>
> We were seeing some weird issues with the balancer in our cluster:
> 1) it can get stuck during an iteration and only restarting it helps
> 2) the iterations are highly inefficient. With 20 minutes iteration it moves 
> 7K blocks a minute for the first 6 minutes and hundreds of blocks in the next 
> 14 minutes
> 3) it can hit namenode and the network pretty hard
> A few improvements we came up with as a result:
> Making balancer more deterministic in terms of running time of iteration, 
> improving the efficiency and making the load configurable:
> Make many of the constants configurable command line parameters: Iteration 
> length, number of blocks to move in parallel to a given node and in cluster 
> overall.
> Terminate transfers that are still in progress after iteration is over.
> Previously iteration time was the time window in which the balancer was 
> scheduling the moves and then it would wait for the moves to finish 
> indefinitely. Each scheduling task can run up to iteration time or even 
> longer. This means if you have too many of them and they are long your actual 
> iterations are longer than 20 minutes. Now each scheduling task has a time of 
> the start of iteration and it should schedule the moves only if it did not 
> run out of time. So the tasks that have started after the iteration is over 
> will not schedule any moves.
> The number of move threads and dispatch threads is configurable so that 
> depending on the load of the cluster you can run it slower.
> I will attach a patch, please let me know what you think and what can be done 
> better.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1105) Balancer improvement

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1105:


Status: Patch Available  (was: Open)

> Balancer improvement
> 
>
> Key: HDFS-1105
> URL: https://issues.apache.org/jira/browse/HDFS-1105
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
> Fix For: 0.23.0
>
> Attachments: HDFS-1105.2.patch, HDFS-1105.3.patch, HDFS-1105.4.patch, 
> HDFS-1105.patch
>
>
> We were seeing some weird issues with the balancer in our cluster:
> 1) it can get stuck during an iteration and only restarting it helps
> 2) the iterations are highly inefficient. With 20 minutes iteration it moves 
> 7K blocks a minute for the first 6 minutes and hundreds of blocks in the next 
> 14 minutes
> 3) it can hit namenode and the network pretty hard
> A few improvements we came up with as a result:
> Making balancer more deterministic in terms of running time of iteration, 
> improving the efficiency and making the load configurable:
> Make many of the constants configurable command line parameters: Iteration 
> length, number of blocks to move in parallel to a given node and in cluster 
> overall.
> Terminate transfers that are still in progress after iteration is over.
> Previously iteration time was the time window in which the balancer was 
> scheduling the moves and then it would wait for the moves to finish 
> indefinitely. Each scheduling task can run up to iteration time or even 
> longer. This means if you have too many of them and they are long your actual 
> iterations are longer than 20 minutes. Now each scheduling task has a time of 
> the start of iteration and it should schedule the moves only if it did not 
> run out of time. So the tasks that have started after the iteration is over 
> will not schedule any moves.
> The number of move threads and dispatch threads is configurable so that 
> depending on the load of the cluster you can run it slower.
> I will attach a patch, please let me know what you think and what can be done 
> better.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1105) Balancer improvement

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1105:


Fix Version/s: 0.23.0
   Status: Open  (was: Patch Available)

> Balancer improvement
> 
>
> Key: HDFS-1105
> URL: https://issues.apache.org/jira/browse/HDFS-1105
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: balancer
>Reporter: Dmytro Molkov
>Assignee: Dmytro Molkov
> Fix For: 0.23.0
>
> Attachments: HDFS-1105.2.patch, HDFS-1105.3.patch, HDFS-1105.4.patch, 
> HDFS-1105.patch
>
>
> We were seeing some weird issues with the balancer in our cluster:
> 1) it can get stuck during an iteration and only restarting it helps
> 2) the iterations are highly inefficient. With 20 minutes iteration it moves 
> 7K blocks a minute for the first 6 minutes and hundreds of blocks in the next 
> 14 minutes
> 3) it can hit namenode and the network pretty hard
> A few improvements we came up with as a result:
> Making balancer more deterministic in terms of running time of iteration, 
> improving the efficiency and making the load configurable:
> Make many of the constants configurable command line parameters: Iteration 
> length, number of blocks to move in parallel to a given node and in cluster 
> overall.
> Terminate transfers that are still in progress after iteration is over.
> Previously iteration time was the time window in which the balancer was 
> scheduling the moves and then it would wait for the moves to finish 
> indefinitely. Each scheduling task can run up to iteration time or even 
> longer. This means if you have too many of them and they are long your actual 
> iterations are longer than 20 minutes. Now each scheduling task has a time of 
> the start of iteration and it should schedule the moves only if it did not 
> run out of time. So the tasks that have started after the iteration is over 
> will not schedule any moves.
> The number of move threads and dispatch threads is configurable so that 
> depending on the load of the cluster you can run it slower.
> I will attach a patch, please let me know what you think and what can be done 
> better.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1070) Speedup NameNode image loading and saving by storing local file names

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1070:


Attachment: trunkLocalNameImage1.patch

> Speedup NameNode image loading and saving by storing local file names
> -
>
> Key: HDFS-1070
> URL: https://issues.apache.org/jira/browse/HDFS-1070
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.22.0
>
> Attachments: trunkLocalNameImage.patch, trunkLocalNameImage1.patch
>
>
> Currently each inode stores its full path in the fsimage. I'd propose to 
> store the local name instead. In order for each inode to identify its parent, 
> all inodes in a directory tree are stored in the image in in-order. This 
> proposal also requires each directory stores the number of its children in 
> image.
> This proposal would bring a few benefits as pointed below and therefore 
> speedup the image loading and saving.
> # Remove the overhead of converting java-UTF8 encoded local name to 
> string-represented full path then to UTF8 encoded full path when saving to an 
> image and vice versa when loading the image.
> # Remove the overhead of traversing the full path when inserting the inode to 
> its parent inode.
> # Reduce the number of temporary java objects during the process of image 
> loading or saving and  therefore reduce the GC overhead.
> # Reduce the size of an image.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1070) Speedup NameNode image loading and saving by storing local file names

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1070:


Attachment: (was: trunkLocalNameImage1.patch)

> Speedup NameNode image loading and saving by storing local file names
> -
>
> Key: HDFS-1070
> URL: https://issues.apache.org/jira/browse/HDFS-1070
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.22.0
>
> Attachments: trunkLocalNameImage.patch, trunkLocalNameImage1.patch
>
>
> Currently each inode stores its full path in the fsimage. I'd propose to 
> store the local name instead. In order for each inode to identify its parent, 
> all inodes in a directory tree are stored in the image in in-order. This 
> proposal also requires each directory stores the number of its children in 
> image.
> This proposal would bring a few benefits as pointed below and therefore 
> speedup the image loading and saving.
> # Remove the overhead of converting java-UTF8 encoded local name to 
> string-represented full path then to UTF8 encoded full path when saving to an 
> image and vice versa when loading the image.
> # Remove the overhead of traversing the full path when inserting the inode to 
> its parent inode.
> # Reduce the number of temporary java objects during the process of image 
> loading or saving and  therefore reduce the GC overhead.
> # Reduce the size of an image.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1070) Speedup NameNode image loading and saving by storing local file names

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1070?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1070:


Attachment: trunkLocalNameImage1.patch

trunkLocalnameImage1.patch fixes a few bugs in the previous patch.

> Speedup NameNode image loading and saving by storing local file names
> -
>
> Key: HDFS-1070
> URL: https://issues.apache.org/jira/browse/HDFS-1070
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.22.0
>
> Attachments: trunkLocalNameImage.patch, trunkLocalNameImage1.patch
>
>
> Currently each inode stores its full path in the fsimage. I'd propose to 
> store the local name instead. In order for each inode to identify its parent, 
> all inodes in a directory tree are stored in the image in in-order. This 
> proposal also requires each directory stores the number of its children in 
> image.
> This proposal would bring a few benefits as pointed below and therefore 
> speedup the image loading and saving.
> # Remove the overhead of converting java-UTF8 encoded local name to 
> string-represented full path then to UTF8 encoded full path when saving to an 
> image and vice versa when loading the image.
> # Remove the overhead of traversing the full path when inserting the inode to 
> its parent inode.
> # Reduce the number of temporary java objects during the process of image 
> loading or saving and  therefore reduce the GC overhead.
> # Reduce the size of an image.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Assigned: (HDFS-1516) mvn-install is broken after 0.22 branch creation

2010-11-24 Thread Konstantin Boudnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1516?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Konstantin Boudnik reassigned HDFS-1516:


Assignee: Konstantin Boudnik

> mvn-install is broken after 0.22 branch creation
> 
>
> Key: HDFS-1516
> URL: https://issues.apache.org/jira/browse/HDFS-1516
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: build
>Affects Versions: 0.23.0
>Reporter: Konstantin Boudnik
>Assignee: Konstantin Boudnik
> Fix For: 0.23.0
>
> Attachments: HDFS-1516.patch, HDFS-1516.patch
>
>
> Version HAS to be bumped for system testing framework artifacts (as mentioned 
> in the build.xml file)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1486) Generalize CLITest structure and interfaces to faciliate upstream adoption (e.g. for web testing)

2010-11-24 Thread Konstantin Boudnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Konstantin Boudnik updated HDFS-1486:
-

Affects Version/s: (was: 0.22.0)
   0.23.0

> Generalize CLITest structure and interfaces to faciliate upstream adoption 
> (e.g. for web testing)
> -
>
> Key: HDFS-1486
> URL: https://issues.apache.org/jira/browse/HDFS-1486
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 0.23.0
>Reporter: Konstantin Boudnik
>Assignee: Konstantin Boudnik
> Attachments: HDFS-1486.patch, HDFS-1486.patch, HDFS-1486.patch, 
> HDFS-1486.patch
>
>
> HDFS part of HADOOP-7014. HDFS side of TestCLI doesn't require any special 
> changes but needs to be aligned with Common

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1486) Generalize CLITest structure and interfaces to faciliate upstream adoption (e.g. for web testing)

2010-11-24 Thread Konstantin Boudnik (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1486?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Konstantin Boudnik updated HDFS-1486:
-

Attachment: HDFS-1486.patch

Changes corresponding to the latest HADOOP-7014. TestStorageRestore is failing 
because of HDFS-1496

> Generalize CLITest structure and interfaces to faciliate upstream adoption 
> (e.g. for web testing)
> -
>
> Key: HDFS-1486
> URL: https://issues.apache.org/jira/browse/HDFS-1486
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>Affects Versions: 0.22.0
>Reporter: Konstantin Boudnik
>Assignee: Konstantin Boudnik
> Attachments: HDFS-1486.patch, HDFS-1486.patch, HDFS-1486.patch, 
> HDFS-1486.patch
>
>
> HDFS part of HADOOP-7014. HDFS side of TestCLI doesn't require any special 
> changes but needs to be aligned with Common

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1103) Replica recovery doesn't distinguish between flushed-but-corrupted last chunk and unflushed last chunk

2010-11-24 Thread Hairong Kuang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935564#action_12935564
 ] 

Hairong Kuang commented on HDFS-1103:
-

RBW B and RBW C's GS gets bumped up by pipeline recovery since the client 
detects that A in pipeline is down.
Even if the client fails before pipeline recovery starts, the lease recovery 
does not takes RWR into consideration when there is RBW or finalized replica 
available.


> Replica recovery doesn't distinguish between flushed-but-corrupted last chunk 
> and unflushed last chunk
> --
>
> Key: HDFS-1103
> URL: https://issues.apache.org/jira/browse/HDFS-1103
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: data-node
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Todd Lipcon
>Priority: Blocker
> Attachments: hdfs-1103-test.txt
>
>
> When the DN creates a replica under recovery, it calls validateIntegrity, 
> which truncates the last checksum chunk off of a replica if it is found to be 
> invalid. Then when the block recovery process happens, this shortened block 
> wins over a longer replica from another node where there was no corruption. 
> Thus, if just one of the DNs has an invalid last checksum chunk, data that 
> has been sync()ed to other datanodes can be lost.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1103) Replica recovery doesn't distinguish between flushed-but-corrupted last chunk and unflushed last chunk

2010-11-24 Thread Thanh Do (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1103?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935560#action_12935560
 ] 

Thanh Do commented on HDFS-1103:


"I do not think that this is the case in 0.21 & the trunk. In our lease 
recovery algorithm in 0.21, If there are 2 RBWs and 1 RWR, 1 RWR is excluded 
from the lease recovery. In the scenario that you described, RBW B and RBW C's 
GS is bumped and the length of recovered two replicas is truncated to MIN( 
len(B), len(C)). "

Hairong, can you explain to me that why RBW B and RBW C's GS are bumped up.
Is that because of the lease recovery protocol?
But from my understanding, from Todd description, NN lease recovery is trigger
after Machine A report...

> Replica recovery doesn't distinguish between flushed-but-corrupted last chunk 
> and unflushed last chunk
> --
>
> Key: HDFS-1103
> URL: https://issues.apache.org/jira/browse/HDFS-1103
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: data-node
>Affects Versions: 0.21.0, 0.22.0
>Reporter: Todd Lipcon
>Priority: Blocker
> Attachments: hdfs-1103-test.txt
>
>
> When the DN creates a replica under recovery, it calls validateIntegrity, 
> which truncates the last checksum chunk off of a replica if it is found to be 
> invalid. Then when the block recovery process happens, this shortened block 
> wins over a longer replica from another node where there was no corruption. 
> Thus, if just one of the DNs has an invalid last checksum chunk, data that 
> has been sync()ed to other datanodes can be lost.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1458) Improve checkpoint performance by avoiding unnecessary image downloads

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1458?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1458:


Resolution: Fixed
Status: Resolved  (was: Patch Available)

I've just committed this.

> Improve checkpoint performance by avoiding unnecessary image downloads
> --
>
> Key: HDFS-1458
> URL: https://issues.apache.org/jira/browse/HDFS-1458
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.22.0
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.23.0
>
> Attachments: checkpoint-checkfsimageissame.patch, 
> trunkNoDownloadImage.patch, trunkNoDownloadImage1.patch, 
> trunkNoDownloadImage2.patch, trunkNoDownloadImage3.patch
>
>
> If secondary namenode could verify that the image it has on its disk is the 
> same as the one in the primary NameNode, it could skip downloading the image 
> from the primary NN, thus completely eliminating the image download overhead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1458) Improve checkpoint performance by avoiding unnecessary image downloads

2010-11-24 Thread Hairong Kuang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935556#action_12935556
 ] 

Hairong Kuang commented on HDFS-1458:
-

I ran the tests on my local machine. Those are the failed test
TestBlockRecovery
TestBlockTokenwithDFS
TestDFSClientRetires
TestStorageRestore
TestSaveNamespace

None of them are related to my patch.

> Improve checkpoint performance by avoiding unnecessary image downloads
> --
>
> Key: HDFS-1458
> URL: https://issues.apache.org/jira/browse/HDFS-1458
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.22.0
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.23.0
>
> Attachments: checkpoint-checkfsimageissame.patch, 
> trunkNoDownloadImage.patch, trunkNoDownloadImage1.patch, 
> trunkNoDownloadImage2.patch, trunkNoDownloadImage3.patch
>
>
> If secondary namenode could verify that the image it has on its disk is the 
> same as the one in the primary NameNode, it could skip downloading the image 
> from the primary NN, thus completely eliminating the image download overhead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Updated: (HDFS-1458) Improve checkpoint performance by avoiding unnecessary image downloads

2010-11-24 Thread Hairong Kuang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-1458?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Hairong Kuang updated HDFS-1458:


Fix Version/s: (was: 0.22.0)
   0.23.0
 Hadoop Flags: [Reviewed]
   Status: Patch Available  (was: Open)

> Improve checkpoint performance by avoiding unnecessary image downloads
> --
>
> Key: HDFS-1458
> URL: https://issues.apache.org/jira/browse/HDFS-1458
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.22.0
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.23.0
>
> Attachments: checkpoint-checkfsimageissame.patch, 
> trunkNoDownloadImage.patch, trunkNoDownloadImage1.patch, 
> trunkNoDownloadImage2.patch, trunkNoDownloadImage3.patch
>
>
> If secondary namenode could verify that the image it has on its disk is the 
> same as the one in the primary NameNode, it could skip downloading the image 
> from the primary NN, thus completely eliminating the image download overhead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1503) TestSaveNamespace fails

2010-11-24 Thread Konstantin Shvachko (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935546#action_12935546
 ] 

Konstantin Shvachko commented on HDFS-1503:
---

+1. Tested on my box: testCrashWhileSavingSecondImage() fails without the 
patch, but passes with it.

> TestSaveNamespace fails
> ---
>
> Key: HDFS-1503
> URL: https://issues.apache.org/jira/browse/HDFS-1503
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Reporter: Eli Collins
>Assignee: Todd Lipcon
>Priority: Minor
> Attachments: hdfs-1503.txt, 
> TEST-org.apache.hadoop.hdfs.server.namenode.TestSaveNamespace.txt
>
>
> Will attach the full log. Here's the relevant snippet:
> {noformat}
> Exception in thread "FSImageSaver for 
> /home/eli/src/hdfs4/build/test/data/dfs/na
> me1 of type IMAGE_AND_EDITS" java.lang.RuntimeException: Injected fault: 
> saveFSI
> mage second time
> 
> at java.lang.Thread.run(Thread.java:619)
> Exception in thread "FSImageSaver for 
> /home/eli/src/hdfs4/build/test/data/dfs/na
> me2 of type IMAGE_AND_EDITS" java.lang.StackOverflowError
> at java.util.Arrays.copyOf(Arrays.java:2882)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1503) TestSaveNamespace fails

2010-11-24 Thread Konstantin Shvachko (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1503?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935516#action_12935516
 ] 

Konstantin Shvachko commented on HDFS-1503:
---

Thanks for clarifications, Todd. I was not sure if the new code broke the test 
itself or the feature it is testing. Now I understand it the test problem.

> TestSaveNamespace fails
> ---
>
> Key: HDFS-1503
> URL: https://issues.apache.org/jira/browse/HDFS-1503
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>Reporter: Eli Collins
>Assignee: Todd Lipcon
>Priority: Minor
> Attachments: hdfs-1503.txt, 
> TEST-org.apache.hadoop.hdfs.server.namenode.TestSaveNamespace.txt
>
>
> Will attach the full log. Here's the relevant snippet:
> {noformat}
> Exception in thread "FSImageSaver for 
> /home/eli/src/hdfs4/build/test/data/dfs/na
> me1 of type IMAGE_AND_EDITS" java.lang.RuntimeException: Injected fault: 
> saveFSI
> mage second time
> 
> at java.lang.Thread.run(Thread.java:619)
> Exception in thread "FSImageSaver for 
> /home/eli/src/hdfs4/build/test/data/dfs/na
> me2 of type IMAGE_AND_EDITS" java.lang.StackOverflowError
> at java.util.Arrays.copyOf(Arrays.java:2882)
> {noformat}

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1391) Exiting safemode takes a long time when there are lots of blocks in the HDFS

2010-11-24 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1391?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935508#action_12935508
 ] 

Hadoop QA commented on HDFS-1391:
-

-1 overall.  Here are the results of testing the latest attachment 
  
http://issues.apache.org/jira/secure/attachment/12454967/excessReplicas.1_trunk.txt
  against trunk revision 1038370.

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 3 new or modified tests.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

-1 release audit.  The applied patch generated 101 release audit warnings 
(more than the trunk's current 98 warnings).

-1 core tests.  The patch failed these core unit tests:
  org.apache.hadoop.hdfs.server.balancer.TestBalancer
  org.apache.hadoop.hdfs.server.datanode.TestBlockRecovery
  org.apache.hadoop.hdfs.server.namenode.TestBlockTokenWithDFS
  
org.apache.hadoop.hdfs.server.namenode.TestLargeDirectoryDelete
  org.apache.hadoop.hdfs.server.namenode.TestSaveNamespace
  org.apache.hadoop.hdfs.server.namenode.TestStorageRestore
  org.apache.hadoop.hdfs.TestFileConcurrentReader
  org.apache.hadoop.hdfs.TestLargeBlock

-1 contrib tests.  The patch failed contrib unit tests.

+1 system test framework.  The patch passed system test framework compile.

Test results: 
https://hudson.apache.org/hudson/job/PreCommit-HDFS-Build/5//testReport/
Release audit warnings: 
https://hudson.apache.org/hudson/job/PreCommit-HDFS-Build/5//artifact/trunk/patchprocess/patchReleaseAuditProblems.txt
Findbugs warnings: 
https://hudson.apache.org/hudson/job/PreCommit-HDFS-Build/5//artifact/trunk/build/test/findbugs/newPatchFindbugsWarnings.html
Console output: 
https://hudson.apache.org/hudson/job/PreCommit-HDFS-Build/5//console

This message is automatically generated.

> Exiting safemode takes a long time when there are lots of blocks in the HDFS
> 
>
> Key: HDFS-1391
> URL: https://issues.apache.org/jira/browse/HDFS-1391
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: name-node
>Reporter: dhruba borthakur
>Assignee: dhruba borthakur
> Attachments: excessReplicas.1_trunk.txt
>
>
> When the namenode decides to exit safemode,  it acquires the FSNamesystem 
> lock and then iterates over all blocks in the blocksmap to determine if any 
> block has any excess replicas. This call takes upwards of 5 minutes on a 
> cluster that has 100 million blocks. This delays namenode restart to a good 
> extent.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-900) Corrupt replicas are not tracked correctly through block report from DN

2010-11-24 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935471#action_12935471
 ] 

Todd Lipcon commented on HDFS-900:
--

Yes, I believe this reproduces in 0.21 and 0.22 (the to-reproduce.patch was 
against trunk at the time though likely is out of date now)

> Corrupt replicas are not tracked correctly through block report from DN
> ---
>
> Key: HDFS-900
> URL: https://issues.apache.org/jira/browse/HDFS-900
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.22.0
>Reporter: Todd Lipcon
>Priority: Critical
> Attachments: log-commented, to-reproduce.patch
>
>
> This one is tough to describe, but essentially the following order of events 
> is seen to occur:
> # A client marks one replica of a block to be corrupt by telling the NN about 
> it
> # Replication is then scheduled to make a new replica of this node
> # The replication completes, such that there are now 3 good replicas and 1 
> corrupt replica
> # The DN holding the corrupt replica sends a block report. Rather than 
> telling this DN to delete the node, the NN instead marks this as a new *good* 
> replica of the block, and schedules deletion on one of the good replicas.
> I don't know if this is a dataloss bug in the case of 1 corrupt replica with 
> dfs.replication=2, but it seems feasible. I will attach a debug log with some 
> commentary marked by '>', plus a unit test patch which I can get 
> to reproduce this behavior reliably. (it's not a proper unit test, just some 
> edits to an existing one to show it)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-900) Corrupt replicas are not tracked correctly through block report from DN

2010-11-24 Thread Thanh Do (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935390#action_12935390
 ] 

Thanh Do commented on HDFS-900:
---

Todd, do you see this on 0.21.0?
Is this a bug of handling corrupt replicas at NN?

> Corrupt replicas are not tracked correctly through block report from DN
> ---
>
> Key: HDFS-900
> URL: https://issues.apache.org/jira/browse/HDFS-900
> Project: Hadoop HDFS
>  Issue Type: Bug
>Affects Versions: 0.22.0
>Reporter: Todd Lipcon
>Priority: Critical
> Attachments: log-commented, to-reproduce.patch
>
>
> This one is tough to describe, but essentially the following order of events 
> is seen to occur:
> # A client marks one replica of a block to be corrupt by telling the NN about 
> it
> # Replication is then scheduled to make a new replica of this node
> # The replication completes, such that there are now 3 good replicas and 1 
> corrupt replica
> # The DN holding the corrupt replica sends a block report. Rather than 
> telling this DN to delete the node, the NN instead marks this as a new *good* 
> replica of the block, and schedules deletion on one of the good replicas.
> I don't know if this is a dataloss bug in the case of 1 corrupt replica with 
> dfs.replication=2, but it seems feasible. I will attach a debug log with some 
> commentary marked by '>', plus a unit test patch which I can get 
> to reproduce this behavior reliably. (it's not a proper unit test, just some 
> edits to an existing one to show it)

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.



[jira] Commented: (HDFS-1458) Improve checkpoint performance by avoiding unnecessary image downloads

2010-11-24 Thread Hairong Kuang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-1458?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12935242#action_12935242
 ] 

Hairong Kuang commented on HDFS-1458:
-

 [exec] +1 overall.  
 [exec] 
 [exec] +1 @author.  The patch does not contain any @author tags.
 [exec] 
 [exec] +1 tests included.  The patch appears to i
 [exec] nclude 3 new or modified tests.
 [exec] 
 [exec] +1 javadoc.  The javadoc tool did not generate any warning 
messages.
 [exec] 
 [exec] +1 javac.  The applied patch does not increase the total number 
of javac compiler warnings.
 [exec] 
 [exec] +1 findbugs.  The patch does not introduce any new Findbugs 
(version 1.3.9) warnings.
 [exec] 
 [exec] +1 release audit.  The applied patch does not increase the 
total number of release audit warnings.
 [exec] 
 [exec] +1 system test framework.  The patch passed system test 
framework compile.
 [exec] 


> Improve checkpoint performance by avoiding unnecessary image downloads
> --
>
> Key: HDFS-1458
> URL: https://issues.apache.org/jira/browse/HDFS-1458
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: name-node
>Affects Versions: 0.22.0
>Reporter: Hairong Kuang
>Assignee: Hairong Kuang
> Fix For: 0.22.0
>
> Attachments: checkpoint-checkfsimageissame.patch, 
> trunkNoDownloadImage.patch, trunkNoDownloadImage1.patch, 
> trunkNoDownloadImage2.patch, trunkNoDownloadImage3.patch
>
>
> If secondary namenode could verify that the image it has on its disk is the 
> same as the one in the primary NameNode, it could skip downloading the image 
> from the primary NN, thus completely eliminating the image download overhead.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.