[jira] [Commented] (HDFS-7116) Add a metric to expose the bandwidth of balancer

2015-08-23 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708464#comment-14708464
 ] 

Allen Wittenauer commented on HDFS-7116:


Could we please stop adding camel case options to command lines?

 Add a metric to expose the bandwidth of balancer
 

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA
Assignee: Rakesh R
 Attachments: HDFS-7116-00.patch, HDFS-7116-01.patch, 
 HDFS-7116-02.patch, HDFS-7116-03.patch, HDFS-7116-04.patch, HDFS-7116-05.patch


 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the value of the same. 
 This jira to discuss  implement the way to access the balancer bandwidth 
 value of the datanode.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a metric to expose the bandwidth of balancer

2015-08-23 Thread Tsz Wo Nicholas Sze (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708546#comment-14708546
 ] 

Tsz Wo Nicholas Sze commented on HDFS-7116:
---

Command options are case insensitive.

 Add a metric to expose the bandwidth of balancer
 

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA
Assignee: Rakesh R
 Attachments: HDFS-7116-00.patch, HDFS-7116-01.patch, 
 HDFS-7116-02.patch, HDFS-7116-03.patch, HDFS-7116-04.patch, HDFS-7116-05.patch


 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the value of the same. 
 This jira to discuss  implement the way to access the balancer bandwidth 
 value of the datanode.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8931) Erasure Coding: Notify exception to client side from ParityGenerator.

2015-08-23 Thread Kai Sasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8931?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Sasaki updated HDFS-8931:
-
Summary: Erasure Coding: Notify exception to client side from 
ParityGenerator.  (was: Notify exception to client side from ParityGenerator.)

 Erasure Coding: Notify exception to client side from ParityGenerator.
 -

 Key: HDFS-8931
 URL: https://issues.apache.org/jira/browse/HDFS-8931
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: HDFS
Affects Versions: HDFS-7285
Reporter: Kai Sasaki
Assignee: Kai Sasaki
  Labels: EC
 Fix For: HDFS-7285


 Following HDFS-8287. 
 Current client thread catch up the exception from {{ParityGenerator}}. In 
 order to handle properly, 
 1. Put together handling logic into UncaughtExceptionHandler.
 2. Notify exception to client side from UncaughtExceptionHandler.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-7285) Erasure Coding Support inside HDFS

2015-08-23 Thread GAO Rui (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-7285?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

GAO Rui updated HDFS-7285:
--
Attachment: HDFSErasureCodingSystemTestPlan-20150824.pdf

Inspired  by [^HDFSErasureCodingPhaseITestPlan.pdf] created by [~zhz].  We has 
implemented system tests(Functional Tests) according to this test plan 
[^HDFSErasureCodingSystemTestPlan-20150824.pdf]. Latest test report will be 
posted separately soon.Please feel free to discuss and add more test cases 
or scenarios to this system test plan.

 Erasure Coding Support inside HDFS
 --

 Key: HDFS-7285
 URL: https://issues.apache.org/jira/browse/HDFS-7285
 Project: Hadoop HDFS
  Issue Type: New Feature
Reporter: Weihua Jiang
Assignee: Zhe Zhang
 Attachments: Consolidated-20150707.patch, 
 Consolidated-20150806.patch, Consolidated-20150810.patch, ECAnalyzer.py, 
 ECParser.py, HDFS-7285-initial-PoC.patch, 
 HDFS-7285-merge-consolidated-01.patch, 
 HDFS-7285-merge-consolidated-trunk-01.patch, 
 HDFS-7285-merge-consolidated.trunk.03.patch, 
 HDFS-7285-merge-consolidated.trunk.04.patch, 
 HDFS-EC-Merge-PoC-20150624.patch, HDFS-EC-merge-consolidated-01.patch, 
 HDFS-bistriped.patch, HDFSErasureCodingDesign-20141028.pdf, 
 HDFSErasureCodingDesign-20141217.pdf, HDFSErasureCodingDesign-20150204.pdf, 
 HDFSErasureCodingDesign-20150206.pdf, HDFSErasureCodingPhaseITestPlan.pdf, 
 HDFSErasureCodingSystemTestPlan-20150824.pdf, fsimage-analysis-20150105.pdf


 Erasure Coding (EC) can greatly reduce the storage overhead without sacrifice 
 of data reliability, comparing to the existing HDFS 3-replica approach. For 
 example, if we use a 10+4 Reed Solomon coding, we can allow loss of 4 blocks, 
 with storage overhead only being 40%. This makes EC a quite attractive 
 alternative for big data storage, particularly for cold data. 
 Facebook had a related open source project called HDFS-RAID. It used to be 
 one of the contribute packages in HDFS but had been removed since Hadoop 2.0 
 for maintain reason. The drawbacks are: 1) it is on top of HDFS and depends 
 on MapReduce to do encoding and decoding tasks; 2) it can only be used for 
 cold files that are intended not to be appended anymore; 3) the pure Java EC 
 coding implementation is extremely slow in practical use. Due to these, it 
 might not be a good idea to just bring HDFS-RAID back.
 We (Intel and Cloudera) are working on a design to build EC into HDFS that 
 gets rid of any external dependencies, makes it self-contained and 
 independently maintained. This design lays the EC feature on the storage type 
 support and considers compatible with existing HDFS features like caching, 
 snapshot, encryption, high availability and etc. This design will also 
 support different EC coding schemes, implementations and policies for 
 different deployment scenarios. By utilizing advanced libraries (e.g. Intel 
 ISA-L library), an implementation can greatly improve the performance of EC 
 encoding/decoding and makes the EC solution even more attractive. We will 
 post the design document soon. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8763) Incremental blockreport order may replicate unnecessary block

2015-08-23 Thread Walter Su (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8763?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708726#comment-14708726
 ] 

Walter Su commented on HDFS-8763:
-

The description gives good logs but not a good reason. It's a race condition 
when you close the small file too early. After file closed, a race between IBR 
of 3rd replica and replication scheduler.

You have closed the file before x.x.7.73 node reports the 3rd finalized replica 
of lastBlock. So the lastBlock is scheduled to replicate. After x.x.7.73 
reported the 3rd replica. The lastBlock should be removed from 
{{needReplications}}. But if the replication command is sent to a new 
DN(x.x.4.65 node). There is no way to cancel the command from x.x.4.65 node.

The solution is to postpone scheduling replication, and wait x.x.7.73 to report 
the 3rd replica.

 Incremental blockreport order may replicate unnecessary block
 -

 Key: HDFS-8763
 URL: https://issues.apache.org/jira/browse/HDFS-8763
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Affects Versions: 2.4.0
Reporter: jiangyu
Priority: Minor

 For our cluster, the NameNode is always very busy, so for every incremental 
 block report , the contention of lock is heavy.
 The logic of incremental block report is as follow, client send block to dn1 
 and dn1 mirror to dn2, dn2 mirror to dn3. After finish this block, all 
 datanode will report the newly received block to namenode. In NameNode side, 
 all will go to the method processIncrementalBlockReport in BlockManager 
 class. But the status of the block reported from dn2,dn3 is RECEIVING_BLOCK, 
 for dn1 is RECEIED_BLOCK. It is okay if dn2, dn3 report before dn1(that is 
 common), but in some busy environment, it is easy to find dn1 report before 
 dn2 or dn3, let’s assume dn2 report first, dn1 report second, and dn3 report 
 third.
 So dn1 will addStoredBlock and find the replica of this block is not reach 
 the the original number(which is 3), and the block will add to 
 neededReplications construction and soon ask some node in pipeline (dn1 or 
 dn2)to replica it dn4 . After sometime, dn4 and dn3 all report this block, 
 then choose one node to invalidate.
 Here is one log i found in our cluster:
 2015-07-08 01:05:34,675 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
 allocateBlock: 
 /logs/***_bigdata_spam/logs/application_1435099124107_470749/xx.xx.4.62_45454.tmp.
  BP-1386326728-xx.xx.2.131-1382089338395 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-a7c0f8f6-2399-4980-9479-efa08487b7b3:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-c75145a0-ed63-4180-87ee-d48ccaa647c5:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW]]}
 2015-07-08 01:05:34,689 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.7.75:50010 is added to 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-74ed264f-da43-4cc3-9fa9-164ba99f752a:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-56121ce1-8991-45b3-95bc-2a5357991512:NORMAL|RBW]]}
  size 0
 2015-07-08 01:05:34,689 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.4.62:50010 is added to 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-74ed264f-da43-4cc3-9fa9-164ba99f752a:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-56121ce1-8991-45b3-95bc-2a5357991512:NORMAL|RBW]]}
  size 0
 2015-07-08 01:05:35,003 INFO BlockStateChange: BLOCK* ask xx.xx.4.62:50010 to 
 replicate blk_3194502674_2121080184 to datanode(s) xx.xx.4.65:50010
 2015-07-08 01:05:35,403 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.7.73:50010 is added to blk_3194502674_2121080184 size 
 67750
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.4.65:50010 is added to blk_3194502674_2121080184 size 
 67750
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* InvalidateBlocks: add 
 blk_3194502674_2121080184 to xx.xx.7.75:50010
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* chooseExcessReplicates: 
 (xx.xx.7.75:50010, blk_3194502674_2121080184) is added to invalidated blocks 
 set
 2015-07-08 01:05:35,852 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
 InvalidateBlocks: ask xx.xx.7.75:50010 to delete [blk_3194502674_2121080184, 
 blk_3194497594_2121075104]
 Some day, the number of this situation can be 40, that is not good for 

[jira] [Commented] (HDFS-8388) Time and Date format need to be in sync in Namenode UI page

2015-08-23 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8388?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708784#comment-14708784
 ] 

Vinayakumar B commented on HDFS-8388:
-

1. {code} var HELPERS = {
   'helper_date_tostring' : function (chunk, ctx, bodies, params) {
 var value = dust.helpers.tap(params.value, chunk, ctx);
-return chunk.write('' + new Date(Number(value)).toLocaleString());
+return chunk.write('' + moment(Number(v)).format('ddd MMM DD HH:mm:ss 
ZZ '));
   }
 };{code}
Here, {{Number(v)}}, there is no {{v}}. You need to replace with {{value}}.

2. Need to fix 2 checkstyles out of 3 shown
3. You can clean lines with whitespace endings, though this can be done during 
commit.


In addition to the above changes, just remove duplicate line in 
{{dfshealth.html}}. Its not from this issue, but i think that also can go along 
with this patch.
{{/scriptscript type=text/javascript src=/static/moment.min.js}}

 Time and Date format need to be in sync in Namenode UI page
 ---

 Key: HDFS-8388
 URL: https://issues.apache.org/jira/browse/HDFS-8388
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Archana T
Assignee: Surendra Singh Lilhore
Priority: Minor
 Attachments: HDFS-8388-002.patch, HDFS-8388-003.patch, 
 HDFS-8388-004.patch, HDFS-8388-005.patch, HDFS-8388.patch, HDFS-8388_1.patch, 
 ScreenShot-InvalidDate.png


 In NameNode UI Page, Date and Time FORMAT  displayed on the page are not in 
 sync currently.
 Started:Wed May 13 12:28:02 IST 2015
 Compiled:23 Apr 2015 12:22:59 
 Block Deletion Start Time   13 May 2015 12:28:02
 We can keep a common format in all the above places.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8943) Read apis in ByteRangeInputStream does not read all the bytes specified when chunked transfer-encoding is used in the server

2015-08-23 Thread Shradha Revankar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shradha Revankar updated HDFS-8943:
---
Description: 
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(byte b[], int off, int len)}}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}

  was:
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(long position, byte[] buffer, int offset, int length)}}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}


 Read apis in ByteRangeInputStream does not read all the bytes specified when 
 chunked transfer-encoding is used in the server
 

 Key: HDFS-8943
 URL: https://issues.apache.org/jira/browse/HDFS-8943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar

 With the default Webhdfs server implementation the read apis in 
 ByteRangeInputStream work as expected reading the correct number of bytes for 
 these apis :
 {{public int read(byte b[], int off, int len)}}
 {{public int read(long position, byte[] buffer, int offset, int length)}}
 But when a custom Webhdfs server implementation is plugged in which uses 
 chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
 would be to loop and read till bytes specified similar to {{readfully()}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8943) Read apis in ByteRangeInputStream does not read all the bytes specified when chunked transfer-encoding is used in the server

2015-08-23 Thread Shradha Revankar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shradha Revankar updated HDFS-8943:
---
Description: 
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(long position, byte[] buffer, int offset, int length) }}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}

  was:
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{ public int read(long position, byte[] buffer, int offset, int length)
  throws IOException }}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}


 Read apis in ByteRangeInputStream does not read all the bytes specified when 
 chunked transfer-encoding is used in the server
 

 Key: HDFS-8943
 URL: https://issues.apache.org/jira/browse/HDFS-8943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar

 With the default Webhdfs server implementation the read apis in 
 ByteRangeInputStream work as expected reading the correct number of bytes for 
 these apis :
 {{public int read(long position, byte[] buffer, int offset, int length) }}
 {{public int read(long position, byte[] buffer, int offset, int length)}}
 But when a custom Webhdfs server implementation is plugged in which uses 
 chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
 would be to loop and read till bytes specified similar to {{readfully()}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8943) Read apis in ByteRangeInputStream does not read all the bytes specified when chunked transfer-encoding is used in the server

2015-08-23 Thread Shradha Revankar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shradha Revankar updated HDFS-8943:
---
Description: 
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(long position, byte[] buffer, int offset, int length)}}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}

  was:
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(long position, byte[] buffer, int offset, int length) }}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}


 Read apis in ByteRangeInputStream does not read all the bytes specified when 
 chunked transfer-encoding is used in the server
 

 Key: HDFS-8943
 URL: https://issues.apache.org/jira/browse/HDFS-8943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar

 With the default Webhdfs server implementation the read apis in 
 ByteRangeInputStream work as expected reading the correct number of bytes for 
 these apis :
 {{public int read(long position, byte[] buffer, int offset, int length)}}
 {{public int read(long position, byte[] buffer, int offset, int length)}}
 But when a custom Webhdfs server implementation is plugged in which uses 
 chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
 would be to loop and read till bytes specified similar to {{readfully()}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-8942:

   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

I've committed this to trunk and branch-2. Thanks [~iwasakims] for the 
contribution.

 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Fix For: 2.8.0

 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Rakesh R (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708788#comment-14708788
 ] 

Rakesh R commented on HDFS-8944:


Thanks [~ajisakaa] for the quick updates. With this change we are removing the 
restriction of case sensitiveness for the command options. IMHO, its good to 
discuss about backward compatibility part and update {{Release Note:}} field in 
the jira.

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
Priority: Minor
 Attachments: HDFS-8944.001.patch


 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-23 Thread Yi Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Yi Liu updated HDFS-8900:
-
Attachment: HDFS-8900.003.patch

 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch, 
 HDFS-8900.003.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8935) Erasure Coding: createErasureCodingZone api should accept the policyname as argument instead of ErasureCodingPolicy

2015-08-23 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8935?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HDFS-8935:

Parent Issue: HDFS-8031  (was: HDFS-7285)

 Erasure Coding: createErasureCodingZone api should accept the policyname as 
 argument instead of ErasureCodingPolicy
 ---

 Key: HDFS-8935
 URL: https://issues.apache.org/jira/browse/HDFS-8935
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: J.Andreina
Assignee: J.Andreina

 Current behavior : User has to specify ErasureCodingPolicy as an argument for 
 createErasureCodingZone api .
 This can be made in sync with creation of EC zone through CLI , where user 
 need to specify only the policy name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8935) Erasure Coding: createErasureCodingZone api should accept the policyname as argument instead of ErasureCodingPolicy

2015-08-23 Thread Zhe Zhang (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8935?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708822#comment-14708822
 ] 

Zhe Zhang commented on HDFS-8935:
-

I agree with Vinay. Moving it under HDFS-8031 since multi-policy support 
belongs to follow-on work.

 Erasure Coding: createErasureCodingZone api should accept the policyname as 
 argument instead of ErasureCodingPolicy
 ---

 Key: HDFS-8935
 URL: https://issues.apache.org/jira/browse/HDFS-8935
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: J.Andreina
Assignee: J.Andreina

 Current behavior : User has to specify ErasureCodingPolicy as an argument for 
 createErasureCodingZone api .
 This can be made in sync with creation of EC zone through CLI , where user 
 need to specify only the policy name.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8287) DFSStripedOutputStream.writeChunk should not wait for writing parity

2015-08-23 Thread Kai Sasaki (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8287?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708691#comment-14708691
 ] 

Kai Sasaki commented on HDFS-8287:
--

[~rakeshr] Thank you for reviewing! Can other someone check this JIRA's patch?

And I created follow on work of this (HDFS-8931). 

 DFSStripedOutputStream.writeChunk should not wait for writing parity 
 -

 Key: HDFS-8287
 URL: https://issues.apache.org/jira/browse/HDFS-8287
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client
Reporter: Tsz Wo Nicholas Sze
Assignee: Kai Sasaki
 Attachments: HDFS-8287-HDFS-7285.00.patch, 
 HDFS-8287-HDFS-7285.01.patch, HDFS-8287-HDFS-7285.02.patch, 
 HDFS-8287-HDFS-7285.03.patch, HDFS-8287-HDFS-7285.04.patch, 
 HDFS-8287-HDFS-7285.05.patch


 When a stripping cell is full, writeChunk computes and generates parity 
 packets.  It sequentially calls waitAndQueuePacket so that user client cannot 
 continue to write data until it finishes.
 We should allow user client to continue writing instead but not blocking it 
 when writing parity.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a metric to expose the bandwidth of balancer

2015-08-23 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708696#comment-14708696
 ] 

Akira AJISAKA commented on HDFS-7116:
-

bq. This seems wrong to show it in metric since it is a mostly constant value. 
Just like that we won't show other conf values in metric.
I understand. Thanks [~szetszwo] for the comment.

bq. Now, your approach is avoiding this problem. I'll start preparing a patch 
if others agree to this approach.
Agree.

bq. Command options are case insensitive.
{code:title=DFSAdmin.java}
try {
  if (-report.equals(cmd)) {
report(argv, i);
  } else if (-safemode.equals(cmd)) {
setSafeMode(argv, i);
  } else if (-allowSnapshot.equalsIgnoreCase(cmd)) {
allowSnapshot(argv);
...
{code}
dfsadmin command options are case sensitive except allowSnapshot and 
disallowSnapshot. I'm thinking we should make them case insensitive in a 
separate jira.

 Add a metric to expose the bandwidth of balancer
 

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA
Assignee: Rakesh R
 Attachments: HDFS-7116-00.patch, HDFS-7116-01.patch, 
 HDFS-7116-02.patch, HDFS-7116-03.patch, HDFS-7116-04.patch, HDFS-7116-05.patch


 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the value of the same. 
 This jira to discuss  implement the way to access the balancer bandwidth 
 value of the datanode.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8943) Read apis in ByteRangeInputStream does not read all the bytes specified when chunked transfer-encoding is used in the server

2015-08-23 Thread Shradha Revankar (JIRA)
Shradha Revankar created HDFS-8943:
--

 Summary: Read apis in ByteRangeInputStream does not read all the 
bytes specified when chunked transfer-encoding is used in the server
 Key: HDFS-8943
 URL: https://issues.apache.org/jira/browse/HDFS-8943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar


With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(long position, byte[] buffer, int offset, int length)
  throws IOException }}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8943) Read apis in ByteRangeInputStream does not read all the bytes specified when chunked transfer-encoding is used in the server

2015-08-23 Thread Shradha Revankar (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8943?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shradha Revankar updated HDFS-8943:
---
Description: 
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{ public int read(long position, byte[] buffer, int offset, int length)
  throws IOException }}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}

  was:
With the default Webhdfs server implementation the read apis in 
ByteRangeInputStream work as expected reading the correct number of bytes for 
these apis :

{{public int read(long position, byte[] buffer, int offset, int length)
  throws IOException }}

{{public int read(long position, byte[] buffer, int offset, int length)}}

But when a custom Webhdfs server implementation is plugged in which uses 
chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
would be to loop and read till bytes specified similar to {{readfully()}}


 Read apis in ByteRangeInputStream does not read all the bytes specified when 
 chunked transfer-encoding is used in the server
 

 Key: HDFS-8943
 URL: https://issues.apache.org/jira/browse/HDFS-8943
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: webhdfs
Affects Versions: 2.7.1
Reporter: Shradha Revankar
Assignee: Shradha Revankar

 With the default Webhdfs server implementation the read apis in 
 ByteRangeInputStream work as expected reading the correct number of bytes for 
 these apis :
 {{ public int read(long position, byte[] buffer, int offset, int length)
   throws IOException }}
 {{public int read(long position, byte[] buffer, int offset, int length)}}
 But when a custom Webhdfs server implementation is plugged in which uses 
 chunked Transfer-encoding, these apis read only the first chunk. Simple fix 
 would be to loop and read till bytes specified similar to {{readfully()}}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-8944:

Attachment: HDFS-8944.001.patch

Attaching a patch that replaces String.equals with String.equalsIgnorecase.

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
Priority: Minor
 Attachments: HDFS-8944.001.patch


 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-8944:

Target Version/s: 2.8.0
  Status: Patch Available  (was: Open)

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
Priority: Minor
 Attachments: HDFS-8944.001.patch


 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-7116) Add a metric to expose the bandwidth of balancer

2015-08-23 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-7116?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708762#comment-14708762
 ] 

Akira AJISAKA commented on HDFS-7116:
-

Filed HDFS-8944 for make them case insensitive.

 Add a metric to expose the bandwidth of balancer
 

 Key: HDFS-7116
 URL: https://issues.apache.org/jira/browse/HDFS-7116
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: balancer  mover
Reporter: Akira AJISAKA
Assignee: Rakesh R
 Attachments: HDFS-7116-00.patch, HDFS-7116-01.patch, 
 HDFS-7116-02.patch, HDFS-7116-03.patch, HDFS-7116-04.patch, HDFS-7116-05.patch


 Now reading logs is the only way to check how the balancer bandwidth is set. 
 It would be useful for administrators if they can get the value of the same. 
 This jira to discuss  implement the way to access the balancer bandwidth 
 value of the datanode.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708779#comment-14708779
 ] 

Hudson commented on HDFS-8942:
--

FAILURE: Integrated in Hadoop-trunk-Commit #8342 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/8342/])
HDFS-8942. Update hyperlink to rack awareness page in HDFS Architecture 
documentation. Contributed by Masatake Iwasaki. (aajisaka: rev 
bcaf83902aa4d1e3e2cd26442df0a253eae7f633)
* hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsDesign.md
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Fix For: 2.8.0

 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-8944:

Priority: Major  (was: Minor)
Release Note: Now hdfs dfsadmin command options are case insensitive. For 
example, both hdfs dfsadmin -saveNamespace and hdfs dfsadmin -savenamespace 
are accepted.

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
 Attachments: HDFS-8944.001.patch


 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708803#comment-14708803
 ] 

Hudson commented on HDFS-8942:
--

FAILURE: Integrated in Hadoop-Mapreduce-trunk-Java8 #298 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Java8/298/])
HDFS-8942. Update hyperlink to rack awareness page in HDFS Architecture 
documentation. Contributed by Masatake Iwasaki. (aajisaka: rev 
bcaf83902aa4d1e3e2cd26442df0a253eae7f633)
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt
* hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsDesign.md


 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Fix For: 2.8.0

 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-8944:

Release Note: Now hdfs dfsadmin command options are case insensitive. For 
example, both hdfs dfsadmin -saveNamespace and hdfs dfsadmin -savenamespace 
are valid.  (was: Now hdfs dfsadmin command options are case insensitive. For 
example, both hdfs dfsadmin -saveNamespace and hdfs dfsadmin -savenamespace 
are accepted.)

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
 Attachments: HDFS-8944.001.patch


 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708801#comment-14708801
 ] 

Akira AJISAKA commented on HDFS-8944:
-

Thanks [~rakeshr] for the comment. Updated the release note. I don't think this 
issue breaks backward compatibility because camel case is supported after the 
patch, however, this issue and HADOOP-10901 (Thanks Kengo for the link!) should 
be discussed. Obviously I'm +1 for make them case insensitive for usability.

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
 Attachments: HDFS-8944.001.patch


 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708817#comment-14708817
 ] 

Hudson commented on HDFS-8942:
--

FAILURE: Integrated in Hadoop-Yarn-trunk-Java8 #302 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk-Java8/302/])
HDFS-8942. Update hyperlink to rack awareness page in HDFS Architecture 
documentation. Contributed by Masatake Iwasaki. (aajisaka: rev 
bcaf83902aa4d1e3e2cd26442df0a253eae7f633)
* hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsDesign.md
* hadoop-hdfs-project/hadoop-hdfs/CHANGES.txt


 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Fix For: 2.8.0

 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8900) Compact XAttrs to optimize memory footprint.

2015-08-23 Thread Yi Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8900?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708819#comment-14708819
 ] 

Yi Liu commented on HDFS-8900:
--

Thanks [~andrew.wang] and [~cmccabe]!
I will create a follow-on JIRA targeted on trunk to remove the max size config.

Failure of {{TestStartup.testXattrConfiguration}} is related, the new patch 
fixes it, and also fixes the checkstyle and whitespace. The findbugs actually 
don't exist. 

 Compact XAttrs to optimize memory footprint.
 

 Key: HDFS-8900
 URL: https://issues.apache.org/jira/browse/HDFS-8900
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Yi Liu
Assignee: Yi Liu
 Attachments: HDFS-8900.001.patch, HDFS-8900.002.patch


 {code}
 private final ImmutableListXAttr xAttrs;
 {code}
 Currently we use above in XAttrFeature, it's not efficient from memory point 
 of view, since {{ImmutableList}} and {{XAttr}} have object memory overhead, 
 and each object has memory alignment. 
 We can use a {{byte[]}} in XAttrFeature and do some compact in {{XAttr}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Masatake Iwasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HDFS-8942:
---
Status: Patch Available  (was: Open)

 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8763) Incremental blockreport order may replicate unnecessary block

2015-08-23 Thread Walter Su (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8763?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708732#comment-14708732
 ] 

Walter Su commented on HDFS-8763:
-

Another solution is to increase {{dfs.namenode.replication.interval}} to slow 
down checker/scheduler.

 Incremental blockreport order may replicate unnecessary block
 -

 Key: HDFS-8763
 URL: https://issues.apache.org/jira/browse/HDFS-8763
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: HDFS
Affects Versions: 2.4.0
Reporter: jiangyu
Priority: Minor

 For our cluster, the NameNode is always very busy, so for every incremental 
 block report , the contention of lock is heavy.
 The logic of incremental block report is as follow, client send block to dn1 
 and dn1 mirror to dn2, dn2 mirror to dn3. After finish this block, all 
 datanode will report the newly received block to namenode. In NameNode side, 
 all will go to the method processIncrementalBlockReport in BlockManager 
 class. But the status of the block reported from dn2,dn3 is RECEIVING_BLOCK, 
 for dn1 is RECEIED_BLOCK. It is okay if dn2, dn3 report before dn1(that is 
 common), but in some busy environment, it is easy to find dn1 report before 
 dn2 or dn3, let’s assume dn2 report first, dn1 report second, and dn3 report 
 third.
 So dn1 will addStoredBlock and find the replica of this block is not reach 
 the the original number(which is 3), and the block will add to 
 neededReplications construction and soon ask some node in pipeline (dn1 or 
 dn2)to replica it dn4 . After sometime, dn4 and dn3 all report this block, 
 then choose one node to invalidate.
 Here is one log i found in our cluster:
 2015-07-08 01:05:34,675 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
 allocateBlock: 
 /logs/***_bigdata_spam/logs/application_1435099124107_470749/xx.xx.4.62_45454.tmp.
  BP-1386326728-xx.xx.2.131-1382089338395 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-a7c0f8f6-2399-4980-9479-efa08487b7b3:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-c75145a0-ed63-4180-87ee-d48ccaa647c5:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW]]}
 2015-07-08 01:05:34,689 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.7.75:50010 is added to 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-74ed264f-da43-4cc3-9fa9-164ba99f752a:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-56121ce1-8991-45b3-95bc-2a5357991512:NORMAL|RBW]]}
  size 0
 2015-07-08 01:05:34,689 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.4.62:50010 is added to 
 blk_3194502674_2121080184{blockUCState=UNDER_CONSTRUCTION, 
 primaryNodeIndex=-1, 
 replicas=[ReplicaUnderConstruction[[DISK]DS-15a4dc8e-5b7d-449f-a941-6dced45e6f07:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-74ed264f-da43-4cc3-9fa9-164ba99f752a:NORMAL|RBW],
  
 ReplicaUnderConstruction[[DISK]DS-56121ce1-8991-45b3-95bc-2a5357991512:NORMAL|RBW]]}
  size 0
 2015-07-08 01:05:35,003 INFO BlockStateChange: BLOCK* ask xx.xx.4.62:50010 to 
 replicate blk_3194502674_2121080184 to datanode(s) xx.xx.4.65:50010
 2015-07-08 01:05:35,403 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.7.73:50010 is added to blk_3194502674_2121080184 size 
 67750
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* addStoredBlock: 
 blockMap updated: xx.xx.4.65:50010 is added to blk_3194502674_2121080184 size 
 67750
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* InvalidateBlocks: add 
 blk_3194502674_2121080184 to xx.xx.7.75:50010
 2015-07-08 01:05:35,833 INFO BlockStateChange: BLOCK* chooseExcessReplicates: 
 (xx.xx.7.75:50010, blk_3194502674_2121080184) is added to invalidated blocks 
 set
 2015-07-08 01:05:35,852 INFO org.apache.hadoop.hdfs.StateChange: BLOCK* 
 InvalidateBlocks: ask xx.xx.7.75:50010 to delete [blk_3194502674_2121080184, 
 blk_3194497594_2121075104]
 Some day, the number of this situation can be 40, that is not good for 
 the performance and waste network band.
 Our base version is hadoop 2.4 and i have checked hadoop 2.7.1 didn’t find 
 any difference.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8944) Make dfsadmin command options case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8944?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HDFS-8944:

Assignee: Akira AJISAKA
 Summary: Make dfsadmin command options case insensitive  (was: Make 
dfsadmin command option case insensitive)

 Make dfsadmin command options case insensitive
 --

 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Assignee: Akira AJISAKA
Priority: Minor

 Now dfsadmin command options are case sensitive except allowSnapshot and  
 disallowSnapshot. It would be better to make them case insensitive for 
 usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8944) Make dfsadmin command option case insensitive

2015-08-23 Thread Akira AJISAKA (JIRA)
Akira AJISAKA created HDFS-8944:
---

 Summary: Make dfsadmin command option case insensitive
 Key: HDFS-8944
 URL: https://issues.apache.org/jira/browse/HDFS-8944
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Akira AJISAKA
Priority: Minor


Now dfsadmin command options are case sensitive except allowSnapshot and  
disallowSnapshot. It would be better to make them case insensitive for 
usability and consistency.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Masatake Iwasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HDFS-8942:
---
Attachment: HDFS-8942.001.patch

 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Masatake Iwasaki (JIRA)
Masatake Iwasaki created HDFS-8942:
--

 Summary: Update hyperlink to rack awareness page in HDFS 
Architecture documentation
 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial


The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HDFS-8942) Update hyperlink to rack awareness page in HDFS Architecture documentation

2015-08-23 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HDFS-8942?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14708775#comment-14708775
 ] 

Akira AJISAKA commented on HDFS-8942:
-

LGTM, +1.

 Update hyperlink to rack awareness page in HDFS Architecture documentation
 --

 Key: HDFS-8942
 URL: https://issues.apache.org/jira/browse/HDFS-8942
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Masatake Iwasaki
Assignee: Masatake Iwasaki
Priority: Trivial
 Attachments: HDFS-8942.001.patch


 The hyperlink to rack awareness documentation in HdfsDesign.md is pointing to 
 non-existent target. It should be fixed to point to RackAwareness.html.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)