Jenkins build is back to normal : Hadoop-Hdfs-trunk #1698

2014-03-11 Thread Apache Jenkins Server
See https://builds.apache.org/job/Hadoop-Hdfs-trunk/1698/changes



Re: Thinking ahead to 2.4

2014-03-11 Thread Arun C Murthy
Sorry, the previous link had a bug, the correct one is: 
http://s.apache.org/hadoop-2.4.0-blockers.

We are currently down to 12 blockers; with several PA.

thanks,
Arun

On Mar 6, 2014, at 1:40 PM, Arun C Murthy a...@hortonworks.com wrote:

 Gang,
 
  Most of the big-ticket items are already in, awesome!
 
  I'm thinking we could roll out a 2.4 RC in the next 2-3 weeks after we get 
 through the list of blockers. Here is a handy link: 
 http://s.apache.org/hadoop-2.4-blockers
 
  If you find more, please set Target Version to 2.4.0 and mark it a blocker. 
 I'll try nudging people to start closing these soon, appreciate any help!
 
 thanks,
 Arun
 
  
 On Feb 20, 2014, at 3:45 PM, Arun C Murthy a...@hortonworks.com wrote:
 
 Thanks Azuryy  Suresh. I've updated the roadmap wiki to reflect this.
 
 Arun
 
 On Feb 20, 2014, at 2:01 PM, Suresh Srinivas sur...@hortonworks.com wrote:
 
 Arun,
 
 Some of the previously 2.4 targeted features were made available in 2.3:
 - Heterogeneous storage support
 - Datanode cache
 
 The following are being targeted for 2.4:
 - Use protobuf for fsimge (already in)
 - ACLs (in trunk. In a week or so, this will be merged to branch-2.4)
 - Rolling upgrades (last bunch of jiras being worked in feature branch.
 Will be in 2.4 in around two weeks. Currently testing is in progress)
 
 So HDFS features should be ready in two weeks.
 
 
 On Sat, Feb 15, 2014 at 4:47 PM, Azuryy azury...@gmail.com wrote:
 
 Hi,
 I think you omit some key pieces in 2.4
 
 Protobuf fsimage, rolling upgrade are also targeting 2.4
 
 
 
 Sent from my iPhone5s
 
 On 2014年2月16日, at 6:59, Arun C Murthy a...@hortonworks.com wrote:
 
 Folks,
 
 With hadoop-2.3 nearly done, I think it's time to think ahead to
 hadoop-2.4. I think it was a good idea to expedite release of 2.3 while we
 finished up pieces that didn't make it in such as HDFS Caching  Support
 for Heterogenous Storage.
 
 Now, most of the key pieces incl. Resource Manager Automatic Failover
 (YARN-149), Application History Server (YARN-321)  Application Timeline
 Server (YARN-1530) are either complete or very close to done, and I think
 we will benefit with an extended test-cycle for 2.4 - similar to what
 happened with 2.2. To provide some context: 2.2 went through nearly 6 weeks
 of extended testing and it really helped us push out a very stable release.
 
 I think it will be good to create a 2.4 branch ASAP and start testing.
 As such, I plan to cut the branch early next week. With this, we should be
 good shape sometime to release 2.4 in mid-March.
 
 I've updated https://wiki.apache.org/hadoop/Roadmap to reflect this.
 
 Also, we should start thinking ahead to 2.5 and what folks would like to
 see in it. If we continue our 6-week cycles, we could shoot to get that out
 in April.
 
 Thoughts?
 
 thanks,
 Arun
 
 
 --
 Arun C. Murthy
 Hortonworks Inc.
 http://hortonworks.com/
 
 
 
 --
 CONFIDENTIALITY NOTICE
 NOTICE: This message is intended for the use of the individual or entity
 to
 which it is addressed and may contain information that is confidential,
 privileged and exempt from disclosure under applicable law. If the reader
 of this message is not the intended recipient, you are hereby notified
 that
 any printing, copying, dissemination, distribution, disclosure or
 forwarding of this communication is strictly prohibited. If you have
 received this communication in error, please contact the sender
 immediately
 and delete it from your system. Thank You.
 
 
 
 
 -- 
 http://hortonworks.com/download/
 
 -- 
 CONFIDENTIALITY NOTICE
 NOTICE: This message is intended for the use of the individual or entity to 
 which it is addressed and may contain information that is confidential, 
 privileged and exempt from disclosure under applicable law. If the reader 
 of this message is not the intended recipient, you are hereby notified that 
 any printing, copying, dissemination, distribution, disclosure or 
 forwarding of this communication is strictly prohibited. If you have 
 received this communication in error, please contact the sender immediately 
 and delete it from your system. Thank You.
 
 --
 Arun C. Murthy
 Hortonworks Inc.
 http://hortonworks.com/
 
 
 
 --
 Arun C. Murthy
 Hortonworks Inc.
 http://hortonworks.com/
 
 

--
Arun C. Murthy
Hortonworks Inc.
http://hortonworks.com/



-- 
CONFIDENTIALITY NOTICE
NOTICE: This message is intended for the use of the individual or entity to 
which it is addressed and may contain information that is confidential, 
privileged and exempt from disclosure under applicable law. If the reader 
of this message is not the intended recipient, you are hereby notified that 
any printing, copying, dissemination, distribution, disclosure or 
forwarding of this communication is strictly prohibited. If you have 
received this communication in error, please contact the sender immediately 
and delete it from your system. Thank You.


[jira] [Created] (HDFS-6087) Unify HDFS write/append/truncate

2014-03-11 Thread Guo Ruijing (JIRA)
Guo Ruijing created HDFS-6087:
-

 Summary: Unify HDFS write/append/truncate
 Key: HDFS-6087
 URL: https://issues.apache.org/jira/browse/HDFS-6087
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs-client
Reporter: Guo Ruijing
 Attachments: HDFS Design Proposal.pdf

In existing implementation, HDFS file can be appended and HDFS block can be 
reopened for append. This design will introduce complexity including lease 
recovery. If we design HDFS block as immutable, it will be very simple for 
append  truncate. The idea is that HDFS block is immutable if the block is 
committed to namenode. If the block is not committed to namenode, it is HDFS 
client’s responsibility to re-added with new block ID.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6088) Add configurable maximum block count for datanode

2014-03-11 Thread Kihwal Lee (JIRA)
Kihwal Lee created HDFS-6088:


 Summary: Add configurable maximum block count for datanode
 Key: HDFS-6088
 URL: https://issues.apache.org/jira/browse/HDFS-6088
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Kihwal Lee


Currently datanode resources are protected by the free space check and the 
balancer.  But datanodes can run out of memory simply storing too many blocks. 
If the sizes of blocks are small, datanodes will appear to have plenty of space 
to put more blocks.

I propose adding a configurable max block count to datanode. Since datanodes 
can have different heap configurations, it will make sense to make it 
datanode-level, rather than something enforced by namenode.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6089) Standby NN while transitioning to active throws a connection refused error when the prior active NN process is suspended

2014-03-11 Thread Arpit Gupta (JIRA)
Arpit Gupta created HDFS-6089:
-

 Summary: Standby NN while transitioning to active throws a 
connection refused error when the prior active NN process is suspended
 Key: HDFS-6089
 URL: https://issues.apache.org/jira/browse/HDFS-6089
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: ha
Affects Versions: 2.4.0
Reporter: Arpit Gupta
Assignee: Jing Zhao


The following scenario was tested:

* Determine Active NN and suspend the process (kill -19)
* Wait about 60s to let the standby transition to active
* Get the service state for nn1 and nn2 and make sure nn2 has transitioned to 
active.


What was noticed that some times the call to get the service state of nn2 got a 
socket time out connection.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6090) Use MiniDFSCluster.Builder instead of deprecated constructors

2014-03-11 Thread Akira AJISAKA (JIRA)
Akira AJISAKA created HDFS-6090:
---

 Summary: Use MiniDFSCluster.Builder instead of deprecated 
constructors
 Key: HDFS-6090
 URL: https://issues.apache.org/jira/browse/HDFS-6090
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: test
Reporter: Akira AJISAKA
Priority: Minor


Some test classes are using deprecated constructors such as 
{{MiniDFSCluster(Configuration, int, boolean, String[], String[])}} for 
building a MiniDFSCluster.
These classes should use {{MiniDFSCluster.Builder}} to reduce javac warnings 
and improve code readability.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6091) dfs.journalnode.edits.dir should accept URI

2014-03-11 Thread Allen Wittenauer (JIRA)
Allen Wittenauer created HDFS-6091:
--

 Summary: dfs.journalnode.edits.dir should accept URI
 Key: HDFS-6091
 URL: https://issues.apache.org/jira/browse/HDFS-6091
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: journal-node
Affects Versions: 2.2.0
Reporter: Allen Wittenauer
Priority: Minor


Using a URI in dfs.journalnode.edits.dir (such as file:///foo)  throws a 
Journal dir 'file:/foo' should be an absolute path'. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6092) DistributedFileSystem#getCanonicalServiceName() and DistributedFileSystem#getUri() may return inconsistent results w.r.t. port

2014-03-11 Thread Ted Yu (JIRA)
Ted Yu created HDFS-6092:


 Summary: DistributedFileSystem#getCanonicalServiceName() and 
DistributedFileSystem#getUri() may return inconsistent results w.r.t. port
 Key: HDFS-6092
 URL: https://issues.apache.org/jira/browse/HDFS-6092
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.3.0
Reporter: Ted Yu


I discovered this when working on HBASE-10717
Here is sample code to reproduce the problem:
{code}
Path desPath = new Path(hdfs://127.0.0.1/);
FileSystem desFs = desPath.getFileSystem(conf);

String s = desFs.getCanonicalServiceName();
URI uri = desFs.getUri();
{code}
Canonical name string contains the default port - 8020
But uri doesn't contain port.
This would result in the following exception:
{code}
testIsSameHdfs(org.apache.hadoop.hbase.util.TestFSHDFSUtils)  Time elapsed: 
0.001 sec   ERROR!
java.lang.IllegalArgumentException: port out of range:-1
at java.net.InetSocketAddress.checkPort(InetSocketAddress.java:143)
at java.net.InetSocketAddress.init(InetSocketAddress.java:224)
at 
org.apache.hadoop.hbase.util.FSHDFSUtils.getNNAddresses(FSHDFSUtils.java:88)
{code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6093) Expose more caching information for debugging by users

2014-03-11 Thread Andrew Wang (JIRA)
Andrew Wang created HDFS-6093:
-

 Summary: Expose more caching information for debugging by users
 Key: HDFS-6093
 URL: https://issues.apache.org/jira/browse/HDFS-6093
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: caching
Affects Versions: 2.4.0
Reporter: Andrew Wang
Assignee: Andrew Wang


When users submit a new cache directive, it's unclear if the NN has recognized 
it and is actively trying to cache it, or if it's hung for some other reason. 
It'd be nice to expose a pending caching/uncaching count the same way we 
expose pending replication work.

It'd also be nice to display the aggregate cache capacity and usage in dfsadmin 
-report, since we already have have it as a metric and expose it per-DN in 
report output.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6094) The same block can be counted twice towards safe mode threshold

2014-03-11 Thread Arpit Agarwal (JIRA)
Arpit Agarwal created HDFS-6094:
---

 Summary: The same block can be counted twice towards safe mode 
threshold
 Key: HDFS-6094
 URL: https://issues.apache.org/jira/browse/HDFS-6094
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.4.0
Reporter: Arpit Agarwal
Assignee: Arpit Agarwal


{{BlockManager#addStoredBlock}} can cause the same block can be counted towards 
safe mode threshold. We see this manifest via 
{{TestHASafeMode#testBlocksAddedWhileStandbyIsDown}} failures on Ubuntu. More 
details in a comment.



--
This message was sent by Atlassian JIRA
(v6.2#6252)