[jira] [Created] (HDFS-7568) Support immutability (Write-once-read-many) in HDFS

2014-12-23 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-7568:
-

 Summary: Support immutability (Write-once-read-many) in HDFS
 Key: HDFS-7568
 URL: https://issues.apache.org/jira/browse/HDFS-7568
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: namenode
Affects Versions: 2.7.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Many regulatory compliance requires storage to support WORM functionality to 
protect sensitive data from being modified or deleted. This jira proposes 
adding that feature to HDFS.

See the following comment for more description.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-7408) Add a counter in the log that shows the number of block reports processed

2014-11-18 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-7408:
-

 Summary: Add a counter in the log that shows the number of block 
reports processed
 Key: HDFS-7408
 URL: https://issues.apache.org/jira/browse/HDFS-7408
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Suresh Srinivas


It would be great to have in the info log corresponding to block report 
processing, printing information on how many block reports have been processed. 
This can be useful to debug when namenode is unresponsive especially during 
startup time to understand if datanodes are sending block reports multiple 
times.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-6320) Namenode exits on exception without printing stack trace in AbstractDelegationTokenSecretManager

2014-05-01 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6320:
-

 Summary: Namenode exits on exception without printing stack trace 
in AbstractDelegationTokenSecretManager
 Key: HDFS-6320
 URL: https://issues.apache.org/jira/browse/HDFS-6320
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 1.2.1, 2.0.0-alpha
Reporter: Suresh Srinivas


Not printing the stack trace makes debugging harder.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6274) Cleanup javadoc warnings in HDFS code

2014-04-23 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6274:
-

 Summary: Cleanup javadoc warnings in HDFS code
 Key: HDFS-6274
 URL: https://issues.apache.org/jira/browse/HDFS-6274
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: 2.4.0
Reporter: Suresh Srinivas
 Attachments: HDFS-6274.patch

Javadoc in HDFS code has many issues -  The parameter names are not correct, 
parameters are not documented, exceptions declared as throws are not thrown, 
typos etc.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6275) Fix warnings - type arguments can be inferred and redudant local variable

2014-04-23 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6275:
-

 Summary: Fix warnings - type arguments can be inferred and 
redudant local variable
 Key: HDFS-6275
 URL: https://issues.apache.org/jira/browse/HDFS-6275
 Project: Hadoop HDFS
  Issue Type: Sub-task
Affects Versions: 2.4.0
Reporter: Suresh Srinivas






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6276) Remove unnecessary conditions and null check

2014-04-23 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6276:
-

 Summary: Remove unnecessary conditions and null check
 Key: HDFS-6276
 URL: https://issues.apache.org/jira/browse/HDFS-6276
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas
 Attachments: HDFS-6276.patch

The code has many places where null check and other condition checks are 
unnecessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6271) Rename JournalProtocol to BackupNodeJournalProtocol

2014-04-22 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6271:
-

 Summary: Rename JournalProtocol to BackupNodeJournalProtocol
 Key: HDFS-6271
 URL: https://issues.apache.org/jira/browse/HDFS-6271
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


[~shv], as indicated in earlier comments, JournalProtocol is used only for 
sending journal to backupnode .Renaming it would make it clear how the protocol 
is being used and will not conflict with QuorumJournalProtocol.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6185) HDFS operational and debuggability improvements

2014-04-02 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6185:
-

 Summary: HDFS operational and debuggability improvements
 Key: HDFS-6185
 URL: https://issues.apache.org/jira/browse/HDFS-6185
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


This umbrella jira proposes improvements in HDFS to make operations simpler and 
the system more robust. The areas of improvements includes - ensuring 
configuration is correct and follows the know best practices, make HDFS robust 
against some of the failures observed due to cluster not being monitored 
correctly, better metadata management to avoid lengthy startup time etc.

Subtasks will be filed with more details on individual improvements.

The goal is simplify operations, improve debuggability and robustness. If you 
have ideas on improvements that fall under this umbrella, please file subtasks 
under this jira.




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6186) Pause deletion of blocks when the namenode starts up

2014-04-02 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6186:
-

 Summary: Pause deletion of blocks when the namenode starts up
 Key: HDFS-6186
 URL: https://issues.apache.org/jira/browse/HDFS-6186
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Suresh Srinivas


HDFS namenode can delete blocks very quickly, given the deletion happens as a 
parallel operation spread across many datanodes. One of the frequent anxieties 
I see is that a lot of data can be deleted very quickly, when a cluster is 
brought up, especially when one of the storage directories has failed and 
namenode metadata was copied from another storage. Copying wrong metadata would 
results in some of the newer files (if old metadata was copied) being deleted 
along with their blocks. 

HDFS-5986 now captures the number of pending deletion block on namenode webUI 
and JMX. I propose pausing deletion of blocks for a configured period of time 
(default 1 hour?) after namenode comes out of safemode. This will give enough 
time for the administrator to notice large number of pending deletion blocks 
and take corrective action.

Thoughts?



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6162) Format strings should use platform independent line separator

2014-03-26 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6162:
-

 Summary: Format strings should use platform independent line 
separator
 Key: HDFS-6162
 URL: https://issues.apache.org/jira/browse/HDFS-6162
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Minor


In format strings, it is generally preferable better to use %n, which will 
result in platform-specific line separator. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6155) Fix Boxing/unboxing to parse a primitive findbugs warnings

2014-03-25 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6155:
-

 Summary: Fix Boxing/unboxing to parse a primitive findbugs warnings
 Key: HDFS-6155
 URL: https://issues.apache.org/jira/browse/HDFS-6155
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


There are many instances of such findbugs warnings related to performance. See 
for details - 
http://findbugs.sourceforge.net/bugDescriptions.html#DM_BOXED_PRIMITIVE_FOR_PARSING



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6150) Add inode id information in the logs to make debugging easier

2014-03-24 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6150:
-

 Summary: Add inode id information in the logs to make debugging 
easier
 Key: HDFS-6150
 URL: https://issues.apache.org/jira/browse/HDFS-6150
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Suresh Srinivas
 Attachments: HDFS-6150.patch

Inode information and path information are missing in the logs and exceptions. 
Adding this will help debug multi threading issues related to using incode 
INode ID information.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Reopened] (HDFS-5138) Support HDFS upgrade in HA

2014-03-20 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5138?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas reopened HDFS-5138:
---


 Support HDFS upgrade in HA
 --

 Key: HDFS-5138
 URL: https://issues.apache.org/jira/browse/HDFS-5138
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.1.1-beta
Reporter: Kihwal Lee
Assignee: Aaron T. Myers
Priority: Blocker
 Fix For: 3.0.0

 Attachments: HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
 HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
 HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, HDFS-5138.patch, 
 hdfs-5138-branch-2.txt


 With HA enabled, NN wo't start with -upgrade. Since there has been a layout 
 version change between 2.0.x and 2.1.x, starting NN in upgrade mode was 
 necessary when deploying 2.1.x to an existing 2.0.x cluster. But the only way 
 to get around this was to disable HA and upgrade. 
 The NN and the cluster cannot be flipped back to HA until the upgrade is 
 finalized. If HA is disabled only on NN for layout upgrade and HA is turned 
 back on without involving DNs, things will work, but finaliizeUpgrade won't 
 work (the NN is in HA and it cannot be in upgrade mode) and DN's upgrade 
 snapshots won't get removed.
 We will need a different ways of doing layout upgrade and upgrade snapshot.  
 I am marking this as a 2.1.1-beta blocker based on feedback from others.  If 
 there is a reasonable workaround that does not increase maintenance window 
 greatly, we can lower its priority from blocker to critical.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6117) Print file path information in FileNotFoundException

2014-03-18 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6117:
-

 Summary: Print file path information in FileNotFoundException
 Key: HDFS-6117
 URL: https://issues.apache.org/jira/browse/HDFS-6117
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.2.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Print file path information in FileNotFoundException thrown from 
INodeId#checkId(). This helps debug issues related possible INode Id mismatches 
during namenode fail over.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6118) Code cleanup

2014-03-18 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6118:
-

 Summary: Code cleanup
 Key: HDFS-6118
 URL: https://issues.apache.org/jira/browse/HDFS-6118
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


HDFS code needs cleanup related to many typos, undocumented parameters, unused 
methods, unnecessary cast, imports and exceptions declared as thrown to name a 
few.

I plan on working on cleaning this up as I get time. To keep code review 
manageable, I will create sub tasks and cleanup the code a few classes at a 
time.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6119) FSNamesystem code cleanup

2014-03-18 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6119:
-

 Summary: FSNamesystem code cleanup
 Key: HDFS-6119
 URL: https://issues.apache.org/jira/browse/HDFS-6119
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6124) Add final modifier to class members

2014-03-18 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6124:
-

 Summary: Add final modifier to class members
 Key: HDFS-6124
 URL: https://issues.apache.org/jira/browse/HDFS-6124
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 2.4.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
 Attachments: HDFS-6124.patch

Many of the member variables declaration for classes are missing final modifier 
in HDFS. This jira adds final modifier where possible.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Reopened] (HDFS-6099) HDFS file system limits not enforced on renames.

2014-03-13 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-6099?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas reopened HDFS-6099:
---


Sorry, accidentally resolved this.

 HDFS file system limits not enforced on renames.
 

 Key: HDFS-6099
 URL: https://issues.apache.org/jira/browse/HDFS-6099
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.3.0
Reporter: Chris Nauroth
Assignee: Chris Nauroth
 Fix For: 2.4.0

 Attachments: HDFS-6099.1.patch


 {{dfs.namenode.fs-limits.max-component-length}} and 
 {{dfs.namenode.fs-limits.max-directory-items}} are not enforced on the 
 destination path during rename operations.  This means that it's still 
 possible to create files that violate these limits.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HDFS-6055) Change default configuration to limit file name length in HDFS

2014-03-04 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6055:
-

 Summary: Change default configuration to limit file name length in 
HDFS
 Key: HDFS-6055
 URL: https://issues.apache.org/jira/browse/HDFS-6055
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Currently configuration dfs.namenode.fs-limits.max-component-length is set to 
0. With this HDFS file names have no length limit. However, we see more users 
run into issues where they copy files from HDFS to another file system and the 
copy fails due to the file name length being too long.

I propose changing the default configuration 
dfs.namenode.fs-limits.max-component-length to a reasonable value. This will 
be an incompatible change. However, user who need long file names can override 
this configuration to turn off length limit.

What do folks think?



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Resolved] (HDFS-3225) Revist upgrade snapshots, roll back, finalize to enable rolling upgrades

2014-02-25 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3225?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-3225.
---

Resolution: Done

The rolling upgrade and the related issues have been address in the umbrella 
jiar, HDFS-5535. I do not think there is any work that needs to be done to 
address this. Resolving this as Done

 Revist upgrade snapshots, roll back, finalize to enable rolling upgrades
 

 Key: HDFS-3225
 URL: https://issues.apache.org/jira/browse/HDFS-3225
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Sanjay Radia
Assignee: Sanjay Radia

 With introduction of wire comaptibility, HDFS is ready for rolling upgrades. 
 Various design choices such as version checks and when to do upgrade 
 snaopshots, rollbacks, finalize need to be revisited.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HDFS-6005) Simplify Datanode rollback and downgrade

2014-02-24 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-6005:
-

 Summary: Simplify Datanode rollback and downgrade
 Key: HDFS-6005
 URL: https://issues.apache.org/jira/browse/HDFS-6005
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Problem:
When rolling upgrade fails, the cluster can either be downgraded or rolled 
back. With the current functionality in this feature branch, it is possible to 
downgrade namenode, while datanode is incorrectly rolled back. This does not 
affect the cluster state. The old blocks that appear back on the datanode due 
to rollback will be deleted. Similarly it is also possible to rollback 
namenode, while datanode is not rolled back. This can cause problem where old 
blocks do not appear back on the datanode and can result in missing blocks.

Solution:
I propose making the following changes:
During rollback or downgrade, the entire cluster must be restarted. The 
datanodes always restore the deleted blocks on restart and go back to trash 
disabled mode. There is no need for datanodes to be started up -rollingUpgrade 
-rollback, anymore.
# On namenode downgrade, the restored blocks are deleted.
# On namenode rollback, the restored blocks will be retained and any newly 
created blocks (since the start of rolling upgrade) are deleted.

This is much simpler operationally and solves the problem described above.





--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Created] (HDFS-5986) Capture the number of blocks pending deletion on namenode webUI

2014-02-20 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5986:
-

 Summary: Capture the number of blocks pending deletion on namenode 
webUI
 Key: HDFS-5986
 URL: https://issues.apache.org/jira/browse/HDFS-5986
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Suresh Srinivas


When a directory that has large number of directories and files are deleted, 
the namespace deletes the corresponding inodes immediately. However it is hard 
to to know when the invalidated blocks are actually deleted on the datanodes, 
which could take a while.

I propose adding on namenode webUI, along with under replicated blocks, the 
number of blocks that are pending deletion.



--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HDFS-5720) DataNode always READ_BLOCK or WRITE_BLOCK

2014-01-06 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5720?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-5720.
---

Resolution: Not A Problem

 DataNode always READ_BLOCK or WRITE_BLOCK
 -

 Key: HDFS-5720
 URL: https://issues.apache.org/jira/browse/HDFS-5720
 Project: Hadoop HDFS
  Issue Type: Task
  Components: datanode
Affects Versions: 2.2.0
Reporter: yangping wu
Priority: Trivial
   Original Estimate: 480h
  Remaining Estimate: 480h

 I find many following errors on my datanode's log. What is the reason for 
 this and how can I resolved it?
 {code}
 2014-01-05 00:14:40,589 ERROR 
 org.apache.hadoop.hdfs.server.datanode.DataNode: date51:50010:DataXceiver 
 error processing WRITE_BLOCK operation  src: /192.168.246.57:35455 dest: 
 /192.168.246.75:50010
 java.io.IOException: Premature EOF from inputStream
 at org.apache.hadoop.io.IOUtils.readFully(IOUtils.java:194)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doReadFully(PacketReceiver.java:213)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.doRead(PacketReceiver.java:134)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.PacketReceiver.receiveNextPacket(PacketReceiver.java:109)
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receivePacket(BlockReceiver.java:435)
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockReceiver.receiveBlock(BlockReceiver.java:693)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.writeBlock(DataXceiver.java:569)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opWriteBlock(Receiver.java:115)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:68)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:221)
 at java.lang.Thread.run(Thread.java:744)
 2014-01-05 02:50:52,552 ERROR 
 org.apache.hadoop.hdfs.server.datanode.DataNode: date51:50010:DataXceiver 
 error processing READ_BLOCK operation  src: /192.168.246.39:46693 dest: 
 /192.168.246.75:50010
 org.apache.hadoop.hdfs.server.datanode.ReplicaNotFoundException: Replica not 
 found for 
 BP-1398136447-192.168.246.60-1386067202761:blk_1089253308_1099605945003
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockSender.getReplica(BlockSender.java:418)
 at 
 org.apache.hadoop.hdfs.server.datanode.BlockSender.init(BlockSender.java:228)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.readBlock(DataXceiver.java:327)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.opReadBlock(Receiver.java:101)
 at 
 org.apache.hadoop.hdfs.protocol.datatransfer.Receiver.processOp(Receiver.java:65)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataXceiver.run(DataXceiver.java:221)
 at java.lang.Thread.run(Thread.java:744)
 2014-01-05 07:39:30,939 ERROR 
 org.apache.hadoop.hdfs.server.datanode.DataNode: 
 DataNode{data=FSDataset{dirpath='[/data1/hadoop/dfs/current, 
 /data2/hadoop/dfs/current, /data3/hadoop/dfs/current, 
 /data4/hadoop/dfs/current, /data5/hadoop/dfs/current, 
 /data6/hadoop/dfs/current, /data7/hadoop/dfs/current, 
 /data8/hadoop/dfs/current, /data9/hadoop/dfs/current, 
 /data10/hadoop/dfs/current, /data11/hadoop/dfs/current, 
 /data12/hadoop/dfs/current]'}, localName='date51:50010', 
 storageID='DS-1312530819-192.168.246.75-50010-1388062043867', 
 xmitsInProgress=0}:Exception transfering block 
 BP-1398136447-192.168.246.60-1386067202761:blk_1089401766_1099606093471 to 
 mirror 192.168.242.51:50010: java.io.IOException: Connection reset by peer
 2014-01-05 07:39:30,939 INFO org.apache.hadoop.hdfs.server.datanode.DataNode: 
 opWriteBlock 
 BP-1398136447-192.168.246.60-1386067202761:blk_1089401766_1099606093471 
 received exception java.io.IOException: Connection reset by peer
 2014-01-05 07:39:30,939 ERROR 
 org.apache.hadoop.hdfs.server.datanode.DataNode: date51:50010:DataXceiver 
 error processing WRITE_BLOCK operation  src: /192.168.246.75:44779 dest: 
 /192.168.246.75:50010
 java.io.IOException: Connection reset by peer
 at sun.nio.ch.FileDispatcherImpl.read0(Native Method)
 at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39)
 at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223)
 at sun.nio.ch.IOUtil.read(IOUtil.java:197)
 at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379)
 at 
 org.apache.hadoop.net.SocketInputStream$Reader.performIO(SocketInputStream.java:57)
 at 
 org.apache.hadoop.net.SocketIOWithTimeout.doIO(SocketIOWithTimeout.java:142)
 at 
 org.apache.hadoop.net.SocketInputStream.read(SocketInputStream.java:161)
 at 
 

[jira] [Created] (HDFS-5704) Change OP_UPDATE_BLOCKS with a new OP_ADD_BLOCK

2013-12-28 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5704:
-

 Summary: Change OP_UPDATE_BLOCKS  with a new OP_ADD_BLOCK
 Key: HDFS-5704
 URL: https://issues.apache.org/jira/browse/HDFS-5704
 Project: Hadoop HDFS
  Issue Type: Bug
 Environment: Currently every time a block a allocated, the entire list 
of blocks are written in the editlog in OP_UPDATE_BLOCKS operation. This has 
n^2 growth issue. The total size of editlog records for a file with large 
number of blocks could be huge.

The goal of this jira is discuss adding a different editlog record that only 
records allocation of block and not the entire block list, on every block 
allocation.
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas






--
This message was sent by Atlassian JIRA
(v6.1.5#6160)


[jira] [Resolved] (HDFS-5367) Restoring namenode storage locks namenode due to unnecessary fsimage write

2013-10-17 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5367?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-5367.
---

   Resolution: Fixed
Fix Version/s: 1.3.0
 Hadoop Flags: Reviewed

I committed the patch to branch-1. Thank you John Zhao.

 Restoring namenode storage locks namenode due to unnecessary fsimage write
 --

 Key: HDFS-5367
 URL: https://issues.apache.org/jira/browse/HDFS-5367
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 1.2.1
Reporter: zhaoyunjiong
Assignee: zhaoyunjiong
 Fix For: 1.3.0

 Attachments: HDFS-5367-branch-1.2.patch


 Our cluster have 40G fsimage, we write one copy of edit log to NFS.
 After NFS temporary failed, when doing checkpoint, NameNode try to recover 
 it, and it will save 40G fsimage to NFS, it takes some time ( 40G/128MB/s = 
 320 seconds) , and it locked FSNamesystem, and this bring down our cluster.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Created] (HDFS-5340) Add Hive to the list of projects using AbstractDelegationTokenSecretManager

2013-10-10 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5340:
-

 Summary: Add Hive to the list of projects using 
AbstractDelegationTokenSecretManager
 Key: HDFS-5340
 URL: https://issues.apache.org/jira/browse/HDFS-5340
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: security
Reporter: Suresh Srinivas


org.apache.hadoop.hive.thrift.DelegationTokenSecretManager extends 
AbstractDelegationTokenSecretManager. This should be captured in the 
InterfaceAudience annotation of AbstractDelegationTokenSecretManager.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Created] (HDFS-5316) Namenode ignore the default https port

2013-10-07 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5316:
-

 Summary: Namenode ignore the default https port
 Key: HDFS-5316
 URL: https://issues.apache.org/jira/browse/HDFS-5316
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


When dfs.https.enable is true and dfs.https.port is not configured, namenode 
does not pickup the default https port (50470), instead picks up random port. 
See:
{code}
boolean needClientAuth = conf.getBoolean(dfs.https.need.client.auth, false);
  InetSocketAddress secInfoSocAddr = NetUtils.createSocketAddr(infoHost + 
: + conf.get(
DFSConfigKeys.DFS_NAMENODE_HTTPS_PORT_KEY, 0));
  Configuration sslConf = new Configuration(false);
  if (certSSL) {

sslConf.addResource(conf.get(DFSConfigKeys.DFS_SERVER_HTTPS_KEYSTORE_RESOURCE_KEY,
 ssl-server.xml));
  }
{code}

Unless https port is specifically configured as 0, we should not be picking 
random port. This needs to be documented in hdfs-default.xml



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Created] (HDFS-5317) Go back to DFS Home link does not work on datanode webUI

2013-10-07 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5317:
-

 Summary: Go back to DFS Home link does not work on datanode webUI
 Key: HDFS-5317
 URL: https://issues.apache.org/jira/browse/HDFS-5317
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas
Priority: Critical


Here is how to reproduce this problem:
# Goto https://namenode:https_port/dfshealth.jsp
# Click on Live Nodes link
# On the Live Nodes page, click one of the links for the datanodes. This link 
has always hardcodes the http link instead of http/https based on the context 
of the webpage.

On datanode webUI page, the link Go back to DFS home does not work.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Resolved] (HDFS-3538) TestBlocksWithNotEnoughRacks fails

2013-10-06 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3538?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-3538.
---

Resolution: Duplicate

This is a duplicate of HDFS-3267.

 TestBlocksWithNotEnoughRacks fails
 --

 Key: HDFS-3538
 URL: https://issues.apache.org/jira/browse/HDFS-3538
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 0.24.0
Reporter: Brandon Li

 It failed for a few days in jenkins test.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Created] (HDFS-5310) Add http policy support for YARN daemons

2013-10-05 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5310:
-

 Summary: Add http policy support for YARN daemons
 Key: HDFS-5310
 URL: https://issues.apache.org/jira/browse/HDFS-5310
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.0.0-alpha
Reporter: Suresh Srinivas


This YARN part of HADOOP-10022.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Created] (HDFS-5306) Datanode https port is not available in the namenode

2013-10-04 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5306:
-

 Summary: Datanode https port is not available in the namenode
 Key: HDFS-5306
 URL: https://issues.apache.org/jira/browse/HDFS-5306
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


To enable https access, the datanode http server https port is needed in 
namenode web pages and redirects from the namenode. This jira adds an 
additional optional field to DatanodeIDProto in hdfs.proto and the 
corresponding DatanodeID java class.



--
This message was sent by Atlassian JIRA
(v6.1#6144)


[jira] [Created] (HDFS-5187) Deletion of non-existing cluster succeeds

2013-09-11 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5187:
-

 Summary: Deletion of non-existing cluster succeeds
 Key: HDFS-5187
 URL: https://issues.apache.org/jira/browse/HDFS-5187
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Suresh Srinivas


Following command succeeds even if the cluster of name non-existent has not 
been added:
{noformat}
bin/falcon entity -delete -type cluster -name non-existent
falcon/abc(cluster) removed successfully
{noformat}


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4940) namenode OOMs under Bigtop's TestCLI

2013-08-27 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4940?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4940.
---

Resolution: Cannot Reproduce

[~rvs], I have closed this. We can reopen this, if this problem occurs again.

 namenode OOMs under Bigtop's TestCLI
 

 Key: HDFS-4940
 URL: https://issues.apache.org/jira/browse/HDFS-4940
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.1.0-beta
Reporter: Roman Shaposhnik
Assignee: Roman Shaposhnik
Priority: Blocker
 Fix For: 2.3.0


 Bigtop's TestCLI when executed against Hadoop 2.1.0 seems to make it OOM 
 quite reliably regardless of the heap size settings. I'm attaching a heap 
 dump URL. Alliteratively anybody can just take Bigtop's tests, compiled them 
 against Hadoop 2.1.0 bits and try to reproduce it.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4747) Convert snapshot user guide to APT from XDOC

2013-08-06 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4747?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4747.
---

Resolution: Won't Fix

Marking this as won't fix. Reopen if necessary. 

 Convert snapshot user guide to APT from XDOC
 

 Key: HDFS-4747
 URL: https://issues.apache.org/jira/browse/HDFS-4747
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: documentation
Affects Versions: Snapshot (HDFS-2802)
Reporter: Aaron T. Myers
Assignee: Aaron T. Myers
 Attachments: HDFS-4747.patch, HdfsSnapshots.html


 To be consistent with the rest of the HDFS docs, the snapshots user guide 
 should use APT instead of XDOC.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-5015) Error while running custom writable code - DFSClient_NONMAPREDUCE_549327626_1 doesnot have any open file

2013-07-21 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-5015?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-5015.
---

Resolution: Invalid

Please use forums for help. Jira is for reporting bugs.

 Error while running custom writable code  - 
 DFSClient_NONMAPREDUCE_549327626_1 doesnot have any open file
 -

 Key: HDFS-5015
 URL: https://issues.apache.org/jira/browse/HDFS-5015
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
 Environment: Cloudera 4.3
Reporter: Allan Gonsalves

 hi , 
 We are facing an below error while running custom writable code, our driver 
 code is using below code . logwritable is the name of custom writable 
 interface.
 --Error message at the code execution is as follows
 DFSClient_NONMAPREDUCE_549327626_1 doesnot have any open file
 --Driver code - 
 job.setMapOutputKeyClass(logwritable.class);
 job.setMapOutputValueClass(logwritable.class);
 job.setNumReducerTasks(0);
 --logwritable class implements Writable interface
 --Definition of mapper is as follows
 public class MAPPER extends MapperText,IntWritable,Logwritable, Logwritabe{ 
 
 contaxt.write(new logwritable(), new logwritable());
 }
 Any help to resolve above issue will be great help
 Thanks
 Allan 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-5008) Make ClientProtocol#abandonBlock() idempotent

2013-07-18 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-5008:
-

 Summary: Make ClientProtocol#abandonBlock() idempotent
 Key: HDFS-5008
 URL: https://issues.apache.org/jira/browse/HDFS-5008
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Suresh Srinivas


Currently when abandonBlock is called, if the block does not exist, then the 
namenode throws an exception. I propose making it idempotent, by returning 
success if the block that is being abandoned does not exist.

Thoughts?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4986) Namenode webUI changes to incorporate storage type

2013-07-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4986:
-

 Summary: Namenode webUI changes to incorporate storage type
 Key: HDFS-4986
 URL: https://issues.apache.org/jira/browse/HDFS-4986
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


Namenode needs to change cluster storage summary to show storage information 
per storage type. Datanode list must also include this information.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4985) Add storage type to the protocol and expose it in block report and block locations

2013-07-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4985:
-

 Summary: Add storage type to the protocol and expose it in block 
report and block locations
 Key: HDFS-4985
 URL: https://issues.apache.org/jira/browse/HDFS-4985
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Reporter: Suresh Srinivas


With HDFS-2880 datanode now supports storage abstraction. This is to add 
storage type in to the protocol. Datanodes currently report blocks per storage. 
Storage would include storage type attribute. Namenode also exposes the storage 
type of a block in block locations.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4987) Namenode changes to track multiple storages

2013-07-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4987:
-

 Summary: Namenode changes to track multiple storages
 Key: HDFS-4987
 URL: https://issues.apache.org/jira/browse/HDFS-4987
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Suresh Srinivas


Currently namenode track in BlockInfo, the corresponding DataNodeDescriptor. I 
propose changing this with new abstraction StorageDescriptor. This change will 
also make DatanodeDescriptor a collection of StroageDescriptors. This will 
allow given a block, identify its storage and also Datanode.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4988) Datanode must support all the volumes as individual storages

2013-07-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4988:
-

 Summary: Datanode must support all the volumes as individual 
storages
 Key: HDFS-4988
 URL: https://issues.apache.org/jira/browse/HDFS-4988
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


Currently all the volumes on datanode is reported as a single storage. This 
change proposes reporting them as individual storage. This requires:
# A unique storage ID for each storage
#* This needs to be generated during formatting
# There should be an option to allow existing disks to be reported as single 
storage unit for backward compatibility.
# A functionality is also needed to split the existing all volumes as single 
storage unit to to individual storage units.
# Configuration must allow for each storage unit a storage type attribute.
# Block reports must be sent on a per storage basis. In some cases (such memory 
tier) block reports may need to be sent more frequently. That means block 
reporting period must be on a per storage type basis.

My proposal is for new clusters to configure volumes by default as separate 
storage unit. Lets discuss.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4989) Balancer needs to consider storage type in balancing

2013-07-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4989:
-

 Summary: Balancer needs to consider storage type in balancing
 Key: HDFS-4989
 URL: https://issues.apache.org/jira/browse/HDFS-4989
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: balancer, datanode, namenode
Reporter: Suresh Srinivas


Balancer needs to balance with in a storage tier. Also needs an option to 
balance only a specific storage type.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4990) Block placement for storage types

2013-07-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4990:
-

 Summary: Block placement for storage types
 Key: HDFS-4990
 URL: https://issues.apache.org/jira/browse/HDFS-4990
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Suresh Srinivas


Currently block location for writes are made based on:
# Datanode load (number of transceivers)
# Space left on datanode
# Topology

With storage abstraction, namenode must choose a storage instead of a datanode 
for block placement. It also needs to consider storage type, load on the 
storage etc.

As an additional benefit, currently HDFS support heterogeneous nodes (nodes 
with different number of spindles etc.) poorly. This work should help solve 
that issue as well.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4973) Move the Rpc request call ID generation to client side InvocationHandler

2013-07-10 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4973:
-

 Summary: Move the Rpc request call ID generation to client side 
InvocationHandler
 Key: HDFS-4973
 URL: https://issues.apache.org/jira/browse/HDFS-4973
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Suresh Srinivas


Currently when RetryInvocationHandler is used to retry an RPC request, a new 
RPC request call ID is generated. This jira proposes moving call ID generation 
to InvocationHandler so that retried RPC requests retain the same call ID. This 
is needed for RetryCache functionality proposed in HDFS-4942.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4974) Analyze and add annotations to the Namenode and Datanode protocol methods to enable retry

2013-07-10 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4974:
-

 Summary: Analyze and add annotations to the Namenode and Datanode 
protocol methods to enable retry
 Key: HDFS-4974
 URL: https://issues.apache.org/jira/browse/HDFS-4974
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


This jira is intended for:
# Discussing current @Idempotent annotations in HDFS protocols and adding that 
annotation where it is missing.
# Discuss how retry should be enabled for non-idempotent requests.

I will post the analysis of current methods in subsequent comment.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4976) Rename Client#uuid to Client#clientId

2013-07-10 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4976:
-

 Summary: Rename Client#uuid to Client#clientId
 Key: HDFS-4976
 URL: https://issues.apache.org/jira/browse/HDFS-4976
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Suresh Srinivas


To address the comment - 
https://issues.apache.org/jira/browse/HADOOP-9688?focusedCommentId=13705032page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13705032

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-3092) Enable journal protocol based editlog streaming for standby namenode

2013-07-08 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3092?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-3092.
---

Resolution: Won't Fix

 Enable journal protocol based editlog streaming for standby namenode
 

 Key: HDFS-3092
 URL: https://issues.apache.org/jira/browse/HDFS-3092
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, namenode
Affects Versions: 0.24.0, 0.23.3
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
 Attachments: ComparisonofApproachesforHAJournals.pdf, JNStates.png, 
 MultipleSharedJournals.pdf, MultipleSharedJournals.pdf, 
 MultipleSharedJournals.pdf, Removingfilerdependency.pdf


 Currently standby namenode relies on reading shared editlogs to stay current 
 with the active namenode, for namespace changes. BackupNode used streaming 
 edits from active namenode for doing the same. This jira is to explore using 
 journal protocol based editlog streams for the standby namenode. A daemon in 
 standby will get the editlogs from the active and write it to local edits. To 
 begin with, the existing standby mechanism of reading from a file, will 
 continue to be used, instead of from shared edits, from the local edits.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4955) Use InodeID to improve snapshot diff

2013-07-03 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4955?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4955.
---

Resolution: Duplicate

Duplicate of HDFS-4667

 Use InodeID to improve snapshot diff
 

 Key: HDFS-4955
 URL: https://issues.apache.org/jira/browse/HDFS-4955
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Binglin Chang

 Currently snapshot diff doesn't handle file/dir rename very well, so distcp 
 can't do minimal data transfer in incremental backup. Using InodeID can 
 uniquely identify a file, and can help diff search algorithm.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4956) DistributedFileSystem does not handle CreateFlag.APPEND in create call

2013-07-03 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4956:
-

 Summary: DistributedFileSystem does not handle CreateFlag.APPEND 
in create call
 Key: HDFS-4956
 URL: https://issues.apache.org/jira/browse/HDFS-4956
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Suresh Srinivas


Currently DistributedFileSystem does not handle CreateFlag.APPEND in the 
implementation of FileSystem#create() method. It only support OVERWRITE  
CREATE or just CREATE.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4942) Add retry cache support in Namenode

2013-06-28 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4942:
-

 Summary: Add retry cache support in Namenode
 Key: HDFS-4942
 URL: https://issues.apache.org/jira/browse/HDFS-4942
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: ha, namenode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


In current HA mechanism with FailoverProxyProvider and non HA setups with 
RetryProxy retry a request from the RPC layer. If the retried request has 
already been processed at the namenode, the subsequent attempts fail for 
idempotent operations such as  create, append, delete, rename etc. This will 
cause application failures during HA failover, network issues etc.

This jira proposes adding retry cache at the namenode to handle these failures. 
More details in the comments.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4923) Save namespace when the namenode is stopped

2013-06-20 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4923:
-

 Summary: Save namespace when the namenode is stopped
 Key: HDFS-4923
 URL: https://issues.apache.org/jira/browse/HDFS-4923
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 3.0.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


In rare instances the namenode fails to load editlog due to corruption during 
startup. This has more severe impact if editlog segment to be checkpointed has 
corruption, as checkpointing fails because the editlog with corruption cannot 
be consumed. If an administrator does not notice this and address it by saving 
the namespace, recovering the namespace would involve complex file editing, 
using previous backups or losing last set of modifications.

The other issue that also happens frequently is, checkpointing fails and has 
not happened for a long time, resulting in long editlogs and even corrupt 
editlogs.

To handle these issues, when namenode is stopped, we can put it in safemode and 
save the namespace, before the process is shutdown. As an added benefit, the 
namenode restart would be faster, given there is no editlog to consume.

What do folks think?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4921) Create a web page to display the namenode features and whether they are turned on or off

2013-06-19 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4921:
-

 Summary: Create a web page to display the namenode features and 
whether they are turned on or off
 Key: HDFS-4921
 URL: https://issues.apache.org/jira/browse/HDFS-4921
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 3.0.0
Reporter: Suresh Srinivas


The only way to determine a feature is turned on or off is to look at namenode 
logs or configuration. Some time when features change/modified (short circuit 
for example), the corresponding log message changes or goes from info to debug 
etc. Logs are not a reliable way to programmatically, for example in tests, to 
determine if a feature is turned on/off.

Also when debugging issues, routinely I have to ask for configuration to 
determine if the feature is turned on or off and hope that the configuration 
provided is the correct one loaded in the namenode.

While namenode does provide config it has loaded over http, some times either 
due to poor documentation or lack of understanding of configuration mechanism, 
people may not be able to determine if a state of the feature. I propose adding 
a web page that shows all that we print today in the logs, that feature status 
and the corresponding configuration values and feature status. This should also 
be made available over JMX and JMX http, just like all the other namenode webUI 
info, so that scripts and other programs can consume it.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4912) Cleanup FSNamesystem#startFileInternal

2013-06-17 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4912:
-

 Summary: Cleanup FSNamesystem#startFileInternal
 Key: HDFS-4912
 URL: https://issues.apache.org/jira/browse/HDFS-4912
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


FSNamesystem#startFileInternal is used by both create and append. This results 
in ugly if else conditions to consider append/create scenarios. This method can 
be refactored and the code can be cleaned up.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-2498) TestParallelRead times out consistently on Jenkins

2013-06-17 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-2498?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-2498.
---

Resolution: Not A Problem

 TestParallelRead times out consistently on Jenkins
 --

 Key: HDFS-2498
 URL: https://issues.apache.org/jira/browse/HDFS-2498
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 0.22.0
Reporter: Konstantin Shvachko
Assignee: Jakob Homan
Priority: Blocker
 Attachments: TestParallelRead.txt


 During last several Jenkins builds TestParallelRead consistently fails. See 
 Hadoop-Hdfs-22-branch for logs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4903) Print trash configuration and trash emptier state in namenode log

2013-06-13 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4903:
-

 Summary: Print trash configuration and trash emptier state in 
namenode log
 Key: HDFS-4903
 URL: https://issues.apache.org/jira/browse/HDFS-4903
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Suresh Srinivas
Priority: Minor


Namenode should print an info level log that print trash interval and if trash 
emptier is started or not.

Also fix a typo Cannot start tresh in Namenode.java.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4904) Remove JournalService

2013-06-13 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4904:
-

 Summary: Remove JournalService
 Key: HDFS-4904
 URL: https://issues.apache.org/jira/browse/HDFS-4904
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.3-alpha
Reporter: Suresh Srinivas


JournalService class was added in HDFS-3099. Since it was not used in 
HDFS-3077, which has JournalNodeRpcServer instead, I propose deleting this dead 
code.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4896) dfs -command webhdfs:// gives IllegalStateException for secure cluster

2013-06-07 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4896?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4896.
---

Resolution: Duplicate

Resolving this as duplicate of HDFS-4841.

 dfs -command webhdfs:// gives IllegalStateException for secure cluster
 

 Key: HDFS-4896
 URL: https://issues.apache.org/jira/browse/HDFS-4896
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.1.0-beta
Reporter: yeshavora
 Fix For: 2.1.0-beta


 Running:
 hadoop dfs -copyToLocal webhdfs://node1:port1/File1 /tmp/File1
 13/06/07 21:54:58 WARN util.ShutdownHookManager: ShutdownHook 
 'ClientFinalizer' failed, java.lang.IllegalStateException: Shutdown in 
 progress, cannot add a shutdownHook
 java.lang.IllegalStateException: Shutdown in progress, cannot add a 
 shutdownHook
   at 
 org.apache.hadoop.util.ShutdownHookManager.addShutdownHook(ShutdownHookManager.java:152)
   at 
 org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2401)
   at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2373)
   at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:352)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$DtRenewer.getWebHdfs(WebHdfsFileSystem.java:1003)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem$DtRenewer.cancel(WebHdfsFileSystem.java:1015)
   at org.apache.hadoop.security.token.Token.cancel(Token.java:382)
   at 
 org.apache.hadoop.fs.DelegationTokenRenewer$RenewAction.cancel(DelegationTokenRenewer.java:152)
   at 
 org.apache.hadoop.fs.DelegationTokenRenewer$RenewAction.access$200(DelegationTokenRenewer.java:58)
   at 
 org.apache.hadoop.fs.DelegationTokenRenewer.removeRenewAction(DelegationTokenRenewer.java:241)
   at 
 org.apache.hadoop.hdfs.web.WebHdfsFileSystem.close(WebHdfsFileSystem.java:824)
   at org.apache.hadoop.fs.FileSystem$Cache.closeAll(FileSystem.java:2447)
   at 
 org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer.run(FileSystem.java:2464)
   at 
 org.apache.hadoop.util.ShutdownHookManager$1.run(ShutdownHookManager.java:54)

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4892) Number of transceivers reported by the datanode is incorrect

2013-06-06 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4892:
-

 Summary: Number of transceivers reported by the datanode is 
incorrect
 Key: HDFS-4892
 URL: https://issues.apache.org/jira/browse/HDFS-4892
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.0.0-alpha
Reporter: Suresh Srinivas


Currently a datanode reports the transceiver count to namenode. Namenode 
aggregates this as TotalLoad metrics and is used for monitoring cluster 
activity. It is also used for making block placement decision, to qualify if a 
datanode is a good target.

Currently transceiver count is = 1 (for XceiverServer) + 1 * (number of 
readers) + 2 * (Number of writers in pipeline) + 1 * (number of datanode 
replications) + 1 * (number of recover blocks)

Should the number of transceiver just reflect number of readers + writers, 
instead of reporting it as is currently done. Separately we should perhaps 
report readers and writers as separate count.

This 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4893) Report number of readers and writes from Datanode to Namenode

2013-06-06 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4893:
-

 Summary: Report number of readers and writes from Datanode to 
Namenode
 Key: HDFS-4893
 URL: https://issues.apache.org/jira/browse/HDFS-4893
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 2.0.0-alpha
Reporter: Suresh Srinivas


Currently datanode reports combined number of readers and writers as 
transceiver count to the namenode in heartbeats. This jira proposes reporting 
number of readers and writers as separate fields in heartbeats.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4889) passing -Dfs.trash.interval to command line is not respected.

2013-06-06 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4889?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4889.
---

Resolution: Not A Problem

 passing -Dfs.trash.interval to command line is not respected.
 -

 Key: HDFS-4889
 URL: https://issues.apache.org/jira/browse/HDFS-4889
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.1.0-beta
Reporter: Trupti Dhavle
 Fix For: 2.1.0-beta


 Ran hadoop dfs -Dfs.trash.interval=0 -rm /user/username/README
 DEPRECATED: Use of this script to execute hdfs command is deprecated.
 Instead use the hdfs command for it.
 Moved: 'hdfs://host:port/user/username/README' to trash at: 
 hdfs://host:port/user/username/.Trash/Current
 Expected that the file doesnt go to Trash and gets deleted directly.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4808) Limit snapshot related methods to DistributedFileSystem

2013-05-08 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4808:
-

 Summary: Limit snapshot related methods to DistributedFileSystem
 Key: HDFS-4808
 URL: https://issues.apache.org/jira/browse/HDFS-4808
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Suresh Srinivas


HDFS-2802 introduces snapshots related methods into FileSystem. This jira is 
for discussing whether these snapshots related methods are needed in FileSystem 
or should be limited to only DistributedFileSystem.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4434) Provide a mapping from INodeId to INode

2013-05-02 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4434.
---

   Resolution: Fixed
Fix Version/s: (was: 3.0.0)
   2.0.5-beta

 Provide a mapping from INodeId to INode
 ---

 Key: HDFS-4434
 URL: https://issues.apache.org/jira/browse/HDFS-4434
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Suresh Srinivas
 Fix For: 2.0.5-beta

 Attachments: HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch


 This JIRA is to provide a way to access the INode via its id. The proposed 
 solution is to have an in-memory mapping from INodeId to INode. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4785) Concat operation does not removed concatenated files from InodeMap

2013-05-01 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4785:
-

 Summary: Concat operation does not removed concatenated files from 
InodeMap
 Key: HDFS-4785
 URL: https://issues.apache.org/jira/browse/HDFS-4785
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 3.0.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Concat operations takes a set of src files under a directory and concatenates 
them to a single target file. The inodes corresponding to src files are not 
deleted from the InodeMap.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4610) Move to using common utils FileUtil#setReadable/Writable/Executable and FileUtil#canRead/Write/Execute

2013-04-29 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4610.
---

   Resolution: Fixed
Fix Version/s: 3.0.0
 Hadoop Flags: Reviewed

+1 for the patch. I committed it to trunk.

Thank you Ivan. Thanks to Arpit and Chris for the reviews.

 Move to using common utils FileUtil#setReadable/Writable/Executable and 
 FileUtil#canRead/Write/Execute
 --

 Key: HDFS-4610
 URL: https://issues.apache.org/jira/browse/HDFS-4610
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Ivan Mitic
Assignee: Ivan Mitic
 Fix For: 3.0.0

 Attachments: HDFS-4610.commonfileutils.2.patch, 
 HDFS-4610.commonfileutils.3.patch, HDFS-4610.commonfileutils.patch


 Switch to using common utils described in HADOOP-9413 that work well 
 cross-platform.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HDFS-4610) Move to using common utils FileUtil#setReadable/Writable/Executable and FileUtil#canRead/Write/Execute

2013-04-29 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4610?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas reopened HDFS-4610:
---


 Move to using common utils FileUtil#setReadable/Writable/Executable and 
 FileUtil#canRead/Write/Execute
 --

 Key: HDFS-4610
 URL: https://issues.apache.org/jira/browse/HDFS-4610
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Ivan Mitic
Assignee: Ivan Mitic
 Fix For: 3.0.0

 Attachments: HDFS-4610.commonfileutils.2.patch, 
 HDFS-4610.commonfileutils.3.patch, HDFS-4610.commonfileutils.patch


 Switch to using common utils described in HADOOP-9413 that work well 
 cross-platform.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4777) File creation code in namenode is incorrectly synchronized

2013-04-29 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4777:
-

 Summary: File creation code in namenode is incorrectly synchronized
 Key: HDFS-4777
 URL: https://issues.apache.org/jira/browse/HDFS-4777
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.0-alpha, 0.23.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Blocker


FSNamesystem#startFileInternal calls delete. Delete method releases the write 
lock, making parts of startFileInternal code unintentionally executed without 
write lock being held.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HDFS-4489) Use InodeID as as an identifier of a file in HDFS protocols and APIs

2013-04-25 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4489?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas reopened HDFS-4489:
---


 Use InodeID as as an identifier of a file in HDFS protocols and APIs
 

 Key: HDFS-4489
 URL: https://issues.apache.org/jira/browse/HDFS-4489
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 2.0.5-beta


 The benefit of using InodeID to uniquely identify a file can be multiple 
 folds. Here are a few of them:
 1. uniquely identify a file cross rename, related JIRAs include HDFS-4258, 
 HDFS-4437.
 2. modification checks in tools like distcp. Since a file could have been 
 replaced or renamed to, the file name and size combination is no t reliable, 
 but the combination of file id and size is unique.
 3. id based protocol support (e.g., NFS)
 4. to make the pluggable block placement policy use fileid instead of 
 filename (HDFS-385).

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HDFS-4434) Provide a mapping from INodeId to INode

2013-04-25 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4434?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas reopened HDFS-4434:
---


 Provide a mapping from INodeId to INode
 ---

 Key: HDFS-4434
 URL: https://issues.apache.org/jira/browse/HDFS-4434
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Suresh Srinivas
 Fix For: 3.0.0, 2.0.5-beta

 Attachments: HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, HDFS-4434.patch, 
 HDFS-4434.patch


 This JIRA is to provide a way to access the INode via its id. The proposed 
 solution is to have an in-memory mapping from INodeId to INode. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4703) Permission checker should not expose resolved paths

2013-04-16 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4703:
-

 Summary: Permission checker should not expose resolved paths
 Key: HDFS-4703
 URL: https://issues.apache.org/jira/browse/HDFS-4703
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Suresh Srinivas
 Attachments: HDFS-4703.patch

Namenode currently prints inode information in AccessControlException. When 
path provided by user is different from the actual path is resolved to a 
different path in namenode (some examples as in snapshot paths, HDFS-4434), 
this might expose information about resolved path.

In this jira, in permission checker, I propose adding a separate field for path 
to print in AccessControlException.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4413) Secondary namenode won't start if HDFS isn't the default file system

2013-04-09 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4413?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4413.
---

   Resolution: Fixed
Fix Version/s: 1.2.0
 Hadoop Flags: Reviewed

I have committed this patch to branch-1, 1.2 and 1-win. Thank you Mostafa.

Thank you Chris for reviewing the patch.

 Secondary namenode won't start if HDFS isn't the default file system
 

 Key: HDFS-4413
 URL: https://issues.apache.org/jira/browse/HDFS-4413
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 1-win, 1.2.1
Reporter: Mostafa Elhemali
Assignee: Mostafa Elhemali
 Fix For: 1.2.0

 Attachments: HDFS-4413.branch-1.patch


 If HDFS is not the default file system (fs.default.name is something other 
 than hdfs://...), then secondary namenode throws early on in its 
 initialization. This is a needless check as far as I can tell, and blocks 
 scenarios where HDFS services are up but HDFS is not the default file system.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4676) TestHDFSFileSystemContract should set MiniDFSCluster variable to null to free up memory

2013-04-08 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4676:
-

 Summary: TestHDFSFileSystemContract should set MiniDFSCluster 
variable to null to free up memory
 Key: HDFS-4676
 URL: https://issues.apache.org/jira/browse/HDFS-4676
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Minor


TestHDFSFileSystemContract should reset the cluster member to null in order to  
make garbage collection quickly collect large chunk of memory associated with 
MiniDFSCluster. This avoids OutOfMemory errors.

See 
https://issues.apache.org/jira/browse/HDFS-4434?focusedCommentId=13624246page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13624246
 and the next jenkins tests where the OOM was fixed.


--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4645) Move from randomly generated block ID to sequentially generated block ID

2013-03-27 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4645:
-

 Summary: Move from randomly generated block ID to sequentially 
generated block ID
 Key: HDFS-4645
 URL: https://issues.apache.org/jira/browse/HDFS-4645
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.0.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Currently block IDs are randomly generated. This means there is no pattern to 
block ID generation and no guarantees such as uniqueness of block ID for the 
life time of the system can be made. I propose using SequentialNumber for block 
ID generation.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4635) Move BlockManager#computeCapacity to LightWeightGSet

2013-03-25 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4635:
-

 Summary: Move BlockManager#computeCapacity to LightWeightGSet
 Key: HDFS-4635
 URL: https://issues.apache.org/jira/browse/HDFS-4635
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


The computeCapacity in BlockManager that calculates the LightWeightGSet 
capacity as the percentage of total JVM memory should be moved to 
LightWeightGSet. This helps in other maps that are based on the GSet to make 
use of the same functionality.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4454) Enable offline image viewers to show file Id

2013-03-25 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4454?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4454.
---

Resolution: Duplicate

This could be done as HDFS-4339.

 Enable offline image viewers to show file Id
 

 Key: HDFS-4454
 URL: https://issues.apache.org/jira/browse/HDFS-4454
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: tools
Reporter: Brandon Li
Assignee: Brandon Li



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4469) Inculde file id in these ClientProtocol RPCs which were originally only using path to identify a file

2013-03-25 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4469.
---

Resolution: Invalid

Given the design from HDFS-4489, a variant of protocol methods to include 
INodeID to identify the inode is no longer required. Existing path will be used 
for carrying the inode information and will be part of HDFS-4434.

 Inculde file id in these ClientProtocol RPCs which were originally only using 
 path to identify a file
 -

 Key: HDFS-4469
 URL: https://issues.apache.org/jira/browse/HDFS-4469
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li

 Like HDFS-4340(add fileId to addBlock), this JIRA is to track the change to 
 add fileId to other RPC calls, such as append, complete and etc.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4398) Change WebHDFS to support file ID

2013-03-25 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4398?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4398.
---

Resolution: Invalid

This is no longer needed. Path will be used for carrying inode information as 
proposed in HDFS-4489. So no protocol change or WebHDFS changes required. I am 
also going make this a task so that it does not clutter the subtask list of 
HDFS-4489.

 Change WebHDFS to support file ID
 -

 Key: HDFS-4398
 URL: https://issues.apache.org/jira/browse/HDFS-4398
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: webhdfs
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li



--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4630) Datanode is going OOM due to small files in hdfs

2013-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4630?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4630.
---

Resolution: Invalid

 Datanode is going OOM due to small files in hdfs
 

 Key: HDFS-4630
 URL: https://issues.apache.org/jira/browse/HDFS-4630
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, namenode
Affects Versions: 2.0.0-alpha
 Environment: Ubuntu, Java 1.6
Reporter: Ankush Bhatiya
Priority: Blocker

 Hi, 
 We have very small files(size ranging 10KB-1MB) in our hdfs and no of files 
 are in tens of millions. Due to this namenode and datanode both going out of 
 memory very frequently. When we analyse the head dump of datanode most of the 
 memory was used by ReplicaMap. 
 Can we use EhCache or other to not to store all the data in memory? 
 Thanks
 Ankush

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4627) Fix FSImageFormat#Loader NPE and synchronization issues

2013-03-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4627?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4627.
---

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)
 Hadoop Flags: Reviewed

I committed the patch to HDFS-2802 branch. Thank you Jing.

 Fix FSImageFormat#Loader NPE and synchronization issues
 ---

 Key: HDFS-4627
 URL: https://issues.apache.org/jira/browse/HDFS-4627
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Jing Zhao
Assignee: Jing Zhao
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4627.001.patch, HDFS-4627.002.patch


 1. Currently FSImageFormat#Loader#loadFilesUnderConstruction calls 
 FSDirectory#unprotectedReplaceINodeFile without acquiring FSDirectory's lock. 
 FSDirectory#replaceINodeFile should be called instead.
 2. Currently FSImage#Loader#loadINode does not create block[] object (i.e., 
 keep it as null) when the length of the block is 0. This will cause NPE when 
 the loaded INodeFile is empty, and then an append operation on this inode is 
 applied from edit log.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4622) Remove redundant synchronized from FSNamesystem#rollEditLog in branch-1

2013-03-21 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4622?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4622.
---

   Resolution: Fixed
Fix Version/s: 1.2.1
 Hadoop Flags: Reviewed

+1 for the patch. I committed it to branch-1. Thank you Jing.

 Remove redundant synchronized from FSNamesystem#rollEditLog in branch-1
 ---

 Key: HDFS-4622
 URL: https://issues.apache.org/jira/browse/HDFS-4622
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 1.2.0
Reporter: Jing Zhao
Assignee: Jing Zhao
Priority: Trivial
 Fix For: 1.2.1

 Attachments: HDFS-4622.b1.001.patch


 HDFS-4222 moves checkSuperuserPrivilege out of the namespace lock scope but 
 missed the rollEditLog method.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4602) TestBookKeeperHACheckpoints fails

2013-03-14 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4602:
-

 Summary: TestBookKeeperHACheckpoints fails
 Key: HDFS-4602
 URL: https://issues.apache.org/jira/browse/HDFS-4602
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Suresh Srinivas


See https://builds.apache.org/job/Hadoop-Hdfs-trunk/1344/

Failed tests:   
testStandbyExceptionThrownDuringCheckpoint(org.apache.hadoop.contrib.bkjournal.TestBookKeeperHACheckpoints):
 SBN should have still been checkpointing.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4144) Create test for all snapshot-related metrics

2013-03-13 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4144?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4144.
---

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)

I committed the patch. Thank you Jing.

 Create test for all snapshot-related metrics
 

 Key: HDFS-4144
 URL: https://issues.apache.org/jira/browse/HDFS-4144
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Reporter: Jing Zhao
Assignee: Jing Zhao
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4144.001.patch, HDFS-4144.002.patch, 
 HDFS-4144.003.patch


 Add testcases for all snapshot-related metrics.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4595) When short circuit read is disallowed, DFSClient does disable short circuit and fall back to regular reads

2013-03-12 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4595:
-

 Summary: When short circuit read is disallowed, DFSClient does 
disable short circuit and fall back to regular reads
 Key: HDFS-4595
 URL: https://issues.apache.org/jira/browse/HDFS-4595
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Affects Versions: 2.0.0-alpha
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


In 1.0, when short circuit is disallowed for a user, the DFSClient disables 
short circuit and falls back to the regular socket based reads over 
DataTransferProtocol. In release 2.0, this fallback functionality is not 
working.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4535) port HDFS-1457 to branch-1.1

2013-03-11 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4535?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4535.
---

Resolution: Won't Fix

 port HDFS-1457 to branch-1.1
 

 Key: HDFS-4535
 URL: https://issues.apache.org/jira/browse/HDFS-4535
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 1.1.2
Reporter: Xiaobo Peng
Assignee: Xiaobo Peng
Priority: Minor
 Attachments: HDFS-4535-branch-1.1.patch


 port HDFS-1457 (configuration option to enable limiting the transfer rate 
 used when sending the image and edits for checkpointing) to branch-1.1

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4579) Annotate snapshot tests

2013-03-09 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4579?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4579.
---

  Resolution: Fixed
Hadoop Flags: Reviewed

+1 for the patch. I committed it. Thank you Arpit.

 Annotate snapshot tests
 ---

 Key: HDFS-4579
 URL: https://issues.apache.org/jira/browse/HDFS-4579
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: test
Affects Versions: Snapshot (HDFS-2802)
Reporter: Arpit Agarwal
Assignee: Arpit Agarwal
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4579.patch


 Add annotations to snapshot tests, required to merge into trunk.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-3578) missing hadoop.css

2013-03-07 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-3578?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-3578.
---

   Resolution: Duplicate
Fix Version/s: 2.0.2-alpha

Duplicate of HDFS-1013.

 missing hadoop.css
 --

 Key: HDFS-3578
 URL: https://issues.apache.org/jira/browse/HDFS-3578
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Roman Shaposhnik
Assignee: Roman Shaposhnik
Priority: Minor
 Fix For: 2.0.2-alpha


 From a report by Konstantin Olchanski:
 Web interface of HDFS is missing the file hadoop.css and the web pages
 look funny. This is the fix:
 [root@ladd12 ~]# locate hadoop.css
 /usr/lib/hadoop-0.20-mapreduce/webapps/static/hadoop.css
 [root@ladd12 ~]# mkdir /usr/lib/hadoop-hdfs/webapps/static
 [root@ladd12 ~]# cp `locate hadoop.css` /usr/lib/hadoop-hdfs/webapps/static

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4222) NN is unresponsive and lose heartbeats of DNs when Hadoop is configured to use LDAP and LDAP has issues

2013-02-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4222?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4222.
---

   Resolution: Fixed
Fix Version/s: 1.2.0
 Hadoop Flags: Reviewed

I committed the patch to branch-1 as well. Thank you Xiaobo!

 NN is unresponsive and lose heartbeats of DNs when Hadoop is configured to 
 use LDAP and LDAP has issues
 ---

 Key: HDFS-4222
 URL: https://issues.apache.org/jira/browse/HDFS-4222
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 1.0.0, 0.23.3, 2.0.0-alpha
Reporter: Xiaobo Peng
Assignee: Xiaobo Peng
Priority: Minor
 Fix For: 1.2.0, 0.23.7, 2.0.4-beta

 Attachments: HDFS-4222.23.patch, hdfs-4222-branch-0.23.3.patch, 
 HDFS-4222-branch-1.patch, HDFS-4222.patch, HDFS-4222.patch, 
 hdfs-4222-release-1.0.3.patch


 For Hadoop clusters configured to access directory information by LDAP, the 
 FSNamesystem calls on behave of DFS clients might hang due to LDAP issues 
 (including LDAP access issues caused by networking issues) while holding the 
 single lock of FSNamesystem. That will result in the NN unresponsive and loss 
 of the heartbeats from DNs.
 The places LDAP got accessed by FSNamesystem calls are the instantiation of 
 FSPermissionChecker, which could be moved out of the lock scope since the 
 instantiation does not need the FSNamesystem lock. After the move, a DFS 
 client hang will not affect other threads by hogging the single lock. This is 
 especially helpful when we use separate RPC servers for ClientProtocol and 
 DatanodeProtocol since the calls for DatanodeProtocol do not need to access 
 LDAP. So even if DFS clients hang due to LDAP issues, the NN will still be 
 able to process the requests (including heartbeats) from DNs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4479) logSync() with the FSNamesystem lock held in commitBlockSynchronization

2013-02-24 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4479?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4479.
---

   Resolution: Fixed
Fix Version/s: 1.2.0
 Hadoop Flags: Reviewed

I committed the patch to branch-1. Thank you Jing.

 logSync() with the FSNamesystem lock held in commitBlockSynchronization
 ---

 Key: HDFS-4479
 URL: https://issues.apache.org/jira/browse/HDFS-4479
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 1.2.0
Reporter: Jing Zhao
Assignee: Jing Zhao
 Fix For: 1.2.0

 Attachments: HDFS-4479.b1.001.patch, HDFS-4479.b1.002.patch


 In FSNamesystem#commitBlockSynchronization of branch-1, logSync() may be 
 called when the FSNamesystem lock is held. Similar with HDFS-4186, this may 
 cause some performance issue.
 Since logSync is called right after the synchronization section, we can 
 simply remove the logSync call.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4340) Update addBlock() to inculde inode id as additional argument

2013-02-14 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4340?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4340.
---

Resolution: Fixed

Marking this issue back as resolved.

 Update addBlock() to inculde inode id as additional argument
 

 Key: HDFS-4340
 URL: https://issues.apache.org/jira/browse/HDFS-4340
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client, namenode
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 3.0.0

 Attachments: HDFS-4340.patch, HDFS-4340.patch, HDFS-4340.patch, 
 HDFS-4340.patch, HDFS-4340.patch, HDFS-4340.patch, HDFS-4340.patch, 
 HDFS-4340.patch, HDFS-4340.patch, HDFS-4340.patch, HDFS-4340.patch




--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4466) Remove the deadlock from AbstractDelegationTokenSecretManager

2013-02-13 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4466?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4466.
---

   Resolution: Fixed
Fix Version/s: 1.2.0
 Hadoop Flags: Reviewed

I committed the patch to branch-1. Thank you Brandon!

 Remove the deadlock from AbstractDelegationTokenSecretManager
 -

 Key: HDFS-4466
 URL: https://issues.apache.org/jira/browse/HDFS-4466
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode, security
Affects Versions: 1.2.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 1.2.0

 Attachments: HDFS-4466.branch-1.patch


 In HDFS-3374, new synchronization in 
 AbstractDelegationTokenSecretManager.ExpiredTokenRemover was added to make 
 sure the ExpiredTokenRemover thread can be interrupted in time. Otherwise 
 TestDelegation fails intermittently because the MiniDFScluster thread could 
 be shut down before tokenRemover thread. 
 However, as Todd pointed out in HDFS-3374, a potential deadlock was 
 introduced by its patch:
 {quote}
* FSNamesystem.saveNamespace (holding FSN lock) calls 
 DTSM.saveSecretManagerState (which takes DTSM lock)
* ExpiredTokenRemover.run (holding DTSM lock) calls rollMasterKey calls 
 updateCurrentKey calls logUpdateMasterKey which takes FSN lock
 So if there is a concurrent saveNamespace at the same tie as the expired 
 token remover runs, it might make the NN deadlock. {quote}
 This JIRA is to track the change of removing the possible deadlock from 
 AbstractDelegationTokenSecretManager. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4475) OutOfMemory by BPServiceActor.offerService() takes down DataNode

2013-02-12 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4475?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4475.
---

Resolution: Invalid

 OutOfMemory by BPServiceActor.offerService() takes down DataNode
 

 Key: HDFS-4475
 URL: https://issues.apache.org/jira/browse/HDFS-4475
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0, 2.0.3-alpha
Reporter: Plamen Jeliazkov
Assignee: Plamen Jeliazkov
 Fix For: 3.0.0, 2.0.3-alpha


 In DataNode, there are catchs around BPServiceActor.offerService() call but 
 no catch for OutOfMemory as there is for the DataXeiver as introduced in 
 0.22.0.
 The issue can be replicated like this:
 1) Create a cluster of X DataNodes and 1 NameNode and low memory settings 
 (-Xmx128M or something similar).
 2) Flood HDFS with small file creations (any should work actually).
 3) DataNodes will hit OoM, stop blockpool service, and shutdown.
 The resolution is to catch the OoMException and handle it properly when 
 calling BPServiceActor.offerService() in DataNode.java; like as done in 
 0.22.0 of Hadoop. DataNodes should not shutdown or crash but remain in a sort 
 of frozen state until memory issues are resolved by GC.
 LOG ERROR:
 2013-02-04 11:46:01,854 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: 
 Unexpected exception in block pool Block pool 
 BP-1105714849-10.10.10.110-1360005776467 (storage id 
 DS-1952316202-10.10.10.112-50010-1360005820993) service to 
 vmhost2-vm0/10.10.10.110:8020
 java.lang.OutOfMemoryError: GC overhead limit exceeded
 2013-02-04 11:46:01,854 WARN org.apache.hadoop.hdfs.server.datanode.DataNode: 
 Ending block pool service for: Block pool 
 BP-1105714849-10.10.10.110-1360005776467 (storage id 
 DS-1952316202-10.10.10.112-50010-1360005820993) service to 
 vmhost2-vm0/10.10.10.110:8020

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4414) Add support for getting snapshot diff from DistributedFileSystem

2013-02-02 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4414?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4414.
---

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)
 Hadoop Flags: Reviewed

Committed the patch to HDFS-2802 branch. Thank you Jing!

 Add support for getting snapshot diff from DistributedFileSystem
 

 Key: HDFS-4414
 URL: https://issues.apache.org/jira/browse/HDFS-4414
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Reporter: Jing Zhao
Assignee: Jing Zhao
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4414.001.patch, HDFS-4414.003.patch, 
 HDFS-4414.004.patch, HDFS-4414.005.patch, HDFS-4414+4131.002.patch


 HDFS-4131 computes the difference between two snapshots (or between a 
 snapshot and the current tree). In this jira we create a DiffReport class to 
 represent the diff to end users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4343) When storageID of dfs.data.dir of being inconsistent, restart datanode will be failure.

2013-02-02 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4343?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4343.
---

Resolution: Invalid

Closing this jira as Invalid. The behavior in the description is the expected 
behavior. When storage directories are in inconsistent state, invalid or stale 
directories are expected to be cleaned manually, before restarting a datanode.

If you disagree, feel free to reopen the jira. Please justify why you think it 
is a valid jira, when you reopen the jira.

 When storageID of dfs.data.dir of being inconsistent, restart datanode will 
 be failure.
 ---

 Key: HDFS-4343
 URL: https://issues.apache.org/jira/browse/HDFS-4343
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode
 Environment: namenode datanode 
Reporter: liuyang
 Attachments: hadoop-root-datanode-167-52-0-55.log, VERSION-1, 
 VERSION-2


 A datanode has multiple storage directories configured using dfs.data.dir. 
 When the storageID in the VERSION files in these directories, the datanode 
 fails to startup. Consider a scenario, when old data in a storage directory 
 is not cleared, the storage ID from it will not match with storage ID of in 
 other storage storage directories. In this situation, the DataNode will quit 
 and restart fails.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4465) Optimize datanode ReplicasMap and ReplicaInfo

2013-02-01 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4465:
-

 Summary: Optimize datanode ReplicasMap and ReplicaInfo
 Key: HDFS-4465
 URL: https://issues.apache.org/jira/browse/HDFS-4465
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: datanode
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


In Hadoop a lot of optimization has been done in namenode data structures to be 
memory efficient. Similar optimizations are necessary for Datanode process. 
With the growth in storage per datanode and number of blocks hosted on 
datanode, this jira intends to optimize long lived ReplicasMap and ReplicaInfo 
objects.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4131) Add capability to namenode to get snapshot diff

2013-01-29 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4131?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4131.
---

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)
 Hadoop Flags: Reviewed

I committed the change to the branch.

Thank you Jing!

 Add capability to namenode to get snapshot diff
 ---

 Key: HDFS-4131
 URL: https://issues.apache.org/jira/browse/HDFS-4131
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: Snapshot (HDFS-2802)
Reporter: Suresh Srinivas
Assignee: Jing Zhao
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4131.001.patch, HDFS-4131.002.patch, 
 HDFS-4131.003.patch, HDFS-4131.004.patch, HDFS-4131.005.patch, 
 HDFS-4131.006.patch


 This jira provides internal data structures and computation processes for 
 calculating and representing the diff between two snapshots, or the diff 
 between a snapshot and the current tree. 
 Specifically, a new method getSnapshotDiffReport(Path, String, String) is 
 added to FSNamesystem to compute the snapshot diff, which is then later 
 represented as a SnapshotDiffReport. In later jiras we will add support to 
 represent the SnapshotDiffReport to end users.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4432) Support INodeFileUnderConstructionWithSnapshot in FSImage saving/loading

2013-01-28 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4432?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4432.
---

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)
 Hadoop Flags: Reviewed

I committed the patch. Thank you Jing!

 Support INodeFileUnderConstructionWithSnapshot in FSImage saving/loading
 

 Key: HDFS-4432
 URL: https://issues.apache.org/jira/browse/HDFS-4432
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Reporter: Jing Zhao
Assignee: Jing Zhao
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4432.001.patch, HDFS-4432.002.patch, 
 HDFS-4432.003.patch


 1. FSImage saver/loader need to be able to recognize 
 INodeFileUnderConstruction, INodeFileUnderConstructionWithSnapshot, and 
 INodeFileUnderConstructionSnapshot.
 2. FSImage saver/loader should be able to form the correct circular linked 
 list for INodeFileUnderConstruction(With)Snapshot.
 3. Add new corresponding unit tests and support file appending in 
 TestSnapshot#testSnapshot.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HDFS-4126) Add reading/writing snapshot information to FSImage

2013-01-22 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-4126?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas resolved HDFS-4126.
---

   Resolution: Fixed
Fix Version/s: Snapshot (HDFS-2802)
 Hadoop Flags: Reviewed

I committed the patch to HDFS-2802 branch.

Thank you Jing!

 Add reading/writing snapshot information to FSImage
 ---

 Key: HDFS-4126
 URL: https://issues.apache.org/jira/browse/HDFS-4126
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, namenode
Affects Versions: Snapshot (HDFS-2802)
Reporter: Suresh Srinivas
Assignee: Jing Zhao
 Fix For: Snapshot (HDFS-2802)

 Attachments: HDFS-4126.001.patch, HDFS-4126.002.patch, 
 HDFS-4126.002.patch, HDFS-4126.003.patch


 After the changes proposed in HDFS-4125 is completed, reading and writing 
 snapshot related information from FSImage can be implemented. This jira 
 tracks changes required for:
 # Loading snapshot information from FSImage
 # Loading snapshot related operations from editlog
 # Writing snapshot information in FSImage
 # Unit tests related to this functionality

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4375) Use token request messages defined in hadoop common

2013-01-09 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4375:
-

 Summary: Use token request messages defined in hadoop common
 Key: HDFS-4375
 URL: https://issues.apache.org/jira/browse/HDFS-4375
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode, security
Affects Versions: 2.0.2-alpha
Reporter: Suresh Srinivas


HDFS changes related to HADOOP-9192 to reuse the protobuf messages defined in 
common.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4367) GetDataEncryptionKeyResponseProto does not handle null response

2013-01-08 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4367:
-

 Summary: GetDataEncryptionKeyResponseProto  does not handle null 
response
 Key: HDFS-4367
 URL: https://issues.apache.org/jira/browse/HDFS-4367
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.2-alpha
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Blocker


GetDataEncryptionKeyResponseProto member dataEncryptionKey should be optional 
to handle null response.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4369) GetBlockKeysResponseProto does not handle null response

2013-01-08 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4369:
-

 Summary: GetBlockKeysResponseProto does not handle null response
 Key: HDFS-4369
 URL: https://issues.apache.org/jira/browse/HDFS-4369
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Affects Versions: 2.0.2-alpha
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Blocker


GetBlockKeysResponseProto#keys should be optional to handle null response

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4371) Move token related request/response messages to common

2013-01-08 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4371:
-

 Summary: Move token related request/response messages to common
 Key: HDFS-4371
 URL: https://issues.apache.org/jira/browse/HDFS-4371
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: security
Affects Versions: 2.0.2-alpha
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


Get, Renew and Cancel delegation token requests and responses are repeated in 
HDFS, Yarn and MR. This jira proposes to move these messages into 
Security.proto in common.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Created] (HDFS-4364) GetLinkTargetResponseProto does not handle null path

2013-01-07 Thread Suresh Srinivas (JIRA)
Suresh Srinivas created HDFS-4364:
-

 Summary: GetLinkTargetResponseProto does not handle null path
 Key: HDFS-4364
 URL: https://issues.apache.org/jira/browse/HDFS-4364
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.0.2-alpha
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas


ClientProtocol#getLinkTarget() returns null targetPath. Hence protobuf 
definition GetLinkTargetResponseProto#targetPath should be optional.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


  1   2   3   >