[jira] [Commented] (HADOOP-8817) Backport Network Topology Extension for Virtualization (HADOOP-8468) to branch-1

2012-09-17 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8817?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13456889#comment-13456889
 ] 

Junping Du commented on HADOOP-8817:


Thanks Suresh for the correction here. So branch-1.2 is next feature branch 
(1.1 is next bug-fix branch). Isn't it?

 Backport Network Topology Extension for Virtualization (HADOOP-8468) to 
 branch-1
 

 Key: HADOOP-8817
 URL: https://issues.apache.org/jira/browse/HADOOP-8817
 Project: Hadoop Common
  Issue Type: Sub-task
Affects Versions: 1.0.0
Reporter: Junping Du
Assignee: Junping Du
  Labels: features
 Attachments: HADOOP-8817.patch


 HADOOP-8468 propose network topology changes for running on virtualized 
 infrastructure, which includes:
 1. Add NodeGroup layer in new NetworkTopology (also known as 
 NetworkTopologyWithNodeGroup): HADOOP-8469, HADOOP-8470
 2. Update Replica Placement/Removal Policy to reflect new topology layer: 
 HDFS-3498, HDFS-3601
 3. Update balancer policy:HDFS-3495
 4. Update Task Scheduling Policy to reflect new topology layer and support 
 the case that compute nodes (NodeManager or TaskTracker) and data nodes are 
 separated into different VMs, but still benefit from physical host locality: 
 YARN-18, YARN-19.
 This JIRA will address the backport work on branch-1 which will be divided 
 into 4 issues/patches in related jira issues.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8468) Umbrella of enhancements to support different failure and locality topologies

2012-09-17 Thread Junping Du (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13456891#comment-13456891
 ] 

Junping Du commented on HADOOP-8468:


Thanks for great comments. Konstantin. the doc revised-1.0 already address full 
policy definition.
Hi guys, I am back porting patches to branch-1. Hope I can get your support and 
help on reviewing. :)

 Umbrella of enhancements to support different failure and locality topologies
 -

 Key: HADOOP-8468
 URL: https://issues.apache.org/jira/browse/HADOOP-8468
 Project: Hadoop Common
  Issue Type: Improvement
  Components: ha, io
Affects Versions: 1.0.0, 2.0.0-alpha
Reporter: Junping Du
Assignee: Junping Du
 Attachments: HADOOP-8468-total.patch, HADOOP-8468-total-v3.patch, 
 Proposal for enchanced failure and locality topologies.pdf, Proposal for 
 enchanced failure and locality topologies (revised-1.0).pdf


 The current hadoop network topology (described in some previous issues like: 
 Hadoop-692) works well in classic three-tiers network when it comes out. 
 However, it does not take into account other failure models or changes in the 
 infrastructure that can affect network bandwidth efficiency like: 
 virtualization. 
 Virtualized platform has following genes that shouldn't been ignored by 
 hadoop topology in scheduling tasks, placing replica, do balancing or 
 fetching block for reading: 
 1. VMs on the same physical host are affected by the same hardware failure. 
 In order to match the reliability of a physical deployment, replication of 
 data across two virtual machines on the same host should be avoided.
 2. The network between VMs on the same physical host has higher throughput 
 and lower latency and does not consume any physical switch bandwidth.
 Thus, we propose to make hadoop network topology extend-able and introduce a 
 new level in the hierarchical topology, a node group level, which maps well 
 onto an infrastructure that is based on a virtualized environment.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8818) Should use equals() rather than == to compare String or Text in MD5MD5CRC32FileChecksum and TFileDumper

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8818?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13456980#comment-13456980
 ] 

Hudson commented on HADOOP-8818:


Integrated in Hadoop-Hdfs-trunk #1168 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1168/])
HADOOP-8818. Use equals instead == in MD5MD5CRC32FileChecksum and 
TFileDumper. Contributed by Brandon Li. (Revision 1385374)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1385374
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/MD5MD5CRC32FileChecksum.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/file/tfile/TFileDumper.java


 Should use equals() rather than == to compare String or Text in 
 MD5MD5CRC32FileChecksum and TFileDumper
 ---

 Key: HADOOP-8818
 URL: https://issues.apache.org/jira/browse/HADOOP-8818
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs, io
Affects Versions: 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
Priority: Minor
 Fix For: 3.0.0

 Attachments: HADOOP-8818.patch


 Should use equals() rather than == to compare String or Text in 
 MD5MD5CRC32FileChecksum and TFileDumper.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8821) Findbugs warning Configuration.dumpDeprecatedKeys() concatenates strings using + in a loop

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8821?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13456981#comment-13456981
 ] 

Hudson commented on HADOOP-8821:


Integrated in Hadoop-Hdfs-trunk #1168 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1168/])
HADOOP-8821. Fix findbugs warning related concatenating string in a for 
loop in Configuration#dumpDeprecatedKeys(). Contributed by Suresh Srinivas. 
(Revision 1385389)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1385389
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java


 Findbugs warning Configuration.dumpDeprecatedKeys() concatenates strings 
 using + in a loop
 --

 Key: HADOOP-8821
 URL: https://issues.apache.org/jira/browse/HADOOP-8821
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Trivial
 Fix For: 3.0.0

 Attachments: HADOOP-8821.patch


 Configuration.dumpDeprecatedKeys() concatenates strings using + in a loop. 
 Instead use StringBuilder.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8819) Should use instead of in a few places in FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13456982#comment-13456982
 ] 

Hudson commented on HADOOP-8819:


Integrated in Hadoop-Hdfs-trunk #1168 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1168/])
HADOOP-8819. Incorrectly  is used instead of  in some file system 
implementations. Contributed by Brandon Li. (Revision 1386451)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386451
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPInputStream.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/s3/S3InputStream.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs
 ---

 Key: HADOOP-8819
 URL: https://issues.apache.org/jira/browse/HADOOP-8819
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8819.patch


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8819) Should use instead of in a few places in FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457032#comment-13457032
 ] 

Hudson commented on HADOOP-8819:


Integrated in Hadoop-Mapreduce-trunk #1199 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1199/])
HADOOP-8819. Incorrectly  is used instead of  in some file system 
implementations. Contributed by Brandon Li. (Revision 1386451)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386451
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/ftp/FTPInputStream.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/s3/S3InputStream.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs
 ---

 Key: HADOOP-8819
 URL: https://issues.apache.org/jira/browse/HADOOP-8819
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8819.patch


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8821) Findbugs warning Configuration.dumpDeprecatedKeys() concatenates strings using + in a loop

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8821?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457031#comment-13457031
 ] 

Hudson commented on HADOOP-8821:


Integrated in Hadoop-Mapreduce-trunk #1199 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1199/])
HADOOP-8821. Fix findbugs warning related concatenating string in a for 
loop in Configuration#dumpDeprecatedKeys(). Contributed by Suresh Srinivas. 
(Revision 1385389)

 Result = SUCCESS
suresh : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1385389
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/conf/Configuration.java


 Findbugs warning Configuration.dumpDeprecatedKeys() concatenates strings 
 using + in a loop
 --

 Key: HADOOP-8821
 URL: https://issues.apache.org/jira/browse/HADOOP-8821
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0
Reporter: Suresh Srinivas
Assignee: Suresh Srinivas
Priority: Trivial
 Fix For: 3.0.0

 Attachments: HADOOP-8821.patch


 Configuration.dumpDeprecatedKeys() concatenates strings using + in a loop. 
 Instead use StringBuilder.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8811) Compile hadoop native library in FreeBSD

2012-09-17 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8811?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457092#comment-13457092
 ] 

Hadoop QA commented on HADOOP-8811:
---

-1 overall.  Here are the results of testing the latest attachment 
  
http://issues.apache.org/jira/secure/attachment/12545425/freebsd-native-flags-hardlink.txt
  against trunk revision .

+1 @author.  The patch does not contain any @author tags.

+1 tests included.  The patch appears to include 2 new or modified test 
files.

+1 javac.  The applied patch does not increase the total number of javac 
compiler warnings.

+1 javadoc.  The javadoc tool did not generate any warning messages.

+1 eclipse:eclipse.  The patch built with eclipse:eclipse.

+1 findbugs.  The patch does not introduce any new Findbugs (version 1.3.9) 
warnings.

+1 release audit.  The applied patch does not increase the total number of 
release audit warnings.

-1 core tests.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common:

  org.apache.hadoop.ha.TestZKFailoverController

+1 contrib tests.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/1480//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/1480//console

This message is automatically generated.

 Compile hadoop native library in FreeBSD
 

 Key: HADOOP-8811
 URL: https://issues.apache.org/jira/browse/HADOOP-8811
 Project: Hadoop Common
  Issue Type: Bug
  Components: native
Affects Versions: 0.23.0, 2.0.0-alpha, 3.0.0
Reporter: Radim Kolar
Priority: Critical
  Labels: freebsd
 Attachments: freebsd-native-flags-hardlink.txt, 
 freebsd-native-flags.txt, freebsd-native-trunk.txt, freebsd-native.txt


 Native hadoop library do not compiles in FreeBSD because setnetgrent returns 
 void and assembler do not supports SSE4 instructions.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8816) HTTP Error 413 full HEAD if using kerberos authentication

2012-09-17 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457119#comment-13457119
 ] 

Alejandro Abdelnur commented on HADOOP-8816:


128K as header buffer size seems a bit too big. 

Could this be related to this? 
http://www.novell.com/communities/node/11516/kerberos-authentication-may-fail-access-manager-identity-server-users-large-group-members

Would be possible to get the actual header size that is making things to fail?


 HTTP Error 413 full HEAD if using kerberos authentication
 -

 Key: HADOOP-8816
 URL: https://issues.apache.org/jira/browse/HADOOP-8816
 Project: Hadoop Common
  Issue Type: Bug
  Components: net
Affects Versions: 2.0.1-alpha
 Environment: ubuntu linux with active directory kerberos.
Reporter: Moritz Moeller

 The HTTP Authentication: header is too large if using kerberos and the 
 request is rejected by Jetty because Jetty has a too low default header size 
 limit.
 Can be fixed by adding ret.setHeaderBufferSize(1024*128); in 
 org.apache.hadoop.http.HttpServer.createDefaultChannelConnector

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8816) HTTP Error 413 full HEAD if using kerberos authentication

2012-09-17 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457121#comment-13457121
 ] 

Alejandro Abdelnur commented on HADOOP-8816:


Also, if we tweak the header buffer size, we should doing it in a configurable 
way. 

 HTTP Error 413 full HEAD if using kerberos authentication
 -

 Key: HADOOP-8816
 URL: https://issues.apache.org/jira/browse/HADOOP-8816
 Project: Hadoop Common
  Issue Type: Bug
  Components: net
Affects Versions: 2.0.1-alpha
 Environment: ubuntu linux with active directory kerberos.
Reporter: Moritz Moeller

 The HTTP Authentication: header is too large if using kerberos and the 
 request is rejected by Jetty because Jetty has a too low default header size 
 limit.
 Can be fixed by adding ret.setHeaderBufferSize(1024*128); in 
 org.apache.hadoop.http.HttpServer.createDefaultChannelConnector

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8816) HTTP Error 413 full HEAD if using kerberos authentication

2012-09-17 Thread Moritz Moeller (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457129#comment-13457129
 ] 

Moritz Moeller commented on HADOOP-8816:


No, Kerberos tokens do not contain group membership information, but tend to 
get pretty large, 4-8k base64 encoded.
I guess 16kb header size would be enough.

Making that configurable is your choice, I personally wouldn't as I know no 
things that cause header sizes larger than Kerberos, but then if it was 
configurable already this ticket wouldn't exist.


 HTTP Error 413 full HEAD if using kerberos authentication
 -

 Key: HADOOP-8816
 URL: https://issues.apache.org/jira/browse/HADOOP-8816
 Project: Hadoop Common
  Issue Type: Bug
  Components: net
Affects Versions: 2.0.1-alpha
 Environment: ubuntu linux with active directory kerberos.
Reporter: Moritz Moeller

 The HTTP Authentication: header is too large if using kerberos and the 
 request is rejected by Jetty because Jetty has a too low default header size 
 limit.
 Can be fixed by adding ret.setHeaderBufferSize(1024*128); in 
 org.apache.hadoop.http.HttpServer.createDefaultChannelConnector

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8819) Should use instead of in a few places in FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs

2012-09-17 Thread Brandon Li (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8819?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Brandon Li updated HADOOP-8819:
---

Attachment: HADOOP-8819.branch-1.patch

 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs
 ---

 Key: HADOOP-8819
 URL: https://issues.apache.org/jira/browse/HADOOP-8819
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8819.branch-1.patch, HADOOP-8819.patch


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8819) Should use instead of in a few places in FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs

2012-09-17 Thread Brandon Li (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457148#comment-13457148
 ] 

Brandon Li commented on HADOOP-8819:


The same problem exists in branch-1 (except viewfs). Uploaded a branch-1 patch. 
Removed created  this time. :-)

 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs
 ---

 Key: HADOOP-8819
 URL: https://issues.apache.org/jira/browse/HADOOP-8819
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8819.branch-1.patch, HADOOP-8819.patch


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8819) Should use instead of in a few places in FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs

2012-09-17 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8819?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457150#comment-13457150
 ] 

Hadoop QA commented on HADOOP-8819:
---

-1 overall.  Here are the results of testing the latest attachment 
  
http://issues.apache.org/jira/secure/attachment/12545438/HADOOP-8819.branch-1.patch
  against trunk revision .

-1 patch.  The patch command could not apply the patch.

Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/1481//console

This message is automatically generated.

 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs
 ---

 Key: HADOOP-8819
 URL: https://issues.apache.org/jira/browse/HADOOP-8819
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8819.branch-1.patch, HADOOP-8819.patch


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8733) TestStreamingTaskLog, TestJvmManager, TestLinuxTaskControllerLaunchArgs fail on Windows

2012-09-17 Thread Vinod Kumar Vavilapalli (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8733?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457152#comment-13457152
 ] 

Vinod Kumar Vavilapalli commented on HADOOP-8733:
-

TestJVMManager is simulating the kill code-path in case of linux. So yes, 
unless it is very difficult, please change the test to verify that the 
attemptid that is sent across can be successfully used for killing processes in 
case of Windows.

 TestStreamingTaskLog, TestJvmManager, TestLinuxTaskControllerLaunchArgs fail 
 on Windows
 ---

 Key: HADOOP-8733
 URL: https://issues.apache.org/jira/browse/HADOOP-8733
 Project: Hadoop Common
  Issue Type: Bug
  Components: test
Affects Versions: 1-win
Reporter: Ivan Mitic
Assignee: Ivan Mitic
 Attachments: HADOOP-8733-scripts.2.patch, 
 HADOOP-8733-scripts.2.patch, HADOOP-8733-scripts.patch


 Jira tracking test failures related to test .sh script dependencies. 

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8819) Should use instead of in a few places in FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs

2012-09-17 Thread Suresh Srinivas (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8819?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Suresh Srinivas updated HADOOP-8819:


   Resolution: Fixed
Fix Version/s: 1.2.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

+1 for the the branch-1 patch. 

Thank you Brandon!

 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs
 ---

 Key: HADOOP-8819
 URL: https://issues.apache.org/jira/browse/HADOOP-8819
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs
Affects Versions: 2.0.0-alpha, 3.0.0
Reporter: Brandon Li
Assignee: Brandon Li
 Fix For: 1.2.0, 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8819.branch-1.patch, HADOOP-8819.patch


 Should use  instead of   in a few places in 
 FTPFileSystem,FTPInputStream,S3InputStream,ViewFileSystem,ViewFs.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8615) EOFException in DecompressorStream.java needs to be more verbose

2012-09-17 Thread Andy Isaacson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457199#comment-13457199
 ] 

Andy Isaacson commented on HADOOP-8615:
---

Thomas,
Thank you for the patch!

bq. Please let me know if there is any procedure for making this tested only on 
hadoop common
release 0.20.2

Don't worry about the robots testing against the wrong branch, you're not doing 
anything wrong.

It seems to me this change would be a good thing on trunk as well. Can you port 
the patch to trunk?

bq. When the user uses this method and pass the filename, it would be printed 
in the EOF exception thrown, if any. So I believe the test cases may not be 
necessary. I was able to test it locally by forcefully creating an EOF 
Exception and verifying the new message as java.io.EOFException: Unexpected 
end of input stream in the file = filename

I think this should be fairly easy to test -- just write a compressed stream, 
truncate the compressed stream, then try to read it, catch the EOFException and 
verify that the filename shows up in the exception text.  Or am I missing 
something?

I'm a little worried about the places where your {{fileName}}-using methods add 
new default values, for example:
{code}
+  public CompressionInputStream createInputStream(InputStream in, 
+Decompressor decompressor, String fileName) 
+  throws IOException {
+return new DecompressorStream(in, decompressor, 
+   conf.getInt(io.file.buffer.size, 4*1024),fileName);
+  }
{code}
I'll have to think about it longer, but having a default value of 4k hidden in 
this method seems wrong to me at a first glance.  There are a few other 
instances of this as well.

 EOFException in DecompressorStream.java needs to be more verbose
 

 Key: HADOOP-8615
 URL: https://issues.apache.org/jira/browse/HADOOP-8615
 Project: Hadoop Common
  Issue Type: Bug
  Components: io
Affects Versions: 0.20.2
Reporter: Jeff Lord
  Labels: patch
 Attachments: HADOOP-8615-release-0.20.2.patch


 In ./src/core/org/apache/hadoop/io/compress/DecompressorStream.java
 The following exception should at least pass back the file that it encounters 
 this error in relation to:
   protected void getCompressedData() throws IOException {
 checkStream();
 int n = in.read(buffer, 0, buffer.length);
 if (n == -1) {
   throw new EOFException(Unexpected end of input stream);
 }
 This would help greatly to debug bad/corrupt files.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Reopened] (HADOOP-7688) When a servlet filter throws an exception in init(..), the Jetty server failed silently.

2012-09-17 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-7688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G reopened HADOOP-7688:
-


 When a servlet filter throws an exception in init(..), the Jetty server 
 failed silently. 
 -

 Key: HADOOP-7688
 URL: https://issues.apache.org/jira/browse/HADOOP-7688
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 0.23.0, 0.24.0
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Uma Maheswara Rao G
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: filter-init-exception-test.patch, 
 HADOOP-7688-branch-1.patch, HADOOP-7688-branch-2.patch, HADOOP-7688.patch, 
 org.apache.hadoop.http.TestServletFilter-output.txt


 When a servlet filter throws a ServletException in init(..), the exception is 
 logged by Jetty but not re-throws to the caller.  As a result, the Jetty 
 server failed silently.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-7688) When a servlet filter throws an exception in init(..), the Jetty server failed silently.

2012-09-17 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-7688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HADOOP-7688:


Attachment: HADOOP-7688-branch-1.patch

I have just committed into branch-1

 When a servlet filter throws an exception in init(..), the Jetty server 
 failed silently. 
 -

 Key: HADOOP-7688
 URL: https://issues.apache.org/jira/browse/HADOOP-7688
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 0.23.0, 0.24.0
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Uma Maheswara Rao G
 Fix For: 3.0.0, 2.0.3-alpha

 Attachments: filter-init-exception-test.patch, 
 HADOOP-7688-branch-1.patch, HADOOP-7688-branch-2.patch, HADOOP-7688.patch, 
 org.apache.hadoop.http.TestServletFilter-output.txt


 When a servlet filter throws a ServletException in init(..), the exception is 
 logged by Jetty but not re-throws to the caller.  As a result, the Jetty 
 server failed silently.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HADOOP-7688) When a servlet filter throws an exception in init(..), the Jetty server failed silently.

2012-09-17 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-7688?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G resolved HADOOP-7688.
-

   Resolution: Fixed
Fix Version/s: 1.2.0

 When a servlet filter throws an exception in init(..), the Jetty server 
 failed silently. 
 -

 Key: HADOOP-7688
 URL: https://issues.apache.org/jira/browse/HADOOP-7688
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 0.23.0, 0.24.0
Reporter: Tsz Wo (Nicholas), SZE
Assignee: Uma Maheswara Rao G
 Fix For: 1.2.0, 3.0.0, 2.0.3-alpha

 Attachments: filter-init-exception-test.patch, 
 HADOOP-7688-branch-1.patch, HADOOP-7688-branch-2.patch, HADOOP-7688.patch, 
 org.apache.hadoop.http.TestServletFilter-output.txt


 When a servlet filter throws a ServletException in init(..), the exception is 
 logged by Jetty but not re-throws to the caller.  As a result, the Jetty 
 server failed silently.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HADOOP-8804) Improve Web UIs when the wildcard address is used

2012-09-17 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8804?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins reassigned HADOOP-8804:
---

Assignee: Vinay

 Improve Web UIs when the wildcard address is used
 -

 Key: HADOOP-8804
 URL: https://issues.apache.org/jira/browse/HADOOP-8804
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 1.0.0, 2.0.0-alpha
Reporter: Eli Collins
Assignee: Vinay
Priority: Minor
  Labels: newbie
 Attachments: HADOOP-8804-1.0.patch, HADOOP-8804-trunk.patch


 When IPC addresses are bound to the wildcard (ie the default config) the NN, 
 JT (and probably RM etc) Web UIs are a little goofy. Eg 0 Hadoop Map/Reduce 
 Administration and NameNode '0.0.0.0:18021' (active). Let's improve them.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8805) Move protocol buffer implementation of GetUserMappingProtocol from HDFS to Common

2012-09-17 Thread Bo Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457235#comment-13457235
 ] 

Bo Wang commented on HADOOP-8805:
-

Looked into the testing results of HADOOP-8805-v3.patch. Test failures are not 
related to changes in this patch.

Out of the 4 findbug warnings, 3 of them are in the protocbuf generated code 
(which we can't change), the rest one is from Configuration class which is 
unrelated to this patch.


 Move protocol buffer implementation of GetUserMappingProtocol from HDFS to 
 Common
 -

 Key: HADOOP-8805
 URL: https://issues.apache.org/jira/browse/HADOOP-8805
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Bo Wang
Assignee: Bo Wang
 Attachments: HADOOP-8805.patch, HADOOP-8805-v2.patch, 
 HADOOP-8805-v3.patch


 org.apache.hadoop.tools.GetUserMappingProtocol is used in both HDFS and YARN. 
 We should move the protocol buffer implementation from HDFS to Common so that 
 it can also be used by YARN.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Andy Isaacson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457236#comment-13457236
 ] 

Andy Isaacson commented on HADOOP-8806:
---

I didn't intend my exploratory tarball to derail my Friday +1.  +1.

 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8786) HttpServer continues to start even if AuthenticationFilter fails to init

2012-09-17 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8786?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HADOOP-8786:


Attachment: HADOOP-8786-branch-1.patch

 HttpServer continues to start even if AuthenticationFilter fails to init
 

 Key: HADOOP-8786
 URL: https://issues.apache.org/jira/browse/HADOOP-8786
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 1.2.0, 3.0.0, 2.0.1-alpha
Reporter: Todd Lipcon
Assignee: Todd Lipcon
 Fix For: 1.2.0, 3.0.0, 2.0.3-alpha

 Attachments: HADOOP-8786-branch-1.patch, HADOOP-8786-branch-2.patch, 
 hadoop-8786.txt


 As seen in HDFS-3904, if the AuthenticationFilter fails to initialize, the 
 web server will continue to start up. We need to check for context 
 initialization errors after starting the server.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8763) Set group owner on Windows failed

2012-09-17 Thread Chuan Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8763?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457240#comment-13457240
 ] 

Chuan Liu commented on HADOOP-8763:
---

The following code seems to be an unrelated change? Also, do you mean BUILDIN 
or BUILTIN?

You are right. It should be BUILTIN. This code is relevant in the sense it 
makes the function less prone to potential errors, which is used by 'winutils 
chown'.

Do you see any unexpected behavior for users because of the following?

No. I did not see any unexpected behavior. This is just for future references.

We can just leave around the public constant Shell.SET_GROUP_COMMAND or 
deprecate it. I am okay leaving it around.
I have left out this in new patch.

Not sure of your usage of asserts vs exit-code, but in src/winutils/chown.c, 
instead of asserts for zero-length string, we should log a msg to stderr and 
return an EXIT_FAILURE? Also, if both are empty also you should return 
EXIT_FAILURE?

The assertions assert for previous parsing code. The parsing code will not 
initiate and allocate memory for 'userName' and 'groupName' of zero-length, 
i.e. 'userName' and 'groupName' are initiated or NULL in such cases. We should 
not return error in such cases because 'chown : file' is a correct usage here, 
though no user name or group name is given.

This code won't be invoked on linux, because, ahm, this is winutils? In any 
case, that is not behaviour I know, a chown user: filename shouldn't change 
the group-name

I have tested the Linux behaviors. You can also check out the man page of 
'chown': http://linux.die.net/man/1/chown
Again, as my answer to Bikas's question, this usage pattern is not found in 
Hadoop on Linux or Windows that I am aware of. I think it is good to document 
the difference here for future reference.


 Set group owner on Windows failed
 -

 Key: HADOOP-8763
 URL: https://issues.apache.org/jira/browse/HADOOP-8763
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Chuan Liu
Assignee: Chuan Liu
Priority: Minor
 Fix For: 1-win

 Attachments: HADOOP-8763-branch-1-win-2.patch, 
 HADOOP-8763-branch-1-win.patch


 RawLocalFileSystem.setOwner() method may incorrectly set the group owner of a 
 file on Windows.
 Specifically the following function in RawLocalFileSystem class will fail on 
 Windows when username is null, i.e. only set group ownership.
 {code}
 public void setOwner(Path p, String username, String groupname)
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8763) Set group owner on Windows failed

2012-09-17 Thread Chuan Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8763?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chuan Liu updated HADOOP-8763:
--

Attachment: HADOOP-8763-branch-1-win-3.patch

 Set group owner on Windows failed
 -

 Key: HADOOP-8763
 URL: https://issues.apache.org/jira/browse/HADOOP-8763
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Chuan Liu
Assignee: Chuan Liu
Priority: Minor
 Fix For: 1-win

 Attachments: HADOOP-8763-branch-1-win-2.patch, 
 HADOOP-8763-branch-1-win-3.patch, HADOOP-8763-branch-1-win.patch


 RawLocalFileSystem.setOwner() method may incorrectly set the group owner of a 
 file on Windows.
 Specifically the following function in RawLocalFileSystem class will fail on 
 Windows when username is null, i.e. only set group ownership.
 {code}
 public void setOwner(Path p, String username, String groupname)
 {code}

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HADOOP-8806:


  Component/s: build
 Target Version/s: 2.0.3-alpha
Affects Version/s: 2.0.0-alpha
 Hadoop Flags: Reviewed

+1 lgtm as well. Thanks for the deleted explanation guys.

Also, thanks for testing both the java.library.patch and LD_LIBRARY_PATH 
scenarios Colin. The test failure and findbugs from the earlier test-patch run 
are unrelated to this change. I've tested a native build on my host for sanity.

 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HADOOP-8806:


  Resolution: Fixed
   Fix Version/s: 2.0.3-alpha
Target Version/s:   (was: 2.0.3-alpha)
  Status: Resolved  (was: Patch Available)

I've committed this and merged to branch-2. Thanks Colin.

 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Fix For: 2.0.3-alpha

 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Updated] (HADOOP-8812) ExitUtil#terminate should print Exception#toString

2012-09-17 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8812?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins updated HADOOP-8812:


Target Version/s: 2.0.3-alpha  (was: 2.0.2-alpha)

 ExitUtil#terminate should print Exception#toString 
 ---

 Key: HADOOP-8812
 URL: https://issues.apache.org/jira/browse/HADOOP-8812
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 2.0.2-alpha
Reporter: Eli Collins
Assignee: Eli Collins
Priority: Minor
 Attachments: hadoop-8812.txt, hadoop-8812.txt


 Per Steve's feedback on ExitUtil#terminate should print Exception#toString 
 rather than use getMessage as the latter may return null.  

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457250#comment-13457250
 ] 

Hudson commented on HADOOP-8806:


Integrated in Hadoop-Common-trunk-Commit #2737 (See 
[https://builds.apache.org/job/Hadoop-Common-trunk-Commit/2737/])
HADOOP-8806. libhadoop.so: dlopen should be better at locating 
libsnappy.so, etc. Contributed by Colin Patrick McCabe (Revision 1386784)
HADOOP-8806. libhadoop.so: dlopen should be better at locating libsnappy.so, 
etc. Contributed by Colin Patrick McCabe (Revision 1386780)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386784
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/CMakeLists.txt

eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386780
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/CMakeLists.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ExitUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java


 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Fix For: 2.0.3-alpha

 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8805) Move protocol buffer implementation of GetUserMappingProtocol from HDFS to Common

2012-09-17 Thread Suresh Srinivas (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457249#comment-13457249
 ] 

Suresh Srinivas commented on HADOOP-8805:
-

When committing, please ensure that svn mv is used to preserve the history.

 Move protocol buffer implementation of GetUserMappingProtocol from HDFS to 
 Common
 -

 Key: HADOOP-8805
 URL: https://issues.apache.org/jira/browse/HADOOP-8805
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Bo Wang
Assignee: Bo Wang
 Attachments: HADOOP-8805.patch, HADOOP-8805-v2.patch, 
 HADOOP-8805-v3.patch


 org.apache.hadoop.tools.GetUserMappingProtocol is used in both HDFS and YARN. 
 We should move the protocol buffer implementation from HDFS to Common so that 
 it can also be used by YARN.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457251#comment-13457251
 ] 

Hudson commented on HADOOP-8806:


Integrated in Hadoop-Hdfs-trunk-Commit #2800 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk-Commit/2800/])
HADOOP-8806. libhadoop.so: dlopen should be better at locating 
libsnappy.so, etc. Contributed by Colin Patrick McCabe (Revision 1386784)
HADOOP-8806. libhadoop.so: dlopen should be better at locating libsnappy.so, 
etc. Contributed by Colin Patrick McCabe (Revision 1386780)

 Result = SUCCESS
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386784
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/CMakeLists.txt

eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386780
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/CMakeLists.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ExitUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java


 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Fix For: 2.0.3-alpha

 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457280#comment-13457280
 ] 

Hudson commented on HADOOP-8806:


Integrated in Hadoop-Mapreduce-trunk-Commit #2761 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk-Commit/2761/])
HADOOP-8806. libhadoop.so: dlopen should be better at locating 
libsnappy.so, etc. Contributed by Colin Patrick McCabe (Revision 1386784)
HADOOP-8806. libhadoop.so: dlopen should be better at locating libsnappy.so, 
etc. Contributed by Colin Patrick McCabe (Revision 1386780)

 Result = FAILURE
eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386784
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/CMakeLists.txt

eli : http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1386780
Files : 
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/CMakeLists.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ExitUtil.java
* 
/hadoop/common/trunk/hadoop-hdfs-project/hadoop-hdfs/src/test/java/org/apache/hadoop/hdfs/server/namenode/TestEditLog.java


 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Fix For: 2.0.3-alpha

 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8806) libhadoop.so: dlopen should be better at locating libsnappy.so, etc.

2012-09-17 Thread Roman Shaposhnik (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8806?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457370#comment-13457370
 ] 

Roman Shaposhnik commented on HADOOP-8806:
--

Since Linux (GNU really) documentation leaves quite a bit open for 
interpretation as far as $ORIGIN and dependencies of unbundled products go, 
here's a much more complete treatment of the from Solaris' Linkers and 
Libraries doc: 
http://docs.oracle.com/cd/E19253-01/817-1984/appendixc-4/index.html

 libhadoop.so: dlopen should be better at locating libsnappy.so, etc.
 

 Key: HADOOP-8806
 URL: https://issues.apache.org/jira/browse/HADOOP-8806
 Project: Hadoop Common
  Issue Type: Improvement
  Components: build
Affects Versions: 2.0.0-alpha
Reporter: Colin Patrick McCabe
Assignee: Colin Patrick McCabe
Priority: Minor
 Fix For: 2.0.3-alpha

 Attachments: HADOOP-8806.003.patch, rpathtest2.tar.gz, 
 rpathtest.tar.gz


 libhadoop calls {{dlopen}} to load {{libsnappy.so}} and {{libz.so}}.  These 
 libraries can be bundled in the {{$HADOOP_ROOT/lib/native}} directory.  For 
 example, the {{-Dbundle.snappy}} build option copies {{libsnappy.so}} to this 
 directory.  However, snappy can't be loaded from this directory unless 
 {{LD_LIBRARY_PATH}} is set to include this directory.
 Can we make this configuration just work without needing to rely on 
 {{LD_LIBRARY_PATH}}?

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8805) Move protocol buffer implementation of GetUserMappingProtocol from HDFS to Common

2012-09-17 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457418#comment-13457418
 ] 

Alejandro Abdelnur commented on HADOOP-8805:


On the patch:

* GetUserMappingProtocol.proto: 
** package name 'proto' seems to depart from the naming conventions being use, 
ti should be 'protobuf'
** class name should remain GetUserMappingsProtocolProtos following current 
naming conventions.
* Why the rename of 
GetUserMappingsProtocolClientSideTranslatorPB/GetUserMappingsProtocolServerSideTranslatorPB
 classes? This seems to depart for the current naming convention?
* GetUserMappingsProtocolPB: the InterfaceAudience seems it should be only HDFS 
 YARN, not MapReduce, correct?

Have you build/deployed and tested it works as expected?

 Move protocol buffer implementation of GetUserMappingProtocol from HDFS to 
 Common
 -

 Key: HADOOP-8805
 URL: https://issues.apache.org/jira/browse/HADOOP-8805
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Bo Wang
Assignee: Bo Wang
 Attachments: HADOOP-8805.patch, HADOOP-8805-v2.patch, 
 HADOOP-8805-v3.patch


 org.apache.hadoop.tools.GetUserMappingProtocol is used in both HDFS and YARN. 
 We should move the protocol buffer implementation from HDFS to Common so that 
 it can also be used by YARN.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8805) Move protocol buffer implementation of GetUserMappingProtocol from HDFS to Common

2012-09-17 Thread Bo Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457448#comment-13457448
 ] 

Bo Wang commented on HADOOP-8805:
-

Thanks for the comments, Alejandro.

Replies inline.

* GetUserMappingProtocol.proto:
** package name 'proto' seems to depart from the naming conventions being use, 
ti should be 'protobuf'
Currently all protobuf generated packages use proto as the name. We can just 
follow the convention.
** class name should remain GetUserMappingsProtocolProtos following current 
naming conventions.
Sure.
* Why the rename of 
GetUserMappingsProtocolClientSideTranslatorPB/GetUserMappingsProtocolServerSideTranslatorPB
 classes? This seems to depart for the current naming convention?
I was following the YARN naming conventions. We'll keep HDFS naming convention 
which is along the lines of the naming convention used in Common.
* GetUserMappingsProtocolPB: the InterfaceAudience seems it should be only HDFS 
 YARN, not MapReduce, correct?
Yes.

Will upload a new patch reflecting these feedbacks.


 Move protocol buffer implementation of GetUserMappingProtocol from HDFS to 
 Common
 -

 Key: HADOOP-8805
 URL: https://issues.apache.org/jira/browse/HADOOP-8805
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Bo Wang
Assignee: Bo Wang
 Attachments: HADOOP-8805.patch, HADOOP-8805-v2.patch, 
 HADOOP-8805-v3.patch


 org.apache.hadoop.tools.GetUserMappingProtocol is used in both HDFS and YARN. 
 We should move the protocol buffer implementation from HDFS to Common so that 
 it can also be used by YARN.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Assigned] (HADOOP-8804) Improve Web UIs when the wildcard address is used

2012-09-17 Thread Eli Collins (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8804?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eli Collins reassigned HADOOP-8804:
---

Assignee: Senthil V Kumar  (was: Vinay)

Hi Senthil,

Not sure we want to use the hostname since presumably that's already in the URL 
bar. Ie we're getting additional info (the RPC address and port).

Some options:
- We just have NameNode (active), eg remove the IP:port
- We use the NN hostname's IP and RPC port
- We use the logical NN id (eg dfs.ha.namenodes.nameserviceId1.nn1) from the  
config
- We leave as is because it's useful to indicate RPC is bound to the wildcard

Btw noticed you're patch is using tabs/4 spaces. Please checkout for the style 
guide: 
http://wiki.apache.org/hadoop/HowToContributeAlso please don't re-org all the 
import statements in this change.

Thanks,
Eli




 Improve Web UIs when the wildcard address is used
 -

 Key: HADOOP-8804
 URL: https://issues.apache.org/jira/browse/HADOOP-8804
 Project: Hadoop Common
  Issue Type: Improvement
Affects Versions: 1.0.0, 2.0.0-alpha
Reporter: Eli Collins
Assignee: Senthil V Kumar
Priority: Minor
  Labels: newbie
 Attachments: HADOOP-8804-1.0.patch, HADOOP-8804-trunk.patch


 When IPC addresses are bound to the wildcard (ie the default config) the NN, 
 JT (and probably RM etc) Web UIs are a little goofy. Eg 0 Hadoop Map/Reduce 
 Administration and NameNode '0.0.0.0:18021' (active). Let's improve them.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Resolved] (HADOOP-8250) Investigate uses of FileUtil and functional correctness based on current use cases

2012-09-17 Thread Bikas Saha (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-8250?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Bikas Saha resolved HADOOP-8250.


Resolution: Won't Fix

Multiple other jiras have superceded this.

 Investigate uses of FileUtil and functional correctness based on current use 
 cases
 --

 Key: HADOOP-8250
 URL: https://issues.apache.org/jira/browse/HADOOP-8250
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: native
Affects Versions: 1.1.0
Reporter: Bikas Saha
Assignee: Bikas Saha

 The current Windows patch replaces symlink with copy. This jira tracks 
 understanding the implications of this change and others like it on expected 
 functionality.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8803) Make Hadoop running more secure public cloud envrionment

2012-09-17 Thread Xianqing Yu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8803?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457534#comment-13457534
 ] 

Xianqing Yu commented on HADOOP-8803:
-

Hi Luke,

Thank you for your comment.

To your first, fourth and fifth comments. My security assumption is that one or 
some of tasktracker or datanode machines may be compromised by hacker. If HDFS 
only use one key for all HDFS, even the block tokens are ephemeral and the key 
only be used for a few minutes, as long as the hacker can get root privilege, 
she or he always can get the new key from the namenode and generate arbitrary 
Block Token to access other datanodes, in the end, the whole HDFS is exposed to 
attacker. The reason datanode have to use unique key is that limiting the 
hacker capabilities when she or he compromised one datanode. Using TLS is a 
good way to provide network security. However, you have to assume that all 
machines are secure or all machine are running under isolated network 
environment no matter how many clusters are working together. Otherwise, 
compromised of one machine almost means whole HDFS can be exposed to hacker 
immediately. For HDFS, once its scale gets bigger and bigger, I feel it is 
necessary to create certain level of isolation (without sacrifice too much 
performance of course). Unique key for each datanode is first step for that. 
But we have to find the best way to lower the overhead.

To your second comment, I am working on that test now. Would add more comments 
when I am done.

To your Third comment, thanks for mentioning that, NameNode has to generate 
token for each replica and that indeed degrade the performance.  Another 
possible way is that let Namenode to choose replica for the client, but 
NameNode have to perform ChooseDatanode code. I can test how much of this 
overhead and also check which way has lower overhead. (I am not sure how much 
overhead it would introduce, will that impact performance once scale of cluster 
get larger)

 Make Hadoop running more secure public cloud envrionment
 

 Key: HADOOP-8803
 URL: https://issues.apache.org/jira/browse/HADOOP-8803
 Project: Hadoop Common
  Issue Type: New Feature
  Components: fs, ipc, security
Affects Versions: 0.20.204.0
Reporter: Xianqing Yu
  Labels: hadoop
   Original Estimate: 2m
  Remaining Estimate: 2m

 I am a Ph.D student in North Carolina State University. I am modifying the 
 Hadoop's code (which including most parts of Hadoop, e.g. JobTracker, 
 TaskTracker, NameNode, DataNode) to achieve better security.
  
 My major goal is that make Hadoop running more secure in the Cloud 
 environment, especially for public Cloud environment. In order to achieve 
 that, I redesign the currently security mechanism and achieve following 
 proprieties:
 1. Bring byte-level access control to Hadoop HDFS. Based on 0.20.204, HDFS 
 access control is based on user or block granularity, e.g. HDFS Delegation 
 Token only check if the file can be accessed by certain user or not, Block 
 Token only proof which block or blocks can be accessed. I make Hadoop can do 
 byte-granularity access control, each access party, user or task process can 
 only access the bytes she or he least needed.
 2. I assume that in the public Cloud environment, only Namenode, secondary 
 Namenode, JobTracker can be trusted. A large number of Datanode and 
 TaskTracker may be compromised due to some of them may be running under less 
 secure environment. So I re-design the secure mechanism to make the damage 
 the hacker can do to be minimized.
  
 a. Re-design the Block Access Token to solve wildly shared-key problem of 
 HDFS. In original Block Access Token design, all HDFS (Namenode and Datanode) 
 share one master key to generate Block Access Token, if one DataNode is 
 compromised by hacker, the hacker can get the key and generate any  Block 
 Access Token he or she want.
  
 b. Re-design the HDFS Delegation Token to do fine-grain access control for 
 TaskTracker and Map-Reduce Task process on HDFS. 
  
 In the Hadoop 0.20.204, all TaskTrackers can use their kerberos credentials 
 to access any files for MapReduce on HDFS. So they have the same privilege as 
 JobTracker to do read or write tokens, copy job file, etc.. However, if one 
 of them is compromised, every critical thing in MapReduce directory (job 
 file, Delegation Token) is exposed to attacker. I solve the problem by making 
 JobTracker to decide which TaskTracker can access which file in MapReduce 
 Directory on HDFS.
  
 For Task process, once it get HDFS Delegation Token, it can access everything 
 belong to this job or user on HDFS. By my design, it can only access the 
 bytes it needed from HDFS.
  
 There are some other improvement in the security, 

[jira] [Commented] (HADOOP-8615) EOFException in DecompressorStream.java needs to be more verbose

2012-09-17 Thread thomastechs (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8615?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457584#comment-13457584
 ] 

thomastechs commented on HADOOP-8615:
-

Thanks Andy for your valuable comments.
I will work on the patch in the trunk too.Also on the test case.
The methods having default values were already existing. I used the same 
methods and overloaded with an added parameter of fileName.This will not affect 
existing functionalities or the people who already used the existing methods 
without a fileName.The new methods with fileName , helps the user to use it, if 
the user faces the issue addressed in this bug.
Please let me know your comments.

 EOFException in DecompressorStream.java needs to be more verbose
 

 Key: HADOOP-8615
 URL: https://issues.apache.org/jira/browse/HADOOP-8615
 Project: Hadoop Common
  Issue Type: Bug
  Components: io
Affects Versions: 0.20.2
Reporter: Jeff Lord
  Labels: patch
 Attachments: HADOOP-8615-release-0.20.2.patch


 In ./src/core/org/apache/hadoop/io/compress/DecompressorStream.java
 The following exception should at least pass back the file that it encounters 
 this error in relation to:
   protected void getCompressedData() throws IOException {
 checkStream();
 int n = in.read(buffer, 0, buffer.length);
 if (n == -1) {
   throw new EOFException(Unexpected end of input stream);
 }
 This would help greatly to debug bad/corrupt files.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira


[jira] [Commented] (HADOOP-8805) Move protocol buffer implementation of GetUserMappingProtocol from HDFS to Common

2012-09-17 Thread Bo Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-8805?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=13457618#comment-13457618
 ] 

Bo Wang commented on HADOOP-8805:
-

I realized a problem that YARN RPC requires protocol implementation follows the 
following naming convention: PROTOCOL + PBServiceImpl. This naming convention 
is hard coded in 
org.apache.hadoop.yarn.factories.impl.pb.RpcServerFactoryPBImpl#getPbServiceImplClassName.
 Thus, to make GetUserMappingProtocol usable in both HDFS  YARN, we have to 
follow the convention of YARN rather than HDFS..

In YARN, the protocbuf generated class shares the same name as the protocol 
itself (e.g. RMAdminProtocol). Thus, the current patch follows YARN convention.

I have tested the patch from command line as follows.
(1) start a namenode
(2) bin/hdfs groups username

It returns the correct results.


 Move protocol buffer implementation of GetUserMappingProtocol from HDFS to 
 Common
 -

 Key: HADOOP-8805
 URL: https://issues.apache.org/jira/browse/HADOOP-8805
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Bo Wang
Assignee: Bo Wang
 Attachments: HADOOP-8805.patch, HADOOP-8805-v2.patch, 
 HADOOP-8805-v3.patch


 org.apache.hadoop.tools.GetUserMappingProtocol is used in both HDFS and YARN. 
 We should move the protocol buffer implementation from HDFS to Common so that 
 it can also be used by YARN.

--
This message is automatically generated by JIRA.
If you think it was sent incorrectly, please contact your JIRA administrators
For more information on JIRA, see: http://www.atlassian.com/software/jira