[jira] [Created] (HADOOP-10849) Implement conf substitution with UGI.current/loginUser

2014-07-16 Thread Gera Shegalov (JIRA)
Gera Shegalov created HADOOP-10849:
--

 Summary: Implement conf substitution with UGI.current/loginUser
 Key: HADOOP-10849
 URL: https://issues.apache.org/jira/browse/HADOOP-10849
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 2.4.1
Reporter: Gera Shegalov
Assignee: Gera Shegalov


Many path properties and similar in hadoop code base would be easily configured 
if we had substitutions with {{UserGroupInformation#getCurrentUser}}. Currently 
we often use less elegant concatenation code if we want to express currentUser 
as opposed to ${user.name} system property representing the user owning the JVM.

This JIRA proposes the corresponding substitution support for keys 
{{ugi.current.user}} and {{ugi.login.user}}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10849) Implement conf substitution with UGI.current/loginUser

2014-07-16 Thread Gera Shegalov (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10849?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Gera Shegalov updated HADOOP-10849:
---

Attachment: HADOOP-10849.v01.patch

v01 to give an idea.

 Implement conf substitution with UGI.current/loginUser
 --

 Key: HADOOP-10849
 URL: https://issues.apache.org/jira/browse/HADOOP-10849
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 2.4.1
Reporter: Gera Shegalov
Assignee: Gera Shegalov
 Attachments: HADOOP-10849.v01.patch


 Many path properties and similar in hadoop code base would be easily 
 configured if we had substitutions with 
 {{UserGroupInformation#getCurrentUser}}. Currently we often use less elegant 
 concatenation code if we want to express currentUser as opposed to 
 ${user.name} system property representing the user owning the JVM.
 This JIRA proposes the corresponding substitution support for keys 
 {{ugi.current.user}} and {{ugi.login.user}}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10720) KMS: Implement generateEncryptedKey and decryptEncryptedKey in the REST API

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063216#comment-14063216
 ] 

Hadoop QA commented on HADOOP-10720:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12655994/HADOOP-10720.11.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 2 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common hadoop-common-project/hadoop-kms:

  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4288//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4288//console

This message is automatically generated.

 KMS: Implement generateEncryptedKey and decryptEncryptedKey in the REST API
 ---

 Key: HADOOP-10720
 URL: https://issues.apache.org/jira/browse/HADOOP-10720
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Arun Suresh
 Attachments: COMBO.patch, COMBO.patch, COMBO.patch, COMBO.patch, 
 COMBO.patch, HADOOP-10720.1.patch, HADOOP-10720.10.patch, 
 HADOOP-10720.11.patch, HADOOP-10720.2.patch, HADOOP-10720.3.patch, 
 HADOOP-10720.4.patch, HADOOP-10720.5.patch, HADOOP-10720.6.patch, 
 HADOOP-10720.7.patch, HADOOP-10720.8.patch, HADOOP-10720.9.patch, 
 HADOOP-10720.patch, HADOOP-10720.patch, HADOOP-10720.patch, 
 HADOOP-10720.patch, HADOOP-10720.patch


 KMS client/server should implement support for generating encrypted keys and 
 decrypting them via the REST API being introduced by HADOOP-10719.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry

2014-07-16 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063236#comment-14063236
 ] 

Vinayakumar B commented on HADOOP-10830:


Hi [~benoyantony], thanks for the patch.

There is no need of double try-catch-finally blocks. existing itself you can 
re-use.
Like following..
{code}  @Override
  public CredentialEntry createCredentialEntry(String alias, char[] credential)
throws IOException {
writeLock.lock();
try {
  if (keyStore.containsAlias(alias) || cache.containsKey(alias)) {
throw new IOException(Credential  + alias +  already exists in 
+ this);
  }
  return innerSetCredential(alias, credential);
} catch (KeyStoreException e) {
  throw new IOException(Problem looking up credential  + alias +  in 
  + this, e);
} finally {
  writeLock.unlock();
}
  }{code}

anyway {{innerSetCredential(..)}} have its own try-catch block. So its fine to 
use one try-catch-finally block itself.

 Missing lock in JavaKeyStoreProvider.createCredentialEntry
 --

 Key: HADOOP-10830
 URL: https://issues.apache.org/jira/browse/HADOOP-10830
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10830.patch


 _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_  to provide thread 
 safety.
 The {{createCredentialEntry}} should hold _writeLock_ before adding the entry.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063238#comment-14063238
 ] 

Vinayakumar B commented on HADOOP-9921:
---

Hi [~umamaheswararao], can you review the patch?

 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10131) NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if excluded nodes passed are not part of the cluster tree

2014-07-16 Thread Vinayakumar B (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063241#comment-14063241
 ] 

Vinayakumar B commented on HADOOP-10131:


Hi [~umamaheswararao], can you review the patch? Thanks In advance

 NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if 
 excluded nodes passed are not part of the cluster tree
 -

 Key: HADOOP-10131
 URL: https://issues.apache.org/jira/browse/HADOOP-10131
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.0.5-alpha
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Attachments: HADOOP-10131.patch, HDFS-5112.patch


 I got File /hdfs_COPYING_ could only be replicated to 0 nodes instead of 
 minReplication (=1).  There are 1 datanode(s) running and 1 node(s) are 
 excluded in this operation. in the following case
 1. 2 DNs cluster,
 2. One of the datanodes was not responding from last 10 min, but about to 
 detect as dead at NN.
 3. Tried to write one file, for the block NN allocated both DNs.
 4. Client While creating the pipeline took some time to detect one node 
 failure.
 5. Before client detects pipeline failure, NN side dead node was removed from 
 cluster map.
 6. Now, client has abandoned previous block and asked for new block with dead 
 node in excluded list and got above exception even though one more node was 
 available live.
 When I dig this more, found that,
 {{NetWorkTopology#countNumOfAvailableNodes()}} is not giving correct count 
 when the excludeNodes passed from client are not part of the cluster map.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10794) A hadoop cluster needs clock synchronization

2014-07-16 Thread Zhijie Shen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063270#comment-14063270
 ] 

Zhijie Shen commented on HADOOP-10794:
--

I'm afraid each component may more or less have made the assumption that clock 
is synchronized across the cluster. Unfortunately,  the problem that made us be 
aware of this issue came from MR: an elapsed time that based the the diff of 
the timestamps on two hosts became negative. Having a try with a cluster whose 
hosts are completely out of sync in terms of clock may give us the idea how bad 
the situation is.

According to the prior discussion, it seems to be good to have a mechanism to 
check the clocks across a cluster, report unacceptable asynchronization, 
doesn't it? On the other hand, it is arguable whether this mechanism should be 
part of Hadoop or other cluster monitoring tools.

Anyway, IMHO, it's good to at least let users be aware of the clock 
synchronization assumption (via jira, mailing list or etc.), and the risks of 
being out of sync, such that they can be more careful about this factor when 
setting up a cluster.

 A hadoop cluster needs clock synchronization
 

 Key: HADOOP-10794
 URL: https://issues.apache.org/jira/browse/HADOOP-10794
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Zhijie Shen

 As a distributed system, a hadoop cluster wants the clock on all the 
 participating hosts synchronized. Otherwise, some problems might happen. For 
 example, in YARN-2251, due to the clock on the host for the task container 
 falls behind that on the host of the AM container, the computed elapsed time 
 (the diff between the timestamps produced on two hosts) becomes negative.
 In YARN-2251, we tried to mask the negative elapsed time. However, we should 
 seek for a decent long term solution, such as providing mechanism to do and 
 check clock synchronization.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063275#comment-14063275
 ] 

Hadoop QA commented on HADOOP-9921:
---

{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12600746/HADOOP-9921.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common:

  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4289//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4289//console

This message is automatically generated.

 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10131) NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if excluded nodes passed are not part of the cluster tree

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063280#comment-14063280
 ] 

Hadoop QA commented on HADOOP-10131:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12616044/HADOOP-10131.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common:

  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem
  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4290//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4290//console

This message is automatically generated.

 NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if 
 excluded nodes passed are not part of the cluster tree
 -

 Key: HADOOP-10131
 URL: https://issues.apache.org/jira/browse/HADOOP-10131
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.0.5-alpha
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Attachments: HADOOP-10131.patch, HDFS-5112.patch


 I got File /hdfs_COPYING_ could only be replicated to 0 nodes instead of 
 minReplication (=1).  There are 1 datanode(s) running and 1 node(s) are 
 excluded in this operation. in the following case
 1. 2 DNs cluster,
 2. One of the datanodes was not responding from last 10 min, but about to 
 detect as dead at NN.
 3. Tried to write one file, for the block NN allocated both DNs.
 4. Client While creating the pipeline took some time to detect one node 
 failure.
 5. Before client detects pipeline failure, NN side dead node was removed from 
 cluster map.
 6. Now, client has abandoned previous block and asked for new block with dead 
 node in excluded list and got above exception even though one more node was 
 available live.
 When I dig this more, found that,
 {{NetWorkTopology#countNumOfAvailableNodes()}} is not giving correct count 
 when the excludeNodes passed from client are not part of the cluster map.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-7664) o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same.

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063366#comment-14063366
 ] 

Hudson commented on HADOOP-7664:


FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/614/])
HADOOP-7664. Move attribution to release 2.6.0 section in CHANGES.txt. 
(cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610873)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


 o.a.h.conf.Configuration complains of overriding final parameter even if the 
 value with which its attempting to override is the same. 
 --

 Key: HADOOP-7664
 URL: https://issues.apache.org/jira/browse/HADOOP-7664
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 0.20.205.0, 0.23.0, 2.5.0
 Environment: commit a2f64ee8d9312fe24780ec53b15af439a315796d
Reporter: Ravi Prakash
Assignee: Ravi Prakash
Priority: Minor
 Fix For: 1.0.0, 0.23.0, 2.6.0

 Attachments: HADOOP-7664.branch-0.20-security.patch, 
 HADOOP-7664.patch, HADOOP-7664.patch, HADOOP-7664.patch

   Original Estimate: 1m
  Time Spent: 1m
  Remaining Estimate: 0h

 o.a.h.conf.Configuration complains of overriding final parameter even if the 
 value with which its attempting to override is the same. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Uma Maheswara Rao G (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063358#comment-14063358
 ] 

Uma Maheswara Rao G commented on HADOOP-9921:
-

+1 patch looks good to me.

 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10831) UserProvider is not thread safe

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063368#comment-14063368
 ] 

Hudson commented on HADOOP-10831:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/614/])
HADOOP-10831. UserProvider is not thread safe. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610789)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialProvider.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/UserProvider.java


 UserProvider is not thread safe
 ---

 Key: HADOOP-10831
 URL: https://issues.apache.org/jira/browse/HADOOP-10831
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10831.patch


 While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread 
 safe.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10468) TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063359#comment-14063359
 ] 

Hudson commented on HADOOP-10468:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/614/])
Addendum patch for HADOOP-10468 TestMetricsSystemImpl.testMultiThreadedPublish 
fails intermediately. Contributed by Akira AJISAKA (jlowe: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610829)
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestGangliaMetrics.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java


 TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately
 ---

 Key: HADOOP-10468
 URL: https://issues.apache.org/jira/browse/HADOOP-10468
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Haohui Mai
Assignee: Akira AJISAKA
Priority: Blocker
 Fix For: 2.5.0

 Attachments: HADOOP-10468.000.patch, HADOOP-10468.001.patch, 
 HADOOP-10468.2.patch


 {{TestMetricsSystemImpl.testMultiThreadedPublish}} can fail intermediately 
 due to the insufficient size of the sink queue:
 {code}
 2014-04-06 21:34:55,269 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 2014-04-06 21:34:55,270 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 2014-04-06 21:34:55,271 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 {code}
 The unit test should increase the default queue size to avoid intermediate 
 failure.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10673) Update rpc metrics when the call throws an exception

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063365#comment-14063365
 ] 

Hudson commented on HADOOP-10673:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/614/])
HADOOP-10673. Update rpc metrics when the call throws an exception. Contributed 
by Ming Ma. (jing9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610879)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/ProtobufRpcEngine.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/WritableRpcEngine.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestRPC.java


 Update rpc metrics when the call throws an exception
 

 Key: HADOOP-10673
 URL: https://issues.apache.org/jira/browse/HADOOP-10673
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Ming Ma
Assignee: Ming Ma
 Fix For: 2.6.0

 Attachments: HADOOP-10673-2.patch, HADOOP-10673-4.patch, 
 HADOOP-10673-5.patch, HADOOP-10673-6.patch, HADOOP-10673.patch


 Currently RPC metrics isn't updated when the call throws an exception. We can 
 either update the existing metrics or have a new set of metrics in the case 
 of exception.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10845) Add common tests for ACLs in combination with viewfs.

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063367#comment-14063367
 ] 

Hudson commented on HADOOP-10845:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/614/])
HADOOP-10845. Add common tests for ACLs in combination with viewfs. Contributed 
by Stephen Chu. (cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610911)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsBaseTest.java


 Add common tests for ACLs in combination with viewfs.
 -

 Key: HADOOP-10845
 URL: https://issues.apache.org/jira/browse/HADOOP-10845
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs, test
Affects Versions: 3.0.0, 2.5.0
Reporter: Chris Nauroth
Assignee: Stephen Chu
 Fix For: 3.0.0, 2.6.0

 Attachments: HADOOP-10845.1.patch


 Add tests in Hadoop Common for the ACL APIs in combination with viewfs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10834) Typo in CredentialShell usage

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063363#comment-14063363
 ] 

Hudson commented on HADOOP-10834:
-

FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See 
[https://builds.apache.org/job/Hadoop-Yarn-trunk/614/])
HADOOP-10834. Typo in CredentialShell usage. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610794)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialShell.java


 Typo in CredentialShell usage
 -

 Key: HADOOP-10834
 URL: https://issues.apache.org/jira/browse/HADOOP-10834
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
Priority: Minor
 Fix For: 3.0.0

 Attachments: HADOOP-10834.patch


 There is a typo in one of the informational messages in _ CredentialShell_
 {code}
 out.println(Nothing has been be deleted.);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Vinayakumar B (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinayakumar B updated HADOOP-9921:
--

   Resolution: Fixed
Fix Version/s: 2.6.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thank You [~umamaheswararao] for the review.
Committed to trunk and branch-2

 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Fix For: 2.6.0

 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063379#comment-14063379
 ] 

Hudson commented on HADOOP-9921:


FAILURE: Integrated in Hadoop-trunk-Commit #5890 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5890/])
HADOOP-9921. daemon scripts should remove pid file on stop call after stop or 
process is found not running ( Contributed by Vinayakumar B) (vinayakumarb: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610964)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh
* /hadoop/common/trunk/hadoop-mapreduce-project/bin/mr-jobhistory-daemon.sh
* /hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/bin/yarn-daemon.sh


 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Fix For: 2.6.0

 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10824) Refactor KMSACLs to avoid locking

2014-07-16 Thread Uma Maheswara Rao G (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Uma Maheswara Rao G updated HADOOP-10824:
-

   Resolution: Fixed
Fix Version/s: 3.0.0
 Hadoop Flags: Reviewed
   Status: Resolved  (was: Patch Available)

Thanks a lot, Benoy for the patch. I have committed this to trunk.

 Refactor KMSACLs to avoid locking
 -

 Key: HADOOP-10824
 URL: https://issues.apache.org/jira/browse/HADOOP-10824
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10824.patch, HADOOP-10824.patch


 Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible 
 to safely publish the _acls_ collection using _volatile_.
 Similar refactoring has been done in 
 [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112]



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10824) Refactor KMSACLs to avoid locking

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063395#comment-14063395
 ] 

Hudson commented on HADOOP-10824:
-

FAILURE: Integrated in Hadoop-trunk-Commit #5891 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5891/])
HADOOP-10824. Refactor KMSACLs to avoid locking. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610969)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSACLs.java


 Refactor KMSACLs to avoid locking
 -

 Key: HADOOP-10824
 URL: https://issues.apache.org/jira/browse/HADOOP-10824
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10824.patch, HADOOP-10824.patch


 Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible 
 to safely publish the _acls_ collection using _volatile_.
 Similar refactoring has been done in 
 [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112]



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HADOOP-10799:


Attachment: HADOOP-10799.patch

adding a missing constructor to {{DelegationTokenAuthenticatedURL}}.

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HADOOP-10850) KerberosAuthenticator should not do the SPNEGO handshake

2014-07-16 Thread Alejandro Abdelnur (JIRA)
Alejandro Abdelnur created HADOOP-10850:
---

 Summary: KerberosAuthenticator should not do the SPNEGO handshake
 Key: HADOOP-10850
 URL: https://issues.apache.org/jira/browse/HADOOP-10850
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur


As mentioned in HADOOP-10453, the JDK automatically does a SPNEGO handshake 
when opening a connection with a URL within a Kerberos login context, there is 
no need to do the SPNEGO handshake in the {{KerberosAuthenticator}}, simply 
extract the auth token (hadoop-auth cookie) and do the fallback if necessary.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10834) Typo in CredentialShell usage

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063476#comment-14063476
 ] 

Hudson commented on HADOOP-10834:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-10834. Typo in CredentialShell usage. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610794)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialShell.java


 Typo in CredentialShell usage
 -

 Key: HADOOP-10834
 URL: https://issues.apache.org/jira/browse/HADOOP-10834
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
Priority: Minor
 Fix For: 3.0.0

 Attachments: HADOOP-10834.patch


 There is a typo in one of the informational messages in _ CredentialShell_
 {code}
 out.println(Nothing has been be deleted.);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-7664) o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same.

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063480#comment-14063480
 ] 

Hudson commented on HADOOP-7664:


FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-7664. Move attribution to release 2.6.0 section in CHANGES.txt. 
(cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610873)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


 o.a.h.conf.Configuration complains of overriding final parameter even if the 
 value with which its attempting to override is the same. 
 --

 Key: HADOOP-7664
 URL: https://issues.apache.org/jira/browse/HADOOP-7664
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 0.20.205.0, 0.23.0, 2.5.0
 Environment: commit a2f64ee8d9312fe24780ec53b15af439a315796d
Reporter: Ravi Prakash
Assignee: Ravi Prakash
Priority: Minor
 Fix For: 1.0.0, 0.23.0, 2.6.0

 Attachments: HADOOP-7664.branch-0.20-security.patch, 
 HADOOP-7664.patch, HADOOP-7664.patch, HADOOP-7664.patch

   Original Estimate: 1m
  Time Spent: 1m
  Remaining Estimate: 0h

 o.a.h.conf.Configuration complains of overriding final parameter even if the 
 value with which its attempting to override is the same. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10845) Add common tests for ACLs in combination with viewfs.

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063482#comment-14063482
 ] 

Hudson commented on HADOOP-10845:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-10845. Add common tests for ACLs in combination with viewfs. Contributed 
by Stephen Chu. (cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610911)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsBaseTest.java


 Add common tests for ACLs in combination with viewfs.
 -

 Key: HADOOP-10845
 URL: https://issues.apache.org/jira/browse/HADOOP-10845
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs, test
Affects Versions: 3.0.0, 2.5.0
Reporter: Chris Nauroth
Assignee: Stephen Chu
 Fix For: 3.0.0, 2.6.0

 Attachments: HADOOP-10845.1.patch


 Add tests in Hadoop Common for the ACL APIs in combination with viewfs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10673) Update rpc metrics when the call throws an exception

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063478#comment-14063478
 ] 

Hudson commented on HADOOP-10673:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-10673. Update rpc metrics when the call throws an exception. Contributed 
by Ming Ma. (jing9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610879)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/ProtobufRpcEngine.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/WritableRpcEngine.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestRPC.java


 Update rpc metrics when the call throws an exception
 

 Key: HADOOP-10673
 URL: https://issues.apache.org/jira/browse/HADOOP-10673
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Ming Ma
Assignee: Ming Ma
 Fix For: 2.6.0

 Attachments: HADOOP-10673-2.patch, HADOOP-10673-4.patch, 
 HADOOP-10673-5.patch, HADOOP-10673-6.patch, HADOOP-10673.patch


 Currently RPC metrics isn't updated when the call throws an exception. We can 
 either update the existing metrics or have a new set of metrics in the case 
 of exception.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10831) UserProvider is not thread safe

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063483#comment-14063483
 ] 

Hudson commented on HADOOP-10831:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-10831. UserProvider is not thread safe. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610789)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialProvider.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/UserProvider.java


 UserProvider is not thread safe
 ---

 Key: HADOOP-10831
 URL: https://issues.apache.org/jira/browse/HADOOP-10831
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10831.patch


 While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread 
 safe.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10824) Refactor KMSACLs to avoid locking

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063475#comment-14063475
 ] 

Hudson commented on HADOOP-10824:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-10824. Refactor KMSACLs to avoid locking. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610969)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSACLs.java


 Refactor KMSACLs to avoid locking
 -

 Key: HADOOP-10824
 URL: https://issues.apache.org/jira/browse/HADOOP-10824
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10824.patch, HADOOP-10824.patch


 Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible 
 to safely publish the _acls_ collection using _volatile_.
 Similar refactoring has been done in 
 [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112]



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063479#comment-14063479
 ] 

Hudson commented on HADOOP-9921:


FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
HADOOP-9921. daemon scripts should remove pid file on stop call after stop or 
process is found not running ( Contributed by Vinayakumar B) (vinayakumarb: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610964)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh
* /hadoop/common/trunk/hadoop-mapreduce-project/bin/mr-jobhistory-daemon.sh
* /hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/bin/yarn-daemon.sh


 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Fix For: 2.6.0

 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10468) TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063471#comment-14063471
 ] 

Hudson commented on HADOOP-10468:
-

FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See 
[https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/])
Addendum patch for HADOOP-10468 TestMetricsSystemImpl.testMultiThreadedPublish 
fails intermediately. Contributed by Akira AJISAKA (jlowe: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610829)
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestGangliaMetrics.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java


 TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately
 ---

 Key: HADOOP-10468
 URL: https://issues.apache.org/jira/browse/HADOOP-10468
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Haohui Mai
Assignee: Akira AJISAKA
Priority: Blocker
 Fix For: 2.5.0

 Attachments: HADOOP-10468.000.patch, HADOOP-10468.001.patch, 
 HADOOP-10468.2.patch


 {{TestMetricsSystemImpl.testMultiThreadedPublish}} can fail intermediately 
 due to the insufficient size of the sink queue:
 {code}
 2014-04-06 21:34:55,269 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 2014-04-06 21:34:55,270 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 2014-04-06 21:34:55,271 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 {code}
 The unit test should increase the default queue size to avoid intermediate 
 failure.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063490#comment-14063490
 ] 

Alejandro Abdelnur commented on HADOOP-10799:
-

[~zjshen], thanks for pointing out HADOOP-10453, I was not aware of it, I've 
just commented there.

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10453) Do not use AuthenticatedURL in hadoop core

2014-07-16 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063487#comment-14063487
 ] 

Alejandro Abdelnur commented on HADOOP-10453:
-

The double SPNEGO authentication issue Daryn refers to is a problem of the 
KerberosAuthenticator, he is correct when stating that JDK does automatically 
SPNEGO when within a Kerberos login context, just created HADOOP-10850 for it.

Regarding removing {{AuthenticatedURL}} altogether, I disagree on that. The 
fallback functionality is useful, HttpFS uses it for testing and now KMS is 
using it as well, plus YARN RM and TimelineServer. So we should work on fixing 
it, rather than trashing it.

In addition, with the work that I'm doing in HADOOP-10799 (adding delegation 
token support)  HADOOP-10835 (adding proxy user support) it simplifies 
significantly the client side code eliminating a big chunk of repetitive 
security related code that currently is done by the component doing HTTP calls 
instead by a security library. 

 Do not use AuthenticatedURL in hadoop core
 --

 Key: HADOOP-10453
 URL: https://issues.apache.org/jira/browse/HADOOP-10453
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Haohui Mai
Priority: Blocker

 As [~daryn] has suggested in HDFS-4564:
 {quote}
 AuthenticatedURL is not used because it is buggy in part to causing replay 
 attacks, double attempts to kerberos authenticate with the fallback 
 authenticator if the TGT is expired, incorrectly uses the fallback 
 authenticator (required by oozie servers) to add the username parameter which 
 webhdfs has already included in the uri.
 AuthenticatedURL's attempt to do SPNEGO auth is a no-op because the JDK 
 transparently does SPNEGO when the user's Subject (UGI) contains kerberos 
 principals. Since AuthenticatedURL is now not used, webhdfs has to check the 
 TGT itself for token operations.
 Bottom line is AuthenticatedURL is unnecessary and introduces nothing but 
 problems for webhdfs. It's only useful for oozie's anon/non-anon support.
 {quote}
 However, several functionalities that relies on SPNEGO in secure mode suffer 
 from the same problem. For example, NNs / JNs create HTTP connections to 
 exchange fsimage and edit logs. Currently all of them are through 
 {{AuthenticatedURL}}. This needs to be fixed to avoid security 
 vulnerabilities.
 This jira purposes to remove {{AuthenticatedURL}} from hadoop core and to 
 move it to oozie.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10834) Typo in CredentialShell usage

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063505#comment-14063505
 ] 

Hudson commented on HADOOP-10834:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-10834. Typo in CredentialShell usage. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610794)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialShell.java


 Typo in CredentialShell usage
 -

 Key: HADOOP-10834
 URL: https://issues.apache.org/jira/browse/HADOOP-10834
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
Priority: Minor
 Fix For: 3.0.0

 Attachments: HADOOP-10834.patch


 There is a typo in one of the informational messages in _ CredentialShell_
 {code}
 out.println(Nothing has been be deleted.);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10831) UserProvider is not thread safe

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063511#comment-14063511
 ] 

Hudson commented on HADOOP-10831:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-10831. UserProvider is not thread safe. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610789)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialProvider.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/UserProvider.java


 UserProvider is not thread safe
 ---

 Key: HADOOP-10831
 URL: https://issues.apache.org/jira/browse/HADOOP-10831
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10831.patch


 While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread 
 safe.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10673) Update rpc metrics when the call throws an exception

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063507#comment-14063507
 ] 

Hudson commented on HADOOP-10673:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-10673. Update rpc metrics when the call throws an exception. Contributed 
by Ming Ma. (jing9: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610879)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/ProtobufRpcEngine.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/WritableRpcEngine.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestRPC.java


 Update rpc metrics when the call throws an exception
 

 Key: HADOOP-10673
 URL: https://issues.apache.org/jira/browse/HADOOP-10673
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Ming Ma
Assignee: Ming Ma
 Fix For: 2.6.0

 Attachments: HADOOP-10673-2.patch, HADOOP-10673-4.patch, 
 HADOOP-10673-5.patch, HADOOP-10673-6.patch, HADOOP-10673.patch


 Currently RPC metrics isn't updated when the call throws an exception. We can 
 either update the existing metrics or have a new set of metrics in the case 
 of exception.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-7664) o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same.

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063509#comment-14063509
 ] 

Hudson commented on HADOOP-7664:


FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-7664. Move attribution to release 2.6.0 section in CHANGES.txt. 
(cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610873)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt


 o.a.h.conf.Configuration complains of overriding final parameter even if the 
 value with which its attempting to override is the same. 
 --

 Key: HADOOP-7664
 URL: https://issues.apache.org/jira/browse/HADOOP-7664
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf
Affects Versions: 0.20.205.0, 0.23.0, 2.5.0
 Environment: commit a2f64ee8d9312fe24780ec53b15af439a315796d
Reporter: Ravi Prakash
Assignee: Ravi Prakash
Priority: Minor
 Fix For: 1.0.0, 0.23.0, 2.6.0

 Attachments: HADOOP-7664.branch-0.20-security.patch, 
 HADOOP-7664.patch, HADOOP-7664.patch, HADOOP-7664.patch

   Original Estimate: 1m
  Time Spent: 1m
  Remaining Estimate: 0h

 o.a.h.conf.Configuration complains of overriding final parameter even if the 
 value with which its attempting to override is the same. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10824) Refactor KMSACLs to avoid locking

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063504#comment-14063504
 ] 

Hudson commented on HADOOP-10824:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-10824. Refactor KMSACLs to avoid locking. (Benoy Antony via umamahesh) 
(umamahesh: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610969)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSACLs.java


 Refactor KMSACLs to avoid locking
 -

 Key: HADOOP-10824
 URL: https://issues.apache.org/jira/browse/HADOOP-10824
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10824.patch, HADOOP-10824.patch


 Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible 
 to safely publish the _acls_ collection using _volatile_.
 Similar refactoring has been done in 
 [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112]



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10845) Add common tests for ACLs in combination with viewfs.

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063510#comment-14063510
 ] 

Hudson commented on HADOOP-10845:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-10845. Add common tests for ACLs in combination with viewfs. Contributed 
by Stephen Chu. (cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610911)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsBaseTest.java


 Add common tests for ACLs in combination with viewfs.
 -

 Key: HADOOP-10845
 URL: https://issues.apache.org/jira/browse/HADOOP-10845
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs, test
Affects Versions: 3.0.0, 2.5.0
Reporter: Chris Nauroth
Assignee: Stephen Chu
 Fix For: 3.0.0, 2.6.0

 Attachments: HADOOP-10845.1.patch


 Add tests in Hadoop Common for the ACL APIs in combination with viewfs.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10468) TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063500#comment-14063500
 ] 

Hudson commented on HADOOP-10468:
-

FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
Addendum patch for HADOOP-10468 TestMetricsSystemImpl.testMultiThreadedPublish 
fails intermediately. Contributed by Akira AJISAKA (jlowe: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610829)
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestGangliaMetrics.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java


 TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately
 ---

 Key: HADOOP-10468
 URL: https://issues.apache.org/jira/browse/HADOOP-10468
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.5.0
Reporter: Haohui Mai
Assignee: Akira AJISAKA
Priority: Blocker
 Fix For: 2.5.0

 Attachments: HADOOP-10468.000.patch, HADOOP-10468.001.patch, 
 HADOOP-10468.2.patch


 {{TestMetricsSystemImpl.testMultiThreadedPublish}} can fail intermediately 
 due to the insufficient size of the sink queue:
 {code}
 2014-04-06 21:34:55,269 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 2014-04-06 21:34:55,270 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 2014-04-06 21:34:55,271 WARN  impl.MetricsSinkAdapter 
 (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full 
 queue and can't consume the given metrics.
 {code}
 The unit test should increase the default queue size to avoid intermediate 
 failure.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063508#comment-14063508
 ] 

Hudson commented on HADOOP-9921:


FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See 
[https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/])
HADOOP-9921. daemon scripts should remove pid file on stop call after stop or 
process is found not running ( Contributed by Vinayakumar B) (vinayakumarb: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610964)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh
* /hadoop/common/trunk/hadoop-mapreduce-project/bin/mr-jobhistory-daemon.sh
* /hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/bin/yarn-daemon.sh


 daemon scripts should remove pid file on stop call after stop or process is 
 found not running
 -

 Key: HADOOP-9921
 URL: https://issues.apache.org/jira/browse/HADOOP-9921
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 3.0.0, 2.1.0-beta
Reporter: Vinayakumar B
Assignee: Vinayakumar B
 Fix For: 2.6.0

 Attachments: HADOOP-9921.patch


 daemon scripts should remove the pid file on stop call using daemon script.
 Should remove the pid file, even though process is not running.
 same pid file will be used by start command. At that time, if the same pid is 
 assigned to some other process, then start may fail.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063519#comment-14063519
 ] 

Hadoop QA commented on HADOOP-10799:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656034/HADOOP-10799.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 4 new 
or modified test files.

  {color:red}-1 javac{color}.  The applied patch generated 1268 javac 
compiler warnings (more than the trunk's current 1260 warnings).

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-auth hadoop-common-project/hadoop-common:

  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4291//testReport/
Javac warnings: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4291//artifact/trunk/patchprocess/diffJavacWarnings.txt
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4291//console

This message is automatically generated.

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10778) Use NativeCrc32 only if it is faster

2014-07-16 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063577#comment-14063577
 ] 

Todd Lipcon commented on HADOOP-10778:
--

Sure, here's the whole table (Native is always faster):

Direct Buffer Performance Table (bpc = byte-per-crc in MB/sec; #T = #Theads)
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|32 |  1 | 422.8 | 802.0 |  89.7% |1908.2 | 351.3% | 137.9% |
|32 |  2 | 316.7 | 622.1 |  96.4% |1853.9 | 485.3% | 198.0% |
|32 |  4 | 285.8 | 516.9 |  80.9% |1806.3 | 532.1% | 249.5% |
|32 |  8 | 171.5 | 304.2 |  77.3% |1052.5 | 513.5% | 246.0% |
|32 | 16 |  95.8 | 189.7 |  97.9% | 640.3 | 568.1% | 237.6% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|64 |  1 | 604.4 |1034.8 |  71.2% |1795.6 | 197.1% |  73.5% |
|64 |  2 | 506.4 | 865.6 |  70.9% |1752.2 | 246.0% | 102.4% |
|64 |  4 | 530.5 | 923.6 |  74.1% |1642.4 | 209.6% |  77.8% |
|64 |  8 | 300.5 | 468.2 |  55.8% |1222.6 | 306.8% | 161.1% |
|64 | 16 | 163.2 | 284.6 |  74.4% | 669.8 | 310.4% | 135.3% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|   128 |  1 | 782.8 |1184.3 |  51.3% |1694.8 | 116.5% |  43.1% |
|   128 |  2 | 759.5 |1147.2 |  51.1% |1690.3 | 122.6% |  47.3% |
|   128 |  4 | 737.2 |1110.1 |  50.6% |1634.1 | 121.7% |  47.2% |
|   128 |  8 | 488.2 | 602.1 |  23.3% |1317.1 | 169.8% | 118.8% |
|   128 | 16 | 281.2 | 375.9 |  33.7% | 794.8 | 182.6% | 111.4% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|   256 |  1 | 921.1 |1242.1 |  34.9% |1711.7 |  85.8% |  37.8% |
|   256 |  2 | 890.4 |1231.2 |  38.3% |1618.5 |  81.8% |  31.5% |
|   256 |  4 | 868.9 |1162.0 |  33.7% |1597.3 |  83.8% |  37.5% |
|   256 |  8 | 651.1 | 673.0 |   3.4% |1321.4 | 102.9% |  96.3% |
|   256 | 16 | 369.6 | 397.5 |   7.5% | 730.2 |  97.6% |  83.7% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|   512 |  1 | 993.0 |1292.9 |  30.2% |1653.6 |  66.5% |  27.9% |
|   512 |  2 | 926.7 |1274.4 |  37.5% |1635.4 |  76.5% |  28.3% |
|   512 |  4 | 918.0 |1217.7 |  32.6% |1589.3 |  73.1% |  30.5% |
|   512 |  8 | 778.0 | 711.6 |  -8.5% |1365.8 |  75.5% |  91.9% |
|   512 | 16 | 457.2 | 406.2 | -11.2% | 790.6 |  72.9% |  94.6% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|  1024 |  1 |1016.0 |1303.2 |  28.3% |1673.8 |  64.7% |  28.4% |
|  1024 |  2 | 987.0 |1264.9 |  28.2% |1637.3 |  65.9% |  29.4% |
|  1024 |  4 | 956.7 |1239.4 |  29.5% |1593.3 |  66.5% |  28.6% |
|  1024 |  8 | 837.3 | 716.4 | -14.4% |1336.8 |  59.7% |  86.6% |
|  1024 | 16 | 522.1 | 389.0 | -25.5% | 948.5 |  81.7% | 143.8% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|  2048 |  1 |1006.9 |1271.3 |  26.3% |1695.3 |  68.4% |  33.4% |
|  2048 |  2 | 989.1 |1234.1 |  24.8% |1652.3 |  67.1% |  33.9% |
|  2048 |  4 | 959.1 |1200.1 |  25.1% |1581.7 |  64.9% |  31.8% |
|  2048 |  8 | 875.4 | 713.4 | -18.5% |1422.8 |  62.5% |  99.4% |
|  2048 | 16 | 507.5 | 402.2 | -20.7% | 923.7 |  82.0% | 129.6% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|  4096 |  1 |1025.0 |1251.6 |  22.1% |1691.7 |  65.0% |  35.2% |
|  4096 |  2 | 990.7 |1219.8 |  23.1% |1630.2 |  64.6% |  33.6% |
|  4096 |  4 | 961.5 |1178.9 |  22.6% |1582.7 |  64.6% |  34.3% |
|  4096 |  8 | 891.1 | 714.0 | -19.9% |1429.4 |  60.4% | 100.2% |
|  4096 | 16 | 529.1 | 405.3 | -23.4% | 862.6 |  63.0% | 112.8% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
|  8192 |  1 |1015.4 |1290.6 |  27.1% |1685.9 |  66.0% |  30.6% |
|  8192 |  2 |1007.6 |1266.1 |  25.7% |1619.9 |  60.8% |  27.9% |
|  8192 |  4 | 981.5 |1208.1 |  23.1% |1602.3 |  63.3% |  32.6% |
|  8192 |  8 | 903.9 | 711.7 | -21.3% |1426.6 |  57.8% | 100.4% |
|  8192 | 16 | 523.9 | 412.6 | -21.2% | 812.4 |  55.1% |  96.9% |
|  bpc  | #T ||  Zip || PureJava | % diff ||   Native | % diff | % diff |
| 16384 |  1 |1027.2 |1319.0 |  28.4% |1656.1 |  61.2% |  25.6% |
| 16384 |  2 |1009.8 |1259.3 |  24.7% |1647.4 |  63.1% |  30.8% |
| 16384 |  4 | 977.9 |1222.2 |  25.0% |1593.4 |  62.9% |  30.4% |
| 16384 |  8 | 902.6 | 709.6 | -21.4% |1424.7 |  57.8% | 100.8% |
| 16384 

[jira] [Commented] (HADOOP-10778) Use NativeCrc32 only if it is faster

2014-07-16 Thread Todd Lipcon (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063609#comment-14063609
 ] 

Todd Lipcon commented on HADOOP-10778:
--

BTW, what clock rate is your laptop? The best non-hardware CRC32 implementation 
I'm aware of (http://code.google.com/p/crcutil/) goes 1 cycle per byte. So, my 
guess is that your 2.3GB/sec is probably from a 2.3 (or so) Ghz machine using 
an algorithm like crcutil's. Perhaps Apple patched their zlib to use this 
algorithm instead of the very simple implementation in the upstream zlib?

 Use NativeCrc32 only if it is faster
 

 Key: HADOOP-10778
 URL: https://issues.apache.org/jira/browse/HADOOP-10778
 Project: Hadoop Common
  Issue Type: Improvement
  Components: util
Reporter: Tsz Wo Nicholas Sze
Assignee: Tsz Wo Nicholas Sze
 Attachments: c10778_20140702.patch


 From the benchmark post in [this 
 comment|https://issues.apache.org/jira/browse/HDFS-6560?focusedCommentId=14044060page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14044060],
  NativeCrc32 is slower than java.util.zip.CRC32 for Java 7 and above when 
 bytesPerChecksum  512.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed

2014-07-16 Thread Dmitry Sivachenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitry Sivachenko updated HADOOP-10783:
---

Status: Patch Available  (was: Open)

I am attaching another patch (trying to solve problems found by Jenkins)

 apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
 ---

 Key: HADOOP-10783
 URL: https://issues.apache.org/jira/browse/HADOOP-10783
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.4.1
Reporter: Dmitry Sivachenko
Assignee: Steve Loughran
 Attachments: commons-lang3_1.patch


 Hadoop-2.4.1 ships with apache-commons.jar version 2.6.
 It does not support FreeBSD (IS_OS_UNIX returns False).
 This is fixed in recent versions of apache-commons.jar
 Please update apache-commons.jar to recent version so it correctly recognizes 
 FreeBSD as UNIX-like system.
 Right now I get in datanode's log:
 2014-07-04 11:58:10,459 DEBUG 
 org.apache.hadoop.hdfs.server.datanode.ShortCircui
 tRegistry: Disabling ShortCircuitRegistry
 java.io.IOException: The OS is not UNIX.
 at 
 org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed

2014-07-16 Thread Dmitry Sivachenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitry Sivachenko updated HADOOP-10783:
---

Attachment: commons-lang3_1.patch

 apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
 ---

 Key: HADOOP-10783
 URL: https://issues.apache.org/jira/browse/HADOOP-10783
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.4.1
Reporter: Dmitry Sivachenko
Assignee: Steve Loughran
 Attachments: commons-lang3_1.patch


 Hadoop-2.4.1 ships with apache-commons.jar version 2.6.
 It does not support FreeBSD (IS_OS_UNIX returns False).
 This is fixed in recent versions of apache-commons.jar
 Please update apache-commons.jar to recent version so it correctly recognizes 
 FreeBSD as UNIX-like system.
 Right now I get in datanode's log:
 2014-07-04 11:58:10,459 DEBUG 
 org.apache.hadoop.hdfs.server.datanode.ShortCircui
 tRegistry: Disabling ShortCircuitRegistry
 java.io.IOException: The OS is not UNIX.
 at 
 org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed

2014-07-16 Thread Dmitry Sivachenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitry Sivachenko updated HADOOP-10783:
---

Status: Open  (was: Patch Available)

 apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
 ---

 Key: HADOOP-10783
 URL: https://issues.apache.org/jira/browse/HADOOP-10783
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.4.1
Reporter: Dmitry Sivachenko
Assignee: Steve Loughran
 Attachments: commons-lang3_1.patch


 Hadoop-2.4.1 ships with apache-commons.jar version 2.6.
 It does not support FreeBSD (IS_OS_UNIX returns False).
 This is fixed in recent versions of apache-commons.jar
 Please update apache-commons.jar to recent version so it correctly recognizes 
 FreeBSD as UNIX-like system.
 Right now I get in datanode's log:
 2014-07-04 11:58:10,459 DEBUG 
 org.apache.hadoop.hdfs.server.datanode.ShortCircui
 tRegistry: Disabling ShortCircuitRegistry
 java.io.IOException: The OS is not UNIX.
 at 
 org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed

2014-07-16 Thread Dmitry Sivachenko (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Dmitry Sivachenko updated HADOOP-10783:
---

Attachment: (was: commons-lang3.patch)

 apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
 ---

 Key: HADOOP-10783
 URL: https://issues.apache.org/jira/browse/HADOOP-10783
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.4.1
Reporter: Dmitry Sivachenko
Assignee: Steve Loughran
 Attachments: commons-lang3_1.patch


 Hadoop-2.4.1 ships with apache-commons.jar version 2.6.
 It does not support FreeBSD (IS_OS_UNIX returns False).
 This is fixed in recent versions of apache-commons.jar
 Please update apache-commons.jar to recent version so it correctly recognizes 
 FreeBSD as UNIX-like system.
 Right now I get in datanode's log:
 2014-07-04 11:58:10,459 DEBUG 
 org.apache.hadoop.hdfs.server.datanode.ShortCircui
 tRegistry: Disabling ShortCircuitRegistry
 java.io.IOException: The OS is not UNIX.
 at 
 org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77)
 at 
 org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041)
 at 
 org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065)



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10824) Refactor KMSACLs to avoid locking

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10824:
--

Affects Version/s: 2.4.1

 Refactor KMSACLs to avoid locking
 -

 Key: HADOOP-10824
 URL: https://issues.apache.org/jira/browse/HADOOP-10824
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10824.patch, HADOOP-10824.patch


 Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible 
 to safely publish the _acls_ collection using _volatile_.
 Similar refactoring has been done in 
 [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112]



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10831) UserProvider is not thread safe

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10831:
--

Affects Version/s: 2.4.1

 UserProvider is not thread safe
 ---

 Key: HADOOP-10831
 URL: https://issues.apache.org/jira/browse/HADOOP-10831
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Fix For: 3.0.0

 Attachments: HADOOP-10831.patch


 While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread 
 safe.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10834) Typo in CredentialShell usage

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10834:
--

Affects Version/s: 2.4.1

 Typo in CredentialShell usage
 -

 Key: HADOOP-10834
 URL: https://issues.apache.org/jira/browse/HADOOP-10834
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
Priority: Minor
 Fix For: 3.0.0

 Attachments: HADOOP-10834.patch


 There is a typo in one of the informational messages in _ CredentialShell_
 {code}
 out.println(Nothing has been be deleted.);
 {code}



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10830:
--

Attachment: HADOOP-10830.patch

Thanks for the review, [~vinayrpet]. I have updated the patch accordingly.

 Missing lock in JavaKeyStoreProvider.createCredentialEntry
 --

 Key: HADOOP-10830
 URL: https://issues.apache.org/jira/browse/HADOOP-10830
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10830.patch, HADOOP-10830.patch


 _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_  to provide thread 
 safety.
 The {{createCredentialEntry}} should hold _writeLock_ before adding the entry.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063801#comment-14063801
 ] 

Hadoop QA commented on HADOOP-10830:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656072/HADOOP-10830.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common:

  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem
  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4293//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4293//console

This message is automatically generated.

 Missing lock in JavaKeyStoreProvider.createCredentialEntry
 --

 Key: HADOOP-10830
 URL: https://issues.apache.org/jira/browse/HADOOP-10830
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10830.patch, HADOOP-10830.patch


 _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_  to provide thread 
 safety.
 The {{createCredentialEntry}} should hold _writeLock_ before adding the entry.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Alejandro Abdelnur (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063807#comment-14063807
 ] 

Alejandro Abdelnur commented on HADOOP-10799:
-

the javac warnings are warnings from Maven regarding the maven-jar-plugin 
defined twice (it does not seem to be the case). I'm not able to reproduce the 
warnings locally with Maven 3.0.4

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10735) Fall back AesCtrCryptoCodec implementation from OpenSSL to JCE if non native support.

2014-07-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063823#comment-14063823
 ] 

Colin Patrick McCabe commented on HADOOP-10735:
---

Thanks, Yi.

 Fall back AesCtrCryptoCodec implementation from OpenSSL to JCE if non native 
 support.
 -

 Key: HADOOP-10735
 URL: https://issues.apache.org/jira/browse/HADOOP-10735
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: fs-encryption (HADOOP-10150 and HDFS-6134)
Reporter: Yi Liu
Assignee: Yi Liu
 Fix For: fs-encryption (HADOOP-10150 and HDFS-6134)

 Attachments: HADOOP-10735.001.patch, HADOOP-10735.002.patch, 
 HADOOP-10735.003.patch, HADOOP-10735.004.patch, HADOOP-10735.005.patch, 
 HADOOP-10735.006.patch, HADOOP-10735.007.patch, HADOOP-10735.008.patch


 If there is no native support or OpenSSL version is too low not supporting 
 AES-CTR, but {{OpensslAesCtrCryptoCodec}} is configured, we need to fall back 
 it to JCE implementation.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HADOOP-10799:


Attachment: HADOOP-10799.patch

ok, found the issue, a commit from yesterday YARN-2233, was overlapping POM 
changes with the patch, patch rebased.

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10830:
--

Affects Version/s: 2.4.1

 Missing lock in JavaKeyStoreProvider.createCredentialEntry
 --

 Key: HADOOP-10830
 URL: https://issues.apache.org/jira/browse/HADOOP-10830
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10830.patch, HADOOP-10830.patch


 _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_  to provide thread 
 safety.
 The {{createCredentialEntry}} should hold _writeLock_ before adding the entry.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10833) Unused cache in UserProvider

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10833:
--

Affects Version/s: 2.4.1

 Unused cache in UserProvider
 

 Key: HADOOP-10833
 URL: https://issues.apache.org/jira/browse/HADOOP-10833
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10833.patch


 _UserProvider_ contains the field cache.
 {code}
   private final MapString, CredentialEntry cache = new HashMapString, 
   CredentialEntry();
 {code}
   
 It is referenced only in {{deleteCredentialEntry}} and so there is no real 
 usage of {{cache}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10833) Unused cache in UserProvider

2014-07-16 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063854#comment-14063854
 ] 

Benoy Antony commented on HADOOP-10833:
---

[~vinayrpet], Could you please review this patch also ?

 Unused cache in UserProvider
 

 Key: HADOOP-10833
 URL: https://issues.apache.org/jira/browse/HADOOP-10833
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10833.patch


 _UserProvider_ contains the field cache.
 {code}
   private final MapString, CredentialEntry cache = new HashMapString, 
   CredentialEntry();
 {code}
   
 It is referenced only in {{deleteCredentialEntry}} and so there is no real 
 usage of {{cache}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-9902) Shell script rewrite

2014-07-16 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-9902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-9902:
-

Attachment: HADOOP-9902-6.patch

vs -5, this patch fixes a kill output not sent to /dev/null, updates some of 
the hadoop-env.sh commentary, and fixes a code style issue in rotate_logs.

If there is interest, I can make a patch for 2.4.1.

 Shell script rewrite
 

 Key: HADOOP-9902
 URL: https://issues.apache.org/jira/browse/HADOOP-9902
 Project: Hadoop Common
  Issue Type: Improvement
  Components: scripts
Affects Versions: 3.0.0
Reporter: Allen Wittenauer
Assignee: Allen Wittenauer
  Labels: releasenotes
 Attachments: HADOOP-9902-2.patch, HADOOP-9902-3.patch, 
 HADOOP-9902-4.patch, HADOOP-9902-5.patch, HADOOP-9902-6.patch, 
 HADOOP-9902.patch, HADOOP-9902.txt, hadoop-9902-1.patch, more-info.txt


 Umbrella JIRA for shell script rewrite.  See more-info.txt for more details.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10833) Unused cache in UserProvider

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10833:
--

Description: 
_UserProvider_ contains the field cache.
{code}
  private final MapString, CredentialEntry cache = new HashMapString, 
  CredentialEntry();
{code}

It is referenced only in {{deleteCredentialEntry}} and so there is no real 
usage of {{cache}}.

  was:
_UserProvider_ contains the field cache.
{code}
  private final MapString, CredentialEntry cache = new HashMapString, 
  CredentialEntry();
{code}

It is referenced only in {{deleteCredentialEntry}} and so there is no real 
usage of {{cache}.


 Unused cache in UserProvider
 

 Key: HADOOP-10833
 URL: https://issues.apache.org/jira/browse/HADOOP-10833
 Project: Hadoop Common
  Issue Type: Improvement
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10833.patch


 _UserProvider_ contains the field cache.
 {code}
   private final MapString, CredentialEntry cache = new HashMapString, 
   CredentialEntry();
 {code}
   
 It is referenced only in {{deleteCredentialEntry}} and so there is no real 
 usage of {{cache}}.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10839) Add unregisterSource() to MetricsSystem API

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10839:
---

Issue Type: Improvement  (was: Bug)

 Add unregisterSource() to MetricsSystem API
 ---

 Key: HADOOP-10839
 URL: https://issues.apache.org/jira/browse/HADOOP-10839
 Project: Hadoop Common
  Issue Type: Improvement
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10839.2.patch, HADOOP-10839.patch


 Currently the MetrisSystem API has register() method to register a 
 MetricsSource but doesn't have unregister() method. This means once a 
 MetricsSource is registered with the MetricsSystem, it will be there forever 
 until the MetricsSystem is shut down. This in some cases can cause Java 
 OutOfMemoryError.
 One such case is in file system metrics implementation. The new 
 AbstractFileSystem/FileContext framework does not implement a cache so every 
 file system access can lead to the creation of a NativeFileSystem instance. 
 (refer to HADOOP-6356). And all these NativeFileSystem needs to share the 
 same instance of MetricsSystemImpl, which means we cannot shut down 
 MetricsSystem to clean up all the MetricsSources that has been registered but 
 no longer active. Over time the MetricsSource instance accumulates and 
 eventually we saw OutOfMemoryError.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10794) A hadoop cluster needs clock synchronization

2014-07-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063901#comment-14063901
 ] 

Colin Patrick McCabe commented on HADOOP-10794:
---

bq. I'm afraid each component may more or less have made the assumption that 
clock is synchronized across the cluster. 

Can you explain which components make this assumption, and what happens if it's 
violated?  So far the only example presented is statistics in YARN.

 A hadoop cluster needs clock synchronization
 

 Key: HADOOP-10794
 URL: https://issues.apache.org/jira/browse/HADOOP-10794
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Zhijie Shen

 As a distributed system, a hadoop cluster wants the clock on all the 
 participating hosts synchronized. Otherwise, some problems might happen. For 
 example, in YARN-2251, due to the clock on the host for the task container 
 falls behind that on the host of the AM container, the computed elapsed time 
 (the diff between the timestamps produced on two hosts) becomes negative.
 In YARN-2251, we tried to mask the negative elapsed time. However, we should 
 seek for a decent long term solution, such as providing mechanism to do and 
 check clock synchronization.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10839) Add unregisterSource() to MetricsSystem API

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10839:
---

   Resolution: Fixed
Fix Version/s: 2.6.0
   3.0.0
   Status: Resolved  (was: Patch Available)

I committed this to trunk and branch-2.  Shanyu, thank you for contributing 
this patch.

 Add unregisterSource() to MetricsSystem API
 ---

 Key: HADOOP-10839
 URL: https://issues.apache.org/jira/browse/HADOOP-10839
 Project: Hadoop Common
  Issue Type: Improvement
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Fix For: 3.0.0, 2.6.0

 Attachments: HADOOP-10839.2.patch, HADOOP-10839.patch


 Currently the MetrisSystem API has register() method to register a 
 MetricsSource but doesn't have unregister() method. This means once a 
 MetricsSource is registered with the MetricsSystem, it will be there forever 
 until the MetricsSystem is shut down. This in some cases can cause Java 
 OutOfMemoryError.
 One such case is in file system metrics implementation. The new 
 AbstractFileSystem/FileContext framework does not implement a cache so every 
 file system access can lead to the creation of a NativeFileSystem instance. 
 (refer to HADOOP-6356). And all these NativeFileSystem needs to share the 
 same instance of MetricsSystemImpl, which means we cannot shut down 
 MetricsSystem to clean up all the MetricsSources that has been registered but 
 no longer active. Over time the MetricsSource instance accumulates and 
 eventually we saw OutOfMemoryError.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10839) Add unregisterSource() to MetricsSystem API

2014-07-16 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063907#comment-14063907
 ] 

Hudson commented on HADOOP-10839:
-

FAILURE: Integrated in Hadoop-trunk-Commit #5893 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/5893/])
HADOOP-10839. Add unregisterSource() to MetricsSystem API. Contributed by 
Shanyu Zhao. (cnauroth: 
http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1611134)
* /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java
* 
/hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java


 Add unregisterSource() to MetricsSystem API
 ---

 Key: HADOOP-10839
 URL: https://issues.apache.org/jira/browse/HADOOP-10839
 Project: Hadoop Common
  Issue Type: Improvement
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Fix For: 3.0.0, 2.6.0

 Attachments: HADOOP-10839.2.patch, HADOOP-10839.patch


 Currently the MetrisSystem API has register() method to register a 
 MetricsSource but doesn't have unregister() method. This means once a 
 MetricsSource is registered with the MetricsSystem, it will be there forever 
 until the MetricsSystem is shut down. This in some cases can cause Java 
 OutOfMemoryError.
 One such case is in file system metrics implementation. The new 
 AbstractFileSystem/FileContext framework does not implement a cache so every 
 file system access can lead to the creation of a NativeFileSystem instance. 
 (refer to HADOOP-6356). And all these NativeFileSystem needs to share the 
 same instance of MetricsSystemImpl, which means we cannot shut down 
 MetricsSystem to clean up all the MetricsSources that has been registered but 
 no longer active. Over time the MetricsSource instance accumulates and 
 eventually we saw OutOfMemoryError.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10840:
---

Target Version/s: 3.0.0, 2.6.0

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10840:
---

Status: Patch Available  (was: Open)

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10840:
---

Target Version/s: 3.0.0  (was: 3.0.0, 2.6.0)

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063923#comment-14063923
 ] 

Hadoop QA commented on HADOOP-10799:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656090/HADOOP-10799.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 4 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-auth hadoop-common-project/hadoop-common:

  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem
  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4294//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4294//console

This message is automatically generated.

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063933#comment-14063933
 ] 

Chris Nauroth commented on HADOOP-10840:


Hi, [~shanyu].  I'm seeing some test failures when I run the tests under 
hadoop-tools/hadoop-azure.  Could you please investigate?  Here is a stack 
trace from one example:

{code}
Running org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck
Tests run: 4, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.871 sec  
FAILURE! - in org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck
testMkdirs(org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck)  
Time elapsed: 0.04 sec   ERROR!
java.lang.NullPointerException: null
at 
org.apache.hadoop.fs.azure.metrics.AzureFileSystemMetricsSystem.unregisterSource(AzureFileSystemMetricsSystem.java:65)
at 
org.apache.hadoop.fs.azure.NativeAzureFileSystem.close(NativeAzureFileSystem.java:1352)
at 
org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.cleanup(AzureBlobStorageTestAccount.java:745)
at 
org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck.tearDown(TestNativeAzureFileSystemFileNameCheck.java:55)
{code}


 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10840:
---

Status: Open  (was: Patch Available)

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Benoy Antony (JIRA)
Benoy Antony created HADOOP-10851:
-

 Summary: NetgroupCache does not remove group memberships
 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony


_NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based 
on net groups.
But it has a serious flaw in that once a user to group membership is 
established, it remains forever even if user is actually removed from the 
netgroup and cache is cleared.  It is cleared only if the server is restarted.

To reproduce this: 

* Cache a group with a set of users.
* Test membership correctness.
* Clear cache, remove a user from the group and cache the group again
* Expected result : user’s groups should not include the group from which 
he/she is removed. 
* Actual result : user’s groups includes the group from which he/she was 
removed.

It is also noted that _NetgroupCache_ has concurrency issues and a separate 
jira is filed to rectify them.




--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10851:
--

Attachment: HADOOP-10851.patch

The attached patch adds 3 test cases 
* a basic test for _NetgroupCache_
* test where a user is initially part of a group and later removed from that 
group.
* test where a group is initially added to cache, but later removed.

_NetgroupCache_ is fixed by clearing the cache properly.

Code is slightly refactored :
* uses {{addAll}} instead of looping through the set of entries.
* removed call to {{isCached}} inside {{add}} since the callers call 
{{isCached}} before calling {{add}}.
* two unused imports are removed.

 NetgroupCache does not remove group memberships
 ---

 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10851.patch


 _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations 
 based on net groups.
 But it has a serious flaw in that once a user to group membership is 
 established, it remains forever even if user is actually removed from the 
 netgroup and cache is cleared.  It is cleared only if the server is restarted.
 To reproduce this: 
 * Cache a group with a set of users.
 * Test membership correctness.
 * Clear cache, remove a user from the group and cache the group again
 * Expected result : user’s groups should not include the group from which 
 he/she is removed. 
 * Actual result : user’s groups includes the group from which he/she was 
 removed.
 It is also noted that _NetgroupCache_ has concurrency issues and a separate 
 jira is filed to rectify them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10851:
--

Status: Patch Available  (was: Open)

 NetgroupCache does not remove group memberships
 ---

 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10851.patch


 _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations 
 based on net groups.
 But it has a serious flaw in that once a user to group membership is 
 established, it remains forever even if user is actually removed from the 
 netgroup and cache is cleared.  It is cleared only if the server is restarted.
 To reproduce this: 
 * Cache a group with a set of users.
 * Test membership correctness.
 * Clear cache, remove a user from the group and cache the group again
 * Expected result : user’s groups should not include the group from which 
 he/she is removed. 
 * Actual result : user’s groups includes the group from which he/she was 
 removed.
 It is also noted that _NetgroupCache_ has concurrency issues and a separate 
 jira is filed to rectify them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed

2014-07-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063951#comment-14063951
 ] 

Colin Patrick McCabe commented on HADOOP-10591:
---

bq. Any reason we can't make CompressionOutputStream.trackedCompressor private?

Yeah, let's make it private.

bq. The javadoc for createInputStreamWithCodecPool says The codec to use to 
create the output stream.

Fixed.

 Compression codecs must used pooled direct buffers or deallocate direct 
 buffers when stream is closed
 -

 Key: HADOOP-10591
 URL: https://issues.apache.org/jira/browse/HADOOP-10591
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Hari Shreedharan
Assignee: Colin Patrick McCabe
 Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch


 Currently direct buffers allocated by compression codecs like Gzip (which 
 allocates 2 direct buffers per instance) are not deallocated when the stream 
 is closed. Eventually for long running processes which create a huge number 
 of files, these direct buffers are left hanging till a full gc, which may or 
 may not happen in a reasonable amount of time - especially if the process 
 does not use a whole lot of heap.
 Either these buffers should be pooled or they should be deallocated when the 
 stream is closed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed

2014-07-16 Thread Colin Patrick McCabe (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Colin Patrick McCabe updated HADOOP-10591:
--

Attachment: HADOOP-10591.002.patch

 Compression codecs must used pooled direct buffers or deallocate direct 
 buffers when stream is closed
 -

 Key: HADOOP-10591
 URL: https://issues.apache.org/jira/browse/HADOOP-10591
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Hari Shreedharan
Assignee: Colin Patrick McCabe
 Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch


 Currently direct buffers allocated by compression codecs like Gzip (which 
 allocates 2 direct buffers per instance) are not deallocated when the stream 
 is closed. Eventually for long running processes which create a huge number 
 of files, these direct buffers are left hanging till a full gc, which may or 
 may not happen in a reasonable amount of time - especially if the process 
 does not use a whole lot of heap.
 Either these buffers should be pooled or they should be deallocated when the 
 stream is closed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Created] (HADOOP-10852) NetgroupCache is not thread-safe

2014-07-16 Thread Benoy Antony (JIRA)
Benoy Antony created HADOOP-10852:
-

 Summary: NetgroupCache is not thread-safe
 Key: HADOOP-10852
 URL: https://issues.apache.org/jira/browse/HADOOP-10852
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony


_NetgroupCache_ internally uses two _ConcurrentHashMap_s and a boolean variable 
to signal updates on one of the _ConcurrentHashMap_s 
None of the functions are synchronized  and hence is possible to have 
unexpected results due to race condition between different threads.

As an example, consider the following sequence:

Thread1 :
{{add}} a group
{{netgroupToUsersMap}} is updated.
{{netgroupToUsersMapUpdated}} is set to true.
Thread 2:
calls {{getNetgroups}} for a user
Due to re-ordering, {{netgroupToUsersMapUpdated=true}} is visible, but updates 
in {{netgroupToUsersMap}} is not visible.
Does a wrong update with older {{netgroupToUsersMap}} values. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10851:
--

Attachment: HADOOP-10852.patch

The _NetgroupCache_ is fixed by keeping state in only one _ConcurrentHashMap_ - 
{{userToNetgroupsMap}}.
{{add}} updates {{userToNetgroupsMap}} directly.

There is a slight performance loss when invoking {{isCached}} and 
{{{userToNetgroupsMap}}
But these are invoked only during {{refresh}} and adding a new group. These 
invocations should be rare compared to {{getNetgroups}} invocation.

 NetgroupCache does not remove group memberships
 ---

 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10851.patch, HADOOP-10852.patch


 _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations 
 based on net groups.
 But it has a serious flaw in that once a user to group membership is 
 established, it remains forever even if user is actually removed from the 
 netgroup and cache is cleared.  It is cleared only if the server is restarted.
 To reproduce this: 
 * Cache a group with a set of users.
 * Test membership correctness.
 * Clear cache, remove a user from the group and cache the group again
 * Expected result : user’s groups should not include the group from which 
 he/she is removed. 
 * Actual result : user’s groups includes the group from which he/she was 
 removed.
 It is also noted that _NetgroupCache_ has concurrency issues and a separate 
 jira is filed to rectify them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063943#comment-14063943
 ] 

Hadoop QA commented on HADOOP-10783:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656057/commons-lang3_1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 31 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:red}-1 findbugs{color}.  The patch appears to cause Findbugs 
(version 2.0.3) to fail.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-auth hadoop-common-project/hadoop-common 
hadoop-common-project/hadoop-minikdc hadoop-hdfs-project/hadoop-hdfs 
hadoop-hdfs-project/hadoop-hdfs-nfs 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient
 hadoop-tools/hadoop-gridmix hadoop-tools/hadoop-rumen 
hadoop-tools/hadoop-streaming hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell
 hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager
 
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager:

  org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl
  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem
  org.apache.hadoop.streaming.TestMultipleArchiveFiles
  org.apache.hadoop.streaming.TestStreamingBadRecords
  org.apache.hadoop.streaming.TestStreamingOutputKeyValueTypes
  org.apache.hadoop.streaming.TestStreamAggregate
  org.apache.hadoop.streaming.TestStreamReduceNone
  org.apache.hadoop.streaming.TestUnconsumedInput
  
org.apache.hadoop.streaming.mapreduce.TestStreamXmlRecordReader
  org.apache.hadoop.streaming.TestGzipInput
  org.apache.hadoop.streaming.TestStreaming
  org.apache.hadoop.streaming.TestStreamingFailure
  org.apache.hadoop.streaming.TestStreamingSeparator
  org.apache.hadoop.streaming.TestStreamingCounters
  org.apache.hadoop.streaming.TestFileArgs
  org.apache.hadoop.streaming.TestStreamDataProtocol
  org.apache.hadoop.streaming.TestStreamingExitStatus
  org.apache.hadoop.streaming.TestSymLink
  org.apache.hadoop.streaming.TestRawBytesStreaming
  org.apache.hadoop.streaming.TestStreamingCombiner
  org.apache.hadoop.streaming.TestStreamingStderr
  org.apache.hadoop.streaming.TestTypedBytesStreaming
  org.apache.hadoop.streaming.TestStreamingBackground
  org.apache.hadoop.streaming.TestStreamXmlRecordReader
  org.apache.hadoop.streaming.TestStreamingKeyValue
  org.apache.hadoop.streaming.TestStreamXmlMultipleRecords
  org.apache.hadoop.streaming.TestMultipleCachefiles
  org.apache.hadoop.streaming.TestStreamingOutputOnlyKeys
  org.apache.hadoop.yarn.util.TestFSDownload
  
org.apache.hadoop.yarn.server.resourcemanager.TestClientRMService
  
org.apache.hadoop.yarn.server.resourcemanager.webapp.TestRMWebServicesApps
  org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart
  
org.apache.hadoop.yarn.server.resourcemanager.TestApplicationACLs
  org.apache.hadoop.yarn.server.resourcemanager.TestRM
  org.apache.hadoop.hdfs.tools.TestDFSAdminWithHA
  
org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover

  

[jira] [Commented] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Benoy Antony (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063959#comment-14063959
 ] 

Benoy Antony commented on HADOOP-10851:
---

The above comment is intended for a another jira.

 NetgroupCache does not remove group memberships
 ---

 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10851.patch


 _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations 
 based on net groups.
 But it has a serious flaw in that once a user to group membership is 
 established, it remains forever even if user is actually removed from the 
 netgroup and cache is cleared.  It is cleared only if the server is restarted.
 To reproduce this: 
 * Cache a group with a set of users.
 * Test membership correctness.
 * Clear cache, remove a user from the group and cache the group again
 * Expected result : user’s groups should not include the group from which 
 he/she is removed. 
 * Actual result : user’s groups includes the group from which he/she was 
 removed.
 It is also noted that _NetgroupCache_ has concurrency issues and a separate 
 jira is filed to rectify them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10851:
--

Attachment: (was: HADOOP-10852.patch)

 NetgroupCache does not remove group memberships
 ---

 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10851.patch


 _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations 
 based on net groups.
 But it has a serious flaw in that once a user to group membership is 
 established, it remains forever even if user is actually removed from the 
 netgroup and cache is cleared.  It is cleared only if the server is restarted.
 To reproduce this: 
 * Cache a group with a set of users.
 * Test membership correctness.
 * Clear cache, remove a user from the group and cache the group again
 * Expected result : user’s groups should not include the group from which 
 he/she is removed. 
 * Actual result : user’s groups includes the group from which he/she was 
 removed.
 It is also noted that _NetgroupCache_ has concurrency issues and a separate 
 jira is filed to rectify them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10852) NetgroupCache is not thread-safe

2014-07-16 Thread Benoy Antony (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Benoy Antony updated HADOOP-10852:
--

Attachment: HADOOP-10852.patch

The _NetgroupCache_ is fixed by keeping state in only one _ConcurrentHashMap_ - 
{{userToNetgroupsMap}}.
{{add}} updates {{userToNetgroupsMap}} directly.

There is a slight performance loss when invoking {{isCached}} and 
{{{userToNetgroupsMap}}
But these are invoked only during {{refresh}} and adding a new group. These 
invocations should be rare compared to {{getNetgroups}} invocation.

 NetgroupCache is not thread-safe
 

 Key: HADOOP-10852
 URL: https://issues.apache.org/jira/browse/HADOOP-10852
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10852.patch


 _NetgroupCache_ internally uses two _ConcurrentHashMap_s and a boolean 
 variable to signal updates on one of the _ConcurrentHashMap_s 
 None of the functions are synchronized  and hence is possible to have 
 unexpected results due to race condition between different threads.
 As an example, consider the following sequence:
 Thread1 :
 {{add}} a group
 {{netgroupToUsersMap}} is updated.
 {{netgroupToUsersMapUpdated}} is set to true.
 Thread 2:
 calls {{getNetgroups}} for a user
 Due to re-ordering, {{netgroupToUsersMapUpdated=true}} is visible, but 
 updates in {{netgroupToUsersMap}} is not visible.
 Does a wrong update with older {{netgroupToUsersMap}} values. 



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064015#comment-14064015
 ] 

Hadoop QA commented on HADOOP-10591:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  
http://issues.apache.org/jira/secure/attachment/12656112/HADOOP-10591.002.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
Please justify why no new tests are needed for this 
patch.
Also please list what manual steps were performed to 
verify this patch.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common:

  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC
  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4295//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4295//console

This message is automatically generated.

 Compression codecs must used pooled direct buffers or deallocate direct 
 buffers when stream is closed
 -

 Key: HADOOP-10591
 URL: https://issues.apache.org/jira/browse/HADOOP-10591
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Hari Shreedharan
Assignee: Colin Patrick McCabe
 Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch


 Currently direct buffers allocated by compression codecs like Gzip (which 
 allocates 2 direct buffers per instance) are not deallocated when the stream 
 is closed. Eventually for long running processes which create a huge number 
 of files, these direct buffers are left hanging till a full gc, which may or 
 may not happen in a reasonable amount of time - especially if the process 
 does not use a whole lot of heap.
 Either these buffers should be pooled or they should be deallocated when the 
 stream is closed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10851) NetgroupCache does not remove group memberships

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064048#comment-14064048
 ] 

Hadoop QA commented on HADOOP-10851:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656111/HADOOP-10851.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:red}-1 core tests{color}.  The patch failed these unit tests in 
hadoop-common-project/hadoop-common:

  org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem
  org.apache.hadoop.fs.TestSymlinkLocalFSFileContext
  org.apache.hadoop.ipc.TestIPC

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4296//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4296//console

This message is automatically generated.

 NetgroupCache does not remove group memberships
 ---

 Key: HADOOP-10851
 URL: https://issues.apache.org/jira/browse/HADOOP-10851
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 2.4.1
Reporter: Benoy Antony
Assignee: Benoy Antony
 Attachments: HADOOP-10851.patch


 _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations 
 based on net groups.
 But it has a serious flaw in that once a user to group membership is 
 established, it remains forever even if user is actually removed from the 
 netgroup and cache is cleared.  It is cleared only if the server is restarted.
 To reproduce this: 
 * Cache a group with a set of users.
 * Test membership correctness.
 * Clear cache, remove a user from the group and cache the group again
 * Expected result : user’s groups should not include the group from which 
 he/she is removed. 
 * Actual result : user’s groups includes the group from which he/she was 
 removed.
 It is also noted that _NetgroupCache_ has concurrency issues and a separate 
 jira is filed to rectify them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10504) Document proxy server support

2014-07-16 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated HADOOP-10504:
--

Target Version/s: 2.6.0  (was: 2.5.0)

Moving this out to 2.6 since no one is actively looking at it. If anyone feels 
strongly about targeting 2.5, we can target 2.5. 

 Document proxy server support
 -

 Key: HADOOP-10504
 URL: https://issues.apache.org/jira/browse/HADOOP-10504
 Project: Hadoop Common
  Issue Type: Improvement
  Components: documentation
Affects Versions: 3.0.0, 2.5.0
Reporter: Daryn Sharp
Priority: Blocker

 Document http proxy support introduced by HADOOP-10498.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10453) Do not use AuthenticatedURL in hadoop core

2014-07-16 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla updated HADOOP-10453:
--

Target Version/s: 2.6.0  (was: 2.5.0)

(Moving this out of 2.5)

We can continue this conversation and handle this in 2.6. 

 Do not use AuthenticatedURL in hadoop core
 --

 Key: HADOOP-10453
 URL: https://issues.apache.org/jira/browse/HADOOP-10453
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Haohui Mai
Priority: Blocker

 As [~daryn] has suggested in HDFS-4564:
 {quote}
 AuthenticatedURL is not used because it is buggy in part to causing replay 
 attacks, double attempts to kerberos authenticate with the fallback 
 authenticator if the TGT is expired, incorrectly uses the fallback 
 authenticator (required by oozie servers) to add the username parameter which 
 webhdfs has already included in the uri.
 AuthenticatedURL's attempt to do SPNEGO auth is a no-op because the JDK 
 transparently does SPNEGO when the user's Subject (UGI) contains kerberos 
 principals. Since AuthenticatedURL is now not used, webhdfs has to check the 
 TGT itself for token operations.
 Bottom line is AuthenticatedURL is unnecessary and introduces nothing but 
 problems for webhdfs. It's only useful for oozie's anon/non-anon support.
 {quote}
 However, several functionalities that relies on SPNEGO in secure mode suffer 
 from the same problem. For example, NNs / JNs create HTTP connections to 
 exchange fsimage and edit logs. Currently all of them are through 
 {{AuthenticatedURL}}. This needs to be fixed to avoid security 
 vulnerabilities.
 This jira purposes to remove {{AuthenticatedURL}} from hadoop core and to 
 move it to oozie.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Assigned] (HADOOP-10821) Prepare the release notes for Hadoop 2.5.0

2014-07-16 Thread Karthik Kambatla (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Karthik Kambatla reassigned HADOOP-10821:
-

Assignee: Karthik Kambatla

 Prepare the release notes for Hadoop 2.5.0
 --

 Key: HADOOP-10821
 URL: https://issues.apache.org/jira/browse/HADOOP-10821
 Project: Hadoop Common
  Issue Type: Task
Affects Versions: 2.5.0
Reporter: Akira AJISAKA
Assignee: Karthik Kambatla
Priority: Blocker

 The release notes for 2.3.0+ 
 (http://hadoop.apache.org/docs/r2.4.1/index.html) still talk about federation 
 and MRv2
 being new features. We should update them.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread shanyu zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

shanyu zhao updated HADOOP-10840:
-

Attachment: HADOOP-10840.1.patch

[~cnauroth] thanks for the findings. I couldn't reproduce the specific failure 
you posted, but I think it is caused by NativeAzureFileSystem.close() being 
called multiple times. I verified that NativeAzureFileSystemStore.close() can 
be called multiple times, but not NativeAzureFileSystem. And for this class, 
since the metrics system keeps a reference counting, we cannot call its close() 
multiple times, so I introduced Boolean isClosed to prevent it from being 
called multiple times. I added a new test case to call 
NativeAzureFileSystem.close() twice to verify this scenario.

New patch attached.

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread shanyu zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

shanyu zhao updated HADOOP-10840:
-

Attachment: (was: HADOOP-10840.1.patch)

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Chris Nauroth updated HADOOP-10840:
---

Status: Patch Available  (was: Open)

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread shanyu zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

shanyu zhao updated HADOOP-10840:
-

Attachment: HADOOP-10840.1.patch

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.

2014-07-16 Thread Alejandro Abdelnur (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Alejandro Abdelnur updated HADOOP-10799:


Status: Open  (was: Patch Available)

canceling patch as some minor tweaks seems to be required in the DT client API 
to make DT wiring transparent.

 Refactor HTTP delegation token logic from httpfs into reusable code in 
 hadoop-common.
 -

 Key: HADOOP-10799
 URL: https://issues.apache.org/jira/browse/HADOOP-10799
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: security
Affects Versions: 3.0.0
Reporter: Alejandro Abdelnur
Assignee: Alejandro Abdelnur
 Attachments: HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, 
 HADOOP-10799.patch, HADOOP-10799.patch






--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064258#comment-14064258
 ] 

Hadoop QA commented on HADOOP-10840:


{color:green}+1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656152/HADOOP-10840.1.patch
  against trunk revision .

{color:green}+1 @author{color}.  The patch does not contain any @author 
tags.

{color:green}+1 tests included{color}.  The patch appears to include 1 new 
or modified test files.

{color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

{color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

{color:green}+1 eclipse:eclipse{color}.  The patch built with 
eclipse:eclipse.

{color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

{color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

{color:green}+1 core tests{color}.  The patch passed unit tests in 
hadoop-tools/hadoop-azure.

{color:green}+1 contrib tests{color}.  The patch passed contrib unit tests.

Test results: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4297//testReport/
Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4297//console

This message is automatically generated.

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Chris Nauroth (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064274#comment-14064274
 ] 

Chris Nauroth commented on HADOOP-10840:


[~shanyu], nice find.  I think your theory was correct.  The last patch mostly 
fixed things, but I still see a few test failures.  With an Azure storage key 
configured for testing against the live service, I get a failure in 
{{TestAzureConcurrentOutOfBandIo}}:

{code}
Running org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo
Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.81 sec  
FAILURE! - in org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo
testReadOOBWrites(org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo)  
Time elapsed: 0.765 sec   ERROR!
org.apache.hadoop.metrics2.MetricsException: Metrics source 
AzureFileSystemMetrics already exists!
at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.newSourceName(DefaultMetricsSystem.java:143)
at 
org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.sourceName(DefaultMetricsSystem.java:120)
at 
org.apache.hadoop.metrics2.impl.MetricsSystemImpl.register(MetricsSystemImpl.java:218)
at 
org.apache.hadoop.fs.azure.metrics.AzureFileSystemMetricsSystem.registerSource(AzureFileSystemMetricsSystem.java:58)
at 
org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.createOutOfBandStore(AzureBlobStorageTestAccount.java:331)
at 
org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo.setUp(TestAzureConcurrentOutOfBandIo.java:51)
{code}

I tested on both Mac and Windows.  On the Windows VM only, I also get failures 
in {{TestRollingWindowAverage}} and {{TestNativeAzureFileSystemMocked}}:

{code}
Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.149 sec  
FAILURE! - in org.apache.hadoop.fs.azure.metrics.TestRollingWindowAverage
testBasicFunctionality(org.apache.hadoop.fs.azure.metrics.TestRollingWindowAverage)
  Time elapsed: 0.112 sec   FAILURE!
java.lang.AssertionError: expected:15 but was:10
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.fs.azure.metrics.TestRollingWindowAverage.testBasicFunctionality(TestRollingWindowAverage.java:38)
{code}

{code}
Tests run: 27, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.431 sec  
FAILURE! - in org.apache.hadoop.fs.azure.TestNativeAzureFileSystemMocked
testFolderLastModifiedTime(org.apache.hadoop.fs.azure.TestNativeAzureFileSystemMocked)
  Time elapsed: 3.24 sec   FAILURE!
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertFalse(Assert.java:64)
at org.junit.Assert.assertFalse(Assert.java:74)
at 
org.apache.hadoop.fs.azure.NativeAzureFileSystemBaseTest.testFolderLastModifiedTime(NativeAzureFileSystemBaseTest.java:479)
{code}

Can you explain why the following code was removed from 
{{AzureFileSystemMetricsSystem#fileSystemClosed}}?  My understanding is that 
this code is important to guarantee timely publishing of metrics for an 
instance when it gets closed.  I expect your new checks against double close 
are also sufficient to protect against extraneous publishing of metrics.

{code}
-if (instance != null) {
-  instance.publishMetricsNow();
-}
{code}


 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10794) A hadoop cluster needs clock synchronization

2014-07-16 Thread eric baldeschwieler (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064326#comment-14064326
 ] 

eric baldeschwieler commented on HADOOP-10794:
--

Agreed. Doing math between server clocks is an anti-pattern to be avoided.
We should file JIRAs to fix any instances of that , because the result is
not a usable number given the error bars today.

On Wednesday, July 16, 2014, Colin Patrick McCabe (JIRA) j...@apache.org



-- 
---
E14 - typing on glass


 A hadoop cluster needs clock synchronization
 

 Key: HADOOP-10794
 URL: https://issues.apache.org/jira/browse/HADOOP-10794
 Project: Hadoop Common
  Issue Type: Bug
Reporter: Zhijie Shen

 As a distributed system, a hadoop cluster wants the clock on all the 
 participating hosts synchronized. Otherwise, some problems might happen. For 
 example, in YARN-2251, due to the clock on the host for the task container 
 falls behind that on the host of the AM container, the computed elapsed time 
 (the diff between the timestamps produced on two hosts) becomes negative.
 In YARN-2251, we tried to mask the negative elapsed time. However, we should 
 seek for a decent long term solution, such as providing mechanism to do and 
 check clock synchronization.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread shanyu zhao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

shanyu zhao updated HADOOP-10840:
-

Attachment: HADOOP-10840.2.patch

Hi [~cnauroth], sorry I missed the live test scenarios which I have to manually 
configure a storage account. The problem with TestAzureConcurrentOutOfBandIo is 
I removed the registration code with DefaultFileSystem in production code but 
not the unit test files. I fixed that in the new patch (v2).

Regarding the two windows test failure, looks like they are not related to my 
changes. It has something to do with system clock. Do they always fail or just 
transient? I cannot reproduce in my Ubuntu box. Let me try it on Windows box.

I removed the publishMetricsNow() from fileSystemClosed() because I moved it to 
unregisterSource(). The expectation is the caller calls unregisterSource() 
before calling fileSystemClosed() which will push the last metrics out. To 
prevent confusion when someone doesn't call unregisterSource() but just call 
fileSystemClosed() I just added publishMetricsNow() to fileSystemClosed() under 
the condition that numFileSystems==1, this way before we are going to 
shutdown the metrics system, we'll do another safety push, usually at this 
point the metrics sources should be empty.



 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.2.patch, 
 HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System

2014-07-16 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064383#comment-14064383
 ] 

Hadoop QA commented on HADOOP-10840:


{color:red}-1 overall{color}.  Here are the results of testing the latest 
attachment 
  http://issues.apache.org/jira/secure/attachment/12656178/HADOOP-10840.2.patch
  against trunk revision .

{color:red}-1 patch{color}.  Trunk compilation may be broken.

Console output: 
https://builds.apache.org/job/PreCommit-HADOOP-Build/4298//console

This message is automatically generated.

 Fix OutOfMemoryError caused by metrics system in Azure File System
 --

 Key: HADOOP-10840
 URL: https://issues.apache.org/jira/browse/HADOOP-10840
 Project: Hadoop Common
  Issue Type: Bug
  Components: metrics
Affects Versions: 2.4.1
Reporter: shanyu zhao
Assignee: shanyu zhao
 Attachments: HADOOP-10840.1.patch, HADOOP-10840.2.patch, 
 HADOOP-10840.patch


 In Hadoop 2.x the Hadoop File System framework changed and no cache is 
 implemented (refer to HADOOP-6356). This means for every WASB access, a new 
 NativeAzureFileSystem is created, along which a Metrics source created and 
 added to MetricsSystemImpl. Over time the sources accumulated, eating memory 
 and causing Java OutOfMemoryError.
 The fix is to utilize the unregisterSource() method added to MetricsSystem in 
 HADOOP-10839.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed

2014-07-16 Thread Colin Patrick McCabe (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064390#comment-14064390
 ] 

Colin Patrick McCabe commented on HADOOP-10591:
---

Test failures are unrelated.  TestIPC is timing out resolving a hostname (looks 
like a jenkins problem), and the symlink tests have been failing for some other 
patches too.

 Compression codecs must used pooled direct buffers or deallocate direct 
 buffers when stream is closed
 -

 Key: HADOOP-10591
 URL: https://issues.apache.org/jira/browse/HADOOP-10591
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Hari Shreedharan
Assignee: Colin Patrick McCabe
 Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch


 Currently direct buffers allocated by compression codecs like Gzip (which 
 allocates 2 direct buffers per instance) are not deallocated when the stream 
 is closed. Eventually for long running processes which create a huge number 
 of files, these direct buffers are left hanging till a full gc, which may or 
 may not happen in a reasonable amount of time - especially if the process 
 does not use a whole lot of heap.
 Either these buffers should be pooled or they should be deallocated when the 
 stream is closed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed

2014-07-16 Thread Aaron T. Myers (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064393#comment-14064393
 ] 

Aaron T. Myers commented on HADOOP-10591:
-

Agreed.

+1, the latest patch looks good to me. Thanks, Colin.

 Compression codecs must used pooled direct buffers or deallocate direct 
 buffers when stream is closed
 -

 Key: HADOOP-10591
 URL: https://issues.apache.org/jira/browse/HADOOP-10591
 Project: Hadoop Common
  Issue Type: Bug
Affects Versions: 2.2.0
Reporter: Hari Shreedharan
Assignee: Colin Patrick McCabe
 Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch


 Currently direct buffers allocated by compression codecs like Gzip (which 
 allocates 2 direct buffers per instance) are not deallocated when the stream 
 is closed. Eventually for long running processes which create a huge number 
 of files, these direct buffers are left hanging till a full gc, which may or 
 may not happen in a reasonable amount of time - especially if the process 
 does not use a whole lot of heap.
 Either these buffers should be pooled or they should be deallocated when the 
 stream is closed.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Commented] (HADOOP-10816) key shell returns -1 to the shell on error, should be 1

2014-07-16 Thread Andrew Wang (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-10816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064425#comment-14064425
 ] 

Andrew Wang commented on HADOOP-10816:
--

Your pedantry has bested me, let's go with with the v2 patch :) Thanks for 
being a good sport Mike, I'll commit this shortly.

 key shell returns -1 to the shell on error, should be 1
 ---

 Key: HADOOP-10816
 URL: https://issues.apache.org/jira/browse/HADOOP-10816
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 3.0.0
Reporter: Mike Yoder
Assignee: Mike Yoder
 Fix For: 3.0.0

 Attachments: HADOOP-10816.001.patch, HADOOP-10816.002.patch


 I've seen this in several places now - commands returning -1 on failure to 
 the shell. It's a bug. Someone confused their posix style returns (0 on 
 success,  0 on failure) with program returns, which are an unsigned 
 character. Thus, a return of -1 actually becomes 255 to the shell.
 {noformat}
 $ hadoop key create happykey2 --provider kms://http@localhost:16000/kms 
 --attr a=a --attr a=b
 Each attribute must correspond to only one value:
 atttribute a was repeated
 ...
 $ echo $?
 255
 {noformat}
 A return value of 1 instead of -1 does the right thing.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


[jira] [Updated] (HADOOP-10816) KeyShell returns -1 on error to the shell, should be 1

2014-07-16 Thread Andrew Wang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-10816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Andrew Wang updated HADOOP-10816:
-

Summary: KeyShell returns -1 on error to the shell, should be 1  (was: key 
shell returns -1 to the shell on error, should be 1)

 KeyShell returns -1 on error to the shell, should be 1
 --

 Key: HADOOP-10816
 URL: https://issues.apache.org/jira/browse/HADOOP-10816
 Project: Hadoop Common
  Issue Type: Bug
  Components: security
Affects Versions: 3.0.0
Reporter: Mike Yoder
Assignee: Mike Yoder
 Fix For: 3.0.0

 Attachments: HADOOP-10816.001.patch, HADOOP-10816.002.patch


 I've seen this in several places now - commands returning -1 on failure to 
 the shell. It's a bug. Someone confused their posix style returns (0 on 
 success,  0 on failure) with program returns, which are an unsigned 
 character. Thus, a return of -1 actually becomes 255 to the shell.
 {noformat}
 $ hadoop key create happykey2 --provider kms://http@localhost:16000/kms 
 --attr a=a --attr a=b
 Each attribute must correspond to only one value:
 atttribute a was repeated
 ...
 $ echo $?
 255
 {noformat}
 A return value of 1 instead of -1 does the right thing.



--
This message was sent by Atlassian JIRA
(v6.2#6252)


  1   2   >