[jira] [Created] (HADOOP-10849) Implement conf substitution with UGI.current/loginUser
Gera Shegalov created HADOOP-10849: -- Summary: Implement conf substitution with UGI.current/loginUser Key: HADOOP-10849 URL: https://issues.apache.org/jira/browse/HADOOP-10849 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 2.4.1 Reporter: Gera Shegalov Assignee: Gera Shegalov Many path properties and similar in hadoop code base would be easily configured if we had substitutions with {{UserGroupInformation#getCurrentUser}}. Currently we often use less elegant concatenation code if we want to express currentUser as opposed to ${user.name} system property representing the user owning the JVM. This JIRA proposes the corresponding substitution support for keys {{ugi.current.user}} and {{ugi.login.user}} -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10849) Implement conf substitution with UGI.current/loginUser
[ https://issues.apache.org/jira/browse/HADOOP-10849?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Gera Shegalov updated HADOOP-10849: --- Attachment: HADOOP-10849.v01.patch v01 to give an idea. Implement conf substitution with UGI.current/loginUser -- Key: HADOOP-10849 URL: https://issues.apache.org/jira/browse/HADOOP-10849 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 2.4.1 Reporter: Gera Shegalov Assignee: Gera Shegalov Attachments: HADOOP-10849.v01.patch Many path properties and similar in hadoop code base would be easily configured if we had substitutions with {{UserGroupInformation#getCurrentUser}}. Currently we often use less elegant concatenation code if we want to express currentUser as opposed to ${user.name} system property representing the user owning the JVM. This JIRA proposes the corresponding substitution support for keys {{ugi.current.user}} and {{ugi.login.user}} -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10720) KMS: Implement generateEncryptedKey and decryptEncryptedKey in the REST API
[ https://issues.apache.org/jira/browse/HADOOP-10720?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063216#comment-14063216 ] Hadoop QA commented on HADOOP-10720: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12655994/HADOOP-10720.11.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 2 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common hadoop-common-project/hadoop-kms: org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4288//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4288//console This message is automatically generated. KMS: Implement generateEncryptedKey and decryptEncryptedKey in the REST API --- Key: HADOOP-10720 URL: https://issues.apache.org/jira/browse/HADOOP-10720 Project: Hadoop Common Issue Type: Improvement Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Arun Suresh Attachments: COMBO.patch, COMBO.patch, COMBO.patch, COMBO.patch, COMBO.patch, HADOOP-10720.1.patch, HADOOP-10720.10.patch, HADOOP-10720.11.patch, HADOOP-10720.2.patch, HADOOP-10720.3.patch, HADOOP-10720.4.patch, HADOOP-10720.5.patch, HADOOP-10720.6.patch, HADOOP-10720.7.patch, HADOOP-10720.8.patch, HADOOP-10720.9.patch, HADOOP-10720.patch, HADOOP-10720.patch, HADOOP-10720.patch, HADOOP-10720.patch, HADOOP-10720.patch KMS client/server should implement support for generating encrypted keys and decrypting them via the REST API being introduced by HADOOP-10719. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry
[ https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063236#comment-14063236 ] Vinayakumar B commented on HADOOP-10830: Hi [~benoyantony], thanks for the patch. There is no need of double try-catch-finally blocks. existing itself you can re-use. Like following.. {code} @Override public CredentialEntry createCredentialEntry(String alias, char[] credential) throws IOException { writeLock.lock(); try { if (keyStore.containsAlias(alias) || cache.containsKey(alias)) { throw new IOException(Credential + alias + already exists in + this); } return innerSetCredential(alias, credential); } catch (KeyStoreException e) { throw new IOException(Problem looking up credential + alias + in + this, e); } finally { writeLock.unlock(); } }{code} anyway {{innerSetCredential(..)}} have its own try-catch block. So its fine to use one try-catch-finally block itself. Missing lock in JavaKeyStoreProvider.createCredentialEntry -- Key: HADOOP-10830 URL: https://issues.apache.org/jira/browse/HADOOP-10830 Project: Hadoop Common Issue Type: Bug Components: security Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10830.patch _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_ to provide thread safety. The {{createCredentialEntry}} should hold _writeLock_ before adding the entry. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063238#comment-14063238 ] Vinayakumar B commented on HADOOP-9921: --- Hi [~umamaheswararao], can you review the patch? daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10131) NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if excluded nodes passed are not part of the cluster tree
[ https://issues.apache.org/jira/browse/HADOOP-10131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063241#comment-14063241 ] Vinayakumar B commented on HADOOP-10131: Hi [~umamaheswararao], can you review the patch? Thanks In advance NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if excluded nodes passed are not part of the cluster tree - Key: HADOOP-10131 URL: https://issues.apache.org/jira/browse/HADOOP-10131 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.0.5-alpha Reporter: Vinayakumar B Assignee: Vinayakumar B Attachments: HADOOP-10131.patch, HDFS-5112.patch I got File /hdfs_COPYING_ could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. in the following case 1. 2 DNs cluster, 2. One of the datanodes was not responding from last 10 min, but about to detect as dead at NN. 3. Tried to write one file, for the block NN allocated both DNs. 4. Client While creating the pipeline took some time to detect one node failure. 5. Before client detects pipeline failure, NN side dead node was removed from cluster map. 6. Now, client has abandoned previous block and asked for new block with dead node in excluded list and got above exception even though one more node was available live. When I dig this more, found that, {{NetWorkTopology#countNumOfAvailableNodes()}} is not giving correct count when the excludeNodes passed from client are not part of the cluster map. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10794) A hadoop cluster needs clock synchronization
[ https://issues.apache.org/jira/browse/HADOOP-10794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063270#comment-14063270 ] Zhijie Shen commented on HADOOP-10794: -- I'm afraid each component may more or less have made the assumption that clock is synchronized across the cluster. Unfortunately, the problem that made us be aware of this issue came from MR: an elapsed time that based the the diff of the timestamps on two hosts became negative. Having a try with a cluster whose hosts are completely out of sync in terms of clock may give us the idea how bad the situation is. According to the prior discussion, it seems to be good to have a mechanism to check the clocks across a cluster, report unacceptable asynchronization, doesn't it? On the other hand, it is arguable whether this mechanism should be part of Hadoop or other cluster monitoring tools. Anyway, IMHO, it's good to at least let users be aware of the clock synchronization assumption (via jira, mailing list or etc.), and the risks of being out of sync, such that they can be more careful about this factor when setting up a cluster. A hadoop cluster needs clock synchronization Key: HADOOP-10794 URL: https://issues.apache.org/jira/browse/HADOOP-10794 Project: Hadoop Common Issue Type: Bug Reporter: Zhijie Shen As a distributed system, a hadoop cluster wants the clock on all the participating hosts synchronized. Otherwise, some problems might happen. For example, in YARN-2251, due to the clock on the host for the task container falls behind that on the host of the AM container, the computed elapsed time (the diff between the timestamps produced on two hosts) becomes negative. In YARN-2251, we tried to mask the negative elapsed time. However, we should seek for a decent long term solution, such as providing mechanism to do and check clock synchronization. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063275#comment-14063275 ] Hadoop QA commented on HADOOP-9921: --- {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12600746/HADOOP-9921.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common: org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4289//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4289//console This message is automatically generated. daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10131) NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if excluded nodes passed are not part of the cluster tree
[ https://issues.apache.org/jira/browse/HADOOP-10131?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063280#comment-14063280 ] Hadoop QA commented on HADOOP-10131: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12616044/HADOOP-10131.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common: org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4290//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4290//console This message is automatically generated. NetWorkTopology#countNumOfAvailableNodes() is returning wrong value if excluded nodes passed are not part of the cluster tree - Key: HADOOP-10131 URL: https://issues.apache.org/jira/browse/HADOOP-10131 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.0.5-alpha Reporter: Vinayakumar B Assignee: Vinayakumar B Attachments: HADOOP-10131.patch, HDFS-5112.patch I got File /hdfs_COPYING_ could only be replicated to 0 nodes instead of minReplication (=1). There are 1 datanode(s) running and 1 node(s) are excluded in this operation. in the following case 1. 2 DNs cluster, 2. One of the datanodes was not responding from last 10 min, but about to detect as dead at NN. 3. Tried to write one file, for the block NN allocated both DNs. 4. Client While creating the pipeline took some time to detect one node failure. 5. Before client detects pipeline failure, NN side dead node was removed from cluster map. 6. Now, client has abandoned previous block and asked for new block with dead node in excluded list and got above exception even though one more node was available live. When I dig this more, found that, {{NetWorkTopology#countNumOfAvailableNodes()}} is not giving correct count when the excludeNodes passed from client are not part of the cluster map. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-7664) o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same.
[ https://issues.apache.org/jira/browse/HADOOP-7664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063366#comment-14063366 ] Hudson commented on HADOOP-7664: FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/614/]) HADOOP-7664. Move attribution to release 2.6.0 section in CHANGES.txt. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610873) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same. -- Key: HADOOP-7664 URL: https://issues.apache.org/jira/browse/HADOOP-7664 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 0.20.205.0, 0.23.0, 2.5.0 Environment: commit a2f64ee8d9312fe24780ec53b15af439a315796d Reporter: Ravi Prakash Assignee: Ravi Prakash Priority: Minor Fix For: 1.0.0, 0.23.0, 2.6.0 Attachments: HADOOP-7664.branch-0.20-security.patch, HADOOP-7664.patch, HADOOP-7664.patch, HADOOP-7664.patch Original Estimate: 1m Time Spent: 1m Remaining Estimate: 0h o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063358#comment-14063358 ] Uma Maheswara Rao G commented on HADOOP-9921: - +1 patch looks good to me. daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10831) UserProvider is not thread safe
[ https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063368#comment-14063368 ] Hudson commented on HADOOP-10831: - FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/614/]) HADOOP-10831. UserProvider is not thread safe. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610789) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialProvider.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/UserProvider.java UserProvider is not thread safe --- Key: HADOOP-10831 URL: https://issues.apache.org/jira/browse/HADOOP-10831 Project: Hadoop Common Issue Type: Bug Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10831.patch While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread safe. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10468) TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately
[ https://issues.apache.org/jira/browse/HADOOP-10468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063359#comment-14063359 ] Hudson commented on HADOOP-10468: - FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/614/]) Addendum patch for HADOOP-10468 TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately. Contributed by Akira AJISAKA (jlowe: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610829) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestGangliaMetrics.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately --- Key: HADOOP-10468 URL: https://issues.apache.org/jira/browse/HADOOP-10468 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.5.0 Reporter: Haohui Mai Assignee: Akira AJISAKA Priority: Blocker Fix For: 2.5.0 Attachments: HADOOP-10468.000.patch, HADOOP-10468.001.patch, HADOOP-10468.2.patch {{TestMetricsSystemImpl.testMultiThreadedPublish}} can fail intermediately due to the insufficient size of the sink queue: {code} 2014-04-06 21:34:55,269 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. 2014-04-06 21:34:55,270 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. 2014-04-06 21:34:55,271 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. {code} The unit test should increase the default queue size to avoid intermediate failure. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10673) Update rpc metrics when the call throws an exception
[ https://issues.apache.org/jira/browse/HADOOP-10673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063365#comment-14063365 ] Hudson commented on HADOOP-10673: - FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/614/]) HADOOP-10673. Update rpc metrics when the call throws an exception. Contributed by Ming Ma. (jing9: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610879) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/ProtobufRpcEngine.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/WritableRpcEngine.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestRPC.java Update rpc metrics when the call throws an exception Key: HADOOP-10673 URL: https://issues.apache.org/jira/browse/HADOOP-10673 Project: Hadoop Common Issue Type: Bug Reporter: Ming Ma Assignee: Ming Ma Fix For: 2.6.0 Attachments: HADOOP-10673-2.patch, HADOOP-10673-4.patch, HADOOP-10673-5.patch, HADOOP-10673-6.patch, HADOOP-10673.patch Currently RPC metrics isn't updated when the call throws an exception. We can either update the existing metrics or have a new set of metrics in the case of exception. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10845) Add common tests for ACLs in combination with viewfs.
[ https://issues.apache.org/jira/browse/HADOOP-10845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063367#comment-14063367 ] Hudson commented on HADOOP-10845: - FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/614/]) HADOOP-10845. Add common tests for ACLs in combination with viewfs. Contributed by Stephen Chu. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610911) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsBaseTest.java Add common tests for ACLs in combination with viewfs. - Key: HADOOP-10845 URL: https://issues.apache.org/jira/browse/HADOOP-10845 Project: Hadoop Common Issue Type: Improvement Components: fs, test Affects Versions: 3.0.0, 2.5.0 Reporter: Chris Nauroth Assignee: Stephen Chu Fix For: 3.0.0, 2.6.0 Attachments: HADOOP-10845.1.patch Add tests in Hadoop Common for the ACL APIs in combination with viewfs. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10834) Typo in CredentialShell usage
[ https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063363#comment-14063363 ] Hudson commented on HADOOP-10834: - FAILURE: Integrated in Hadoop-Yarn-trunk #614 (See [https://builds.apache.org/job/Hadoop-Yarn-trunk/614/]) HADOOP-10834. Typo in CredentialShell usage. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610794) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialShell.java Typo in CredentialShell usage - Key: HADOOP-10834 URL: https://issues.apache.org/jira/browse/HADOOP-10834 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-10834.patch There is a typo in one of the informational messages in _ CredentialShell_ {code} out.println(Nothing has been be deleted.); {code} -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinayakumar B updated HADOOP-9921: -- Resolution: Fixed Fix Version/s: 2.6.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Thank You [~umamaheswararao] for the review. Committed to trunk and branch-2 daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Fix For: 2.6.0 Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063379#comment-14063379 ] Hudson commented on HADOOP-9921: FAILURE: Integrated in Hadoop-trunk-Commit #5890 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/5890/]) HADOOP-9921. daemon scripts should remove pid file on stop call after stop or process is found not running ( Contributed by Vinayakumar B) (vinayakumarb: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610964) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh * /hadoop/common/trunk/hadoop-mapreduce-project/bin/mr-jobhistory-daemon.sh * /hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/bin/yarn-daemon.sh daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Fix For: 2.6.0 Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10824) Refactor KMSACLs to avoid locking
[ https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Uma Maheswara Rao G updated HADOOP-10824: - Resolution: Fixed Fix Version/s: 3.0.0 Hadoop Flags: Reviewed Status: Resolved (was: Patch Available) Thanks a lot, Benoy for the patch. I have committed this to trunk. Refactor KMSACLs to avoid locking - Key: HADOOP-10824 URL: https://issues.apache.org/jira/browse/HADOOP-10824 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10824.patch, HADOOP-10824.patch Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible to safely publish the _acls_ collection using _volatile_. Similar refactoring has been done in [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112] -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10824) Refactor KMSACLs to avoid locking
[ https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063395#comment-14063395 ] Hudson commented on HADOOP-10824: - FAILURE: Integrated in Hadoop-trunk-Commit #5891 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/5891/]) HADOOP-10824. Refactor KMSACLs to avoid locking. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610969) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSACLs.java Refactor KMSACLs to avoid locking - Key: HADOOP-10824 URL: https://issues.apache.org/jira/browse/HADOOP-10824 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10824.patch, HADOOP-10824.patch Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible to safely publish the _acls_ collection using _volatile_. Similar refactoring has been done in [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112] -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alejandro Abdelnur updated HADOOP-10799: Attachment: HADOOP-10799.patch adding a missing constructor to {{DelegationTokenAuthenticatedURL}}. Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Created] (HADOOP-10850) KerberosAuthenticator should not do the SPNEGO handshake
Alejandro Abdelnur created HADOOP-10850: --- Summary: KerberosAuthenticator should not do the SPNEGO handshake Key: HADOOP-10850 URL: https://issues.apache.org/jira/browse/HADOOP-10850 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur As mentioned in HADOOP-10453, the JDK automatically does a SPNEGO handshake when opening a connection with a URL within a Kerberos login context, there is no need to do the SPNEGO handshake in the {{KerberosAuthenticator}}, simply extract the auth token (hadoop-auth cookie) and do the fallback if necessary. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10834) Typo in CredentialShell usage
[ https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063476#comment-14063476 ] Hudson commented on HADOOP-10834: - FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-10834. Typo in CredentialShell usage. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610794) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialShell.java Typo in CredentialShell usage - Key: HADOOP-10834 URL: https://issues.apache.org/jira/browse/HADOOP-10834 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-10834.patch There is a typo in one of the informational messages in _ CredentialShell_ {code} out.println(Nothing has been be deleted.); {code} -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-7664) o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same.
[ https://issues.apache.org/jira/browse/HADOOP-7664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063480#comment-14063480 ] Hudson commented on HADOOP-7664: FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-7664. Move attribution to release 2.6.0 section in CHANGES.txt. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610873) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same. -- Key: HADOOP-7664 URL: https://issues.apache.org/jira/browse/HADOOP-7664 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 0.20.205.0, 0.23.0, 2.5.0 Environment: commit a2f64ee8d9312fe24780ec53b15af439a315796d Reporter: Ravi Prakash Assignee: Ravi Prakash Priority: Minor Fix For: 1.0.0, 0.23.0, 2.6.0 Attachments: HADOOP-7664.branch-0.20-security.patch, HADOOP-7664.patch, HADOOP-7664.patch, HADOOP-7664.patch Original Estimate: 1m Time Spent: 1m Remaining Estimate: 0h o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10845) Add common tests for ACLs in combination with viewfs.
[ https://issues.apache.org/jira/browse/HADOOP-10845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063482#comment-14063482 ] Hudson commented on HADOOP-10845: - FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-10845. Add common tests for ACLs in combination with viewfs. Contributed by Stephen Chu. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610911) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsBaseTest.java Add common tests for ACLs in combination with viewfs. - Key: HADOOP-10845 URL: https://issues.apache.org/jira/browse/HADOOP-10845 Project: Hadoop Common Issue Type: Improvement Components: fs, test Affects Versions: 3.0.0, 2.5.0 Reporter: Chris Nauroth Assignee: Stephen Chu Fix For: 3.0.0, 2.6.0 Attachments: HADOOP-10845.1.patch Add tests in Hadoop Common for the ACL APIs in combination with viewfs. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10673) Update rpc metrics when the call throws an exception
[ https://issues.apache.org/jira/browse/HADOOP-10673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063478#comment-14063478 ] Hudson commented on HADOOP-10673: - FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-10673. Update rpc metrics when the call throws an exception. Contributed by Ming Ma. (jing9: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610879) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/ProtobufRpcEngine.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/WritableRpcEngine.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestRPC.java Update rpc metrics when the call throws an exception Key: HADOOP-10673 URL: https://issues.apache.org/jira/browse/HADOOP-10673 Project: Hadoop Common Issue Type: Bug Reporter: Ming Ma Assignee: Ming Ma Fix For: 2.6.0 Attachments: HADOOP-10673-2.patch, HADOOP-10673-4.patch, HADOOP-10673-5.patch, HADOOP-10673-6.patch, HADOOP-10673.patch Currently RPC metrics isn't updated when the call throws an exception. We can either update the existing metrics or have a new set of metrics in the case of exception. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10831) UserProvider is not thread safe
[ https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063483#comment-14063483 ] Hudson commented on HADOOP-10831: - FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-10831. UserProvider is not thread safe. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610789) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialProvider.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/UserProvider.java UserProvider is not thread safe --- Key: HADOOP-10831 URL: https://issues.apache.org/jira/browse/HADOOP-10831 Project: Hadoop Common Issue Type: Bug Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10831.patch While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread safe. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10824) Refactor KMSACLs to avoid locking
[ https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063475#comment-14063475 ] Hudson commented on HADOOP-10824: - FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-10824. Refactor KMSACLs to avoid locking. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610969) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSACLs.java Refactor KMSACLs to avoid locking - Key: HADOOP-10824 URL: https://issues.apache.org/jira/browse/HADOOP-10824 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10824.patch, HADOOP-10824.patch Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible to safely publish the _acls_ collection using _volatile_. Similar refactoring has been done in [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112] -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063479#comment-14063479 ] Hudson commented on HADOOP-9921: FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) HADOOP-9921. daemon scripts should remove pid file on stop call after stop or process is found not running ( Contributed by Vinayakumar B) (vinayakumarb: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610964) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh * /hadoop/common/trunk/hadoop-mapreduce-project/bin/mr-jobhistory-daemon.sh * /hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/bin/yarn-daemon.sh daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Fix For: 2.6.0 Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10468) TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately
[ https://issues.apache.org/jira/browse/HADOOP-10468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063471#comment-14063471 ] Hudson commented on HADOOP-10468: - FAILURE: Integrated in Hadoop-Mapreduce-trunk #1833 (See [https://builds.apache.org/job/Hadoop-Mapreduce-trunk/1833/]) Addendum patch for HADOOP-10468 TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately. Contributed by Akira AJISAKA (jlowe: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610829) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestGangliaMetrics.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately --- Key: HADOOP-10468 URL: https://issues.apache.org/jira/browse/HADOOP-10468 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.5.0 Reporter: Haohui Mai Assignee: Akira AJISAKA Priority: Blocker Fix For: 2.5.0 Attachments: HADOOP-10468.000.patch, HADOOP-10468.001.patch, HADOOP-10468.2.patch {{TestMetricsSystemImpl.testMultiThreadedPublish}} can fail intermediately due to the insufficient size of the sink queue: {code} 2014-04-06 21:34:55,269 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. 2014-04-06 21:34:55,270 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. 2014-04-06 21:34:55,271 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. {code} The unit test should increase the default queue size to avoid intermediate failure. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063490#comment-14063490 ] Alejandro Abdelnur commented on HADOOP-10799: - [~zjshen], thanks for pointing out HADOOP-10453, I was not aware of it, I've just commented there. Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10453) Do not use AuthenticatedURL in hadoop core
[ https://issues.apache.org/jira/browse/HADOOP-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063487#comment-14063487 ] Alejandro Abdelnur commented on HADOOP-10453: - The double SPNEGO authentication issue Daryn refers to is a problem of the KerberosAuthenticator, he is correct when stating that JDK does automatically SPNEGO when within a Kerberos login context, just created HADOOP-10850 for it. Regarding removing {{AuthenticatedURL}} altogether, I disagree on that. The fallback functionality is useful, HttpFS uses it for testing and now KMS is using it as well, plus YARN RM and TimelineServer. So we should work on fixing it, rather than trashing it. In addition, with the work that I'm doing in HADOOP-10799 (adding delegation token support) HADOOP-10835 (adding proxy user support) it simplifies significantly the client side code eliminating a big chunk of repetitive security related code that currently is done by the component doing HTTP calls instead by a security library. Do not use AuthenticatedURL in hadoop core -- Key: HADOOP-10453 URL: https://issues.apache.org/jira/browse/HADOOP-10453 Project: Hadoop Common Issue Type: Bug Reporter: Haohui Mai Priority: Blocker As [~daryn] has suggested in HDFS-4564: {quote} AuthenticatedURL is not used because it is buggy in part to causing replay attacks, double attempts to kerberos authenticate with the fallback authenticator if the TGT is expired, incorrectly uses the fallback authenticator (required by oozie servers) to add the username parameter which webhdfs has already included in the uri. AuthenticatedURL's attempt to do SPNEGO auth is a no-op because the JDK transparently does SPNEGO when the user's Subject (UGI) contains kerberos principals. Since AuthenticatedURL is now not used, webhdfs has to check the TGT itself for token operations. Bottom line is AuthenticatedURL is unnecessary and introduces nothing but problems for webhdfs. It's only useful for oozie's anon/non-anon support. {quote} However, several functionalities that relies on SPNEGO in secure mode suffer from the same problem. For example, NNs / JNs create HTTP connections to exchange fsimage and edit logs. Currently all of them are through {{AuthenticatedURL}}. This needs to be fixed to avoid security vulnerabilities. This jira purposes to remove {{AuthenticatedURL}} from hadoop core and to move it to oozie. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10834) Typo in CredentialShell usage
[ https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063505#comment-14063505 ] Hudson commented on HADOOP-10834: - FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-10834. Typo in CredentialShell usage. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610794) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialShell.java Typo in CredentialShell usage - Key: HADOOP-10834 URL: https://issues.apache.org/jira/browse/HADOOP-10834 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-10834.patch There is a typo in one of the informational messages in _ CredentialShell_ {code} out.println(Nothing has been be deleted.); {code} -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10831) UserProvider is not thread safe
[ https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063511#comment-14063511 ] Hudson commented on HADOOP-10831: - FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-10831. UserProvider is not thread safe. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610789) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/CredentialProvider.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/alias/UserProvider.java UserProvider is not thread safe --- Key: HADOOP-10831 URL: https://issues.apache.org/jira/browse/HADOOP-10831 Project: Hadoop Common Issue Type: Bug Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10831.patch While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread safe. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10673) Update rpc metrics when the call throws an exception
[ https://issues.apache.org/jira/browse/HADOOP-10673?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063507#comment-14063507 ] Hudson commented on HADOOP-10673: - FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-10673. Update rpc metrics when the call throws an exception. Contributed by Ming Ma. (jing9: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610879) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/ProtobufRpcEngine.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/Server.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/ipc/WritableRpcEngine.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestRPC.java Update rpc metrics when the call throws an exception Key: HADOOP-10673 URL: https://issues.apache.org/jira/browse/HADOOP-10673 Project: Hadoop Common Issue Type: Bug Reporter: Ming Ma Assignee: Ming Ma Fix For: 2.6.0 Attachments: HADOOP-10673-2.patch, HADOOP-10673-4.patch, HADOOP-10673-5.patch, HADOOP-10673-6.patch, HADOOP-10673.patch Currently RPC metrics isn't updated when the call throws an exception. We can either update the existing metrics or have a new set of metrics in the case of exception. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-7664) o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same.
[ https://issues.apache.org/jira/browse/HADOOP-7664?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063509#comment-14063509 ] Hudson commented on HADOOP-7664: FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-7664. Move attribution to release 2.6.0 section in CHANGES.txt. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610873) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same. -- Key: HADOOP-7664 URL: https://issues.apache.org/jira/browse/HADOOP-7664 Project: Hadoop Common Issue Type: Improvement Components: conf Affects Versions: 0.20.205.0, 0.23.0, 2.5.0 Environment: commit a2f64ee8d9312fe24780ec53b15af439a315796d Reporter: Ravi Prakash Assignee: Ravi Prakash Priority: Minor Fix For: 1.0.0, 0.23.0, 2.6.0 Attachments: HADOOP-7664.branch-0.20-security.patch, HADOOP-7664.patch, HADOOP-7664.patch, HADOOP-7664.patch Original Estimate: 1m Time Spent: 1m Remaining Estimate: 0h o.a.h.conf.Configuration complains of overriding final parameter even if the value with which its attempting to override is the same. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10824) Refactor KMSACLs to avoid locking
[ https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063504#comment-14063504 ] Hudson commented on HADOOP-10824: - FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-10824. Refactor KMSACLs to avoid locking. (Benoy Antony via umamahesh) (umamahesh: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610969) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSACLs.java Refactor KMSACLs to avoid locking - Key: HADOOP-10824 URL: https://issues.apache.org/jira/browse/HADOOP-10824 Project: Hadoop Common Issue Type: Improvement Components: security Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10824.patch, HADOOP-10824.patch Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible to safely publish the _acls_ collection using _volatile_. Similar refactoring has been done in [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112] -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10845) Add common tests for ACLs in combination with viewfs.
[ https://issues.apache.org/jira/browse/HADOOP-10845?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063510#comment-14063510 ] Hudson commented on HADOOP-10845: - FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-10845. Add common tests for ACLs in combination with viewfs. Contributed by Stephen Chu. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610911) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ChRootedFs.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFileSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/viewfs/ViewFs.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFileSystemBaseTest.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/viewfs/ViewFsBaseTest.java Add common tests for ACLs in combination with viewfs. - Key: HADOOP-10845 URL: https://issues.apache.org/jira/browse/HADOOP-10845 Project: Hadoop Common Issue Type: Improvement Components: fs, test Affects Versions: 3.0.0, 2.5.0 Reporter: Chris Nauroth Assignee: Stephen Chu Fix For: 3.0.0, 2.6.0 Attachments: HADOOP-10845.1.patch Add tests in Hadoop Common for the ACL APIs in combination with viewfs. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10468) TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately
[ https://issues.apache.org/jira/browse/HADOOP-10468?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063500#comment-14063500 ] Hudson commented on HADOOP-10468: - FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) Addendum patch for HADOOP-10468 TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately. Contributed by Akira AJISAKA (jlowe: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610829) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsConfig.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestGangliaMetrics.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java TestMetricsSystemImpl.testMultiThreadedPublish fails intermediately --- Key: HADOOP-10468 URL: https://issues.apache.org/jira/browse/HADOOP-10468 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.5.0 Reporter: Haohui Mai Assignee: Akira AJISAKA Priority: Blocker Fix For: 2.5.0 Attachments: HADOOP-10468.000.patch, HADOOP-10468.001.patch, HADOOP-10468.2.patch {{TestMetricsSystemImpl.testMultiThreadedPublish}} can fail intermediately due to the insufficient size of the sink queue: {code} 2014-04-06 21:34:55,269 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. 2014-04-06 21:34:55,270 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. 2014-04-06 21:34:55,271 WARN impl.MetricsSinkAdapter (MetricsSinkAdapter.java:putMetricsImmediate(107)) - Collector has a full queue and can't consume the given metrics. {code} The unit test should increase the default queue size to avoid intermediate failure. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-9921) daemon scripts should remove pid file on stop call after stop or process is found not running
[ https://issues.apache.org/jira/browse/HADOOP-9921?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063508#comment-14063508 ] Hudson commented on HADOOP-9921: FAILURE: Integrated in Hadoop-Hdfs-trunk #1806 (See [https://builds.apache.org/job/Hadoop-Hdfs-trunk/1806/]) HADOOP-9921. daemon scripts should remove pid file on stop call after stop or process is found not running ( Contributed by Vinayakumar B) (vinayakumarb: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1610964) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/bin/hadoop-daemon.sh * /hadoop/common/trunk/hadoop-mapreduce-project/bin/mr-jobhistory-daemon.sh * /hadoop/common/trunk/hadoop-yarn-project/hadoop-yarn/bin/yarn-daemon.sh daemon scripts should remove pid file on stop call after stop or process is found not running - Key: HADOOP-9921 URL: https://issues.apache.org/jira/browse/HADOOP-9921 Project: Hadoop Common Issue Type: Bug Affects Versions: 3.0.0, 2.1.0-beta Reporter: Vinayakumar B Assignee: Vinayakumar B Fix For: 2.6.0 Attachments: HADOOP-9921.patch daemon scripts should remove the pid file on stop call using daemon script. Should remove the pid file, even though process is not running. same pid file will be used by start command. At that time, if the same pid is assigned to some other process, then start may fail. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063519#comment-14063519 ] Hadoop QA commented on HADOOP-10799: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656034/HADOOP-10799.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 4 new or modified test files. {color:red}-1 javac{color}. The applied patch generated 1268 javac compiler warnings (more than the trunk's current 1260 warnings). {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-auth hadoop-common-project/hadoop-common: org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4291//testReport/ Javac warnings: https://builds.apache.org/job/PreCommit-HADOOP-Build/4291//artifact/trunk/patchprocess/diffJavacWarnings.txt Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4291//console This message is automatically generated. Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10778) Use NativeCrc32 only if it is faster
[ https://issues.apache.org/jira/browse/HADOOP-10778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063577#comment-14063577 ] Todd Lipcon commented on HADOOP-10778: -- Sure, here's the whole table (Native is always faster): Direct Buffer Performance Table (bpc = byte-per-crc in MB/sec; #T = #Theads) | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | |32 | 1 | 422.8 | 802.0 | 89.7% |1908.2 | 351.3% | 137.9% | |32 | 2 | 316.7 | 622.1 | 96.4% |1853.9 | 485.3% | 198.0% | |32 | 4 | 285.8 | 516.9 | 80.9% |1806.3 | 532.1% | 249.5% | |32 | 8 | 171.5 | 304.2 | 77.3% |1052.5 | 513.5% | 246.0% | |32 | 16 | 95.8 | 189.7 | 97.9% | 640.3 | 568.1% | 237.6% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | |64 | 1 | 604.4 |1034.8 | 71.2% |1795.6 | 197.1% | 73.5% | |64 | 2 | 506.4 | 865.6 | 70.9% |1752.2 | 246.0% | 102.4% | |64 | 4 | 530.5 | 923.6 | 74.1% |1642.4 | 209.6% | 77.8% | |64 | 8 | 300.5 | 468.2 | 55.8% |1222.6 | 306.8% | 161.1% | |64 | 16 | 163.2 | 284.6 | 74.4% | 669.8 | 310.4% | 135.3% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 128 | 1 | 782.8 |1184.3 | 51.3% |1694.8 | 116.5% | 43.1% | | 128 | 2 | 759.5 |1147.2 | 51.1% |1690.3 | 122.6% | 47.3% | | 128 | 4 | 737.2 |1110.1 | 50.6% |1634.1 | 121.7% | 47.2% | | 128 | 8 | 488.2 | 602.1 | 23.3% |1317.1 | 169.8% | 118.8% | | 128 | 16 | 281.2 | 375.9 | 33.7% | 794.8 | 182.6% | 111.4% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 256 | 1 | 921.1 |1242.1 | 34.9% |1711.7 | 85.8% | 37.8% | | 256 | 2 | 890.4 |1231.2 | 38.3% |1618.5 | 81.8% | 31.5% | | 256 | 4 | 868.9 |1162.0 | 33.7% |1597.3 | 83.8% | 37.5% | | 256 | 8 | 651.1 | 673.0 | 3.4% |1321.4 | 102.9% | 96.3% | | 256 | 16 | 369.6 | 397.5 | 7.5% | 730.2 | 97.6% | 83.7% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 512 | 1 | 993.0 |1292.9 | 30.2% |1653.6 | 66.5% | 27.9% | | 512 | 2 | 926.7 |1274.4 | 37.5% |1635.4 | 76.5% | 28.3% | | 512 | 4 | 918.0 |1217.7 | 32.6% |1589.3 | 73.1% | 30.5% | | 512 | 8 | 778.0 | 711.6 | -8.5% |1365.8 | 75.5% | 91.9% | | 512 | 16 | 457.2 | 406.2 | -11.2% | 790.6 | 72.9% | 94.6% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 1024 | 1 |1016.0 |1303.2 | 28.3% |1673.8 | 64.7% | 28.4% | | 1024 | 2 | 987.0 |1264.9 | 28.2% |1637.3 | 65.9% | 29.4% | | 1024 | 4 | 956.7 |1239.4 | 29.5% |1593.3 | 66.5% | 28.6% | | 1024 | 8 | 837.3 | 716.4 | -14.4% |1336.8 | 59.7% | 86.6% | | 1024 | 16 | 522.1 | 389.0 | -25.5% | 948.5 | 81.7% | 143.8% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 2048 | 1 |1006.9 |1271.3 | 26.3% |1695.3 | 68.4% | 33.4% | | 2048 | 2 | 989.1 |1234.1 | 24.8% |1652.3 | 67.1% | 33.9% | | 2048 | 4 | 959.1 |1200.1 | 25.1% |1581.7 | 64.9% | 31.8% | | 2048 | 8 | 875.4 | 713.4 | -18.5% |1422.8 | 62.5% | 99.4% | | 2048 | 16 | 507.5 | 402.2 | -20.7% | 923.7 | 82.0% | 129.6% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 4096 | 1 |1025.0 |1251.6 | 22.1% |1691.7 | 65.0% | 35.2% | | 4096 | 2 | 990.7 |1219.8 | 23.1% |1630.2 | 64.6% | 33.6% | | 4096 | 4 | 961.5 |1178.9 | 22.6% |1582.7 | 64.6% | 34.3% | | 4096 | 8 | 891.1 | 714.0 | -19.9% |1429.4 | 60.4% | 100.2% | | 4096 | 16 | 529.1 | 405.3 | -23.4% | 862.6 | 63.0% | 112.8% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 8192 | 1 |1015.4 |1290.6 | 27.1% |1685.9 | 66.0% | 30.6% | | 8192 | 2 |1007.6 |1266.1 | 25.7% |1619.9 | 60.8% | 27.9% | | 8192 | 4 | 981.5 |1208.1 | 23.1% |1602.3 | 63.3% | 32.6% | | 8192 | 8 | 903.9 | 711.7 | -21.3% |1426.6 | 57.8% | 100.4% | | 8192 | 16 | 523.9 | 412.6 | -21.2% | 812.4 | 55.1% | 96.9% | | bpc | #T || Zip || PureJava | % diff || Native | % diff | % diff | | 16384 | 1 |1027.2 |1319.0 | 28.4% |1656.1 | 61.2% | 25.6% | | 16384 | 2 |1009.8 |1259.3 | 24.7% |1647.4 | 63.1% | 30.8% | | 16384 | 4 | 977.9 |1222.2 | 25.0% |1593.4 | 62.9% | 30.4% | | 16384 | 8 | 902.6 | 709.6 | -21.4% |1424.7 | 57.8% | 100.8% | | 16384
[jira] [Commented] (HADOOP-10778) Use NativeCrc32 only if it is faster
[ https://issues.apache.org/jira/browse/HADOOP-10778?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063609#comment-14063609 ] Todd Lipcon commented on HADOOP-10778: -- BTW, what clock rate is your laptop? The best non-hardware CRC32 implementation I'm aware of (http://code.google.com/p/crcutil/) goes 1 cycle per byte. So, my guess is that your 2.3GB/sec is probably from a 2.3 (or so) Ghz machine using an algorithm like crcutil's. Perhaps Apple patched their zlib to use this algorithm instead of the very simple implementation in the upstream zlib? Use NativeCrc32 only if it is faster Key: HADOOP-10778 URL: https://issues.apache.org/jira/browse/HADOOP-10778 Project: Hadoop Common Issue Type: Improvement Components: util Reporter: Tsz Wo Nicholas Sze Assignee: Tsz Wo Nicholas Sze Attachments: c10778_20140702.patch From the benchmark post in [this comment|https://issues.apache.org/jira/browse/HDFS-6560?focusedCommentId=14044060page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-14044060], NativeCrc32 is slower than java.util.zip.CRC32 for Java 7 and above when bytesPerChecksum 512. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
[ https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dmitry Sivachenko updated HADOOP-10783: --- Status: Patch Available (was: Open) I am attaching another patch (trying to solve problems found by Jenkins) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed --- Key: HADOOP-10783 URL: https://issues.apache.org/jira/browse/HADOOP-10783 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.4.1 Reporter: Dmitry Sivachenko Assignee: Steve Loughran Attachments: commons-lang3_1.patch Hadoop-2.4.1 ships with apache-commons.jar version 2.6. It does not support FreeBSD (IS_OS_UNIX returns False). This is fixed in recent versions of apache-commons.jar Please update apache-commons.jar to recent version so it correctly recognizes FreeBSD as UNIX-like system. Right now I get in datanode's log: 2014-07-04 11:58:10,459 DEBUG org.apache.hadoop.hdfs.server.datanode.ShortCircui tRegistry: Disabling ShortCircuitRegistry java.io.IOException: The OS is not UNIX. at org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77) at org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169) at org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771) at org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065) -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
[ https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dmitry Sivachenko updated HADOOP-10783: --- Attachment: commons-lang3_1.patch apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed --- Key: HADOOP-10783 URL: https://issues.apache.org/jira/browse/HADOOP-10783 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.4.1 Reporter: Dmitry Sivachenko Assignee: Steve Loughran Attachments: commons-lang3_1.patch Hadoop-2.4.1 ships with apache-commons.jar version 2.6. It does not support FreeBSD (IS_OS_UNIX returns False). This is fixed in recent versions of apache-commons.jar Please update apache-commons.jar to recent version so it correctly recognizes FreeBSD as UNIX-like system. Right now I get in datanode's log: 2014-07-04 11:58:10,459 DEBUG org.apache.hadoop.hdfs.server.datanode.ShortCircui tRegistry: Disabling ShortCircuitRegistry java.io.IOException: The OS is not UNIX. at org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77) at org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169) at org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771) at org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065) -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
[ https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dmitry Sivachenko updated HADOOP-10783: --- Status: Open (was: Patch Available) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed --- Key: HADOOP-10783 URL: https://issues.apache.org/jira/browse/HADOOP-10783 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.4.1 Reporter: Dmitry Sivachenko Assignee: Steve Loughran Attachments: commons-lang3_1.patch Hadoop-2.4.1 ships with apache-commons.jar version 2.6. It does not support FreeBSD (IS_OS_UNIX returns False). This is fixed in recent versions of apache-commons.jar Please update apache-commons.jar to recent version so it correctly recognizes FreeBSD as UNIX-like system. Right now I get in datanode's log: 2014-07-04 11:58:10,459 DEBUG org.apache.hadoop.hdfs.server.datanode.ShortCircui tRegistry: Disabling ShortCircuitRegistry java.io.IOException: The OS is not UNIX. at org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77) at org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169) at org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771) at org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065) -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
[ https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Dmitry Sivachenko updated HADOOP-10783: --- Attachment: (was: commons-lang3.patch) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed --- Key: HADOOP-10783 URL: https://issues.apache.org/jira/browse/HADOOP-10783 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.4.1 Reporter: Dmitry Sivachenko Assignee: Steve Loughran Attachments: commons-lang3_1.patch Hadoop-2.4.1 ships with apache-commons.jar version 2.6. It does not support FreeBSD (IS_OS_UNIX returns False). This is fixed in recent versions of apache-commons.jar Please update apache-commons.jar to recent version so it correctly recognizes FreeBSD as UNIX-like system. Right now I get in datanode's log: 2014-07-04 11:58:10,459 DEBUG org.apache.hadoop.hdfs.server.datanode.ShortCircui tRegistry: Disabling ShortCircuitRegistry java.io.IOException: The OS is not UNIX. at org.apache.hadoop.io.nativeio.SharedFileDescriptorFactory.create(SharedFileDescriptorFactory.java:77) at org.apache.hadoop.hdfs.server.datanode.ShortCircuitRegistry.init(ShortCircuitRegistry.java:169) at org.apache.hadoop.hdfs.server.datanode.DataNode.initDataXceiver(DataNode.java:583) at org.apache.hadoop.hdfs.server.datanode.DataNode.startDataNode(DataNode.java:771) at org.apache.hadoop.hdfs.server.datanode.DataNode.init(DataNode.java:289) at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:1931) at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:1818) at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:1865) at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2041) at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2065) -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10824) Refactor KMSACLs to avoid locking
[ https://issues.apache.org/jira/browse/HADOOP-10824?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10824: -- Affects Version/s: 2.4.1 Refactor KMSACLs to avoid locking - Key: HADOOP-10824 URL: https://issues.apache.org/jira/browse/HADOOP-10824 Project: Hadoop Common Issue Type: Improvement Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10824.patch, HADOOP-10824.patch Currently _KMSACLs_ is made thread safe using _ReadWriteLock_. It is possible to safely publish the _acls_ collection using _volatile_. Similar refactoring has been done in [HADOOP-10448|https://issues.apache.org/jira/browse/HADOOP-10448?focusedCommentId=13980112page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-13980112] -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10831) UserProvider is not thread safe
[ https://issues.apache.org/jira/browse/HADOOP-10831?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10831: -- Affects Version/s: 2.4.1 UserProvider is not thread safe --- Key: HADOOP-10831 URL: https://issues.apache.org/jira/browse/HADOOP-10831 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Fix For: 3.0.0 Attachments: HADOOP-10831.patch While _JavaKeyStoreProvider_ is thread safe, _UserProvider_ is not thread safe. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10834) Typo in CredentialShell usage
[ https://issues.apache.org/jira/browse/HADOOP-10834?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10834: -- Affects Version/s: 2.4.1 Typo in CredentialShell usage - Key: HADOOP-10834 URL: https://issues.apache.org/jira/browse/HADOOP-10834 Project: Hadoop Common Issue Type: Improvement Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Priority: Minor Fix For: 3.0.0 Attachments: HADOOP-10834.patch There is a typo in one of the informational messages in _ CredentialShell_ {code} out.println(Nothing has been be deleted.); {code} -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry
[ https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10830: -- Attachment: HADOOP-10830.patch Thanks for the review, [~vinayrpet]. I have updated the patch accordingly. Missing lock in JavaKeyStoreProvider.createCredentialEntry -- Key: HADOOP-10830 URL: https://issues.apache.org/jira/browse/HADOOP-10830 Project: Hadoop Common Issue Type: Bug Components: security Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10830.patch, HADOOP-10830.patch _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_ to provide thread safety. The {{createCredentialEntry}} should hold _writeLock_ before adding the entry. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry
[ https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063801#comment-14063801 ] Hadoop QA commented on HADOOP-10830: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656072/HADOOP-10830.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common: org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4293//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4293//console This message is automatically generated. Missing lock in JavaKeyStoreProvider.createCredentialEntry -- Key: HADOOP-10830 URL: https://issues.apache.org/jira/browse/HADOOP-10830 Project: Hadoop Common Issue Type: Bug Components: security Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10830.patch, HADOOP-10830.patch _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_ to provide thread safety. The {{createCredentialEntry}} should hold _writeLock_ before adding the entry. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063807#comment-14063807 ] Alejandro Abdelnur commented on HADOOP-10799: - the javac warnings are warnings from Maven regarding the maven-jar-plugin defined twice (it does not seem to be the case). I'm not able to reproduce the warnings locally with Maven 3.0.4 Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10735) Fall back AesCtrCryptoCodec implementation from OpenSSL to JCE if non native support.
[ https://issues.apache.org/jira/browse/HADOOP-10735?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063823#comment-14063823 ] Colin Patrick McCabe commented on HADOOP-10735: --- Thanks, Yi. Fall back AesCtrCryptoCodec implementation from OpenSSL to JCE if non native support. - Key: HADOOP-10735 URL: https://issues.apache.org/jira/browse/HADOOP-10735 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: fs-encryption (HADOOP-10150 and HDFS-6134) Reporter: Yi Liu Assignee: Yi Liu Fix For: fs-encryption (HADOOP-10150 and HDFS-6134) Attachments: HADOOP-10735.001.patch, HADOOP-10735.002.patch, HADOOP-10735.003.patch, HADOOP-10735.004.patch, HADOOP-10735.005.patch, HADOOP-10735.006.patch, HADOOP-10735.007.patch, HADOOP-10735.008.patch If there is no native support or OpenSSL version is too low not supporting AES-CTR, but {{OpensslAesCtrCryptoCodec}} is configured, we need to fall back it to JCE implementation. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alejandro Abdelnur updated HADOOP-10799: Attachment: HADOOP-10799.patch ok, found the issue, a commit from yesterday YARN-2233, was overlapping POM changes with the patch, patch rebased. Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10830) Missing lock in JavaKeyStoreProvider.createCredentialEntry
[ https://issues.apache.org/jira/browse/HADOOP-10830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10830: -- Affects Version/s: 2.4.1 Missing lock in JavaKeyStoreProvider.createCredentialEntry -- Key: HADOOP-10830 URL: https://issues.apache.org/jira/browse/HADOOP-10830 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10830.patch, HADOOP-10830.patch _JavaKeyStoreProvider_ uses _ReentrantReadWriteLock_ to provide thread safety. The {{createCredentialEntry}} should hold _writeLock_ before adding the entry. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10833) Unused cache in UserProvider
[ https://issues.apache.org/jira/browse/HADOOP-10833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10833: -- Affects Version/s: 2.4.1 Unused cache in UserProvider Key: HADOOP-10833 URL: https://issues.apache.org/jira/browse/HADOOP-10833 Project: Hadoop Common Issue Type: Improvement Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10833.patch _UserProvider_ contains the field cache. {code} private final MapString, CredentialEntry cache = new HashMapString, CredentialEntry(); {code} It is referenced only in {{deleteCredentialEntry}} and so there is no real usage of {{cache}. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10833) Unused cache in UserProvider
[ https://issues.apache.org/jira/browse/HADOOP-10833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063854#comment-14063854 ] Benoy Antony commented on HADOOP-10833: --- [~vinayrpet], Could you please review this patch also ? Unused cache in UserProvider Key: HADOOP-10833 URL: https://issues.apache.org/jira/browse/HADOOP-10833 Project: Hadoop Common Issue Type: Improvement Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10833.patch _UserProvider_ contains the field cache. {code} private final MapString, CredentialEntry cache = new HashMapString, CredentialEntry(); {code} It is referenced only in {{deleteCredentialEntry}} and so there is no real usage of {{cache}. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-9902) Shell script rewrite
[ https://issues.apache.org/jira/browse/HADOOP-9902?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Allen Wittenauer updated HADOOP-9902: - Attachment: HADOOP-9902-6.patch vs -5, this patch fixes a kill output not sent to /dev/null, updates some of the hadoop-env.sh commentary, and fixes a code style issue in rotate_logs. If there is interest, I can make a patch for 2.4.1. Shell script rewrite Key: HADOOP-9902 URL: https://issues.apache.org/jira/browse/HADOOP-9902 Project: Hadoop Common Issue Type: Improvement Components: scripts Affects Versions: 3.0.0 Reporter: Allen Wittenauer Assignee: Allen Wittenauer Labels: releasenotes Attachments: HADOOP-9902-2.patch, HADOOP-9902-3.patch, HADOOP-9902-4.patch, HADOOP-9902-5.patch, HADOOP-9902-6.patch, HADOOP-9902.patch, HADOOP-9902.txt, hadoop-9902-1.patch, more-info.txt Umbrella JIRA for shell script rewrite. See more-info.txt for more details. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10833) Unused cache in UserProvider
[ https://issues.apache.org/jira/browse/HADOOP-10833?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10833: -- Description: _UserProvider_ contains the field cache. {code} private final MapString, CredentialEntry cache = new HashMapString, CredentialEntry(); {code} It is referenced only in {{deleteCredentialEntry}} and so there is no real usage of {{cache}}. was: _UserProvider_ contains the field cache. {code} private final MapString, CredentialEntry cache = new HashMapString, CredentialEntry(); {code} It is referenced only in {{deleteCredentialEntry}} and so there is no real usage of {{cache}. Unused cache in UserProvider Key: HADOOP-10833 URL: https://issues.apache.org/jira/browse/HADOOP-10833 Project: Hadoop Common Issue Type: Improvement Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10833.patch _UserProvider_ contains the field cache. {code} private final MapString, CredentialEntry cache = new HashMapString, CredentialEntry(); {code} It is referenced only in {{deleteCredentialEntry}} and so there is no real usage of {{cache}}. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10839) Add unregisterSource() to MetricsSystem API
[ https://issues.apache.org/jira/browse/HADOOP-10839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10839: --- Issue Type: Improvement (was: Bug) Add unregisterSource() to MetricsSystem API --- Key: HADOOP-10839 URL: https://issues.apache.org/jira/browse/HADOOP-10839 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10839.2.patch, HADOOP-10839.patch Currently the MetrisSystem API has register() method to register a MetricsSource but doesn't have unregister() method. This means once a MetricsSource is registered with the MetricsSystem, it will be there forever until the MetricsSystem is shut down. This in some cases can cause Java OutOfMemoryError. One such case is in file system metrics implementation. The new AbstractFileSystem/FileContext framework does not implement a cache so every file system access can lead to the creation of a NativeFileSystem instance. (refer to HADOOP-6356). And all these NativeFileSystem needs to share the same instance of MetricsSystemImpl, which means we cannot shut down MetricsSystem to clean up all the MetricsSources that has been registered but no longer active. Over time the MetricsSource instance accumulates and eventually we saw OutOfMemoryError. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10794) A hadoop cluster needs clock synchronization
[ https://issues.apache.org/jira/browse/HADOOP-10794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063901#comment-14063901 ] Colin Patrick McCabe commented on HADOOP-10794: --- bq. I'm afraid each component may more or less have made the assumption that clock is synchronized across the cluster. Can you explain which components make this assumption, and what happens if it's violated? So far the only example presented is statistics in YARN. A hadoop cluster needs clock synchronization Key: HADOOP-10794 URL: https://issues.apache.org/jira/browse/HADOOP-10794 Project: Hadoop Common Issue Type: Bug Reporter: Zhijie Shen As a distributed system, a hadoop cluster wants the clock on all the participating hosts synchronized. Otherwise, some problems might happen. For example, in YARN-2251, due to the clock on the host for the task container falls behind that on the host of the AM container, the computed elapsed time (the diff between the timestamps produced on two hosts) becomes negative. In YARN-2251, we tried to mask the negative elapsed time. However, we should seek for a decent long term solution, such as providing mechanism to do and check clock synchronization. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10839) Add unregisterSource() to MetricsSystem API
[ https://issues.apache.org/jira/browse/HADOOP-10839?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10839: --- Resolution: Fixed Fix Version/s: 2.6.0 3.0.0 Status: Resolved (was: Patch Available) I committed this to trunk and branch-2. Shanyu, thank you for contributing this patch. Add unregisterSource() to MetricsSystem API --- Key: HADOOP-10839 URL: https://issues.apache.org/jira/browse/HADOOP-10839 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Fix For: 3.0.0, 2.6.0 Attachments: HADOOP-10839.2.patch, HADOOP-10839.patch Currently the MetrisSystem API has register() method to register a MetricsSource but doesn't have unregister() method. This means once a MetricsSource is registered with the MetricsSystem, it will be there forever until the MetricsSystem is shut down. This in some cases can cause Java OutOfMemoryError. One such case is in file system metrics implementation. The new AbstractFileSystem/FileContext framework does not implement a cache so every file system access can lead to the creation of a NativeFileSystem instance. (refer to HADOOP-6356). And all these NativeFileSystem needs to share the same instance of MetricsSystemImpl, which means we cannot shut down MetricsSystem to clean up all the MetricsSources that has been registered but no longer active. Over time the MetricsSource instance accumulates and eventually we saw OutOfMemoryError. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10839) Add unregisterSource() to MetricsSystem API
[ https://issues.apache.org/jira/browse/HADOOP-10839?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063907#comment-14063907 ] Hudson commented on HADOOP-10839: - FAILURE: Integrated in Hadoop-trunk-Commit #5893 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/5893/]) HADOOP-10839. Add unregisterSource() to MetricsSystem API. Contributed by Shanyu Zhao. (cnauroth: http://svn.apache.org/viewcvs.cgi/?root=Apache-SVNview=revrev=1611134) * /hadoop/common/trunk/hadoop-common-project/hadoop-common/CHANGES.txt * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/MetricsSystem.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/metrics2/impl/MetricsSystemImpl.java * /hadoop/common/trunk/hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/metrics2/impl/TestMetricsSystemImpl.java Add unregisterSource() to MetricsSystem API --- Key: HADOOP-10839 URL: https://issues.apache.org/jira/browse/HADOOP-10839 Project: Hadoop Common Issue Type: Improvement Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Fix For: 3.0.0, 2.6.0 Attachments: HADOOP-10839.2.patch, HADOOP-10839.patch Currently the MetrisSystem API has register() method to register a MetricsSource but doesn't have unregister() method. This means once a MetricsSource is registered with the MetricsSystem, it will be there forever until the MetricsSystem is shut down. This in some cases can cause Java OutOfMemoryError. One such case is in file system metrics implementation. The new AbstractFileSystem/FileContext framework does not implement a cache so every file system access can lead to the creation of a NativeFileSystem instance. (refer to HADOOP-6356). And all these NativeFileSystem needs to share the same instance of MetricsSystemImpl, which means we cannot shut down MetricsSystem to clean up all the MetricsSources that has been registered but no longer active. Over time the MetricsSource instance accumulates and eventually we saw OutOfMemoryError. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10840: --- Target Version/s: 3.0.0, 2.6.0 Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10840: --- Status: Patch Available (was: Open) Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10840: --- Target Version/s: 3.0.0 (was: 3.0.0, 2.6.0) Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063923#comment-14063923 ] Hadoop QA commented on HADOOP-10799: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656090/HADOOP-10799.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 4 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-auth hadoop-common-project/hadoop-common: org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4294//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4294//console This message is automatically generated. Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063933#comment-14063933 ] Chris Nauroth commented on HADOOP-10840: Hi, [~shanyu]. I'm seeing some test failures when I run the tests under hadoop-tools/hadoop-azure. Could you please investigate? Here is a stack trace from one example: {code} Running org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck Tests run: 4, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.871 sec FAILURE! - in org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck testMkdirs(org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck) Time elapsed: 0.04 sec ERROR! java.lang.NullPointerException: null at org.apache.hadoop.fs.azure.metrics.AzureFileSystemMetricsSystem.unregisterSource(AzureFileSystemMetricsSystem.java:65) at org.apache.hadoop.fs.azure.NativeAzureFileSystem.close(NativeAzureFileSystem.java:1352) at org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.cleanup(AzureBlobStorageTestAccount.java:745) at org.apache.hadoop.fs.azure.TestNativeAzureFileSystemFileNameCheck.tearDown(TestNativeAzureFileSystemFileNameCheck.java:55) {code} Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10840: --- Status: Open (was: Patch Available) Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Created] (HADOOP-10851) NetgroupCache does not remove group memberships
Benoy Antony created HADOOP-10851: - Summary: NetgroupCache does not remove group memberships Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships
[ https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10851: -- Attachment: HADOOP-10851.patch The attached patch adds 3 test cases * a basic test for _NetgroupCache_ * test where a user is initially part of a group and later removed from that group. * test where a group is initially added to cache, but later removed. _NetgroupCache_ is fixed by clearing the cache properly. Code is slightly refactored : * uses {{addAll}} instead of looping through the set of entries. * removed call to {{isCached}} inside {{add}} since the callers call {{isCached}} before calling {{add}}. * two unused imports are removed. NetgroupCache does not remove group memberships --- Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10851.patch _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships
[ https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10851: -- Status: Patch Available (was: Open) NetgroupCache does not remove group memberships --- Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10851.patch _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed
[ https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063951#comment-14063951 ] Colin Patrick McCabe commented on HADOOP-10591: --- bq. Any reason we can't make CompressionOutputStream.trackedCompressor private? Yeah, let's make it private. bq. The javadoc for createInputStreamWithCodecPool says The codec to use to create the output stream. Fixed. Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed - Key: HADOOP-10591 URL: https://issues.apache.org/jira/browse/HADOOP-10591 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.2.0 Reporter: Hari Shreedharan Assignee: Colin Patrick McCabe Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch Currently direct buffers allocated by compression codecs like Gzip (which allocates 2 direct buffers per instance) are not deallocated when the stream is closed. Eventually for long running processes which create a huge number of files, these direct buffers are left hanging till a full gc, which may or may not happen in a reasonable amount of time - especially if the process does not use a whole lot of heap. Either these buffers should be pooled or they should be deallocated when the stream is closed. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed
[ https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Colin Patrick McCabe updated HADOOP-10591: -- Attachment: HADOOP-10591.002.patch Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed - Key: HADOOP-10591 URL: https://issues.apache.org/jira/browse/HADOOP-10591 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.2.0 Reporter: Hari Shreedharan Assignee: Colin Patrick McCabe Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch Currently direct buffers allocated by compression codecs like Gzip (which allocates 2 direct buffers per instance) are not deallocated when the stream is closed. Eventually for long running processes which create a huge number of files, these direct buffers are left hanging till a full gc, which may or may not happen in a reasonable amount of time - especially if the process does not use a whole lot of heap. Either these buffers should be pooled or they should be deallocated when the stream is closed. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Created] (HADOOP-10852) NetgroupCache is not thread-safe
Benoy Antony created HADOOP-10852: - Summary: NetgroupCache is not thread-safe Key: HADOOP-10852 URL: https://issues.apache.org/jira/browse/HADOOP-10852 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony _NetgroupCache_ internally uses two _ConcurrentHashMap_s and a boolean variable to signal updates on one of the _ConcurrentHashMap_s None of the functions are synchronized and hence is possible to have unexpected results due to race condition between different threads. As an example, consider the following sequence: Thread1 : {{add}} a group {{netgroupToUsersMap}} is updated. {{netgroupToUsersMapUpdated}} is set to true. Thread 2: calls {{getNetgroups}} for a user Due to re-ordering, {{netgroupToUsersMapUpdated=true}} is visible, but updates in {{netgroupToUsersMap}} is not visible. Does a wrong update with older {{netgroupToUsersMap}} values. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships
[ https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10851: -- Attachment: HADOOP-10852.patch The _NetgroupCache_ is fixed by keeping state in only one _ConcurrentHashMap_ - {{userToNetgroupsMap}}. {{add}} updates {{userToNetgroupsMap}} directly. There is a slight performance loss when invoking {{isCached}} and {{{userToNetgroupsMap}} But these are invoked only during {{refresh}} and adding a new group. These invocations should be rare compared to {{getNetgroups}} invocation. NetgroupCache does not remove group memberships --- Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10851.patch, HADOOP-10852.patch _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10783) apache-commons-lang.jar 2.6 does not support FreeBSD -upgrade to 3.x needed
[ https://issues.apache.org/jira/browse/HADOOP-10783?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063943#comment-14063943 ] Hadoop QA commented on HADOOP-10783: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656057/commons-lang3_1.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 31 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:red}-1 findbugs{color}. The patch appears to cause Findbugs (version 2.0.3) to fail. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-auth hadoop-common-project/hadoop-common hadoop-common-project/hadoop-minikdc hadoop-hdfs-project/hadoop-hdfs hadoop-hdfs-project/hadoop-hdfs-nfs hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-app hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-hs hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-jobclient hadoop-tools/hadoop-gridmix hadoop-tools/hadoop-rumen hadoop-tools/hadoop-streaming hadoop-yarn-project/hadoop-yarn/hadoop-yarn-api hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-distributedshell hadoop-yarn-project/hadoop-yarn/hadoop-yarn-client hadoop-yarn-project/hadoop-yarn/hadoop-yarn-common hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-common hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager: org.apache.hadoop.metrics2.impl.TestMetricsSystemImpl org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem org.apache.hadoop.streaming.TestMultipleArchiveFiles org.apache.hadoop.streaming.TestStreamingBadRecords org.apache.hadoop.streaming.TestStreamingOutputKeyValueTypes org.apache.hadoop.streaming.TestStreamAggregate org.apache.hadoop.streaming.TestStreamReduceNone org.apache.hadoop.streaming.TestUnconsumedInput org.apache.hadoop.streaming.mapreduce.TestStreamXmlRecordReader org.apache.hadoop.streaming.TestGzipInput org.apache.hadoop.streaming.TestStreaming org.apache.hadoop.streaming.TestStreamingFailure org.apache.hadoop.streaming.TestStreamingSeparator org.apache.hadoop.streaming.TestStreamingCounters org.apache.hadoop.streaming.TestFileArgs org.apache.hadoop.streaming.TestStreamDataProtocol org.apache.hadoop.streaming.TestStreamingExitStatus org.apache.hadoop.streaming.TestSymLink org.apache.hadoop.streaming.TestRawBytesStreaming org.apache.hadoop.streaming.TestStreamingCombiner org.apache.hadoop.streaming.TestStreamingStderr org.apache.hadoop.streaming.TestTypedBytesStreaming org.apache.hadoop.streaming.TestStreamingBackground org.apache.hadoop.streaming.TestStreamXmlRecordReader org.apache.hadoop.streaming.TestStreamingKeyValue org.apache.hadoop.streaming.TestStreamXmlMultipleRecords org.apache.hadoop.streaming.TestMultipleCachefiles org.apache.hadoop.streaming.TestStreamingOutputOnlyKeys org.apache.hadoop.yarn.util.TestFSDownload org.apache.hadoop.yarn.server.resourcemanager.TestClientRMService org.apache.hadoop.yarn.server.resourcemanager.webapp.TestRMWebServicesApps org.apache.hadoop.yarn.server.resourcemanager.TestRMRestart org.apache.hadoop.yarn.server.resourcemanager.TestApplicationACLs org.apache.hadoop.yarn.server.resourcemanager.TestRM org.apache.hadoop.hdfs.tools.TestDFSAdminWithHA org.apache.hadoop.hdfs.server.namenode.ha.TestPipelinesFailover
[jira] [Commented] (HADOOP-10851) NetgroupCache does not remove group memberships
[ https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14063959#comment-14063959 ] Benoy Antony commented on HADOOP-10851: --- The above comment is intended for a another jira. NetgroupCache does not remove group memberships --- Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10851.patch _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10851) NetgroupCache does not remove group memberships
[ https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10851: -- Attachment: (was: HADOOP-10852.patch) NetgroupCache does not remove group memberships --- Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10851.patch _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10852) NetgroupCache is not thread-safe
[ https://issues.apache.org/jira/browse/HADOOP-10852?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Benoy Antony updated HADOOP-10852: -- Attachment: HADOOP-10852.patch The _NetgroupCache_ is fixed by keeping state in only one _ConcurrentHashMap_ - {{userToNetgroupsMap}}. {{add}} updates {{userToNetgroupsMap}} directly. There is a slight performance loss when invoking {{isCached}} and {{{userToNetgroupsMap}} But these are invoked only during {{refresh}} and adding a new group. These invocations should be rare compared to {{getNetgroups}} invocation. NetgroupCache is not thread-safe Key: HADOOP-10852 URL: https://issues.apache.org/jira/browse/HADOOP-10852 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10852.patch _NetgroupCache_ internally uses two _ConcurrentHashMap_s and a boolean variable to signal updates on one of the _ConcurrentHashMap_s None of the functions are synchronized and hence is possible to have unexpected results due to race condition between different threads. As an example, consider the following sequence: Thread1 : {{add}} a group {{netgroupToUsersMap}} is updated. {{netgroupToUsersMapUpdated}} is set to true. Thread 2: calls {{getNetgroups}} for a user Due to re-ordering, {{netgroupToUsersMapUpdated=true}} is visible, but updates in {{netgroupToUsersMap}} is not visible. Does a wrong update with older {{netgroupToUsersMap}} values. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed
[ https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064015#comment-14064015 ] Hadoop QA commented on HADOOP-10591: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656112/HADOOP-10591.002.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:red}-1 tests included{color}. The patch doesn't appear to include any new or modified tests. Please justify why no new tests are needed for this patch. Also please list what manual steps were performed to verify this patch. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common: org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4295//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4295//console This message is automatically generated. Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed - Key: HADOOP-10591 URL: https://issues.apache.org/jira/browse/HADOOP-10591 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.2.0 Reporter: Hari Shreedharan Assignee: Colin Patrick McCabe Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch Currently direct buffers allocated by compression codecs like Gzip (which allocates 2 direct buffers per instance) are not deallocated when the stream is closed. Eventually for long running processes which create a huge number of files, these direct buffers are left hanging till a full gc, which may or may not happen in a reasonable amount of time - especially if the process does not use a whole lot of heap. Either these buffers should be pooled or they should be deallocated when the stream is closed. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10851) NetgroupCache does not remove group memberships
[ https://issues.apache.org/jira/browse/HADOOP-10851?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064048#comment-14064048 ] Hadoop QA commented on HADOOP-10851: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656111/HADOOP-10851.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:red}-1 core tests{color}. The patch failed these unit tests in hadoop-common-project/hadoop-common: org.apache.hadoop.fs.TestSymlinkLocalFSFileSystem org.apache.hadoop.fs.TestSymlinkLocalFSFileContext org.apache.hadoop.ipc.TestIPC {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4296//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4296//console This message is automatically generated. NetgroupCache does not remove group memberships --- Key: HADOOP-10851 URL: https://issues.apache.org/jira/browse/HADOOP-10851 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 2.4.1 Reporter: Benoy Antony Assignee: Benoy Antony Attachments: HADOOP-10851.patch _NetgroupCache_ is used by _GroupMappingServiceProvider_ implementations based on net groups. But it has a serious flaw in that once a user to group membership is established, it remains forever even if user is actually removed from the netgroup and cache is cleared. It is cleared only if the server is restarted. To reproduce this: * Cache a group with a set of users. * Test membership correctness. * Clear cache, remove a user from the group and cache the group again * Expected result : user’s groups should not include the group from which he/she is removed. * Actual result : user’s groups includes the group from which he/she was removed. It is also noted that _NetgroupCache_ has concurrency issues and a separate jira is filed to rectify them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10504) Document proxy server support
[ https://issues.apache.org/jira/browse/HADOOP-10504?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla updated HADOOP-10504: -- Target Version/s: 2.6.0 (was: 2.5.0) Moving this out to 2.6 since no one is actively looking at it. If anyone feels strongly about targeting 2.5, we can target 2.5. Document proxy server support - Key: HADOOP-10504 URL: https://issues.apache.org/jira/browse/HADOOP-10504 Project: Hadoop Common Issue Type: Improvement Components: documentation Affects Versions: 3.0.0, 2.5.0 Reporter: Daryn Sharp Priority: Blocker Document http proxy support introduced by HADOOP-10498. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10453) Do not use AuthenticatedURL in hadoop core
[ https://issues.apache.org/jira/browse/HADOOP-10453?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla updated HADOOP-10453: -- Target Version/s: 2.6.0 (was: 2.5.0) (Moving this out of 2.5) We can continue this conversation and handle this in 2.6. Do not use AuthenticatedURL in hadoop core -- Key: HADOOP-10453 URL: https://issues.apache.org/jira/browse/HADOOP-10453 Project: Hadoop Common Issue Type: Bug Reporter: Haohui Mai Priority: Blocker As [~daryn] has suggested in HDFS-4564: {quote} AuthenticatedURL is not used because it is buggy in part to causing replay attacks, double attempts to kerberos authenticate with the fallback authenticator if the TGT is expired, incorrectly uses the fallback authenticator (required by oozie servers) to add the username parameter which webhdfs has already included in the uri. AuthenticatedURL's attempt to do SPNEGO auth is a no-op because the JDK transparently does SPNEGO when the user's Subject (UGI) contains kerberos principals. Since AuthenticatedURL is now not used, webhdfs has to check the TGT itself for token operations. Bottom line is AuthenticatedURL is unnecessary and introduces nothing but problems for webhdfs. It's only useful for oozie's anon/non-anon support. {quote} However, several functionalities that relies on SPNEGO in secure mode suffer from the same problem. For example, NNs / JNs create HTTP connections to exchange fsimage and edit logs. Currently all of them are through {{AuthenticatedURL}}. This needs to be fixed to avoid security vulnerabilities. This jira purposes to remove {{AuthenticatedURL}} from hadoop core and to move it to oozie. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Assigned] (HADOOP-10821) Prepare the release notes for Hadoop 2.5.0
[ https://issues.apache.org/jira/browse/HADOOP-10821?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Karthik Kambatla reassigned HADOOP-10821: - Assignee: Karthik Kambatla Prepare the release notes for Hadoop 2.5.0 -- Key: HADOOP-10821 URL: https://issues.apache.org/jira/browse/HADOOP-10821 Project: Hadoop Common Issue Type: Task Affects Versions: 2.5.0 Reporter: Akira AJISAKA Assignee: Karthik Kambatla Priority: Blocker The release notes for 2.3.0+ (http://hadoop.apache.org/docs/r2.4.1/index.html) still talk about federation and MRv2 being new features. We should update them. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] shanyu zhao updated HADOOP-10840: - Attachment: HADOOP-10840.1.patch [~cnauroth] thanks for the findings. I couldn't reproduce the specific failure you posted, but I think it is caused by NativeAzureFileSystem.close() being called multiple times. I verified that NativeAzureFileSystemStore.close() can be called multiple times, but not NativeAzureFileSystem. And for this class, since the metrics system keeps a reference counting, we cannot call its close() multiple times, so I introduced Boolean isClosed to prevent it from being called multiple times. I added a new test case to call NativeAzureFileSystem.close() twice to verify this scenario. New patch attached. Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] shanyu zhao updated HADOOP-10840: - Attachment: (was: HADOOP-10840.1.patch) Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Chris Nauroth updated HADOOP-10840: --- Status: Patch Available (was: Open) Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] shanyu zhao updated HADOOP-10840: - Attachment: HADOOP-10840.1.patch Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10799) Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common.
[ https://issues.apache.org/jira/browse/HADOOP-10799?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Alejandro Abdelnur updated HADOOP-10799: Status: Open (was: Patch Available) canceling patch as some minor tweaks seems to be required in the DT client API to make DT wiring transparent. Refactor HTTP delegation token logic from httpfs into reusable code in hadoop-common. - Key: HADOOP-10799 URL: https://issues.apache.org/jira/browse/HADOOP-10799 Project: Hadoop Common Issue Type: Sub-task Components: security Affects Versions: 3.0.0 Reporter: Alejandro Abdelnur Assignee: Alejandro Abdelnur Attachments: HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch, HADOOP-10799.patch -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064258#comment-14064258 ] Hadoop QA commented on HADOOP-10840: {color:green}+1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656152/HADOOP-10840.1.patch against trunk revision . {color:green}+1 @author{color}. The patch does not contain any @author tags. {color:green}+1 tests included{color}. The patch appears to include 1 new or modified test files. {color:green}+1 javac{color}. The applied patch does not increase the total number of javac compiler warnings. {color:green}+1 javadoc{color}. There were no new javadoc warning messages. {color:green}+1 eclipse:eclipse{color}. The patch built with eclipse:eclipse. {color:green}+1 findbugs{color}. The patch does not introduce any new Findbugs (version 2.0.3) warnings. {color:green}+1 release audit{color}. The applied patch does not increase the total number of release audit warnings. {color:green}+1 core tests{color}. The patch passed unit tests in hadoop-tools/hadoop-azure. {color:green}+1 contrib tests{color}. The patch passed contrib unit tests. Test results: https://builds.apache.org/job/PreCommit-HADOOP-Build/4297//testReport/ Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4297//console This message is automatically generated. Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064274#comment-14064274 ] Chris Nauroth commented on HADOOP-10840: [~shanyu], nice find. I think your theory was correct. The last patch mostly fixed things, but I still see a few test failures. With an Azure storage key configured for testing against the live service, I get a failure in {{TestAzureConcurrentOutOfBandIo}}: {code} Running org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo Tests run: 1, Failures: 0, Errors: 1, Skipped: 0, Time elapsed: 0.81 sec FAILURE! - in org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo testReadOOBWrites(org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo) Time elapsed: 0.765 sec ERROR! org.apache.hadoop.metrics2.MetricsException: Metrics source AzureFileSystemMetrics already exists! at org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.newSourceName(DefaultMetricsSystem.java:143) at org.apache.hadoop.metrics2.lib.DefaultMetricsSystem.sourceName(DefaultMetricsSystem.java:120) at org.apache.hadoop.metrics2.impl.MetricsSystemImpl.register(MetricsSystemImpl.java:218) at org.apache.hadoop.fs.azure.metrics.AzureFileSystemMetricsSystem.registerSource(AzureFileSystemMetricsSystem.java:58) at org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.createOutOfBandStore(AzureBlobStorageTestAccount.java:331) at org.apache.hadoop.fs.azure.TestAzureConcurrentOutOfBandIo.setUp(TestAzureConcurrentOutOfBandIo.java:51) {code} I tested on both Mac and Windows. On the Windows VM only, I also get failures in {{TestRollingWindowAverage}} and {{TestNativeAzureFileSystemMocked}}: {code} Tests run: 1, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.149 sec FAILURE! - in org.apache.hadoop.fs.azure.metrics.TestRollingWindowAverage testBasicFunctionality(org.apache.hadoop.fs.azure.metrics.TestRollingWindowAverage) Time elapsed: 0.112 sec FAILURE! java.lang.AssertionError: expected:15 but was:10 at org.junit.Assert.fail(Assert.java:88) at org.junit.Assert.failNotEquals(Assert.java:743) at org.junit.Assert.assertEquals(Assert.java:118) at org.junit.Assert.assertEquals(Assert.java:555) at org.junit.Assert.assertEquals(Assert.java:542) at org.apache.hadoop.fs.azure.metrics.TestRollingWindowAverage.testBasicFunctionality(TestRollingWindowAverage.java:38) {code} {code} Tests run: 27, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 4.431 sec FAILURE! - in org.apache.hadoop.fs.azure.TestNativeAzureFileSystemMocked testFolderLastModifiedTime(org.apache.hadoop.fs.azure.TestNativeAzureFileSystemMocked) Time elapsed: 3.24 sec FAILURE! java.lang.AssertionError: null at org.junit.Assert.fail(Assert.java:86) at org.junit.Assert.assertTrue(Assert.java:41) at org.junit.Assert.assertFalse(Assert.java:64) at org.junit.Assert.assertFalse(Assert.java:74) at org.apache.hadoop.fs.azure.NativeAzureFileSystemBaseTest.testFolderLastModifiedTime(NativeAzureFileSystemBaseTest.java:479) {code} Can you explain why the following code was removed from {{AzureFileSystemMetricsSystem#fileSystemClosed}}? My understanding is that this code is important to guarantee timely publishing of metrics for an instance when it gets closed. I expect your new checks against double close are also sufficient to protect against extraneous publishing of metrics. {code} -if (instance != null) { - instance.publishMetricsNow(); -} {code} Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10794) A hadoop cluster needs clock synchronization
[ https://issues.apache.org/jira/browse/HADOOP-10794?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064326#comment-14064326 ] eric baldeschwieler commented on HADOOP-10794: -- Agreed. Doing math between server clocks is an anti-pattern to be avoided. We should file JIRAs to fix any instances of that , because the result is not a usable number given the error bars today. On Wednesday, July 16, 2014, Colin Patrick McCabe (JIRA) j...@apache.org -- --- E14 - typing on glass A hadoop cluster needs clock synchronization Key: HADOOP-10794 URL: https://issues.apache.org/jira/browse/HADOOP-10794 Project: Hadoop Common Issue Type: Bug Reporter: Zhijie Shen As a distributed system, a hadoop cluster wants the clock on all the participating hosts synchronized. Otherwise, some problems might happen. For example, in YARN-2251, due to the clock on the host for the task container falls behind that on the host of the AM container, the computed elapsed time (the diff between the timestamps produced on two hosts) becomes negative. In YARN-2251, we tried to mask the negative elapsed time. However, we should seek for a decent long term solution, such as providing mechanism to do and check clock synchronization. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] shanyu zhao updated HADOOP-10840: - Attachment: HADOOP-10840.2.patch Hi [~cnauroth], sorry I missed the live test scenarios which I have to manually configure a storage account. The problem with TestAzureConcurrentOutOfBandIo is I removed the registration code with DefaultFileSystem in production code but not the unit test files. I fixed that in the new patch (v2). Regarding the two windows test failure, looks like they are not related to my changes. It has something to do with system clock. Do they always fail or just transient? I cannot reproduce in my Ubuntu box. Let me try it on Windows box. I removed the publishMetricsNow() from fileSystemClosed() because I moved it to unregisterSource(). The expectation is the caller calls unregisterSource() before calling fileSystemClosed() which will push the last metrics out. To prevent confusion when someone doesn't call unregisterSource() but just call fileSystemClosed() I just added publishMetricsNow() to fileSystemClosed() under the condition that numFileSystems==1, this way before we are going to shutdown the metrics system, we'll do another safety push, usually at this point the metrics sources should be empty. Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.2.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10840) Fix OutOfMemoryError caused by metrics system in Azure File System
[ https://issues.apache.org/jira/browse/HADOOP-10840?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064383#comment-14064383 ] Hadoop QA commented on HADOOP-10840: {color:red}-1 overall{color}. Here are the results of testing the latest attachment http://issues.apache.org/jira/secure/attachment/12656178/HADOOP-10840.2.patch against trunk revision . {color:red}-1 patch{color}. Trunk compilation may be broken. Console output: https://builds.apache.org/job/PreCommit-HADOOP-Build/4298//console This message is automatically generated. Fix OutOfMemoryError caused by metrics system in Azure File System -- Key: HADOOP-10840 URL: https://issues.apache.org/jira/browse/HADOOP-10840 Project: Hadoop Common Issue Type: Bug Components: metrics Affects Versions: 2.4.1 Reporter: shanyu zhao Assignee: shanyu zhao Attachments: HADOOP-10840.1.patch, HADOOP-10840.2.patch, HADOOP-10840.patch In Hadoop 2.x the Hadoop File System framework changed and no cache is implemented (refer to HADOOP-6356). This means for every WASB access, a new NativeAzureFileSystem is created, along which a Metrics source created and added to MetricsSystemImpl. Over time the sources accumulated, eating memory and causing Java OutOfMemoryError. The fix is to utilize the unregisterSource() method added to MetricsSystem in HADOOP-10839. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed
[ https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064390#comment-14064390 ] Colin Patrick McCabe commented on HADOOP-10591: --- Test failures are unrelated. TestIPC is timing out resolving a hostname (looks like a jenkins problem), and the symlink tests have been failing for some other patches too. Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed - Key: HADOOP-10591 URL: https://issues.apache.org/jira/browse/HADOOP-10591 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.2.0 Reporter: Hari Shreedharan Assignee: Colin Patrick McCabe Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch Currently direct buffers allocated by compression codecs like Gzip (which allocates 2 direct buffers per instance) are not deallocated when the stream is closed. Eventually for long running processes which create a huge number of files, these direct buffers are left hanging till a full gc, which may or may not happen in a reasonable amount of time - especially if the process does not use a whole lot of heap. Either these buffers should be pooled or they should be deallocated when the stream is closed. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10591) Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed
[ https://issues.apache.org/jira/browse/HADOOP-10591?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064393#comment-14064393 ] Aaron T. Myers commented on HADOOP-10591: - Agreed. +1, the latest patch looks good to me. Thanks, Colin. Compression codecs must used pooled direct buffers or deallocate direct buffers when stream is closed - Key: HADOOP-10591 URL: https://issues.apache.org/jira/browse/HADOOP-10591 Project: Hadoop Common Issue Type: Bug Affects Versions: 2.2.0 Reporter: Hari Shreedharan Assignee: Colin Patrick McCabe Attachments: HADOOP-10591.001.patch, HADOOP-10591.002.patch Currently direct buffers allocated by compression codecs like Gzip (which allocates 2 direct buffers per instance) are not deallocated when the stream is closed. Eventually for long running processes which create a huge number of files, these direct buffers are left hanging till a full gc, which may or may not happen in a reasonable amount of time - especially if the process does not use a whole lot of heap. Either these buffers should be pooled or they should be deallocated when the stream is closed. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Commented] (HADOOP-10816) key shell returns -1 to the shell on error, should be 1
[ https://issues.apache.org/jira/browse/HADOOP-10816?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanelfocusedCommentId=14064425#comment-14064425 ] Andrew Wang commented on HADOOP-10816: -- Your pedantry has bested me, let's go with with the v2 patch :) Thanks for being a good sport Mike, I'll commit this shortly. key shell returns -1 to the shell on error, should be 1 --- Key: HADOOP-10816 URL: https://issues.apache.org/jira/browse/HADOOP-10816 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 3.0.0 Reporter: Mike Yoder Assignee: Mike Yoder Fix For: 3.0.0 Attachments: HADOOP-10816.001.patch, HADOOP-10816.002.patch I've seen this in several places now - commands returning -1 on failure to the shell. It's a bug. Someone confused their posix style returns (0 on success, 0 on failure) with program returns, which are an unsigned character. Thus, a return of -1 actually becomes 255 to the shell. {noformat} $ hadoop key create happykey2 --provider kms://http@localhost:16000/kms --attr a=a --attr a=b Each attribute must correspond to only one value: atttribute a was repeated ... $ echo $? 255 {noformat} A return value of 1 instead of -1 does the right thing. -- This message was sent by Atlassian JIRA (v6.2#6252)
[jira] [Updated] (HADOOP-10816) KeyShell returns -1 on error to the shell, should be 1
[ https://issues.apache.org/jira/browse/HADOOP-10816?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Andrew Wang updated HADOOP-10816: - Summary: KeyShell returns -1 on error to the shell, should be 1 (was: key shell returns -1 to the shell on error, should be 1) KeyShell returns -1 on error to the shell, should be 1 -- Key: HADOOP-10816 URL: https://issues.apache.org/jira/browse/HADOOP-10816 Project: Hadoop Common Issue Type: Bug Components: security Affects Versions: 3.0.0 Reporter: Mike Yoder Assignee: Mike Yoder Fix For: 3.0.0 Attachments: HADOOP-10816.001.patch, HADOOP-10816.002.patch I've seen this in several places now - commands returning -1 on failure to the shell. It's a bug. Someone confused their posix style returns (0 on success, 0 on failure) with program returns, which are an unsigned character. Thus, a return of -1 actually becomes 255 to the shell. {noformat} $ hadoop key create happykey2 --provider kms://http@localhost:16000/kms --attr a=a --attr a=b Each attribute must correspond to only one value: atttribute a was repeated ... $ echo $? 255 {noformat} A return value of 1 instead of -1 does the right thing. -- This message was sent by Atlassian JIRA (v6.2#6252)