[jira] [Commented] (HADOOP-15426) Make S3guard client resilient to DDB throttle events and network failures

2018-08-30 Thread genericqa (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15426?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598236#comment-16598236
 ] 

genericqa commented on HADOOP-15426:


| (x) *{color:red}-1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 16m 
33s{color} | {color:blue} Docker mode activated. {color} |
|| || || || {color:brown} Prechecks {color} ||
| {color:green}+1{color} | {color:green} @author {color} | {color:green}  0m  
0s{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green}  0m 
 0s{color} | {color:green} The patch appears to include 6 new or modified test 
files. {color} |
|| || || || {color:brown} trunk Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  5m 
45s{color} | {color:blue} Maven dependency ordering for branch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 17m 
39s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 14m 
36s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green}  2m 
58s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
57s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
16m 25s{color} | {color:green} branch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
27s{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
36s{color} | {color:green} trunk passed {color} |
|| || || || {color:brown} Patch Compile Tests {color} ||
| {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue}  0m 
20s{color} | {color:blue} Maven dependency ordering for patch {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green}  1m 
18s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 13m 
40s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 13m 
40s{color} | {color:green} the patch passed {color} |
| {color:orange}-0{color} | {color:orange} checkstyle {color} | {color:orange}  
2m 56s{color} | {color:orange} root: The patch generated 15 new + 32 unchanged 
- 2 fixed = 47 total (was 34) {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green}  1m 
51s{color} | {color:green} the patch passed {color} |
| {color:red}-1{color} | {color:red} whitespace {color} | {color:red}  0m  
0s{color} | {color:red} The patch has 23 line(s) that end in whitespace. Use 
git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply {color} |
| {color:green}+1{color} | {color:green} xml {color} | {color:green}  0m  
4s{color} | {color:green} The patch has no ill-formed XML file. {color} |
| {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 
10m 56s{color} | {color:green} patch has no errors when building and testing 
our client artifacts. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green}  2m 
43s{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green}  1m 
29s{color} | {color:green} the patch passed {color} |
|| || || || {color:brown} Other Tests {color} ||
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  8m  
6s{color} | {color:green} hadoop-common in the patch passed. {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green}  4m 
34s{color} | {color:green} hadoop-aws in the patch passed. {color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green}  0m 
41s{color} | {color:green} The patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black}126m 46s{color} | 
{color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:ba1ab08 |
| JIRA Issue | HADOOP-15426 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12937784/HADOOP-15426-008.patch
 |
| Optional Tests |  dupname  asflicense  compile  javac  javadoc  mvninstall  
mvnsite  unit  shadedclient  xml  findbugs  checkstyle  |
| uname | Linux 1b840870a225 4.4.0-133-generic #159-Ubuntu SMP Fri Aug 10 
07:31:43 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | 

[jira] [Updated] (HADOOP-15663) ABFS: Simplify configuration

2018-08-30 Thread Thomas Marquardt (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15663?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Thomas Marquardt updated HADOOP-15663:
--
Resolution: Fixed
Status: Resolved  (was: Patch Available)

commit d58ea7c96abe1dbba2f89019676a2dd869e6b550
Author: Thomas Marquardt 
Date: Fri Aug 31 03:24:42 2018 +

HADOOP-15663. ABFS: Simplify configuration.
 Contributed by Da Zhou.

> ABFS: Simplify configuration
> 
>
> Key: HADOOP-15663
> URL: https://issues.apache.org/jira/browse/HADOOP-15663
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Reporter: Thomas Marquardt
>Assignee: Da Zhou
>Priority: Major
> Attachments: HADOOP-15663-HADOOP-15407-001.patch, 
> HADOOP-15663-HADOOP-15407-002.patch, HADOOP-15663-HADOOP-15407-003.patch, 
> HADOOP-15663-HADOOP-15407-004.patch, HADOOP-15663-HADOOP-15407-005.patch
>
>
> Configuration for WASB and ABFS is too complex.  The current approach is to 
> use four files for test configuration. 
> Both WASB and ABFS have basic test configuration which is committed to the 
> repo (azure-test.xml and azure-bfs-test.xml).  Currently these contain the 
> fs.AbstractFileSystem.[scheme].impl configuration, but otherwise are empty 
> except for an include reference to a file containing the endpoint 
> credentials. 
> Both WASB and ABFS have endpoint credential configuration files 
> (azure-auth-keys.xml and azure-bfs-auth-keys.xml).  These have been added to 
> .gitignore to prevent them from accidentally being submitted in a patch, 
> which would leak the developers storage account credentials.  These files 
> contain account names, storage account keys, and service endpoints.
> There is some overlap of the configuration for WASB and ABFS, where they use 
> the same property name but use different values.  
> 1) Let's reduce the number of test configuration files to one, if possible.
> 2) Let's simplify the account name, key, and endpoint configuration for WASB 
> and ABFS if possible, but still support the legacy way of doing it, which is 
> very error prone.
> 3) Let's improve error handling, so that typos or misconfiguration are not so 
> difficult to troubleshoot.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15663) ABFS: Simplify configuration

2018-08-30 Thread Thomas Marquardt (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15663?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598179#comment-16598179
 ] 

Thomas Marquardt commented on HADOOP-15663:
---

+1, thanks for addressing the feedback. All tests pass against my US storage 
account:

  *mvn -T 1C -Dparallel-tests -Dscale -DtestsThreadCount=8 clean verify*

  *Tests run: 265, Failures: 0, Errors: 0, Skipped: 11*
  *Tests run: 1, Failures: 0, Errors: 0, Skipped: 0*
  *Tests run: 863, Failures: 0, Errors: 0, Skipped: 262*
  *Tests run: 186, Failures: 0, Errors: 0, Skipped: 10*

I made edits to testing_azure.md to clarify the configuration for running the 
tests.
 There is a section named *"Testing the Azure ABFS Client"* near the end of the 
document,
 and readers can copy paste the contents to the file 
*src/test/resources/azure-auth-keys.xml*
 to get started. The necessary contents are also below:
{noformat}


http://www.w3.org/2001/XInclude;>
  
fs.azure.abfs.account.name
{ACCOUNT_NAME}.dfs.core.windows.net
  

  
fs.azure.account.key.{ACCOUNT_NAME}.dfs.core.windows.net
{ACCOUNT_ACCESS_KEY}
  

  
fs.azure.wasb.account.name
{ACCOUNT_NAME}.blob.core.windows.net
  
  
  
fs.azure.account.key.{ACCOUNT_NAME}.blob.core.windows.net
{ACCOUNT_ACCESS_KEY}
  

  
fs.contract.test.fs.abfs
abfs://{CONTAINER_NAME}@{ACCOUNT_NAME}.dfs.core.windows.net
A file system URI to be used by the contract 
tests.
  

  
fs.contract.test.fs.wasb
wasb://{CONTAINER_NAME}@{ACCOUNT_NAME}.blob.core.windows.net
A file system URI to be used by the contract 
tests.
  

{noformat}

> ABFS: Simplify configuration
> 
>
> Key: HADOOP-15663
> URL: https://issues.apache.org/jira/browse/HADOOP-15663
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Reporter: Thomas Marquardt
>Assignee: Da Zhou
>Priority: Major
> Attachments: HADOOP-15663-HADOOP-15407-001.patch, 
> HADOOP-15663-HADOOP-15407-002.patch, HADOOP-15663-HADOOP-15407-003.patch, 
> HADOOP-15663-HADOOP-15407-004.patch, HADOOP-15663-HADOOP-15407-005.patch
>
>
> Configuration for WASB and ABFS is too complex.  The current approach is to 
> use four files for test configuration. 
> Both WASB and ABFS have basic test configuration which is committed to the 
> repo (azure-test.xml and azure-bfs-test.xml).  Currently these contain the 
> fs.AbstractFileSystem.[scheme].impl configuration, but otherwise are empty 
> except for an include reference to a file containing the endpoint 
> credentials. 
> Both WASB and ABFS have endpoint credential configuration files 
> (azure-auth-keys.xml and azure-bfs-auth-keys.xml).  These have been added to 
> .gitignore to prevent them from accidentally being submitted in a patch, 
> which would leak the developers storage account credentials.  These files 
> contain account names, storage account keys, and service endpoints.
> There is some overlap of the configuration for WASB and ABFS, where they use 
> the same property name but use different values.  
> 1) Let's reduce the number of test configuration files to one, if possible.
> 2) Let's simplify the account name, key, and endpoint configuration for WASB 
> and ABFS if possible, but still support the legacy way of doing it, which is 
> very error prone.
> 3) Let's improve error handling, so that typos or misconfiguration are not so 
> difficult to troubleshoot.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15107) Stabilize/tune S3A committers; review correctness & docs

2018-08-30 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598148#comment-16598148
 ] 

Hudson commented on HADOOP-15107:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14855/])
HADOOP-15107. Stabilize/tune S3A committers; review correctness & docs. 
(stevel: rev 5a0babf76550f63dad4c17173c4da2bf335c6532)
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/Invoker.java
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/S3ACommitterFactory.java
* (edit) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/staging/integration/ITestStagingCommitProtocol.java
* (edit) 
hadoop-mapreduce-project/hadoop-mapreduce-client/hadoop-mapreduce-client-core/src/main/java/org/apache/hadoop/mapreduce/lib/output/PathOutputCommitter.java
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/staging/PartitionedStagingCommitter.java
* (edit) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/magic/ITestMagicCommitProtocol.java
* (edit) 
hadoop-tools/hadoop-aws/src/site/markdown/tools/hadoop-aws/committers.md
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/staging/DirectoryStagingCommitter.java
* (edit) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/magic/ITMagicCommitMRJob.java
* (add) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/staging/integration/ITStagingCommitMRJobBadDest.java
* (edit) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/AbstractCommitITest.java
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/staging/Paths.java
* (edit) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/AbstractITCommitMRJob.java
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/magic/MagicS3GuardCommitter.java
* (edit) 
hadoop-tools/hadoop-aws/src/site/markdown/tools/hadoop-aws/committer_architecture.md
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/staging/StagingCommitter.java
* (add) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/ITestS3ACommitterFactory.java
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/commit/AbstractS3ACommitter.java
* (edit) 
hadoop-tools/hadoop-aws/src/test/java/org/apache/hadoop/fs/s3a/commit/AbstractITCommitProtocol.java


> Stabilize/tune S3A committers; review correctness & docs
> 
>
> Key: HADOOP-15107
> URL: https://issues.apache.org/jira/browse/HADOOP-15107
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.1.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Blocker
> Fix For: 3.1.2
>
> Attachments: HADOOP-15107-001.patch, HADOOP-15107-002.patch, 
> HADOOP-15107-003.patch, HADOOP-15107-004.patch
>
>
> I'm writing about the paper on the committers, one which, being a proper 
> paper, requires me to show the committers work.
> # define the requirements of a "Correct" committed job (this applies to the 
> FileOutputCommitter too)
> # show that the Staging committer meets these requirements (most of this is 
> implicit in that it uses the V1 FileOutputCommitter to marshall .pendingset 
> lists from committed tasks to the final destination, where they are read and 
> committed.
> # Show the magic committer also works.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15680) ITestNativeAzureFileSystemConcurrencyLive times out

2018-08-30 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15680?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598147#comment-16598147
 ] 

Hudson commented on HADOOP-15680:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14855/])
HADOOP-15680. ITestNativeAzureFileSystemConcurrencyLive times out. (stevel: rev 
e8d138ca7c1b695688515d816ac693437c87df62)
* (edit) 
hadoop-tools/hadoop-azure/src/test/java/org/apache/hadoop/fs/azure/ITestNativeAzureFileSystemConcurrencyLive.java


> ITestNativeAzureFileSystemConcurrencyLive times out
> ---
>
> Key: HADOOP-15680
> URL: https://issues.apache.org/jira/browse/HADOOP-15680
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Andras Bokor
>Assignee: Andras Bokor
>Priority: Major
> Fix For: 3.1.2
>
> Attachments: HADOOP-15680.001.patch, HADOOP-15680.002.patch
>
>
> When I am running tests locally ITestNativeAzureFileSystemConcurrencyLive 
> sometimes times out.
> I would like to increase the timeout to avoid unnecessary noise.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15705) Typo in the definition of "stable" in the interface classification

2018-08-30 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15705?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598143#comment-16598143
 ] 

Hudson commented on HADOOP-15705:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14855/])
HADOOP-15705. Typo in the definition of "stable" in the interface (templedf: 
rev d53a10b0a552155de700e396fd7f450a4c5f9c22)
* (edit) 
hadoop-common-project/hadoop-common/src/site/markdown/InterfaceClassification.md


> Typo in the definition of "stable" in the interface classification
> --
>
> Key: HADOOP-15705
> URL: https://issues.apache.org/jira/browse/HADOOP-15705
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
>Priority: Minor
> Fix For: 3.2.0, 3.0.4, 3.1.2
>
> Attachments: HADOOP-15705.001.patch
>
>
> "Compatible changes allowed: maintenance (x.Y.0)"
> should be 
> "Compatible changes allowed: maintenance (x.y.Z)"



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15698) KMS log4j is not initialized properly at startup

2018-08-30 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15698?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598145#comment-16598145
 ] 

Hudson commented on HADOOP-15698:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14855/])
HADOOP-15698. KMS log4j is not initialized properly at startup. (xiao: rev 
781437c219dc3422797a32dc7ba72cd4f5ee38e2)
* (edit) 
hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSWebApp.java
* (edit) 
hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSConfiguration.java
* (edit) 
hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/KMSWebServer.java


> KMS log4j is not initialized properly at startup
> 
>
> Key: HADOOP-15698
> URL: https://issues.apache.org/jira/browse/HADOOP-15698
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.1.0
>Reporter: Kitti Nanasi
>Assignee: Kitti Nanasi
>Priority: Major
> Fix For: 3.2.0, 3.0.4, 3.1.2
>
> Attachments: HADOOP-15698.001.patch
>
>
> During KMs startup, log4j logs don't show up resulting in important logs 
> getting omitted. This happens because log4 initialisation only happens in 
> KMSWebApp#contextInitialized and logs written before that don't show up.
> For example the following log never shows up:
> [https://github.com/apache/hadoop/blob/a55d6bba71c81c1c4e9d8cd11f55c78f10a548b0/hadoop-common-project/hadoop-auth/src/main/java/org/apache/hadoop/security/authentication/util/ZKSignerSecretProvider.java#L197-L199]
> Another example is that the KMS startup message never shows up in the kms 
> logs.
> Note that this works in the unit tests, because MiniKMS sets the log4j system 
> property.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15667) FileSystemMultipartUploader should verify that UploadHandle has non-0 length

2018-08-30 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15667?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598146#comment-16598146
 ] 

Hudson commented on HADOOP-15667:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14855/])
HADOOP-15667. FileSystemMultipartUploader should verify that (stevel: rev 
2e6c1109dcdeedb59a3345047e9201271c9a0b27)
* (edit) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/MultipartUploader.java
* (edit) 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/fs/FileSystemMultipartUploader.java
* (edit) 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/fs/contract/AbstractContractMultipartUploaderTest.java
* (edit) 
hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AMultipartUploader.java


> FileSystemMultipartUploader should verify that UploadHandle has non-0 length
> 
>
> Key: HADOOP-15667
> URL: https://issues.apache.org/jira/browse/HADOOP-15667
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.2.0
>Reporter: Ewan Higgs
>Assignee: Ewan Higgs
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HADOOP-15667.001.patch
>
>
> The S3AMultipartUploader has a good check on the length of the UploadHandle. 
> This should be moved to MultipartUploader, made protected, and called in the 
> various implementations.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15706) Typo in compatibility doc: SHOUD -> SHOULD

2018-08-30 Thread Hudson (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598149#comment-16598149
 ] 

Hudson commented on HADOOP-15706:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14855 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/14855/])
HADOOP-15706. Typo in compatibility doc: SHOUD -> SHOULD (Contributed by 
(templedf: rev f2c2a68ec208f640e778fc41f95f0284fcc44729)
* (edit) hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md


> Typo in compatibility doc: SHOUD -> SHOULD
> --
>
> Key: HADOOP-15706
> URL: https://issues.apache.org/jira/browse/HADOOP-15706
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Daniel Templeton
>Assignee: Laszlo Kollar
>Priority: Trivial
> Fix For: 3.2.0, 3.0.4, 3.1.2
>
> Attachments: HADOOP-15706.001.patch
>
>
> {quote}% grep SHOUD 
> ./hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md
> Apache Hadoop revisions SHOUD retain binary compatability such that 
> end-user{quote}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-14833) Remove s3a user:secret authentication

2018-08-30 Thread Mingliang Liu (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-14833?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16595664#comment-16595664
 ] 

Mingliang Liu edited comment on HADOOP-14833 at 8/31/18 1:45 AM:
-

Thanks Steve. I'm +1 on the proposal, and will finish the review this week 
(EDIT: hopefully). Don't get blocked by me if there is any binding +1.


was (Author: liuml07):
Thanks Steve. I'm +1 on the proposal, and will finish the review this week. 
Don't get blocked by me if there is any binding +1.

> Remove s3a user:secret authentication
> -
>
> Key: HADOOP-14833
> URL: https://issues.apache.org/jira/browse/HADOOP-14833
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-14833-001.patch
>
>
> Remove the s3a://user:secret@host auth mechanism from S3a. 
> As well as being insecure, it causes problems with S3Guard's URI matching 
> code.
> Proposed: cull it utterly. We've been telling people to stop using it since 
> HADOOP-3733



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HADOOP-15707:

Description: 
Hadoop has a few services with HA setups and it is common to set them behind 
Load Balancers.
We should add a way for the Load Balancers to understand what should be the UI 
to show.
For example, the standby RM just redirects the requests to the active RM.
However, if both RMs are behind a Load Balancer the IP might not be reachable.

Most Load balancers have probes to check if a server reports HTTP code 200:
* 
[Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
* 
[AWS|https://docs.aws.amazon.com/elasticloadbalancing/latest/classic/elb-healthchecks.html]

Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
report if they are active.

  was:
Hadoop has a few services with HA setups and it is common to set them behind 
Load Balancers.
We should add a way for the Load Balancers to understand what should be the UI 
to show.
For example, the standby RM just redirects the requests to the active RM.
However, if both RMs are behind a Load Balancer the IP might not be reachable.

Most Load balancers have probes to check if a server reports HTTP code 200:
* 
[Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
* 
[AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]

Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
report if they are active.


> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch, HADOOP-15707.003.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.aws.amazon.com/elasticloadbalancing/latest/classic/elb-healthchecks.html]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598109#comment-16598109
 ] 

Lukas Majercak commented on HADOOP-15707:
-

Updated documentation in HA pages for this in patch003.

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch, HADOOP-15707.003.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HADOOP-15707:

Attachment: HADOOP-15707.003.patch

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch, HADOOP-15707.003.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15696) KMS performance regression due to too many open file descriptors after Jetty migration

2018-08-30 Thread Misha Dmitriev (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15696?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598091#comment-16598091
 ] 

Misha Dmitriev commented on HADOOP-15696:
-

[~xiaochen] I agree that any work on connection reusing is more complex and 
should be done under another jira.

Can you elaborate on "the connection object is created in relation to the 
caller ugi" statement? What's the exact condition for when a connection can be 
reused or not?

> KMS performance regression due to too many open file descriptors after Jetty 
> migration
> --
>
> Key: HADOOP-15696
> URL: https://issues.apache.org/jira/browse/HADOOP-15696
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: kms
>Affects Versions: 3.0.0-alpha2
>Reporter: Wei-Chiu Chuang
>Assignee: Wei-Chiu Chuang
>Priority: Blocker
> Attachments: HADOOP-15696.001.patch, Screen Shot 2018-08-22 at 
> 11.36.16 AM.png, Screen Shot 2018-08-22 at 4.26.51 PM.png, Screen Shot 
> 2018-08-22 at 4.26.51 PM.png, Screen Shot 2018-08-22 at 4.27.02 PM.png, 
> Screen Shot 2018-08-22 at 4.30.32 PM.png, Screen Shot 2018-08-22 at 4.30.39 
> PM.png, Screen Shot 2018-08-24 at 7.08.16 PM.png
>
>
> We recently found KMS performance regressed in Hadoop 3.0, possibly linking 
> to the migration from Tomcat to Jetty in HADOOP-13597.
> Symptoms:
> # Hadoop 3.x KMS open file descriptors quickly rises to more than 10 thousand 
> under stress, sometimes even exceeds 32K, which is the system limit, causing 
> failures for any access to encryption zones. Our internal testing shows the 
> openfd number was in the range of a few hundred in Hadoop 2.x, and it 
> increases by almost 100x in Hadoop 3.
> # Hadoop 3.x KMS as much as twice the heap size than in Hadoop 2.x. The same 
> heap size can go OOM in Hadoop 3.x. Jxray analysis suggests most of them are 
> temporary byte arrays associated with open SSL connections.
> # Due to the heap usage, Hadoop 3.x KMS has more frequent GC activities, and 
> we observed up to 20% performance reduction due to GC.
> A possible solution is to reduce the idle timeout setting in HttpServer2. It 
> is currently hard-coded 10 seconds. By setting it to 1 second, open fds 
> dropped from 20 thousand down to 3 thousand in my experiment.
> File this jira to invite open discussion for a solution.
> Credit: [~mi...@cloudera.com] for the proposed Jetty idle timeout remedy; 
> [~xiaochen] for digging into this problem.
> Screenshots:
> CDH5 (Hadoop 2) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.39 PM.png! 
> CDH6 (Hadoop 3) KMS CPU utilization, resident memory and file descriptor 
> chart.
>  !Screen Shot 2018-08-22 at 4.30.32 PM.png! 
> CDH5 (Hadoop 2) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.26.51 PM.png! 
> CDH6 (Hadoop 3) GC activities on the KMS process
>  !Screen Shot 2018-08-22 at 4.27.02 PM.png! 
> JXray report
>  !Screen Shot 2018-08-22 at 11.36.16 AM.png! 
> open fd drops from 20 k down to 3k after the proposed change.
>  !Screen Shot 2018-08-24 at 7.08.16 PM.png! 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread JIRA


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598029#comment-16598029
 ] 

Íñigo Goiri edited comment on HADOOP-15707 at 8/30/18 11:12 PM:


Thanks [~virajith] for the comments.
We have been using it for a couple weeks with our internal Load Balancers.
Today tested it in Azure VMSS and it seems to do the trick, this is the example 
for how to set it for the RM:
{code}
az network lb probe create -g XXX --lb-name YYY -n RMProbe --protocol http 
--port 8088 --path /isActive
{code}


was (Author: elgoiri):
Thanks [~virajith] for the comments.
We have been using it for a couple weeks with our internal Load Balancers.
Today we had tested it in Azure VMSS and it seems to do the trick, this is the 
example for how to set it for the RM:
{code}
az network lb probe create -g XXX --lb-name YYY -n RMProbe --protocol http 
--port 8088 --path /isActive
{code}

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread JIRA


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598029#comment-16598029
 ] 

Íñigo Goiri commented on HADOOP-15707:
--

Thanks [~virajith] for the comments.
We have been using it for a couple weeks with our internal Load Balancers.
Today we had tested it in Azure VMSS and it seems to do the trick, this is the 
example for how to set it for the RM:
{code}
az network lb probe create -g XXX --lb-name YYY -n RMProbe --protocol http 
--port 8088 --path /isActive
{code}

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Virajith Jalaparti (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16598019#comment-16598019
 ] 

Virajith Jalaparti commented on HADOOP-15707:
-

I don't know this part of the code much but the patch itself looks mostly good 
to me. Has this been tested with Azure/AWS load balancers? If so, would be 
great to see any info on that.

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15694) ABFS: Allow OAuth credentials to not be tied to accounts

2018-08-30 Thread Sean Mackrory (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15694?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597993#comment-16597993
 ] 

Sean Mackrory commented on HADOOP-15694:


Attaching a patch that's basically what I want to do. Feedback welcome, but I'm 
not entirely finished with this so I'm not proposing it for inclusion yet. 
Integration tests are running now, but so far so good. I also still need to add 
my own tests and do my own code review for silly mistakes, but I was able to 
connect to ABFS by just specifying "fs.azure.account.key" without any account 
context (although that'll be more useful with ABFS), and that's basically what 
I wanted to achieve.

> ABFS: Allow OAuth credentials to not be tied to accounts
> 
>
> Key: HADOOP-15694
> URL: https://issues.apache.org/jira/browse/HADOOP-15694
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Sean Mackrory
>Assignee: Sean Mackrory
>Priority: Major
> Attachments: HADOOP-15694.001.patch
>
>
> Now that there's OAuth support, it's possible to have a notion of identity 
> that's distinct from the account itself. If a cluster is configured via OAuth 
> with it's own identity, it's likely operators will want to use that identity 
> regardless of which storage account a job uses.
> So OAuth configs right now (and probably others) are looked up with 
> .. I propose that we add a function for looking up these 
> configs that returns an account-specific value if it exists, but in the event 
> it does not will also try to return , if that exists.
> I can work on a patch for this if nobody has any objections.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15694) ABFS: Allow OAuth credentials to not be tied to accounts

2018-08-30 Thread Sean Mackrory (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15694?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Sean Mackrory updated HADOOP-15694:
---
Attachment: HADOOP-15694.001.patch

> ABFS: Allow OAuth credentials to not be tied to accounts
> 
>
> Key: HADOOP-15694
> URL: https://issues.apache.org/jira/browse/HADOOP-15694
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Sean Mackrory
>Assignee: Sean Mackrory
>Priority: Major
> Attachments: HADOOP-15694.001.patch
>
>
> Now that there's OAuth support, it's possible to have a notion of identity 
> that's distinct from the account itself. If a cluster is configured via OAuth 
> with it's own identity, it's likely operators will want to use that identity 
> regardless of which storage account a job uses.
> So OAuth configs right now (and probably others) are looked up with 
> .. I propose that we add a function for looking up these 
> configs that returns an account-specific value if it exists, but in the event 
> it does not will also try to return , if that exists.
> I can work on a patch for this if nobody has any objections.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597972#comment-16597972
 ] 

Lukas Majercak edited comment on HADOOP-15707 at 8/30/18 10:07 PM:
---

Add patch002 for setting up IsRouterActiveServlet in RouterHttpServer


was (Author: lukmajercak):
Add -HADOOP-1570+7+-.002.patch for setting up IsRouterActiveServlet in 
RouterHttpServer

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597972#comment-16597972
 ] 

Lukas Majercak commented on HADOOP-15707:
-

Add HADOOP-15705.002.patch for setting up IsRouterActiveServlet in 
RouterHttpServer

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Comment Edited] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597972#comment-16597972
 ] 

Lukas Majercak edited comment on HADOOP-15707 at 8/30/18 10:06 PM:
---

Add -HADOOP-1570+7+-.002.patch for setting up IsRouterActiveServlet in 
RouterHttpServer


was (Author: lukmajercak):
Add HADOOP-15705.002.patch for setting up IsRouterActiveServlet in 
RouterHttpServer

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HADOOP-15707:

Attachment: HADOOP-15707.002.patch

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch, 
> HADOOP-15707.002.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15703) ABFS - Implement client-side throttling

2018-08-30 Thread Da Zhou (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15703?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597969#comment-16597969
 ] 

Da Zhou commented on HADOOP-15703:
--

AbfsClientThrottlingIntercept:
 L76, please remove "java.net.", it doesn't need to be full qualified name.
 L99: ABFS Gen2 doesn't depend on Azure Storage SDK, please update the comments.
  

> ABFS - Implement client-side throttling 
> 
>
> Key: HADOOP-15703
> URL: https://issues.apache.org/jira/browse/HADOOP-15703
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Sneha Varma
>Assignee: Sneha Varma
>Priority: Major
> Attachments: HADOOP-15703-HADOOP-15407-001.patch
>
>
> Big data workloads frequently exceed the AzureBlobFS max ingress and egress 
> limits 
> (https://docs.microsoft.com/en-us/azure/storage/common/storage-scalability-targets).
>  For example, the max ingress limit for a GRS account in the United States is 
> currently 10 Gbps. When the limit is exceeded, the AzureBlobFS service fails 
> a percentage of incoming requests, and this causes the client to initiate the 
> retry policy. The retry policy delays requests by sleeping, but the sleep 
> duration is independent of the client throughput and account limit. This 
> results in low throughput, due to the high number of failed requests and 
> thrashing causes by the retry policy.
> To fix this, we introduce a client-side throttle which minimizes failed 
> requests and maximizes throughput. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597966#comment-16597966
 ] 

Lukas Majercak commented on HADOOP-15707:
-

Added patch 001 with updated comments. I'll add the documentation in HA pages 
in a bit.

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HADOOP-15707:

Attachment: HADOOP-15707.001.patch

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch, HADOOP-15707.001.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Íñigo Goiri updated HADOOP-15707:
-
Description: 
Hadoop has a few services with HA setups and it is common to set them behind 
Load Balancers.
We should add a way for the Load Balancers to understand what should be the UI 
to show.
For example, the standby RM just redirects the requests to the active RM.
However, if both RMs are behind a Load Balancer the IP might not be reachable.

Most Load balancers have probes to check if a server reports HTTP code 200:
* 
[Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
* 
[AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]

Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
report if they are active.

  was:
Hadoop has a few services with HA setups and it is common to set them behind 
Load Balancers.
We should add a way for the Load Balancers to understand what should be the UI 
to show.
For example, the standby RM just redirects the requests to the active RM.
However, if both RMs are behind a Load Balancer the IP might not be reachable.

Most Load balancers have probes to check if a server reports HTTP code 200:
* 
[Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
* 
[AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
report if they are active.


> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread JIRA


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Íñigo Goiri updated HADOOP-15707:
-
Description: 
Hadoop has a few services with HA setups and it is common to set them behind 
Load Balancers.
We should add a way for the Load Balancers to understand what should be the UI 
to show.
For example, the standby RM just redirects the requests to the active RM.
However, if both RMs are behind a Load Balancer the IP might not be reachable.

Most Load balancers have probes to check if a server reports HTTP code 200:
* 
[Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
* 
[AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
report if they are active.

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch
>
>
> Hadoop has a few services with HA setups and it is common to set them behind 
> Load Balancers.
> We should add a way for the Load Balancers to understand what should be the 
> UI to show.
> For example, the standby RM just redirects the requests to the active RM.
> However, if both RMs are behind a Load Balancer the IP might not be reachable.
> Most Load balancers have probes to check if a server reports HTTP code 200:
> * 
> [Azure|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> * 
> [AWS|https://docs.microsoft.com/en-us/azure/load-balancer/load-balancer-custom-probe-overview]
> Components in Hadoop (e.g., NN, RM, Router,...) should have a unified way to 
> report if they are active.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread JIRA


[ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597942#comment-16597942
 ] 

Íñigo Goiri commented on HADOOP-15707:
--

Thanks [~lukmajercak] for working on this.
Right now,  [^HADOOP-15707.000.patch] touches Commons, YARN (RM), and HDFS (NN 
and Router).
I think this is small enough that it could go in this single patch.

We may want to add some documentation maybe in the HA pages.
I'll update the description with the use cases.

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15707?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lukas Majercak updated HADOOP-15707:

Attachment: HADOOP-15707.000.patch

> Add IsActiveServlet to be used for Load Balancers
> -
>
> Key: HADOOP-15707
> URL: https://issues.apache.org/jira/browse/HADOOP-15707
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: common
>Reporter: Lukas Majercak
>Assignee: Lukas Majercak
>Priority: Major
> Attachments: HADOOP-15707.000.patch
>
>




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15707) Add IsActiveServlet to be used for Load Balancers

2018-08-30 Thread Lukas Majercak (JIRA)
Lukas Majercak created HADOOP-15707:
---

 Summary: Add IsActiveServlet to be used for Load Balancers
 Key: HADOOP-15707
 URL: https://issues.apache.org/jira/browse/HADOOP-15707
 Project: Hadoop Common
  Issue Type: New Feature
  Components: common
Reporter: Lukas Majercak
Assignee: Lukas Majercak






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-10219) ipc.Client.setupIOstreams() needs to check for ClientCache.stopClient requested shutdowns

2018-08-30 Thread Lukas Majercak (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-10219?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597910#comment-16597910
 ] 

Lukas Majercak commented on HADOOP-10219:
-

Kindly ping.

> ipc.Client.setupIOstreams() needs to check for ClientCache.stopClient 
> requested shutdowns 
> --
>
> Key: HADOOP-10219
> URL: https://issues.apache.org/jira/browse/HADOOP-10219
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: ipc
>Affects Versions: 2.2.0, 2.6.0
>Reporter: Steve Loughran
>Assignee: Kihwal Lee
>Priority: Major
> Attachments: HADOOP-10219.patch, HADOOP-10219.v1.patch, 
> HADOOP-10219.v2.patch, HADOOP-10219.v3.patch, HADOOP-10219.v4.patch
>
>
> When {{ClientCache.stopClient()}} is called to stop the IPC client, if the 
> client
>  is blocked spinning due to a connectivity problem, it does not exit until 
> the policy has timed out -so the stopClient() operation can hang for an 
> extended period of time.
> This can surface in the shutdown hook of FileSystem.cache.closeAll()
> Also, Client.stop() is for used in NN switch from Standby to Active, and can 
> therefore have very bad consequences and cause downtime.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-14734) add option to tag DDB table(s) created

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-14734?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran reassigned HADOOP-14734:
---

Assignee: Gabor Bota  (was: Abraham Fine)

> add option to tag DDB table(s) created
> --
>
> Key: HADOOP-14734
> URL: https://issues.apache.org/jira/browse/HADOOP-14734
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.0.0-beta1
>Reporter: Steve Loughran
>Assignee: Gabor Bota
>Priority: Minor
> Attachments: HADOOP-14734-001.patch, HADOOP-14734-002.patch, 
> HADOOP-14734-003.patch
>
>
> Many organisations have a "no untagged" resource policy; s3guard runs into 
> this when a table is created untagged. If there's a strict "delete untagged 
> resources" policy, the tables will go without warning.
> Proposed: we add an option which can be used to declare the tags for a table 
> when created, use it in creation. No need to worry about updating/viewing 
> tags, as the AWS console can do that



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15426) Make S3guard client resilient to DDB throttle events and network failures

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15426:

Status: Patch Available  (was: Open)

patch 008: patch 007 with checkstyles corrected.

looks like we've been slowly adding legit checkstyle errors to the code; needs 
a fixup patch which I'm not going to do here, as it will complicate cherry 
picking this back

> Make S3guard client resilient to DDB throttle events and network failures
> -
>
> Key: HADOOP-15426
> URL: https://issues.apache.org/jira/browse/HADOOP-15426
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.1.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Blocker
> Attachments: HADOOP-15426-001.patch, HADOOP-15426-002.patch, 
> HADOOP-15426-003.patch, HADOOP-15426-004.patch, HADOOP-15426-005.patch, 
> HADOOP-15426-006.patch, HADOOP-15426-007.patch, HADOOP-15426-008.patch, 
> Screen Shot 2018-07-24 at 15.16.46.png, Screen Shot 2018-07-25 at 
> 16.22.10.png, Screen Shot 2018-07-25 at 16.28.53.png, Screen Shot 2018-07-27 
> at 14.07.38.png, 
> org.apache.hadoop.fs.s3a.s3guard.ITestDynamoDBMetadataStoreScale-output.txt
>
>
> managed to create on a parallel test run
> {code}
> org.apache.hadoop.fs.s3a.AWSServiceThrottledException: delete on 
> s3a://hwdev-steve-ireland-new/fork-0005/test/existing-dir/existing-file: 
> com.amazonaws.services.dynamodbv2.model.ProvisionedThroughputExceededException:
>  The level of configured provisioned throughput for the table was exceeded. 
> Consider increasing your provisioning level with the UpdateTable API. 
> (Service: AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ProvisionedThroughputExceededException; Request ID: 
> RDM3370REDBBJQ0SLCLOFC8G43VV4KQNSO5AEMVJF66Q9ASUAAJG): The level of 
> configured provisioned throughput for the table was exceeded. Consider 
> increasing your provisioning level with the UpdateTable API. (Service: 
> AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ProvisionedThroughputExceededException; Request ID: 
> RDM3370REDBBJQ0SLCLOFC8G43VV4KQNSO5AEMVJF66Q9ASUAAJG)
>   at 
> {code}
> We should be able to handle this. 400 "bad things happened" error though, not 
> the 503 from S3.
> h3. We need a retry handler for DDB throttle operations



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15426) Make S3guard client resilient to DDB throttle events and network failures

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15426:

Attachment: HADOOP-15426-008.patch

> Make S3guard client resilient to DDB throttle events and network failures
> -
>
> Key: HADOOP-15426
> URL: https://issues.apache.org/jira/browse/HADOOP-15426
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.1.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Blocker
> Attachments: HADOOP-15426-001.patch, HADOOP-15426-002.patch, 
> HADOOP-15426-003.patch, HADOOP-15426-004.patch, HADOOP-15426-005.patch, 
> HADOOP-15426-006.patch, HADOOP-15426-007.patch, HADOOP-15426-008.patch, 
> Screen Shot 2018-07-24 at 15.16.46.png, Screen Shot 2018-07-25 at 
> 16.22.10.png, Screen Shot 2018-07-25 at 16.28.53.png, Screen Shot 2018-07-27 
> at 14.07.38.png, 
> org.apache.hadoop.fs.s3a.s3guard.ITestDynamoDBMetadataStoreScale-output.txt
>
>
> managed to create on a parallel test run
> {code}
> org.apache.hadoop.fs.s3a.AWSServiceThrottledException: delete on 
> s3a://hwdev-steve-ireland-new/fork-0005/test/existing-dir/existing-file: 
> com.amazonaws.services.dynamodbv2.model.ProvisionedThroughputExceededException:
>  The level of configured provisioned throughput for the table was exceeded. 
> Consider increasing your provisioning level with the UpdateTable API. 
> (Service: AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ProvisionedThroughputExceededException; Request ID: 
> RDM3370REDBBJQ0SLCLOFC8G43VV4KQNSO5AEMVJF66Q9ASUAAJG): The level of 
> configured provisioned throughput for the table was exceeded. Consider 
> increasing your provisioning level with the UpdateTable API. (Service: 
> AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ProvisionedThroughputExceededException; Request ID: 
> RDM3370REDBBJQ0SLCLOFC8G43VV4KQNSO5AEMVJF66Q9ASUAAJG)
>   at 
> {code}
> We should be able to handle this. 400 "bad things happened" error though, not 
> the 503 from S3.
> h3. We need a retry handler for DDB throttle operations



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15426) Make S3guard client resilient to DDB throttle events and network failures

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15426?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15426:

Status: Open  (was: Patch Available)

> Make S3guard client resilient to DDB throttle events and network failures
> -
>
> Key: HADOOP-15426
> URL: https://issues.apache.org/jira/browse/HADOOP-15426
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.1.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Blocker
> Attachments: HADOOP-15426-001.patch, HADOOP-15426-002.patch, 
> HADOOP-15426-003.patch, HADOOP-15426-004.patch, HADOOP-15426-005.patch, 
> HADOOP-15426-006.patch, HADOOP-15426-007.patch, Screen Shot 2018-07-24 at 
> 15.16.46.png, Screen Shot 2018-07-25 at 16.22.10.png, Screen Shot 2018-07-25 
> at 16.28.53.png, Screen Shot 2018-07-27 at 14.07.38.png, 
> org.apache.hadoop.fs.s3a.s3guard.ITestDynamoDBMetadataStoreScale-output.txt
>
>
> managed to create on a parallel test run
> {code}
> org.apache.hadoop.fs.s3a.AWSServiceThrottledException: delete on 
> s3a://hwdev-steve-ireland-new/fork-0005/test/existing-dir/existing-file: 
> com.amazonaws.services.dynamodbv2.model.ProvisionedThroughputExceededException:
>  The level of configured provisioned throughput for the table was exceeded. 
> Consider increasing your provisioning level with the UpdateTable API. 
> (Service: AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ProvisionedThroughputExceededException; Request ID: 
> RDM3370REDBBJQ0SLCLOFC8G43VV4KQNSO5AEMVJF66Q9ASUAAJG): The level of 
> configured provisioned throughput for the table was exceeded. Consider 
> increasing your provisioning level with the UpdateTable API. (Service: 
> AmazonDynamoDBv2; Status Code: 400; Error Code: 
> ProvisionedThroughputExceededException; Request ID: 
> RDM3370REDBBJQ0SLCLOFC8G43VV4KQNSO5AEMVJF66Q9ASUAAJG)
>   at 
> {code}
> We should be able to handle this. 400 "bad things happened" error though, not 
> the 503 from S3.
> h3. We need a retry handler for DDB throttle operations



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15657) Registering MutableQuantiles via Metric annotation

2018-08-30 Thread Vrushali C (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15657?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597625#comment-16597625
 ] 

Vrushali C commented on HADOOP-15657:
-

+1 LGTM

thanks for the patch [~Sushil-K-S]! 

I will commit this later today. 

> Registering MutableQuantiles via Metric annotation
> --
>
> Key: HADOOP-15657
> URL: https://issues.apache.org/jira/browse/HADOOP-15657
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: metrics
>Reporter: Sushil Ks
>Priority: Major
> Attachments: HADOOP-15657.001.patch
>
>
> Currently when creating new metrics we use @Metric annotation for registering 
> the MutableMetric i.e
> {code:java}
> @Metric
> private MutableInt foobarMetricCount
> {code}
> However,  there's no support for registering MutableQuantiles via Metric 
> annotation, hence creating this Jira to register MutableQuantiles via Metric 
> annotation. 
> Example:
>  
> {code:java}
> @Metric(about = "async PUT entities latency", valueName = "latency", interval 
> = 10)
> private MutableQuantiles foobarAsyncLatency;
>  
> {code}
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15706) Typo in compatibility doc: SHOUD -> SHOULD

2018-08-30 Thread Daniel Templeton (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton updated HADOOP-15706:
--
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 3.1.2
   3.0.4
   3.2.0
   Status: Resolved  (was: Patch Available)

Thanks for the patch!  Committed to trunk, branch-3.1, and branch-3.0.

> Typo in compatibility doc: SHOUD -> SHOULD
> --
>
> Key: HADOOP-15706
> URL: https://issues.apache.org/jira/browse/HADOOP-15706
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Daniel Templeton
>Assignee: Laszlo Kollar
>Priority: Trivial
> Fix For: 3.2.0, 3.0.4, 3.1.2
>
> Attachments: HADOOP-15706.001.patch
>
>
> {quote}% grep SHOUD 
> ./hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md
> Apache Hadoop revisions SHOUD retain binary compatability such that 
> end-user{quote}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15706) Typo in compatibility doc: SHOUD -> SHOULD

2018-08-30 Thread Daniel Templeton (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15706?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597615#comment-16597615
 ] 

Daniel Templeton commented on HADOOP-15706:
---

+1, thanks for the patch.

> Typo in compatibility doc: SHOUD -> SHOULD
> --
>
> Key: HADOOP-15706
> URL: https://issues.apache.org/jira/browse/HADOOP-15706
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Daniel Templeton
>Assignee: Laszlo Kollar
>Priority: Trivial
> Attachments: HADOOP-15706.001.patch
>
>
> {quote}% grep SHOUD 
> ./hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md
> Apache Hadoop revisions SHOUD retain binary compatability such that 
> end-user{quote}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15706) Typo in compatibility doc: SHOUD -> SHOULD

2018-08-30 Thread Laszlo Kollar (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Kollar updated HADOOP-15706:
---
Status: Patch Available  (was: Open)

> Typo in compatibility doc: SHOUD -> SHOULD
> --
>
> Key: HADOOP-15706
> URL: https://issues.apache.org/jira/browse/HADOOP-15706
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Daniel Templeton
>Assignee: Laszlo Kollar
>Priority: Trivial
> Attachments: HADOOP-15706.001.patch
>
>
> {quote}% grep SHOUD 
> ./hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md
> Apache Hadoop revisions SHOUD retain binary compatability such that 
> end-user{quote}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15706) Typo in compatibility doc: SHOUD -> SHOULD

2018-08-30 Thread Laszlo Kollar (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Laszlo Kollar updated HADOOP-15706:
---
Attachment: HADOOP-15706.001.patch

> Typo in compatibility doc: SHOUD -> SHOULD
> --
>
> Key: HADOOP-15706
> URL: https://issues.apache.org/jira/browse/HADOOP-15706
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Daniel Templeton
>Assignee: Laszlo Kollar
>Priority: Trivial
> Attachments: HADOOP-15706.001.patch
>
>
> {quote}% grep SHOUD 
> ./hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md
> Apache Hadoop revisions SHOUD retain binary compatability such that 
> end-user{quote}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-15706) Typo in compatibility doc: SHOUD -> SHOULD

2018-08-30 Thread Daniel Templeton (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15706?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Daniel Templeton reassigned HADOOP-15706:
-

Assignee: Laszlo Kollar  (was: Daniel Templeton)

> Typo in compatibility doc: SHOUD -> SHOULD
> --
>
> Key: HADOOP-15706
> URL: https://issues.apache.org/jira/browse/HADOOP-15706
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Daniel Templeton
>Assignee: Laszlo Kollar
>Priority: Trivial
>
> {quote}% grep SHOUD 
> ./hadoop-common-project/hadoop-common/src/site/markdown/Compatibility.md
> Apache Hadoop revisions SHOUD retain binary compatability such that 
> end-user{quote}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15639) Classloader issue with hadoop cmd -libjars option

2018-08-30 Thread Qinghui Xu (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15639?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597575#comment-16597575
 ] 

Qinghui Xu commented on HADOOP-15639:
-

Here is a patch to fix it: 
https://github.com/criteo-forks/hadoop-common/pull/87/commits/da6fdfa53410906068383550b6214b7a81344083

> Classloader issue with hadoop cmd -libjars option 
> --
>
> Key: HADOOP-15639
> URL: https://issues.apache.org/jira/browse/HADOOP-15639
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Qinghui Xu
>Priority: Major
>
> '-libjars' option gives the possibility to add extra libraries in a hadoop 
> command by providing a custom URLClassLoader to fetch jars regarding to the 
> option.
> The classloader is provided to both hadoop Configuration and the main 
> thread's context classloader, so that when calling Configuration#getClass or 
> using Class.forName the class can be loaded by the URLClassLoader.
> But two instances of URLClassLoader are provided to Configuration and to 
> thread context classloader respectively, which makes a same class possible to 
> be loaded twice by two classloaders. And class loaded by different 
> classloaders is considered different.
> So the following assertion will fail:
> {code:java}
> // CustomFileSystem is provided by -libjars option
> Class clazz1 = conf.getClass("CustomFileSystem");
> Class clazz2 = Class.forName("CustomFileSystem");
> // The two classes are different
> assert clazz1 == clazz2{code}



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15107) Stabilize/tune S3A committers; review correctness & docs

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15107?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15107:

   Resolution: Fixed
Fix Version/s: 3.1.2
   Status: Resolved  (was: Patch Available)

> Stabilize/tune S3A committers; review correctness & docs
> 
>
> Key: HADOOP-15107
> URL: https://issues.apache.org/jira/browse/HADOOP-15107
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.1.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Blocker
> Fix For: 3.1.2
>
> Attachments: HADOOP-15107-001.patch, HADOOP-15107-002.patch, 
> HADOOP-15107-003.patch, HADOOP-15107-004.patch
>
>
> I'm writing about the paper on the committers, one which, being a proper 
> paper, requires me to show the committers work.
> # define the requirements of a "Correct" committed job (this applies to the 
> FileOutputCommitter too)
> # show that the Staging committer meets these requirements (most of this is 
> implicit in that it uses the V1 FileOutputCommitter to marshall .pendingset 
> lists from committed tasks to the final destination, where they are read and 
> committed.
> # Show the magic committer also works.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15107) Stabilize/tune S3A committers; review correctness & docs

2018-08-30 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15107?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597511#comment-16597511
 ] 

Steve Loughran commented on HADOOP-15107:
-

Thanks, committed to 3.1.x & trunk!



bq. New commit attempts always get the same attempt id +1? (I don't know how 
those are allocated)

its how they know to recover from the previous attempt. Yarn App ID is used to 
guarantee uniqueness over apps. Spark always starts off with 0 as its app 
Id/attempt ID, so >1 query from different spark instances can clash.

bq. The mergePathsV1 seems pretty straightforward. Not sure why the actual code 
is so complicated.

unplanned evolution is my guess, possibly with a goal of not breaking any 
explict subclasses of the FileOutputCommitter. I didn't try that for the new 
commit stuff.

v2 resilience? 

It is broken in that nothing can handle a task which fails during commit: its 
(non-atomic) state is unknown. Neither MapReduce nor Spark are aware 
of/resilient to this issue. There's also the partitioning failure mode: task 
doesn't fail during commit, it merely hangs for a while (GC?) then completes 
its commit when it resumes, without noticing that it's been superceded by a 
second task attempt, or indeed, that the entire job has now completed. Oops. 
Bear in mind though that outside object stores with slow renames the 
probability of a failure during task commit is likely to low.

Really committers should expose their semantics here & MR & Spark can handle 
this failure condition. 

V1 doesn't have this problem as the task commit is atomic; job commit is not, 
but as {{isCommitJobRepeatable()}} returns false for that, MR AM restart knows 
to give up then (something is saved to HDFS indicate in-job-commit). Spark 
doesn't restart failed AM/driver, so it's moot there.

S3A committers

* Staging: relies on V1 semantics in cluster HDFS
* Magic. Task commit writes {{PendingSet}} of all files to commit to task in an 
atomic PUT; task commit is therefore also atomic. After a job completes we 
purge all pending uploads under $dest, so any failed tasks' output is deleted.


> Stabilize/tune S3A committers; review correctness & docs
> 
>
> Key: HADOOP-15107
> URL: https://issues.apache.org/jira/browse/HADOOP-15107
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.1.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Blocker
> Fix For: 3.1.2
>
> Attachments: HADOOP-15107-001.patch, HADOOP-15107-002.patch, 
> HADOOP-15107-003.patch, HADOOP-15107-004.patch
>
>
> I'm writing about the paper on the committers, one which, being a proper 
> paper, requires me to show the committers work.
> # define the requirements of a "Correct" committed job (this applies to the 
> FileOutputCommitter too)
> # show that the Staging committer meets these requirements (most of this is 
> implicit in that it uses the V1 FileOutputCommitter to marshall .pendingset 
> lists from committed tasks to the final destination, where they are read and 
> committed.
> # Show the magic committer also works.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15628) S3A Filesystem does not check return from AmazonS3Client deleteObjects

2018-08-30 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15628?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597497#comment-16597497
 ] 

Steve Loughran commented on HADOOP-15628:
-

we are retrying for network failures but not for failure of partial delete 
except when an exception is raised.

But in the AWS SDK, I see the check and uprate to an exception
{Code}
DeleteObjectsResponse response = invoke(request, responseHandler, 
deleteObjectsRequest.getBucketName(), null);
if ( !response.getErrors().isEmpty() ) {
..
MultiObjectDeleteException ex = new MultiObjectDeleteException(
response.getErrors(),
response.getDeletedObjects());
...
throw ex;
}
{Code}

it's relying on the errors list being non-empty. So if your store is failing 
with some other error code rather than returning a list of rejected files, we 
aren't picking it up

> S3A Filesystem does not check return from AmazonS3Client deleteObjects
> --
>
> Key: HADOOP-15628
> URL: https://issues.apache.org/jira/browse/HADOOP-15628
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 2.9.1, 2.8.4, 3.1.1, 3.0.3
> Environment: Hadoop 3.0.2 / Hadoop 2.8.3
> Hive 2.3.2 / Hive 2.3.3 / Hive 3.0.0
> Non-AWS S3 implementation
>Reporter: Steve Jacobs
>Priority: Minor
>
> Deletes in S3A that use the Multi-Delete functionality in the Amazon S3 api 
> do not check to see if all objects have been succesfully delete. In the event 
> of a failure, the api will still return a 200 OK (which isn't checked 
> currently):
> [Delete Code from Hadoop 
> 2.8|https://github.com/apache/hadoop/blob/a0da1ec01051108b77f86799dd5e97563b2a3962/hadoop-tools/hadoop-aws/src/main/java/org/apache/hadoop/fs/s3a/S3AFileSystem.java#L574]
>  
> {code:java}
> if (keysToDelete.size() == MAX_ENTRIES_TO_DELETE) {
> DeleteObjectsRequest deleteRequest =
> new DeleteObjectsRequest(bucket).withKeys(keysToDelete);
> s3.deleteObjects(deleteRequest);
> statistics.incrementWriteOps(1);
> keysToDelete.clear();
> }
> {code}
> This should be converted to use the DeleteObjectsResult class from the 
> S3Client: 
> [Amazon Code 
> Example|https://docs.aws.amazon.com/AmazonS3/latest/dev/DeletingMultipleObjectsUsingJava.htm]
> {code:java}
> // Verify that the objects were deleted successfully.
> DeleteObjectsResult delObjRes = 
> s3Client.deleteObjects(multiObjectDeleteRequest); int successfulDeletes = 
> delObjRes.getDeletedObjects().size();
> System.out.println(successfulDeletes + " objects successfully deleted.");
> {code}
> Bucket policies can be misconfigured, and deletes will fail without warning 
> by S3A clients.
>  
>  
>  



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-14630) Contract Tests to verify create, mkdirs and rename under a file is forbidden

2018-08-30 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-14630?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597491#comment-16597491
 ] 

Steve Loughran commented on HADOOP-14630:
-

I'm not sure this is ready to go in; because of the 
{{TestLocalFSContractRename}} failure. I was just rebasing it to trunk to see 
if it still builds before going near it again

> Contract Tests to verify create, mkdirs and rename under a file is forbidden
> 
>
> Key: HADOOP-14630
> URL: https://issues.apache.org/jira/browse/HADOOP-14630
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs, fs/azure, fs/s3, fs/swift
>Affects Versions: 2.9.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
> Attachments: HADOOP-14630-001.patch, HADOOP-14630-002.patch, 
> HADOOP-14630-003.patch, HADOOP-14630-004.patch
>
>
> Object stores can get into trouble in ways which an FS would never, do, ways 
> so obvious we've never done tests for them. We know what the problems are: 
> test for file and dir creation directly/indirectly under other files
> * mkdir(file/file)
> * mkdir(file/subdir)
> * dir under file/subdir/subdir
> * dir/dir2/file, verify dir & dir2 exist
> * dir/dir2/dir3, verify dir & dir2 exist 
> * rename(src, file/dest)
> * rename(src, file/dir/dest)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15680) ITestNativeAzureFileSystemConcurrencyLive times out

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15680?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15680:

   Resolution: Fixed
Fix Version/s: 3.1.2
   Status: Resolved  (was: Patch Available)

committed to branch-3.1 & trunk: thanks!

> ITestNativeAzureFileSystemConcurrencyLive times out
> ---
>
> Key: HADOOP-15680
> URL: https://issues.apache.org/jira/browse/HADOOP-15680
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Andras Bokor
>Assignee: Andras Bokor
>Priority: Major
> Fix For: 3.1.2
>
> Attachments: HADOOP-15680.001.patch, HADOOP-15680.002.patch
>
>
> When I am running tests locally ITestNativeAzureFileSystemConcurrencyLive 
> sometimes times out.
> I would like to increase the timeout to avoid unnecessary noise.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15667) FileSystemMultipartUploader should verify that UploadHandle has non-0 length

2018-08-30 Thread Steve Loughran (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15667?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-15667:

   Resolution: Fixed
Fix Version/s: 3.2.0
   Status: Resolved  (was: Patch Available)

ran all tests against s3 ireland w s3guard/ddb, all well.

committed to trunk

> FileSystemMultipartUploader should verify that UploadHandle has non-0 length
> 
>
> Key: HADOOP-15667
> URL: https://issues.apache.org/jira/browse/HADOOP-15667
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.2.0
>Reporter: Ewan Higgs
>Assignee: Ewan Higgs
>Priority: Major
> Fix For: 3.2.0
>
> Attachments: HADOOP-15667.001.patch
>
>
> The S3AMultipartUploader has a good check on the length of the UploadHandle. 
> This should be moved to MultipartUploader, made protected, and called in the 
> various implementations.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15704) ABFS: Consider passing FS URI to CustomDelegationTokenManager

2018-08-30 Thread Steve Loughran (JIRA)


[ 
https://issues.apache.org/jira/browse/HADOOP-15704?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16597367#comment-16597367
 ] 

Steve Loughran commented on HADOOP-15704:
-

bq. We currently have implementations of CustomDelegationTokenManager, and need 
to do a little leg work here, but it may be possible to update before GA.

should be trivial in intellij; it's not like the property needs to be used —yet—

> ABFS: Consider passing FS URI to CustomDelegationTokenManager
> -
>
> Key: HADOOP-15704
> URL: https://issues.apache.org/jira/browse/HADOOP-15704
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/azure
>Reporter: Thomas Marquardt
>Priority: Major
>
> Refer to Steve's comments in HADOOP-15692.  Passing the FS or FS URI to the 
> CustomDelegationTokenManager would allow it to provide per-filesystem tokens. 
>  We currently have implementations of CustomDelegationTokenManager, and need 
> to do a little leg work here, but it may be possible to update before GA.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15687) Credentials class should allow access to aliases

2018-08-30 Thread Lars Francke (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Francke updated HADOOP-15687:
--
Status: Patch Available  (was: Open)

Attached is a patch that adds two new methods to get to the full map of tokens 
and secret keys. It also changes the test (which was working around the lack of 
these methods) to make use of them.

> Credentials class should allow access to aliases
> 
>
> Key: HADOOP-15687
> URL: https://issues.apache.org/jira/browse/HADOOP-15687
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.1.0
>Reporter: Lars Francke
>Assignee: Lars Francke
>Priority: Trivial
> Attachments: HADOOP-15687.patch
>
>
> The credentials class can read token files from disk which are keyed by an 
> alias. It also allows to retrieve tokens by alias and it also allows to list 
> all tokens.
> It does not - however - allow to get the full map of all tokens including the 
> aliases (or at least a list of all aliases).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Assigned] (HADOOP-15687) Credentials class should allow access to aliases

2018-08-30 Thread Lars Francke (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Francke reassigned HADOOP-15687:
-

Assignee: Lars Francke

> Credentials class should allow access to aliases
> 
>
> Key: HADOOP-15687
> URL: https://issues.apache.org/jira/browse/HADOOP-15687
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.1.0
>Reporter: Lars Francke
>Assignee: Lars Francke
>Priority: Trivial
> Attachments: HADOOP-15687.patch
>
>
> The credentials class can read token files from disk which are keyed by an 
> alias. It also allows to retrieve tokens by alias and it also allows to list 
> all tokens.
> It does not - however - allow to get the full map of all tokens including the 
> aliases (or at least a list of all aliases).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15687) Credentials class should allow access to aliases

2018-08-30 Thread Lars Francke (JIRA)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15687?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Lars Francke updated HADOOP-15687:
--
Attachment: HADOOP-15687.patch

> Credentials class should allow access to aliases
> 
>
> Key: HADOOP-15687
> URL: https://issues.apache.org/jira/browse/HADOOP-15687
> Project: Hadoop Common
>  Issue Type: Improvement
>Affects Versions: 3.1.0
>Reporter: Lars Francke
>Priority: Trivial
> Attachments: HADOOP-15687.patch
>
>
> The credentials class can read token files from disk which are keyed by an 
> alias. It also allows to retrieve tokens by alias and it also allows to list 
> all tokens.
> It does not - however - allow to get the full map of all tokens including the 
> aliases (or at least a list of all aliases).



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org