[jira] [Commented] (HADOOP-16739) Fix native build failure of hadoop-pipes on CentOS 8

2020-02-09 Thread Akira Ajisaka (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033391#comment-17033391
 ] 

Akira Ajisaka commented on HADOOP-16739:


Thank you, [~iwasakims]!

> Fix native build failure of hadoop-pipes on CentOS 8
> 
>
> Key: HADOOP-16739
> URL: https://issues.apache.org/jira/browse/HADOOP-16739
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: tools/pipes
>Affects Versions: 2.10.0, 3.2.1
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Major
> Fix For: 3.3.0
>
> Attachments: HADOOP-16739.001.patch
>
>
> Native build fails in hadoop-tools/hadoop-pips on CentOS 8 due to lack of 
> rpc.h which was removed from glibc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-16739) Fix native build failure of hadoop-pipes on CentOS 8

2020-02-09 Thread Masatake Iwasaki (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-16739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HADOOP-16739:
--
Fix Version/s: 3.3.0
 Hadoop Flags: Reviewed
   Resolution: Fixed
   Status: Resolved  (was: Patch Available)

I committed this to trunk. The contents added to BUILDING.txt is only 
applidcable to trunk (protobuf-3.7.1).

If you want to build branch-2.10 with {{-Pnative}} on CentOS 8, you need the 
fix of HADOOP-14597 too. The patches of HADOOP-16739 and HADOOP-14597 are 
cleanly applicable on today's branch-2.10.

> Fix native build failure of hadoop-pipes on CentOS 8
> 
>
> Key: HADOOP-16739
> URL: https://issues.apache.org/jira/browse/HADOOP-16739
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: tools/pipes
>Affects Versions: 2.10.0, 3.2.1
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Major
> Fix For: 3.3.0
>
> Attachments: HADOOP-16739.001.patch
>
>
> Native build fails in hadoop-tools/hadoop-pips on CentOS 8 due to lack of 
> rpc.h which was removed from glibc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13951) Precommit builds do not adequately protect against test malformed fs permissions.

2020-02-09 Thread Akira Ajisaka (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-13951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033384#comment-17033384
 ] 

Akira Ajisaka commented on HADOOP-13951:


I noticed the cleanup job only cleans "Precommit-HDFS-build". Added a parameter 
to specify the job name to clean up other Jenkins jobs.
https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-clean-hdfs-workspace/16/console

> Precommit builds do not adequately protect against test malformed fs 
> permissions.
> -
>
> Key: HADOOP-13951
> URL: https://issues.apache.org/jira/browse/HADOOP-13951
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: test
>Reporter: Sean Busbey
>Priority: Critical
>
> Right now this is expressed as failed Precommit-YARN-build jobs when they run 
> on H5 / H6 (see INFRA-13148), but the problem exists for all of the hadoop 
> related precommit jobs.
> The issue is that we have some tests in Common (and maybe HDFS) that 
> purposefully set permissions within the {{target/}} directory to simulate a 
> failure to interact with underlying fs data. The test sets some 
> subdirectories to have permissions such that we can no longer delete their 
> contents.
> Right now our precommit jobs include a step post-yetus-test-patch that 
> traverses the target directories and ensures that all subdirectories are 
> modifiable:
> {code}
> find ${WORKSPACE} -name target | xargs chmod -R u+w
> {code}
> Unfortunately, if we don't get to that line (say due to an aborted build, or 
> if the call to yetus test-patch exceeds the job timeout), then we are left in 
> a state where there are still subdirectories that can't be modified 
> (including deleted).
> Our builds also currently attempt to run a {{git clean}} at the very start of 
> the build after the repo is updated. If we have one of the aforementioned 
> timeouts that leaves a can't-be-deleted test directory, then all future 
> builds on that machine will fail attempting to run the {{git clean}} command.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13951) Precommit builds do not adequately protect against test malformed fs permissions.

2020-02-09 Thread Akira Ajisaka (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-13951?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033378#comment-17033378
 ] 

Akira Ajisaka commented on HADOOP-13951:


H17 is failing 
([https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1406/console)|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1406/console]
 and ran 
[https://builds.apache.org/view/H-L/view/Hadoop/job/hadoop-clean-hdfs-workspace/]
 on the node.

Thanks Sean and Andrew for the work.

> Precommit builds do not adequately protect against test malformed fs 
> permissions.
> -
>
> Key: HADOOP-13951
> URL: https://issues.apache.org/jira/browse/HADOOP-13951
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: test
>Reporter: Sean Busbey
>Priority: Critical
>
> Right now this is expressed as failed Precommit-YARN-build jobs when they run 
> on H5 / H6 (see INFRA-13148), but the problem exists for all of the hadoop 
> related precommit jobs.
> The issue is that we have some tests in Common (and maybe HDFS) that 
> purposefully set permissions within the {{target/}} directory to simulate a 
> failure to interact with underlying fs data. The test sets some 
> subdirectories to have permissions such that we can no longer delete their 
> contents.
> Right now our precommit jobs include a step post-yetus-test-patch that 
> traverses the target directories and ensures that all subdirectories are 
> modifiable:
> {code}
> find ${WORKSPACE} -name target | xargs chmod -R u+w
> {code}
> Unfortunately, if we don't get to that line (say due to an aborted build, or 
> if the call to yetus test-patch exceeds the job timeout), then we are left in 
> a state where there are still subdirectories that can't be modified 
> (including deleted).
> Our builds also currently attempt to run a {{git clean}} at the very start of 
> the build after the repo is updated. If we have one of the aforementioned 
> timeouts that leaves a can't-be-deleted test directory, then all future 
> builds on that machine will fail attempting to run the {{git clean}} command.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-16739) Fix native build failure of hadoop-pipes on CentOS 8

2020-02-09 Thread Hudson (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033370#comment-17033370
 ] 

Hudson commented on HADOOP-16739:
-

SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #17936 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/17936/])
HADOOP-16739. Fix native build failure of hadoop-pipes on CentOS 8. (iwasakims: 
rev d5467d299db3dddcbcd2f77a281d5fa82c4a9e4b)
* (edit) BUILDING.txt
* (edit) hadoop-tools/hadoop-pipes/src/CMakeLists.txt


> Fix native build failure of hadoop-pipes on CentOS 8
> 
>
> Key: HADOOP-16739
> URL: https://issues.apache.org/jira/browse/HADOOP-16739
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: tools/pipes
>Affects Versions: 2.10.0, 3.2.1
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Major
> Attachments: HADOOP-16739.001.patch
>
>
> Native build fails in hadoop-tools/hadoop-pips on CentOS 8 due to lack of 
> rpc.h which was removed from glibc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] aajisaka commented on issue #1832: HDFS-13989. RBF: Add FSCK to the Router

2020-02-09 Thread GitBox
aajisaka commented on issue #1832: HDFS-13989. RBF: Add FSCK to the Router
URL: https://github.com/apache/hadoop/pull/1832#issuecomment-583946905
 
 
   @goiri Thank you for your review. Fixed checkstyle issues.
   
   > Yetus does not look very happy but it seems unrelated.
   
   Agreed.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-16739) Fix native build failure of hadoop-pipes on CentOS 8

2020-02-09 Thread Masatake Iwasaki (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033361#comment-17033361
 ] 

Masatake Iwasaki commented on HADOOP-16739:
---

Thanks [~aajisaka]. I'm committing this.

> Fix native build failure of hadoop-pipes on CentOS 8
> 
>
> Key: HADOOP-16739
> URL: https://issues.apache.org/jira/browse/HADOOP-16739
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: tools/pipes
>Affects Versions: 2.10.0, 3.2.1
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Major
> Attachments: HADOOP-16739.001.patch
>
>
> Native build fails in hadoop-tools/hadoop-pips on CentOS 8 due to lack of 
> rpc.h which was removed from glibc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-16739) Fix native build failure of hadoop-pipes on CentOS 8

2020-02-09 Thread Akira Ajisaka (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16739?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1703#comment-1703
 ] 

Akira Ajisaka commented on HADOOP-16739:


+1, I could compile hadoop-pipes successfully with your patch.

> Fix native build failure of hadoop-pipes on CentOS 8
> 
>
> Key: HADOOP-16739
> URL: https://issues.apache.org/jira/browse/HADOOP-16739
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: tools/pipes
>Affects Versions: 2.10.0, 3.2.1
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Major
> Attachments: HADOOP-16739.001.patch
>
>
> Native build fails in hadoop-tools/hadoop-pips on CentOS 8 due to lack of 
> rpc.h which was removed from glibc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-16739) Fix native build failure of hadoop-pipes on CentOS 8

2020-02-09 Thread Masatake Iwasaki (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-16739?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HADOOP-16739:
--
Affects Version/s: 2.10.0
   3.2.1

> Fix native build failure of hadoop-pipes on CentOS 8
> 
>
> Key: HADOOP-16739
> URL: https://issues.apache.org/jira/browse/HADOOP-16739
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: tools/pipes
>Affects Versions: 2.10.0, 3.2.1
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Major
> Attachments: HADOOP-16739.001.patch
>
>
> Native build fails in hadoop-tools/hadoop-pips on CentOS 8 due to lack of 
> rpc.h which was removed from glibc.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] hadoop-yetus commented on issue #1839: HADOOP-16848. Refactoring: initial layering

2020-02-09 Thread GitBox
hadoop-yetus commented on issue #1839: HADOOP-16848. Refactoring: initial 
layering
URL: https://github.com/apache/hadoop/pull/1839#issuecomment-583900610
 
 
   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |::|--:|:|:|
   | +0 :ok: |  reexec  |   1m 30s |  Docker mode activated.  |
   ||| _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  No case conflicting files 
found.  |
   | +1 :green_heart: |  @author  |   0m  0s |  The patch does not contain any 
@author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  The patch appears to include 
1 new or modified test files.  |
   ||| _ trunk Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 29s |  Maven dependency ordering for branch  |
   | -1 :x: |  mvninstall  |   0m 26s |  root in trunk failed.  |
   | -1 :x: |  compile  |   0m 28s |  root in trunk failed.  |
   | -0 :warning: |  checkstyle  |   2m 58s |  The patch fails to run 
checkstyle in root  |
   | -1 :x: |  mvnsite  |   0m 59s |  hadoop-aws in trunk failed.  |
   | -1 :x: |  shadedclient  |  10m 39s |  branch has errors when building and 
testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   1m 31s |  trunk passed  |
   | +0 :ok: |  spotbugs  |   1m 15s |  Used deprecated FindBugs config; 
considering switching to SpotBugs.  |
   | +1 :green_heart: |  findbugs  |   4m  6s |  trunk passed  |
   ||| _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 54s |  Maven dependency ordering for patch  |
   | -1 :x: |  mvninstall  |   0m 29s |  hadoop-aws in the patch failed.  |
   | -1 :x: |  compile  |  16m 23s |  root in the patch failed.  |
   | -1 :x: |  javac  |  16m 23s |  root in the patch failed.  |
   | -0 :warning: |  checkstyle  |   2m 37s |  root: The patch generated 36 new 
+ 0 unchanged - 0 fixed = 36 total (was 0)  |
   | -1 :x: |  mvnsite  |   0m 47s |  hadoop-aws in the patch failed.  |
   | +1 :green_heart: |  whitespace  |   0m  0s |  The patch has no whitespace 
issues.  |
   | +1 :green_heart: |  shadedclient  |  14m 19s |  patch has no errors when 
building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   1m 43s |  the patch passed  |
   | -1 :x: |  findbugs  |   0m 47s |  hadoop-aws in the patch failed.  |
   ||| _ Other Tests _ |
   | -1 :x: |  unit  |   9m 37s |  hadoop-common in the patch passed.  |
   | -1 :x: |  unit  |   0m 47s |  hadoop-aws in the patch failed.  |
   | +1 :green_heart: |  asflicense  |   0m 52s |  The patch does not generate 
ASF License warnings.  |
   |  |   |  78m 12s |   |
   
   
   | Reason | Tests |
   |---:|:--|
   | Failed junit tests | hadoop.fs.viewfs.TestViewFsTrash |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | Client=19.03.5 Server=19.03.5 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/1839 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient findbugs checkstyle |
   | uname | Linux e7211e3ac278 4.15.0-60-generic #67-Ubuntu SMP Thu Aug 22 
16:55:30 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | personality/hadoop.sh |
   | git revision | trunk / 6191d4b |
   | Default Java | 1.8.0_242 |
   | mvninstall | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/branch-mvninstall-root.txt
 |
   | compile | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/branch-compile-root.txt
 |
   | checkstyle | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out//home/jenkins/jenkins-slave/workspace/hadoop-multibranch_PR-1839/out/maven-branch-checkstyle-root.txt
 |
   | mvnsite | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/branch-mvnsite-hadoop-tools_hadoop-aws.txt
 |
   | mvninstall | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/patch-mvninstall-hadoop-tools_hadoop-aws.txt
 |
   | compile | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/patch-compile-root.txt
 |
   | javac | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/patch-compile-root.txt
 |
   | checkstyle | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/diff-checkstyle-root.txt
 |
   | mvnsite | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/patch-mvnsite-hadoop-tools_hadoop-aws.txt
 |
   | findbugs | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/patch-findbugs-hadoop-tools_hadoop-aws.txt
 |
   | unit | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1839/1/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt
 |
   | unit | 

[GitHub] [hadoop] steveloughran commented on issue #1839: HADOOP-16848. Refactoring: initial layering

2020-02-09 Thread GitBox
steveloughran commented on issue #1839: HADOOP-16848. Refactoring: initial 
layering
URL: https://github.com/apache/hadoop/pull/1839#issuecomment-583892914
 
 
   FYI @bgaborg @sidseth 
   
   not quite ready for review, but you get the idea. Split up the S3AFileSystem 
class into layers with interface and impl, async chained init from S3A FS, 
which creates them, knows about DelegationTokens, and how to bind its 
operations to the store


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] steveloughran opened a new pull request #1839: HADOOP-16848. Refactoring: initial layering

2020-02-09 Thread GitBox
steveloughran opened a new pull request #1839: HADOOP-16848. Refactoring: 
initial layering
URL: https://github.com/apache/hadoop/pull/1839
 
 
   First PoC of My planned layout model of the S3A FS.
   
   * There's a raw layer and a guarded layer
   * which are instantiated in sequence in a separate executor from 
S3AFileSystem.initalize
   
   And whose accessors block until completed or rethrow failures.
   
   The layers are being handed in all their dependencies from FS.initialize()
   and we currently block until started.
   
   What I plan to do, in a future iteration, is:
   
   * each layer extracts their own settings from the config and stores locally
 (list version, upload size etc)
   * have each layer instantiate their internal classes (AWS S3 client, 
transfer manager) internally
   * Also async create: metastore, DT binding
   * And all startup actions (check bucket, init multipart, ...)
   
   Then
   * move ops to the layers, raw* -> rawStore; inner -> S3AStore
   * move WriteOperationHelper, SelectBinding, etc, to all work against 
S3AStore rather than FS.
   
   S3AStore will become where most of the code moves to; S3AFilesystem more of 
the init and binding to hadoop FS API.
   RawS3A will be the accessor through which all AWS client access goes.
   
   Not going to change: all accessors on S3AFileSystem...not just tests use it 
but some external code (cloudstore) needs it to get at low level S3A, etc.
   
   Change-Id: I998c0d61cce2ee7fd0be804bf21da6b68fd69a6f
   
   ## NOTICE
   
   Please create an issue in ASF JIRA before opening a pull request,
   and you need to set the title of the pull request which starts with
   the corresponding JIRA issue number. (e.g. HADOOP-X. Fix a typo in YYY.)
   For more details, please see 
https://cwiki.apache.org/confluence/display/HADOOP/How+To+Contribute
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-16848) add initial S3A layering + async init

2020-02-09 Thread Gabor Bota (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16848?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033292#comment-17033292
 ] 

Gabor Bota commented on HADOOP-16848:
-

Good idea, happy to do the review and help with the work.

> add initial S3A layering + async init
> -
>
> Key: HADOOP-16848
> URL: https://issues.apache.org/jira/browse/HADOOP-16848
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Affects Versions: 3.3.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>
> Split the S3A code into layers
> * S3AFileSystem
> * S3AStore + Impl
> * RawS3A + Impl
> S3AFS will create the others and start in order: DelegationTokens, RawS3A, 
> S3AStore, Metastore
> this will involve wrapping all access of DTs, s3client, Metastore to block 
> until that layer is complete, or raise an exception if instantiation of 
> it/predecessor failed.
> New layers will all be subclasses of Service, split into Interface and Impl, 
> so we can manage the init/start/stop lifecycle with existing code



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13447) Refactor S3AFileSystem to support introduction of separate metadata repository and tests.

2020-02-09 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-13447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-13447:

Parent: (was: HADOOP-16456)
Issue Type: Improvement  (was: Sub-task)

> Refactor S3AFileSystem to support introduction of separate metadata 
> repository and tests.
> -
>
> Key: HADOOP-13447
> URL: https://issues.apache.org/jira/browse/HADOOP-13447
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs/s3
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
>Priority: Major
> Fix For: 2.8.0, 3.0.0-alpha2
>
> Attachments: HADOOP-13447-HADOOP-13446.001.patch, 
> HADOOP-13447-HADOOP-13446.002.patch, HADOOP-13447.003.patch, 
> HADOOP-13447.004.patch, HADOOP-13447.005.patch
>
>
> The scope of this issue is to refactor the existing {{S3AFileSystem}} into 
> multiple coordinating classes.  The goal of this refactoring is to separate 
> the {{FileSystem}} API binding from the AWS SDK integration, make code 
> maintenance easier while we're making changes for S3Guard, and make it easier 
> to mock some implementation details so that tests can simulate eventual 
> consistency behavior in a deterministic way.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-16830) Add public IOStatistics API; S3A to collect and report across threads

2020-02-09 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-16830?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-16830:

Parent Issue: HADOOP-16456  (was: HADOOP-16829)

> Add public IOStatistics API; S3A to collect and report across threads
> -
>
> Key: HADOOP-16830
> URL: https://issues.apache.org/jira/browse/HADOOP-16830
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs, fs/s3
>Affects Versions: 3.3.0
>Reporter: Steve Loughran
>Assignee: Steve Loughran
>Priority: Major
>
> Applications like to collect the statistics which specific operations take, 
> by collecting exactly those operations done during the execution of FS API 
> calls by their individual worker threads, and returning these to their job 
> driver
> * S3A has a statistics API for some streams, but it's a non-standard one; 
> Impala  can't use it
> * FileSystem storage statistics are public, but as they aren't cross-thread, 
> they don't aggregate properly
> Proposed
> # A new IOStatistics interface to serve up statistics
> # S3A to implement
> # other stores to follow
> # Pass-through from the usual wrapper classes (FS data input/output streams)
> It's hard to think about how best to offer an API for operation context 
> stats, and how to actually implement.
> ThreadLocal isn't enough because the helper threads need to update on the 
> thread local value of the instigator
> My Initial PoC doesn't address that issue, but it shows what I'm thinking of



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Created] (HADOOP-16848) add initial S3A layering + async init

2020-02-09 Thread Steve Loughran (Jira)
Steve Loughran created HADOOP-16848:
---

 Summary: add initial S3A layering + async init
 Key: HADOOP-16848
 URL: https://issues.apache.org/jira/browse/HADOOP-16848
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.3.0
Reporter: Steve Loughran
Assignee: Steve Loughran


Split the S3A code into layers

* S3AFileSystem
* S3AStore + Impl
* RawS3A + Impl

S3AFS will create the others and start in order: DelegationTokens, RawS3A, 
S3AStore, Metastore

this will involve wrapping all access of DTs, s3client, Metastore to block 
until that layer is complete, or raise an exception if instantiation of 
it/predecessor failed.

New layers will all be subclasses of Service, split into Interface and Impl, so 
we can manage the init/start/stop lifecycle with existing code



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-13447) Refactor S3AFileSystem to support introduction of separate metadata repository and tests.

2020-02-09 Thread Steve Loughran (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-13447?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Steve Loughran updated HADOOP-13447:

Parent Issue: HADOOP-16456  (was: HADOOP-13345)

> Refactor S3AFileSystem to support introduction of separate metadata 
> repository and tests.
> -
>
> Key: HADOOP-13447
> URL: https://issues.apache.org/jira/browse/HADOOP-13447
> Project: Hadoop Common
>  Issue Type: Sub-task
>  Components: fs/s3
>Reporter: Chris Nauroth
>Assignee: Chris Nauroth
>Priority: Major
> Fix For: 2.8.0, 3.0.0-alpha2
>
> Attachments: HADOOP-13447-HADOOP-13446.001.patch, 
> HADOOP-13447-HADOOP-13446.002.patch, HADOOP-13447.003.patch, 
> HADOOP-13447.004.patch, HADOOP-13447.005.patch
>
>
> The scope of this issue is to refactor the existing {{S3AFileSystem}} into 
> multiple coordinating classes.  The goal of this refactoring is to separate 
> the {{FileSystem}} API binding from the AWS SDK integration, make code 
> maintenance easier while we're making changes for S3Guard, and make it easier 
> to mock some implementation details so that tests can simulate eventual 
> consistency behavior in a deterministic way.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-16838) Support for `fs.s3a.endpoint.region`

2020-02-09 Thread Steve Loughran (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16838?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=17033237#comment-17033237
 ] 

Steve Loughran commented on HADOOP-16838:
-

oh, and if you are wondering why we expect hostnames for endpoint and not  
region names is (a) to support third party stores like minio and (b) so when 
AWS add a new region there's no need to update the SDKs on every shipping 
hadoop version, retest, patch where needed and then rerelease

> Support for `fs.s3a.endpoint.region`
> 
>
> Key: HADOOP-16838
> URL: https://issues.apache.org/jira/browse/HADOOP-16838
> Project: Hadoop Common
>  Issue Type: New Feature
>Reporter: Nitish
>Priority: Major
>
> Currently it is not possible to connect S3 Compatible services like MinIO, 
> Ceph, etc (running with a custom region) to Spark with s3a connector. For 
> example, if MinIO is running on a Server with
>  * IP Address: 192.168.0.100
>  * Region: ap-southeast-1
> The s3a connector can't be configured to use the region `ap-southeast-1`. 
> It would be great to have a configuration field like 
> `fs.s3a.endpoint.region`. This will be very helpful for users deploying 
> Private Cloud and who intend to use S3 like services on premises.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] hadoop-yetus commented on issue #1838: HADOOP-16711 Add way to skip verifyBuckets check in S3A fs init()

2020-02-09 Thread GitBox
hadoop-yetus commented on issue #1838: HADOOP-16711 Add way to skip 
verifyBuckets check in S3A fs init()
URL: https://github.com/apache/hadoop/pull/1838#issuecomment-583830785
 
 
   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |::|--:|:|:|
   | +0 :ok: |  reexec  |   1m 58s |  Docker mode activated.  |
   ||| _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  0s |  No case conflicting files 
found.  |
   | +0 :ok: |  markdownlint  |   0m  0s |  markdownlint was not available.  |
   | +1 :green_heart: |  @author  |   0m  0s |  The patch does not contain any 
@author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  The patch appears to include 
3 new or modified test files.  |
   ||| _ trunk Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |  27m  6s |  trunk passed  |
   | +1 :green_heart: |  compile  |   0m 38s |  trunk passed  |
   | +1 :green_heart: |  checkstyle  |   0m 33s |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   0m 47s |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  20m  8s |  branch has no errors when 
building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   0m 35s |  trunk passed  |
   | +0 :ok: |  spotbugs  |   1m 22s |  Used deprecated FindBugs config; 
considering switching to SpotBugs.  |
   | +1 :green_heart: |  findbugs  |   1m 15s |  trunk passed  |
   ||| _ Patch Compile Tests _ |
   | +1 :green_heart: |  mvninstall  |   0m 39s |  the patch passed  |
   | +1 :green_heart: |  compile  |   0m 34s |  the patch passed  |
   | +1 :green_heart: |  javac  |   0m 34s |  the patch passed  |
   | -0 :warning: |  checkstyle  |   0m 26s |  hadoop-tools/hadoop-aws: The 
patch generated 22 new + 15 unchanged - 0 fixed = 37 total (was 15)  |
   | +1 :green_heart: |  mvnsite  |   0m 42s |  the patch passed  |
   | -1 :x: |  whitespace  |   0m  0s |  The patch has 5 line(s) that end in 
whitespace. Use git apply --whitespace=fix <>. Refer 
https://git-scm.com/docs/git-apply  |
   | +1 :green_heart: |  shadedclient  |  17m 52s |  patch has no errors when 
building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   0m 29s |  the patch passed  |
   | -1 :x: |  findbugs  |   1m 22s |  hadoop-tools/hadoop-aws generated 1 new 
+ 0 unchanged - 0 fixed = 1 total (was 0)  |
   ||| _ Other Tests _ |
   | +1 :green_heart: |  unit  |   1m 34s |  hadoop-aws in the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 36s |  The patch does not generate 
ASF License warnings.  |
   |  |   |  78m 33s |   |
   
   
   | Reason | Tests |
   |---:|:--|
   | FindBugs | module:hadoop-tools/hadoop-aws |
   |  |  Switch statement found in 
org.apache.hadoop.fs.s3a.S3AFileSystem.doBucketProbing() where default case is 
missing  At S3AFileSystem.java:where default case is missing  At 
S3AFileSystem.java:[lines 463-470] |
   
   
   | Subsystem | Report/Notes |
   |--:|:-|
   | Docker | Client=19.03.5 Server=19.03.5 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1838/1/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/1838 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient findbugs checkstyle markdownlint |
   | uname | Linux d8622c0ac2b8 4.15.0-74-generic #84-Ubuntu SMP Thu Dec 19 
08:06:28 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | personality/hadoop.sh |
   | git revision | trunk / d23317b |
   | Default Java | 1.8.0_232 |
   | checkstyle | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1838/1/artifact/out/diff-checkstyle-hadoop-tools_hadoop-aws.txt
 |
   | whitespace | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1838/1/artifact/out/whitespace-eol.txt
 |
   | findbugs | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1838/1/artifact/out/new-findbugs-hadoop-tools_hadoop-aws.html
 |
   |  Test Results | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1838/1/testReport/ |
   | Max. process+thread count | 342 (vs. ulimit of 5500) |
   | modules | C: hadoop-tools/hadoop-aws U: hadoop-tools/hadoop-aws |
   | Console output | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1838/1/console |
   | versions | git=2.7.4 maven=3.3.9 findbugs=3.1.0-RC1 |
   | Powered by | Apache Yetus 0.11.1 https://yetus.apache.org |
   
   
   This message was automatically generated.
   
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: 

[GitHub] [hadoop] mukund-thakur opened a new pull request #1838: HADOOP-16711 Add way to skip verifyBuckets check in S3A fs init()

2020-02-09 Thread GitBox
mukund-thakur opened a new pull request #1838: HADOOP-16711 Add way to skip 
verifyBuckets check in S3A fs init()
URL: https://github.com/apache/hadoop/pull/1838
 
 
   Testing Bucket used : https://mthakur-data.s3.ap-south-1.amazonaws.com/file2
   Ran all the UT's and IT's using default settings with S3guard.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org