[GitHub] [hadoop] lqjack edited a comment on issue #362: YARN-8057 Inadequate information for handling catch clauses

2019-09-08 Thread GitBox
lqjack edited a comment on issue #362: YARN-8057 Inadequate information for 
handling catch clauses
URL: https://github.com/apache/hadoop/pull/362#issuecomment-529286711
 
 
   > Hey Iq you have been in my phone a lot and who ever you been on here with 
is hacking my phone so tell me what an WHY u messing with my phone
   
   I do nothing for your phone, however I will also receive the apple id is 
binding by other person now . 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bharatviswa504 edited a comment on issue #1377: HDDS-2057. Incorrect Default OM Port in Ozone FS URI Error Message. Contributed by Supratim Deka

2019-09-08 Thread GitBox
bharatviswa504 edited a comment on issue #1377: HDDS-2057. Incorrect Default OM 
Port in Ozone FS URI Error Message. Contributed by Supratim Deka
URL: https://github.com/apache/hadoop/pull/1377#issuecomment-529292657
 
 
   Looks like acceptance test failures are related to this, can you once verify 
them, as I have not seen them failing in recent CI runs?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bharatviswa504 commented on issue #1377: HDDS-2057. Incorrect Default OM Port in Ozone FS URI Error Message. Contributed by Supratim Deka

2019-09-08 Thread GitBox
bharatviswa504 commented on issue #1377: HDDS-2057. Incorrect Default OM Port 
in Ozone FS URI Error Message. Contributed by Supratim Deka
URL: https://github.com/apache/hadoop/pull/1377#issuecomment-529292657
 
 
   Looks like acceptance test failures are related to this, can you once verify 
them?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] lzh3636 closed pull request #362: YARN-8057 Inadequate information for handling catch clauses

2019-09-08 Thread GitBox
lzh3636 closed pull request #362: YARN-8057 Inadequate information for handling 
catch clauses
URL: https://github.com/apache/hadoop/pull/362
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bharatviswa504 edited a comment on issue #1411: HDDS-2098 : Ozone shell command prints out ERROR when the log4j file …

2019-09-08 Thread GitBox
bharatviswa504 edited a comment on issue #1411: HDDS-2098 : Ozone shell command 
prints out ERROR when the log4j file …
URL: https://github.com/apache/hadoop/pull/1411#issuecomment-529291245
 
 
   I have a question
   During ozone tarball build, we do copy ozone-shell-log4j.properties to 
etc/hadoop (like we copy log4.properties then why do we see this error or 
something need to be fixed in copying this script?
   
   
https://github.com/apache/hadoop/blob/trunk/hadoop-ozone/dist/dev-support/bin/dist-layout-stitching#L95
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bharatviswa504 commented on issue #1411: HDDS-2098 : Ozone shell command prints out ERROR when the log4j file …

2019-09-08 Thread GitBox
bharatviswa504 commented on issue #1411: HDDS-2098 : Ozone shell command prints 
out ERROR when the log4j file …
URL: https://github.com/apache/hadoop/pull/1411#issuecomment-529291245
 
 
   I have a question
   During ozone tarball build, we do copy ozone-shell-log4j.properties to 
etc/hadoop (like we copy log4.properties then why do we see this error?


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-16551) The changelog*.md seems not generated when create-release

2019-09-08 Thread Zhankun Tang (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-16551?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925330#comment-16925330
 ] 

Zhankun Tang commented on HADOOP-16551:
---

Could you please take a look at this? [~aajisaka]

> The changelog*.md seems not generated when create-release
> -
>
> Key: HADOOP-16551
> URL: https://issues.apache.org/jira/browse/HADOOP-16551
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Zhankun Tang
>Priority: Blocker
>
> Hi,
>  When creating Hadoop 3.1.3 release with "create-release" script, after the 
> mvn site succeeded. But it complains about this and failed:
> {code:java}
> dev-support/bin/create-release --asfrelease --docker --dockercache{code}
> {code:java}
> $ cd /build/source
> $ mv /build/source/target/hadoop-site-3.1.3.tar.gz 
> /build/source/target/artifacts/hadoop-3.1.3-site.tar.gz
> $ cp -p 
> /build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md
>  /build/source/target/artifacts/CHANGES.md
> cp: cannot stat 
> '/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md':
>  No such file or directory
> {code}
> And there's no 3.1.3 release site markdown folder.
> {code:java}
> [ztang@release-vm hadoop]$ ls 
> hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3
> ls: cannot access 
> hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3: No such 
> file or directory
> {code}
> I've checked the HADOOP-14671 but have no idea why this changelog is missing.
> *Update:*
>  Found that the CHANGELOG.md and RELEASENOTES.md are generated but not in 
> directory "3.1.3"
> {code:java}
> [ztang@release-vm hadoop]$ ls 
> hadoop-common-project/hadoop-common/src/site/markdown/release/
> 0.1.0 0.15.2 0.19.2 0.23.2 0.7.2 2.0.1-alpha 2.6.3 3.0.0-alpha3
> 0.10.0 0.15.3 0.2.0 0.23.3 0.8.0 2.0.2-alpha 2.6.4 3.0.0-alpha4
> 0.10.1 0.15.4 0.20.0 0.23.4 0.9.0 2.0.3-alpha 2.6.5 3.0.0-beta1
> 0.1.1 0.16.0 0.20.1 0.23.5 0.9.1 2.0.4-alpha 2.6.6 3.0.1
> 0.11.0 0.16.1 0.20.2 0.23.6 0.9.2 2.0.5-alpha 2.7.0 3.0.3
> 0.11.1 0.16.2 0.20.203.0 0.23.7 1.0.0 2.0.6-alpha 2.7.1 3.1.0
> 0.11.2 0.16.3 0.20.203.1 0.23.8 1.0.1 2.1.0-beta 2.7.2 3.1.1
> 0.12.0 0.16.4 0.20.204.0 0.23.9 1.0.2 2.1.1-beta 2.7.3 3.1.2
> 0.12.1 0.17.0 0.20.205.0 0.24.0 1.0.3 2.2.0 2.7.4 CHANGELOG.md
> 0.12.2 0.17.1 0.20.3 0.3.0 1.0.4 2.2.1 2.7.5 index.md
> 0.12.3 0.17.2 0.2.1 0.3.1 1.1.0 2.3.0 2.8.0 README.md
> 0.13.0 0.17.3 0.21.0 0.3.2 1.1.1 2.4.0 2.8.1 RELEASENOTES.md
> 0.14.0 0.18.0 0.21.1 0.4.0 1.1.2 2.4.1 2.8.2
> 0.14.1 0.18.1 0.22.0 0.5.0 1.1.3 2.5.0 2.8.3
> 0.14.2 0.18.2 0.22.1 0.6.0 1.2.0 2.5.1 2.9.0
> 0.14.3 0.18.3 0.23.0 0.6.1 1.2.1 2.5.2 2.9.1
> 0.14.4 0.18.4 0.23.1 0.6.2 1.2.2 2.6.0 3.0.0
> 0.15.0 0.19.0 0.23.10 0.7.0 1.3.0 2.6.1 3.0.0-alpha1
> 0.15.1 0.19.1 0.23.11 0.7.1 2.0.0-alpha 2.6.2 3.0.0-alpha2{code}



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bharatviswa504 merged pull request #1409: HDDS-2087. Remove the hard coded config key in ChunkManager

2019-09-08 Thread GitBox
bharatviswa504 merged pull request #1409: HDDS-2087. Remove the hard coded 
config key in ChunkManager
URL: https://github.com/apache/hadoop/pull/1409
 
 
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bharatviswa504 commented on issue #1409: HDDS-2087. Remove the hard coded config key in ChunkManager

2019-09-08 Thread GitBox
bharatviswa504 commented on issue #1409: HDDS-2087. Remove the hard coded 
config key in ChunkManager
URL: https://github.com/apache/hadoop/pull/1409#issuecomment-529288559
 
 
   Thank You @vivekratnavel for the contribution and @anuengineer for the 
review.
   I have committed this to the trunk.


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-16551) The changelog*.md seems not generated when create-release

2019-09-08 Thread Zhankun Tang (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-16551?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhankun Tang updated HADOOP-16551:
--
Description: 
Hi,
 When creating Hadoop 3.1.3 release with "create-release" script, after the mvn 
site succeeded. But it complains about this and failed:
{code:java}
dev-support/bin/create-release --asfrelease --docker --dockercache{code}
{code:java}
$ cd /build/source
$ mv /build/source/target/hadoop-site-3.1.3.tar.gz 
/build/source/target/artifacts/hadoop-3.1.3-site.tar.gz
$ cp -p 
/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md
 /build/source/target/artifacts/CHANGES.md
cp: cannot stat 
'/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md':
 No such file or directory
{code}
And there's no 3.1.3 release site markdown folder.
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3
ls: cannot access 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3: No such 
file or directory

{code}
I've checked the HADOOP-14671 but have no idea why this changelog is missing.

*Update:*
 Found that the CHANGELOG.md and RELEASENOTES.md are generated but not in 
directory "3.1.3"
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/
0.1.0 0.15.2 0.19.2 0.23.2 0.7.2 2.0.1-alpha 2.6.3 3.0.0-alpha3
0.10.0 0.15.3 0.2.0 0.23.3 0.8.0 2.0.2-alpha 2.6.4 3.0.0-alpha4
0.10.1 0.15.4 0.20.0 0.23.4 0.9.0 2.0.3-alpha 2.6.5 3.0.0-beta1
0.1.1 0.16.0 0.20.1 0.23.5 0.9.1 2.0.4-alpha 2.6.6 3.0.1
0.11.0 0.16.1 0.20.2 0.23.6 0.9.2 2.0.5-alpha 2.7.0 3.0.3
0.11.1 0.16.2 0.20.203.0 0.23.7 1.0.0 2.0.6-alpha 2.7.1 3.1.0
0.11.2 0.16.3 0.20.203.1 0.23.8 1.0.1 2.1.0-beta 2.7.2 3.1.1
0.12.0 0.16.4 0.20.204.0 0.23.9 1.0.2 2.1.1-beta 2.7.3 3.1.2
0.12.1 0.17.0 0.20.205.0 0.24.0 1.0.3 2.2.0 2.7.4 CHANGELOG.md
0.12.2 0.17.1 0.20.3 0.3.0 1.0.4 2.2.1 2.7.5 index.md
0.12.3 0.17.2 0.2.1 0.3.1 1.1.0 2.3.0 2.8.0 README.md
0.13.0 0.17.3 0.21.0 0.3.2 1.1.1 2.4.0 2.8.1 RELEASENOTES.md
0.14.0 0.18.0 0.21.1 0.4.0 1.1.2 2.4.1 2.8.2
0.14.1 0.18.1 0.22.0 0.5.0 1.1.3 2.5.0 2.8.3
0.14.2 0.18.2 0.22.1 0.6.0 1.2.0 2.5.1 2.9.0
0.14.3 0.18.3 0.23.0 0.6.1 1.2.1 2.5.2 2.9.1
0.14.4 0.18.4 0.23.1 0.6.2 1.2.2 2.6.0 3.0.0
0.15.0 0.19.0 0.23.10 0.7.0 1.3.0 2.6.1 3.0.0-alpha1
0.15.1 0.19.1 0.23.11 0.7.1 2.0.0-alpha 2.6.2 3.0.0-alpha2{code}

  was:
Hi,
 When creating Hadoop 3.1.3 release with "create-release" script, after the mvn 
site succeeded. But it complains about this and failed:
{code:java}
dev-support/bin/create-release --asfrelease --docker --dockercache{code}
{code:java}
$ cd /build/source
$ mv /build/source/target/hadoop-site-3.1.3.tar.gz 
/build/source/target/artifacts/hadoop-3.1.3-site.tar.gz
$ cp -p 
/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md
 /build/source/target/artifacts/CHANGES.md
cp: cannot stat 
'/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md':
 No such file or directory
{code}
And there's no 3.1.3 release site markdown folder.
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3
ls: cannot access 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3: No such 
file or directory

{code}
I've checked the HADOOP-14671 but have no idea why this changelog is missing.

Update:
Found that the CHANGELOG.md and RELEASENOTES.md are generated but not in 
directory "3.1.3"
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/
0.1.0 0.15.2 0.19.2 0.23.2 0.7.2 2.0.1-alpha 2.6.3 3.0.0-alpha3
0.10.0 0.15.3 0.2.0 0.23.3 0.8.0 2.0.2-alpha 2.6.4 3.0.0-alpha4
0.10.1 0.15.4 0.20.0 0.23.4 0.9.0 2.0.3-alpha 2.6.5 3.0.0-beta1
0.1.1 0.16.0 0.20.1 0.23.5 0.9.1 2.0.4-alpha 2.6.6 3.0.1
0.11.0 0.16.1 0.20.2 0.23.6 0.9.2 2.0.5-alpha 2.7.0 3.0.3
0.11.1 0.16.2 0.20.203.0 0.23.7 1.0.0 2.0.6-alpha 2.7.1 3.1.0
0.11.2 0.16.3 0.20.203.1 0.23.8 1.0.1 2.1.0-beta 2.7.2 3.1.1
0.12.0 0.16.4 0.20.204.0 0.23.9 1.0.2 2.1.1-beta 2.7.3 3.1.2
0.12.1 0.17.0 0.20.205.0 0.24.0 1.0.3 2.2.0 2.7.4 CHANGELOG.md
0.12.2 0.17.1 0.20.3 0.3.0 1.0.4 2.2.1 2.7.5 index.md
0.12.3 0.17.2 0.2.1 0.3.1 1.1.0 2.3.0 2.8.0 README.md
0.13.0 0.17.3 0.21.0 0.3.2 1.1.1 2.4.0 2.8.1 RELEASENOTES.md
0.14.0 0.18.0 0.21.1 0.4.0 1.1.2 2.4.1 2.8.2
0.14.1 0.18.1 0.22.0 0.5.0 1.1.3 2.5.0 2.8.3
0.14.2 0.18.2 0.22.1 0.6.0 1.2.0 2.5.1 2.9.0
0.14.3 0.18.3 0.23.0 0.6.1 1.2.1 2.5.2 2.9.1
0.14.4 0.18.4 0.23.1 0.6.2 1.2.2 2.6.0 3.0.0
0.15.0 0.19.0 0.23.10 0.7.0 1.3.0 2.6.1 3.0.0-alpha1
0.15.1 0.19.1 0.23.11 0.7.1 2.0.0-alpha 2.6.2 3.0.0-alpha2{code}


> The changelog*.md seems not generated when create-release
> -
>
> Key: HADOOP-16551
> 

[jira] [Updated] (HADOOP-16551) The changelog*.md seems not generated when create-release

2019-09-08 Thread Zhankun Tang (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-16551?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhankun Tang updated HADOOP-16551:
--
Description: 
Hi,
 When creating Hadoop 3.1.3 release with "create-release" script, after the mvn 
site succeeded. But it complains about this and failed:
{code:java}
dev-support/bin/create-release --asfrelease --docker --dockercache{code}
{code:java}
$ cd /build/source
$ mv /build/source/target/hadoop-site-3.1.3.tar.gz 
/build/source/target/artifacts/hadoop-3.1.3-site.tar.gz
$ cp -p 
/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md
 /build/source/target/artifacts/CHANGES.md
cp: cannot stat 
'/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md':
 No such file or directory
{code}
And there's no 3.1.3 release site markdown folder.
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3
ls: cannot access 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3: No such 
file or directory

{code}
I've checked the HADOOP-14671 but have no idea why this changelog is missing.

Update:
Found that the CHANGELOG.md and RELEASENOTES.md are generated but not in 
directory "3.1.3"
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/
0.1.0 0.15.2 0.19.2 0.23.2 0.7.2 2.0.1-alpha 2.6.3 3.0.0-alpha3
0.10.0 0.15.3 0.2.0 0.23.3 0.8.0 2.0.2-alpha 2.6.4 3.0.0-alpha4
0.10.1 0.15.4 0.20.0 0.23.4 0.9.0 2.0.3-alpha 2.6.5 3.0.0-beta1
0.1.1 0.16.0 0.20.1 0.23.5 0.9.1 2.0.4-alpha 2.6.6 3.0.1
0.11.0 0.16.1 0.20.2 0.23.6 0.9.2 2.0.5-alpha 2.7.0 3.0.3
0.11.1 0.16.2 0.20.203.0 0.23.7 1.0.0 2.0.6-alpha 2.7.1 3.1.0
0.11.2 0.16.3 0.20.203.1 0.23.8 1.0.1 2.1.0-beta 2.7.2 3.1.1
0.12.0 0.16.4 0.20.204.0 0.23.9 1.0.2 2.1.1-beta 2.7.3 3.1.2
0.12.1 0.17.0 0.20.205.0 0.24.0 1.0.3 2.2.0 2.7.4 CHANGELOG.md
0.12.2 0.17.1 0.20.3 0.3.0 1.0.4 2.2.1 2.7.5 index.md
0.12.3 0.17.2 0.2.1 0.3.1 1.1.0 2.3.0 2.8.0 README.md
0.13.0 0.17.3 0.21.0 0.3.2 1.1.1 2.4.0 2.8.1 RELEASENOTES.md
0.14.0 0.18.0 0.21.1 0.4.0 1.1.2 2.4.1 2.8.2
0.14.1 0.18.1 0.22.0 0.5.0 1.1.3 2.5.0 2.8.3
0.14.2 0.18.2 0.22.1 0.6.0 1.2.0 2.5.1 2.9.0
0.14.3 0.18.3 0.23.0 0.6.1 1.2.1 2.5.2 2.9.1
0.14.4 0.18.4 0.23.1 0.6.2 1.2.2 2.6.0 3.0.0
0.15.0 0.19.0 0.23.10 0.7.0 1.3.0 2.6.1 3.0.0-alpha1
0.15.1 0.19.1 0.23.11 0.7.1 2.0.0-alpha 2.6.2 3.0.0-alpha2{code}

  was:
Hi,
 When creating Hadoop 3.1.3 release with "create-release" script, after the mvn 
site succeeded. But it complains about this and failed:

{code:java}
dev-support/bin/create-release --asfrelease --docker --dockercache{code}
{code:java}
$ cd /build/source
$ mv /build/source/target/hadoop-site-3.1.3.tar.gz 
/build/source/target/artifacts/hadoop-3.1.3-site.tar.gz
$ cp -p 
/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md
 /build/source/target/artifacts/CHANGES.md
cp: cannot stat 
'/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md':
 No such file or directory
{code}

And there's no 3.1.3 release site markdown folder.
{code:java}
[ztang@release-vm hadoop]$ ls 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3
ls: cannot access 
hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3: No such 
file or directory

{code}
I've checked the HADOOP-14671 but have no idea why this changelog is not 
generated.


> The changelog*.md seems not generated when create-release
> -
>
> Key: HADOOP-16551
> URL: https://issues.apache.org/jira/browse/HADOOP-16551
> Project: Hadoop Common
>  Issue Type: Task
>Reporter: Zhankun Tang
>Priority: Blocker
>
> Hi,
>  When creating Hadoop 3.1.3 release with "create-release" script, after the 
> mvn site succeeded. But it complains about this and failed:
> {code:java}
> dev-support/bin/create-release --asfrelease --docker --dockercache{code}
> {code:java}
> $ cd /build/source
> $ mv /build/source/target/hadoop-site-3.1.3.tar.gz 
> /build/source/target/artifacts/hadoop-3.1.3-site.tar.gz
> $ cp -p 
> /build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md
>  /build/source/target/artifacts/CHANGES.md
> cp: cannot stat 
> '/build/source/hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3/CHANGES*.md':
>  No such file or directory
> {code}
> And there's no 3.1.3 release site markdown folder.
> {code:java}
> [ztang@release-vm hadoop]$ ls 
> hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3
> ls: cannot access 
> hadoop-common-project/hadoop-common/src/site/markdown/release/3.1.3: No such 
> file or directory
> {code}
> I've checked the HADOOP-14671 but have no idea why this changelog is missing.
> 

[GitHub] [hadoop] lqjack commented on issue #362: YARN-8057 Inadequate information for handling catch clauses

2019-09-08 Thread GitBox
lqjack commented on issue #362: YARN-8057 Inadequate information for handling 
catch clauses
URL: https://github.com/apache/hadoop/pull/362#issuecomment-529286711
 
 
   > Hey Iq you have been in my phone a lot and who ever you been on here with 
is hacking my phone so tell me what an WHY u messing with my phone
   
   I do thing for your phone, however I will also receive the apple id is 
binding by other person now . 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] lqjack commented on issue #1037: HADOOP-15847 limit the r/w capacity

2019-09-08 Thread GitBox
lqjack commented on issue #1037: HADOOP-15847 limit the r/w capacity 
URL: https://github.com/apache/hadoop/pull/1037#issuecomment-529286303
 
 
   @steveloughran 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] lqjack commented on issue #1037: HADOOP-15847 limit the r/w capacity

2019-09-08 Thread GitBox
lqjack commented on issue #1037: HADOOP-15847 limit the r/w capacity 
URL: https://github.com/apache/hadoop/pull/1037#issuecomment-529286270
 
 
   Now I found that the diff is nothing,  I could not to change the email in 
the commit settings. 


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15169) "hadoop.ssl.enabled.protocols" should be considered in httpserver2

2019-09-08 Thread Wei-Chiu Chuang (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-15169?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925304#comment-16925304
 ] 

Wei-Chiu Chuang commented on HADOOP-15169:
--

[~aajisaka] I understand your concern. However, this is merely to achieve 
consistency with other Hadoop components. We've got customers with legacy tools 
that can only support SSLv2Hello, and they aren't able to use it after 
upgrading to Hadoop 3.

 

[~brahmareddy] thanks for the patch. have you tested it? Looking at Jetty's 
SslContextFactory implementation (SslContextFactory#selectProtocols()), after 
included protocols are added, it removes excluded protocols, which contains 

"SSL", "SSLv2", "SSLv2Hello", "SSLv3". I suspect we should reset excluded 
protocols before adding included protocols.

> "hadoop.ssl.enabled.protocols" should be considered in httpserver2
> --
>
> Key: HADOOP-15169
> URL: https://issues.apache.org/jira/browse/HADOOP-15169
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Reporter: Brahma Reddy Battula
>Assignee: Brahma Reddy Battula
>Priority: Major
> Attachments: HADOOP-15169-branch-2.patch, HADOOP-15169.patch
>
>
> As of now *hadoop.ssl.enabled.protocols"* will not take effect for all the 
> http servers( only Datanodehttp server will use this config).



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Updated] (HADOOP-15922) DelegationTokenAuthenticationFilter get wrong doAsUser since it does not decode URL

2019-09-08 Thread Eric Yang (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-15922?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Eric Yang updated HADOOP-15922:
---
Release Note: - Fix DelegationTokenAuthentication filter for incorrectly 
double encode doAs user parameter.

> DelegationTokenAuthenticationFilter get wrong doAsUser since it does not 
> decode URL
> ---
>
> Key: HADOOP-15922
> URL: https://issues.apache.org/jira/browse/HADOOP-15922
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common, kms
>Reporter: He Xiaoqiao
>Assignee: He Xiaoqiao
>Priority: Major
> Fix For: 3.3.0, 3.2.1, 3.1.3
>
> Attachments: HADOOP-15922.001.patch, HADOOP-15922.002.patch, 
> HADOOP-15922.003.patch, HADOOP-15922.004.patch, HADOOP-15922.005.patch, 
> HADOOP-15922.006.patch, HADOOP-15922.007.patch
>
>
> DelegationTokenAuthenticationFilter get wrong doAsUser when proxy user from 
> client is complete kerberos name (e.g., user/hostn...@realm.com, actually it 
> is acceptable), because DelegationTokenAuthenticationFilter does not decode 
> DOAS parameter in URL which is encoded by {{URLEncoder}} at client.
> e.g. KMS as example:
> a. KMSClientProvider creates connection to KMS Server using 
> DelegationTokenAuthenticatedURL#openConnection.
> b. If KMSClientProvider is a doAsUser, KMSClientProvider will put {{doas}} 
> with url encoded user as one parameter of http request. 
> {code:java}
> // proxyuser
> if (doAs != null) {
>   extraParams.put(DO_AS, URLEncoder.encode(doAs, "UTF-8"));
> }
> {code}
> c. when KMS server receives the request, it does not decode the proxy user.
> As result, KMS Server will get the wrong proxy user if this proxy user is 
> complete Kerberos Name or it includes some special character. Some other 
> authentication and authorization exception will throws next to it.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-13363) Upgrade protobuf from 2.5.0 to something newer

2019-09-08 Thread Vinayakumar B (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-13363?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925234#comment-16925234
 ] 

Vinayakumar B commented on HADOOP-13363:


Hi [~anu] and [~stack].

Apologies: Long reply ahead!!!

I looked at both HBase and Ozone's way of shading and relocating the generated 
protobuf classes.

Technically, the latest PR does the samething. Features of both plugins (1. 
downloading appropriate protoc exe, 2. replacing package with shaded prefix in 
generated classes) is merged to hadoop-maven-plugin itself. regex to relocate 
is hard coded only for protobuf, since the goal is specific to protoc.

Anyway I am open to replace the "hadoop-maven-plugin" with 
"protoc-maven-plugin" and "replacer" plugins.

 

Problem is not in the Generated classes, but the dependency with shaded 
protobuf classes.

In HBase and Ozone, shaded dependency is maintained in a separate repo. 
[hbase-thirdparty|https://github.com/apache/hbase-thirdparty]   for hbase and  
[incubator-ratis-thirdparty|https://github.com/apache/incubator-ratis-thirdparty]
 for ozone, its not maintained as a submodule.

 

Above compilation Problem is with only "mvn compile".

"mvn compile" will not download or refer to local repo for submodules from mvn 
execution. It just looks for the compile classes.

But "mvn package" also will not refer to local repo for submodules, but it 
builds and uses compiled jars for dependency. Here, shaded classes will be 
included as well.

So current problem is with only "mvn compile" at root level.

 

To verify this, I have manually built *hadoop-shaded-thirdparty-dependency* and 
deployed to central snapshot repo and excluded this module from normal build by 
moving into separate profile in Latest commit.
 This hack seems to have passed the compile phase in yetus test. But jenkins 
itself timedout after 5hours, due to unit testing of all modules. 
 This is the latest build for PR with 
[https://builds.apache.org/job/hadoop-multibranch/job/PR-1408/5/]

There are many javac warnings due to new protobuf-3.6.1 dependency due to 
deprecated APIs usage.

 

To Reduce the jenkins time and to reduce huge changes together, following may 
be the options.
 # (Current way) Combine protobuf dependency upgrade + shading in a single PR  
to reduce the impact for the other developments, but need to allow yetus to 
take its own time for completion of tests. 
 #  
 ** Upgrade protobuf dependency to 3.6.1 in hadoop-project/pom.xml (version 
bump + compile fix)  in a separate Jira
 ** Create separate module/repo to shade the dependencies. (com.google.protobuf)
 ** Relocate usage of protobuf module-by-module in module specific Jiras.
 ** Bring back protobuf dependency to 2.5.0 in hadoop-project/pom.xml for 
downstreams.
 #  Do entire step #2 in a separate branch (just avoid impact on current 
development) and merge back immediately to trunk once done.

Here I am fine with #2 or even #3. 

 

 Decision needs to be taken whether to keep 
*hadoop-shaded-thirdparty-dependency* as a separate repo (like in hbase and 
ozone/ratis) or separate sub-module within hadoop.
 # Separate repo: comes with its own maintainance cost.
 # Submodule-activated-on-profile, less maintainance as its within same repo, 
but requires build steps changes. i.e. This module shoud be installed/packaged 
first before otrhers. Also need to change yetus script.

Here I am leaning towards #2.

Any thoughts?

> Upgrade protobuf from 2.5.0 to something newer
> --
>
> Key: HADOOP-13363
> URL: https://issues.apache.org/jira/browse/HADOOP-13363
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: build
>Affects Versions: 3.0.0-alpha1, 3.0.0-alpha2
>Reporter: Allen Wittenauer
>Assignee: Vinayakumar B
>Priority: Major
>  Labels: security
> Attachments: HADOOP-13363.001.patch, HADOOP-13363.002.patch, 
> HADOOP-13363.003.patch, HADOOP-13363.004.patch, HADOOP-13363.005.patch
>
>
> Standard protobuf 2.5.0 does not work properly on many platforms.  (See, for 
> example, https://gist.github.com/BennettSmith/7111094 ).  In order for us to 
> avoid crazy work arounds in the build environment and the fact that 2.5.0 is 
> starting to slowly disappear as a standard install-able package for even 
> Linux/x86, we need to either upgrade or self bundle or something else.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] bgaborg commented on issue #1302: HADOOP-16138. hadoop fs mkdir / of nonexistent abfs container raises NPE

2019-09-08 Thread GitBox
bgaborg commented on issue #1302: HADOOP-16138. hadoop fs mkdir / of 
nonexistent abfs container raises NPE
URL: https://github.com/apache/hadoop/pull/1302#issuecomment-529234167
 
 
   I haven't figured out yet where to handle that. We can talk about that
   offline - it will be just a few minutes.
   
   On Fri, 6 Sep 2019 at 22:34, Steve Loughran 
   wrote:
   
   > OK, so this is the test. What about the underlying NPE?
   >
   > —
   > You are receiving this because you authored the thread.
   > Reply to this email directly, view it on GitHub
   > 
,
   > or mute the thread
   > 

   > .
   >
   


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] nandakumar131 commented on a change in pull request #1410: HDDS-2076. Read fails because the block cannot be located in the container

2019-09-08 Thread GitBox
nandakumar131 commented on a change in pull request #1410: HDDS-2076. Read 
fails because the block cannot be located in the container
URL: https://github.com/apache/hadoop/pull/1410#discussion_r322011257
 
 

 ##
 File path: 
hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerReplication.java
 ##
 @@ -0,0 +1,199 @@
+package org.apache.hadoop.ozone.client.rpc;
 
 Review comment:
   License is missing


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] nandakumar131 commented on a change in pull request #1410: HDDS-2076. Read fails because the block cannot be located in the container

2019-09-08 Thread GitBox
nandakumar131 commented on a change in pull request #1410: HDDS-2076. Read 
fails because the block cannot be located in the container
URL: https://github.com/apache/hadoop/pull/1410#discussion_r322011377
 
 

 ##
 File path: 
hadoop-hdds/container-service/src/main/java/org/apache/hadoop/ozone/container/keyvalue/TarContainerPacker.java
 ##
 @@ -235,6 +238,16 @@ private void includePath(String containerPath, String 
subdir,
 }
   }
 
+  private void includeBCSID(ArchiveOutputStream archiveOutputStream, long 
bcsID)
 
 Review comment:
   `includeBCSID` is not used anywhere


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[GitHub] [hadoop] nandakumar131 commented on a change in pull request #1410: HDDS-2076. Read fails because the block cannot be located in the container

2019-09-08 Thread GitBox
nandakumar131 commented on a change in pull request #1410: HDDS-2076. Read 
fails because the block cannot be located in the container
URL: https://github.com/apache/hadoop/pull/1410#discussion_r322011275
 
 

 ##
 File path: 
hadoop-ozone/integration-test/src/test/java/org/apache/hadoop/ozone/client/rpc/TestContainerReplication.java
 ##
 @@ -0,0 +1,199 @@
+package org.apache.hadoop.ozone.client.rpc;
+
+import org.apache.hadoop.hdds.client.BlockID;
 
 Review comment:
   Unused import


This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
us...@infra.apache.org


With regards,
Apache Git Services

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org



[jira] [Commented] (HADOOP-15922) DelegationTokenAuthenticationFilter get wrong doAsUser since it does not decode URL

2019-09-08 Thread Rohith Sharma K S (Jira)


[ 
https://issues.apache.org/jira/browse/HADOOP-15922?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16925127#comment-16925127
 ] 

Rohith Sharma K S commented on HADOOP-15922:


[~eyang] [~tasanuma] [~hexiaoqiao] Could anyone update the release note since 
it is a incompatible issue for release 3.2.1?

> DelegationTokenAuthenticationFilter get wrong doAsUser since it does not 
> decode URL
> ---
>
> Key: HADOOP-15922
> URL: https://issues.apache.org/jira/browse/HADOOP-15922
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common, kms
>Reporter: He Xiaoqiao
>Assignee: He Xiaoqiao
>Priority: Major
> Fix For: 3.3.0, 3.2.1, 3.1.3
>
> Attachments: HADOOP-15922.001.patch, HADOOP-15922.002.patch, 
> HADOOP-15922.003.patch, HADOOP-15922.004.patch, HADOOP-15922.005.patch, 
> HADOOP-15922.006.patch, HADOOP-15922.007.patch
>
>
> DelegationTokenAuthenticationFilter get wrong doAsUser when proxy user from 
> client is complete kerberos name (e.g., user/hostn...@realm.com, actually it 
> is acceptable), because DelegationTokenAuthenticationFilter does not decode 
> DOAS parameter in URL which is encoded by {{URLEncoder}} at client.
> e.g. KMS as example:
> a. KMSClientProvider creates connection to KMS Server using 
> DelegationTokenAuthenticatedURL#openConnection.
> b. If KMSClientProvider is a doAsUser, KMSClientProvider will put {{doas}} 
> with url encoded user as one parameter of http request. 
> {code:java}
> // proxyuser
> if (doAs != null) {
>   extraParams.put(DO_AS, URLEncoder.encode(doAs, "UTF-8"));
> }
> {code}
> c. when KMS server receives the request, it does not decode the proxy user.
> As result, KMS Server will get the wrong proxy user if this proxy user is 
> complete Kerberos Name or it includes some special character. Some other 
> authentication and authorization exception will throws next to it.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org