[jira] [Commented] (HADOOP-15523) Shell command timeout given is in seconds whereas it is taken as millisec while scheduling
[ https://issues.apache.org/jira/browse/HADOOP-15523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510567#comment-16510567 ] genericqa commented on HADOOP-15523: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 23s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 25m 45s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 28m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 19s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 7s{color} | {color:green} trunk passed {color} | | {color:red}-1{color} | {color:red} shadedclient {color} | {color:red} 3m 38s{color} | {color:red} branch has errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 28s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 57s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 45s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 26m 57s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 26m 57s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 20s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 6s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:red}-1{color} | {color:red} shadedclient {color} | {color:red} 2m 6s{color} | {color:red} patch has errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 38s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 54s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 43s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 35s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}105m 5s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15523 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927562/HADOOP-15523-003.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux a5de5a5d44cb 3.13.0-137-generic #186-Ubuntu SMP Mon Dec 4 19:09:19 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 108da85 | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_171 | | findbugs | v3.1.0-RC1 | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14769/testReport/ | | Max. process+thread count | 1716 (vs. ulimit of 1) | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14769/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Shell command timeout given is in seconds whereas it is taken as millisec > while scheduling >
[jira] [Updated] (HADOOP-13837) Always get unable to kill error message even the hadoop process was successfully killed
[ https://issues.apache.org/jira/browse/HADOOP-13837?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Weiwei Yang updated HADOOP-13837: - Resolution: Duplicate Status: Resolved (was: Patch Available) Seems like to be dup of YARN-15527 and it is just resolved. Closing this one... > Always get unable to kill error message even the hadoop process was > successfully killed > --- > > Key: HADOOP-13837 > URL: https://issues.apache.org/jira/browse/HADOOP-13837 > Project: Hadoop Common > Issue Type: Bug > Components: scripts >Reporter: Weiwei Yang >Assignee: Weiwei Yang >Priority: Critical > Attachments: HADOOP-13837.01.patch, HADOOP-13837.02.patch, > HADOOP-13837.03.patch, HADOOP-13837.04.patch, HADOOP-13837.05.patch, > check_proc.sh > > > *Reproduce steps* > # Setup a hadoop cluster > # Stop resource manager : yarn --daemon stop resourcemanager > # Stop node manager : yarn --daemon stop nodemanager > WARNING: nodemanager did not stop gracefully after 5 seconds: Trying to kill > with kill -9 > ERROR: Unable to kill 20325 > it always gets "Unable to kill " error message, this gives user > impression there is something wrong with the node manager process because it > was not able to be forcibly killed. But in fact, the kill command works as > expected. > This was because hadoop-functions.sh did not check process existence after > kill properly. Currently it checks the process liveness right after the kill > command > {code} > ... > kill -9 "${pid}" >/dev/null 2>&1 > if ps -p "${pid}" > /dev/null 2>&1; then > hadoop_error "ERROR: Unable to kill ${pid}" > ... > {code} > when resource manager stopped before node managers, it always takes some > additional time until the process completely terminates. I tried to print > output of {{ps -p }} in a while loop after kill -9, and found > following > {noformat} > 16212 ?00:00:11 java > 0 > PID TTY TIME CMD > 16212 ?00:00:11 java > 0 > PID TTY TIME CMD > 16212 ?00:00:11 java > 0 > PID TTY TIME CMD > 1 > PID TTY TIME CMD > 1 > PID TTY TIME CMD > 1 > PID TTY TIME CMD > ... > {noformat} > in the first 3 times of the loop, the process did not terminate so the exit > code of {{ps -p}} are still {{0}} > *Proposal of a fix* > Firstly I was thinking to add a more comprehensive pid check, it checks the > pid liveness until reaches the HADOOP_STOP_TIMEOUT, but this seems to add too > much complexity. Second fix was to simply add a {{sleep 3}} after {{kill > -9}}, it should fix the error in most cases with relative small changes to > the script. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510496#comment-16510496 ] Eric Yang commented on HADOOP-15527: Jenkins failure is due to protoc version mismatch, not caused by the commit. Thank you [~vinodkv] for the patch. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Fix For: 3.2.0, 3.1.1 > > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Eric Yang updated HADOOP-15527: --- Resolution: Fixed Fix Version/s: 3.1.1 3.2.0 Status: Resolved (was: Patch Available) > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Fix For: 3.2.0, 3.1.1 > > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15523) Shell command timeout given is in seconds whereas it is taken as millisec while scheduling
[ https://issues.apache.org/jira/browse/HADOOP-15523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bilwa S T updated HADOOP-15523: --- Attachment: HADOOP-15523-003.patch > Shell command timeout given is in seconds whereas it is taken as millisec > while scheduling > -- > > Key: HADOOP-15523 > URL: https://issues.apache.org/jira/browse/HADOOP-15523 > Project: Hadoop Common > Issue Type: Bug >Reporter: Bilwa S T >Assignee: Bilwa S T >Priority: Major > Attachments: HADOOP-15523-001.patch, HADOOP-15523-002.patch, > HADOOP-15523-003.patch > > > ShellBasedUnixGroupsMapping has a property > {{hadoop.security.groups.shell.command.timeout}} to control how long to wait > for the fetch groups command which can be configured in seconds. but while > scheduling the time taken is millisecs. so currently if u give value as 60s, > it is taken as 60ms. > {code:java} > timeout = conf.getTimeDuration( > CommonConfigurationKeys. > HADOOP_SECURITY_GROUP_SHELL_COMMAND_TIMEOUT_SECS, > CommonConfigurationKeys. > HADOOP_SECURITY_GROUP_SHELL_COMMAND_TIMEOUT_SECS_DEFAULT, > TimeUnit.SECONDS);{code} > Time unit given is in seconds but it should be millisecs -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510472#comment-16510472 ] Hudson commented on HADOOP-15527: - FAILURE: Integrated in Jenkins build Hadoop-trunk-Commit #14416 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/14416/]) HADOOP-15527. Improve delay check for stopping processes. (eyang: rev 108da85320d65e37fe835de65866b818e5420587) * (edit) hadoop-common-project/hadoop-common/src/main/bin/hadoop-functions.sh * (edit) hadoop-common-project/hadoop-common/src/test/scripts/hadoop_stop_daemon.bats > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510466#comment-16510466 ] Eric Yang commented on HADOOP-15527: +1 looks good to me. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510448#comment-16510448 ] genericqa commented on HADOOP-15527: | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 23m 13s{color} | {color:blue} Docker mode activated. {color} | | {color:blue}0{color} | {color:blue} patch {color} | {color:blue} 0m 2s{color} | {color:blue} The patch file was not named according to hadoop's naming conventions. Please see https://wiki.apache.org/hadoop/HowToContribute for instructions. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 28m 42s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 6s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 58s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 15s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} shellcheck {color} | {color:green} 0m 4s{color} | {color:green} There were no new shellcheck issues. {color} | | {color:green}+1{color} | {color:green} shelldocs {color} | {color:green} 0m 11s{color} | {color:green} There were no new shelldocs issues. {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 17s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 2m 17s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 24s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 80m 58s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15527 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927538/HADOOP-15527.2.txt | | Optional Tests | asflicense mvnsite unit shellcheck shelldocs | | uname | Linux 8a88e0c60072 3.13.0-137-generic #186-Ubuntu SMP Mon Dec 4 19:09:19 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 5670e89 | | maven | version: Apache Maven 3.3.9 | | shellcheck | v0.4.6 | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14767/testReport/ | | Max. process+thread count | 348 (vs. ulimit of 1) | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14767/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian
[jira] [Commented] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510433#comment-16510433 ] genericqa commented on HADOOP-15528: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s{color} | {color:blue} Docker mode activated. {color} | | {color:red}-1{color} | {color:red} patch {color} | {color:red} 0m 5s{color} | {color:red} HADOOP-15528 does not apply to HADOOP-15461. Rebase required? Wrong Branch? See https://wiki.apache.org/hadoop/HowToContribute for help. {color} | \\ \\ || Subsystem || Report/Notes || | JIRA Issue | HADOOP-15528 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927557/HADOOP-15528-HADOOP-15461.v2.patch | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14768/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch, > HADOOP-15528-HADOOP-15461.v2.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510432#comment-16510432 ] Íñigo Goiri commented on HADOOP-15528: -- Thanks [~giovanni.fumarola] for [^HADOOP-15528-HADOOP-15461.v2.patch]. Let's focus to the scope of the scope and do the following: * Keep a writeLaunchEnv with the old parameters and passing null. This will reduce the number of calls where we have to add null manually. * Don't remove the supresswarnings, do a separate JIRA for that; we can commit that to trunk. * Don't remove link just mark it as deprecate. * Remove putEnvIfNotNull and putEnvIfNotNull in the new JIRA for trunk. * Avoid the stetic fixes to parts that are not changed (e.g., TestContainerLaunch #200) * Avoid fixes to close and supresswarning and jarFile.close(). * Don't remove the testWindowsShellScriptBuilderLink. > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch, > HADOOP-15528-HADOOP-15461.v2.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-14178) Move Mockito up to version 2.x
[ https://issues.apache.org/jira/browse/HADOOP-14178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510431#comment-16510431 ] genericqa commented on HADOOP-14178: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 25s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 258 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 1m 38s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 26m 42s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 29m 16s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 19m 31s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 30m 41s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project hadoop-hdfs-project/hadoop-hdfs-native-client hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests hadoop-mapreduce-project/hadoop-mapreduce-client hadoop-mapreduce-project hadoop-client-modules/hadoop-client-minicluster . hadoop-ozone/integration-test {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 1m 13s{color} | {color:red} hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager in trunk has 1 extant Findbugs warnings. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 32s{color} | {color:red} server-scm in trunk failed. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 21s{color} | {color:red} hadoop-ozone in trunk failed. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 5m 27s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 16s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 19s{color} | {color:red} server-scm in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 15s{color} | {color:red} integration-test in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 13s{color} | {color:red} hadoop-ozone in the patch failed. {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 28m 23s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} javac {color} | {color:red} 28m 23s{color} | {color:red} root generated 15 new + 1537 unchanged - 0 fixed = 1552 total (was 1537) {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 21s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 19m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 58s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 6s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project hadoop-hdfs-project/hadoop-hdfs-native-client hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-tests hadoop-mapreduce-project/hadoop-mapreduce-client hadoop-mapreduce-project hadoop-client-modules/hadoop-client-minicluster . hadoop-ozone/integration-test
[jira] [Commented] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510407#comment-16510407 ] Giovanni Matteo Fumarola commented on HADOOP-15528: --- After an offline chat with [~elgoiri] we discussed on how to avoid to break the compatibility. Proper design notes will follow up. Attached [^HADOOP-15528-HADOOP-15461.v2.patch]. Let's see what Yetus will say and then we can continue work on the patch. > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch, > HADOOP-15528-HADOOP-15461.v2.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Created] (HADOOP-15533) Making WASB listStatus error codes consistent
Esfandiar Manii created HADOOP-15533: Summary: Making WASB listStatus error codes consistent Key: HADOOP-15533 URL: https://issues.apache.org/jira/browse/HADOOP-15533 Project: Hadoop Common Issue Type: Bug Components: fs/azure Reporter: Esfandiar Manii Assignee: Esfandiar Manii Very small string updated to make the error messages consistent -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15528: -- Attachment: HADOOP-15528-HADOOP-15461.v2.patch > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch, > HADOOP-15528-HADOOP-15461.v2.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15528: -- Attachment: (was: HADOOP-15528-HADOOP-15461.v2.patch) > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15528: -- Attachment: HADOOP-15528-HADOOP-15461.v2.patch > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch, > HADOOP-15528-HADOOP-15461.v2.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510403#comment-16510403 ] Da Zhou commented on HADOOP-15407: -- The following *independent* unit tests failed in latest Jenkins build: - hadoop.fs.shell.TestCopyFromLocal - hadoop.crypto.key.TestKeyShell - hadoop.crypto.key.TestKeyProviderFactory Because all Azure changes goes under project hadoop-azure, so these unit test failure cannot be caused by the patch. Can someone help? Regards, Da > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail:
[jira] [Commented] (HADOOP-15483) Upgrade jquery to version 3.3.1
[ https://issues.apache.org/jira/browse/HADOOP-15483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510382#comment-16510382 ] genericqa commented on HADOOP-15483: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 14s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 18s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 25m 53s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 27m 52s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 19m 27s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 30m 43s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: . hadoop-ozone hadoop-ozone/acceptance-test {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 32s{color} | {color:red} server-scm in trunk failed. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 24s{color} | {color:red} ozone-manager in trunk failed. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 5m 36s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 18s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 19s{color} | {color:red} server-scm in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 58s{color} | {color:red} hadoop-ozone in the patch failed. {color} | | {color:red}-1{color} | {color:red} mvninstall {color} | {color:red} 0m 20s{color} | {color:red} ozone-manager in the patch failed. {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 28m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 28m 42s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 18m 54s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} whitespace {color} | {color:red} 0m 0s{color} | {color:red} The patch 5861 line(s) with tabs. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 4s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 1s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: . hadoop-ozone hadoop-ozone/acceptance-test {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 22s{color} | {color:red} server-scm in the patch failed. {color} | | {color:red}-1{color} | {color:red} findbugs {color} | {color:red} 0m 21s{color} | {color:red} ozone-manager in the patch failed. {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 5m 19s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red}127m 43s{color} | {color:red} root in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 42s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} |
[jira] [Commented] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510371#comment-16510371 ] genericqa commented on HADOOP-15407: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 16s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 55 new or modified test files. {color} | || || || || {color:brown} HADOOP-15407 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 6m 15s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 26m 36s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 29m 38s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 27s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 20m 39s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 34m 17s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project . {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 4s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 5m 28s{color} | {color:green} HADOOP-15407 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 20s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 27m 58s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 28m 34s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 28m 34s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 18s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 20m 13s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 7s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 13s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project . {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 16s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 5m 26s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 14m 45s{color} | {color:red} root in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 37s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}219m 54s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.fs.shell.TestCopyFromLocal | | | hadoop.crypto.key.TestKeyShell | | | hadoop.crypto.key.TestKeyProviderFactory | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15407 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927526/HADOOP-15407-HADOOP-15407.008.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite
[jira] [Commented] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510334#comment-16510334 ] genericqa commented on HADOOP-15407: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 19s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 55 new or modified test files. {color} | || || || || {color:brown} HADOOP-15407 Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 6m 20s{color} | {color:blue} Maven dependency ordering for branch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 26m 30s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 29m 8s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 17s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 19m 54s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 33m 30s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project . {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 4s{color} | {color:green} HADOOP-15407 passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 6m 1s{color} | {color:green} HADOOP-15407 passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:blue}0{color} | {color:blue} mvndep {color} | {color:blue} 0m 19s{color} | {color:blue} Maven dependency ordering for patch {color} | | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 28m 20s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 28m 35s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 28m 35s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 3m 22s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 19m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} xml {color} | {color:green} 0m 7s{color} | {color:green} The patch has no ill-formed XML file. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 33s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:blue}0{color} | {color:blue} findbugs {color} | {color:blue} 0m 0s{color} | {color:blue} Skipped patched modules with no Java source: hadoop-project . {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 2m 23s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 5m 43s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:red}-1{color} | {color:red} unit {color} | {color:red} 14m 9s{color} | {color:red} root in the patch failed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 38s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}219m 7s{color} | {color:black} {color} | \\ \\ || Reason || Tests || | Failed junit tests | hadoop.fs.shell.TestCopyFromLocal | | | hadoop.crypto.key.TestKeyShell | | | hadoop.crypto.key.TestKeyProviderFactory | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15407 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927516/HADOOP-15407-HADOOP-15407.008.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite
[jira] [Commented] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510277#comment-16510277 ] Hudson commented on HADOOP-15532: - ABORTED: Integrated in Jenkins build Hadoop-trunk-Commit #14414 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/14414/]) HADOOP-15532. TestBasicDiskValidator fails with NoSuchFileException. (inigoiri: rev aeaf9fec62f10699d1c809d66444520fe4533c2c) * (edit) hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/util/TestDiskChecker.java > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Fix For: 2.10.0, 3.2.0, 3.1.1, 2.9.2, 3.0.4 > > Attachments: AfterFix.png, BeforeFix.png, HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Status: Patch Available (was: Open) Addressing shellcheck issue. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Attachment: HADOOP-15527.2.txt > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Status: Open (was: Patch Available) > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.2.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Thomas Marquardt updated HADOOP-15407: -- Status: Open (was: Patch Available) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Thomas Marquardt updated HADOOP-15407: -- Status: Patch Available (was: Open) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Thomas Marquardt updated HADOOP-15407: -- Attachment: (was: HADOOP-15407-patch-atop-patch-007.patch) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HADOOP-15532: - Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.0.4 2.9.2 3.1.1 3.2.0 2.10.0 Status: Resolved (was: Patch Available) Thanks [~giovanni.fumarola] for the fix. Committed to trunk, branch-3.1, branch-3.0, branch-2, and branch-2.9. > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Fix For: 2.10.0, 3.2.0, 3.1.1, 2.9.2, 3.0.4 > > Attachments: AfterFix.png, BeforeFix.png, HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510187#comment-16510187 ] Íñigo Goiri commented on HADOOP-15532: -- The default test dir fix is the same as in YARN-8422. [^HADOOP-15532.v1.patch] LGTM. +1 Committing. > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: AfterFix.png, BeforeFix.png, HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510151#comment-16510151 ] Eric Yang commented on HADOOP-15527: {code} timeout=$(printf "%.0f\n" ${timeout}) {code} Can be simplified to: {code} timeout=$((0 + ${timeout})) {code} To avoid shellcheck warning. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510110#comment-16510110 ] genericqa commented on HADOOP-15527: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 12s{color} | {color:blue} Docker mode activated. {color} | | {color:blue}0{color} | {color:blue} patch {color} | {color:blue} 0m 2s{color} | {color:blue} The patch file was not named according to hadoop's naming conventions. Please see https://wiki.apache.org/hadoop/HowToContribute for instructions. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 27m 32s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 14s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 59s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} shellcheck {color} | {color:red} 0m 4s{color} | {color:red} The patch generated 1 new + 20 unchanged - 0 fixed = 21 total (was 20) {color} | | {color:green}+1{color} | {color:green} shelldocs {color} | {color:green} 0m 12s{color} | {color:green} There were no new shelldocs issues. {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 27s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 2m 23s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 25s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 57m 10s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15527 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927524/HADOOP-15527.1.txt | | Optional Tests | asflicense mvnsite unit shellcheck shelldocs | | uname | Linux b8cd6bdde67d 3.13.0-139-generic #188-Ubuntu SMP Tue Jan 9 14:43:09 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / a16623d | | maven | version: Apache Maven 3.3.9 | | shellcheck | v0.4.6 | | shellcheck | https://builds.apache.org/job/PreCommit-HADOOP-Build/14764/artifact/out/diff-patch-shellcheck.txt | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14764/testReport/ | | Max. process+thread count | 334 (vs. ulimit of 1) | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14764/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other
[jira] [Commented] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510071#comment-16510071 ] genericqa commented on HADOOP-15532: | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 19s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 27m 32s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 30m 0s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 14s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 43s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 37s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 29m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 29m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 20s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 14s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 3s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 51s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 9m 23s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 36s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}130m 13s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15532 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927504/HADOOP-15532.v1.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 758c705cf339 3.13.0-137-generic #186-Ubuntu SMP Mon Dec 4 19:09:19 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 10d0e4b | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_171 | | findbugs | v3.1.0-RC1 | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14761/testReport/ | | Max. process+thread count | 1693 (vs. ulimit of 1) | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14761/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 >
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou updated HADOOP-15407: - Attachment: HADOOP-15407-HADOOP-15407.008.patch > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch, HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou updated HADOOP-15407: - Attachment: (was: HADOOP-15407-HADOOP-15407.008.patch) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch, HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15508) Test and validate Tez 0.8.4/5 jobs run with hadoop 3.x
[ https://issues.apache.org/jira/browse/HADOOP-15508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Suma Shivaprasad updated HADOOP-15508: -- Summary: Test and validate Tez 0.8.4/5 jobs run with hadoop 3.x (was: Test and validate Tez 0.8.5 jobs run with hadoop 3.x) > Test and validate Tez 0.8.4/5 jobs run with hadoop 3.x > -- > > Key: HADOOP-15508 > URL: https://issues.apache.org/jira/browse/HADOOP-15508 > Project: Hadoop Common > Issue Type: Sub-task >Affects Versions: 3.1.0 >Reporter: Suma Shivaprasad >Priority: Major > > Validate that Tez 0.8.5 jobs can run with Hadoop 3.1 through distributed cache -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15508) Test and validate Tez 0.8.5 jobs run with hadoop 3.x
[ https://issues.apache.org/jira/browse/HADOOP-15508?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Suma Shivaprasad updated HADOOP-15508: -- Description: Validate that Tez 0.8.5 jobs can run with Hadoop 3.1 through distributed cache (was: Validate that Tez 0.8.5 jobs can run with hadoop 3.1 through distribute cache) > Test and validate Tez 0.8.5 jobs run with hadoop 3.x > > > Key: HADOOP-15508 > URL: https://issues.apache.org/jira/browse/HADOOP-15508 > Project: Hadoop Common > Issue Type: Sub-task >Affects Versions: 3.1.0 >Reporter: Suma Shivaprasad >Priority: Major > > Validate that Tez 0.8.5 jobs can run with Hadoop 3.1 through distributed cache -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Status: Patch Available (was: Open) Updated new patch addressing the complaints from Jenkins. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Status: Open (was: Patch Available) bq. In the batch file, hadoop_stop_daemon is renamed to hadoop_stop_daemon_changing_pid. Is this change necessary? It's not necessary, but earlier there was only one test and so it was okay to have a generic name. We now have two tests so modified to disambiguate what each test is doing. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Attachment: HADOOP-15527.1.txt > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.1.txt, HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15307) NFS: flavor AUTH_SYS should use VerifierNone
[ https://issues.apache.org/jira/browse/HADOOP-15307?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16510025#comment-16510025 ] Hudson commented on HADOOP-15307: - SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14412 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/14412/]) HADOOP-15307. NFS: flavor AUTH_SYS should use VerifierNone. Contributed (xiao: rev a16623df2148e59e1b7dcf98939b03cb7440d143) * (edit) hadoop-common-project/hadoop-nfs/src/main/java/org/apache/hadoop/oncrpc/security/Verifier.java > NFS: flavor AUTH_SYS should use VerifierNone > > > Key: HADOOP-15307 > URL: https://issues.apache.org/jira/browse/HADOOP-15307 > Project: Hadoop Common > Issue Type: Bug > Components: nfs > Environment: CentOS 7.4, CDH5.13.1, Kerberized Hadoop cluster >Reporter: Wei-Chiu Chuang >Assignee: Gabor Bota >Priority: Major > Fix For: 3.2.0 > > Attachments: HADOOP-15307.001.patch, HADOOP-15307.002.patch, > HADOOP-15307.003.patch, HADOOP-15307.004.patch, HADOOP-15307.005.patch > > > When NFS gateway starts and if the portmapper request is denied by rpcbind > for any reason (in our case, /etc/hosts.allow did not have the localhost), > NFS gateway fails with the following obscure exception: > {noformat} > 2018-03-05 12:49:31,976 INFO org.apache.hadoop.oncrpc.SimpleUdpServer: > Started listening to UDP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,988 INFO org.apache.hadoop.oncrpc.SimpleTcpServer: > Started listening to TCP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,993 TRACE org.apache.hadoop.oncrpc.RpcCall: > Xid:692394656, messageType:RPC_CALL, rpcVersion:2, program:10, version:2, > procedure:1, credential:(AuthFlavor:AUTH_NONE), > verifier:(AuthFlavor:AUTH_NONE) > 2018-03-05 12:49:31,998 FATAL org.apache.hadoop.mount.MountdBase: Failed to > start the server. Cause: > java.lang.UnsupportedOperationException: Unsupported verifier flavorAUTH_SYS > at > org.apache.hadoop.oncrpc.security.Verifier.readFlavorAndVerifier(Verifier.java:45) > at org.apache.hadoop.oncrpc.RpcDeniedReply.read(RpcDeniedReply.java:50) > at org.apache.hadoop.oncrpc.RpcReply.read(RpcReply.java:67) > at org.apache.hadoop.oncrpc.SimpleUdpClient.run(SimpleUdpClient.java:71) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:130) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:101) > at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:83) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69) > at > org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:60) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243) > 2018-03-05 12:49:32,007 INFO org.apache.hadoop.util.ExitUtil: Exiting with > status 1{noformat} > Reading the code comment for class Verifier, I think this bug existed since > its inception > {code:java} > /** > * Base class for verifier. Currently our authentication only supports 3 types > * of auth flavors: {@link RpcAuthInfo.AuthFlavor#AUTH_NONE}, {@link > RpcAuthInfo.AuthFlavor#AUTH_SYS}, > * and {@link RpcAuthInfo.AuthFlavor#RPCSEC_GSS}. Thus for verifier we only > need to handle > * AUTH_NONE and RPCSEC_GSS > */ > public abstract class Verifier extends RpcAuthInfo {{code} > The verifier should also handle AUTH_SYS too. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou updated HADOOP-15407: - Status: Patch Available (was: Open) Trying to get Jenkins to build with HADOOP-15407-HADOOP-15407.008.patch > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch, HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15307) NFS: flavor AUTH_SYS should use VerifierNone
[ https://issues.apache.org/jira/browse/HADOOP-15307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Chen updated HADOOP-15307: --- Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.2.0 Status: Resolved (was: Patch Available) Committed to trunk. Thanks Gabor for the contribution, and others for review / reporting! > NFS: flavor AUTH_SYS should use VerifierNone > > > Key: HADOOP-15307 > URL: https://issues.apache.org/jira/browse/HADOOP-15307 > Project: Hadoop Common > Issue Type: Bug > Components: nfs > Environment: CentOS 7.4, CDH5.13.1, Kerberized Hadoop cluster >Reporter: Wei-Chiu Chuang >Assignee: Gabor Bota >Priority: Major > Fix For: 3.2.0 > > Attachments: HADOOP-15307.001.patch, HADOOP-15307.002.patch, > HADOOP-15307.003.patch, HADOOP-15307.004.patch, HADOOP-15307.005.patch > > > When NFS gateway starts and if the portmapper request is denied by rpcbind > for any reason (in our case, /etc/hosts.allow did not have the localhost), > NFS gateway fails with the following obscure exception: > {noformat} > 2018-03-05 12:49:31,976 INFO org.apache.hadoop.oncrpc.SimpleUdpServer: > Started listening to UDP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,988 INFO org.apache.hadoop.oncrpc.SimpleTcpServer: > Started listening to TCP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,993 TRACE org.apache.hadoop.oncrpc.RpcCall: > Xid:692394656, messageType:RPC_CALL, rpcVersion:2, program:10, version:2, > procedure:1, credential:(AuthFlavor:AUTH_NONE), > verifier:(AuthFlavor:AUTH_NONE) > 2018-03-05 12:49:31,998 FATAL org.apache.hadoop.mount.MountdBase: Failed to > start the server. Cause: > java.lang.UnsupportedOperationException: Unsupported verifier flavorAUTH_SYS > at > org.apache.hadoop.oncrpc.security.Verifier.readFlavorAndVerifier(Verifier.java:45) > at org.apache.hadoop.oncrpc.RpcDeniedReply.read(RpcDeniedReply.java:50) > at org.apache.hadoop.oncrpc.RpcReply.read(RpcReply.java:67) > at org.apache.hadoop.oncrpc.SimpleUdpClient.run(SimpleUdpClient.java:71) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:130) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:101) > at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:83) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69) > at > org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:60) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243) > 2018-03-05 12:49:32,007 INFO org.apache.hadoop.util.ExitUtil: Exiting with > status 1{noformat} > Reading the code comment for class Verifier, I think this bug existed since > its inception > {code:java} > /** > * Base class for verifier. Currently our authentication only supports 3 types > * of auth flavors: {@link RpcAuthInfo.AuthFlavor#AUTH_NONE}, {@link > RpcAuthInfo.AuthFlavor#AUTH_SYS}, > * and {@link RpcAuthInfo.AuthFlavor#RPCSEC_GSS}. Thus for verifier we only > need to handle > * AUTH_NONE and RPCSEC_GSS > */ > public abstract class Verifier extends RpcAuthInfo {{code} > The verifier should also handle AUTH_SYS too. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15307) NFS: flavor AUTH_SYS should use VerifierNone
[ https://issues.apache.org/jira/browse/HADOOP-15307?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Xiao Chen updated HADOOP-15307: --- Summary: NFS: flavor AUTH_SYS should use VerifierNone (was: Improve NFS error handling: Unsupported verifier flavorAUTH_SYS) > NFS: flavor AUTH_SYS should use VerifierNone > > > Key: HADOOP-15307 > URL: https://issues.apache.org/jira/browse/HADOOP-15307 > Project: Hadoop Common > Issue Type: Bug > Components: nfs > Environment: CentOS 7.4, CDH5.13.1, Kerberized Hadoop cluster >Reporter: Wei-Chiu Chuang >Assignee: Gabor Bota >Priority: Major > Attachments: HADOOP-15307.001.patch, HADOOP-15307.002.patch, > HADOOP-15307.003.patch, HADOOP-15307.004.patch, HADOOP-15307.005.patch > > > When NFS gateway starts and if the portmapper request is denied by rpcbind > for any reason (in our case, /etc/hosts.allow did not have the localhost), > NFS gateway fails with the following obscure exception: > {noformat} > 2018-03-05 12:49:31,976 INFO org.apache.hadoop.oncrpc.SimpleUdpServer: > Started listening to UDP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,988 INFO org.apache.hadoop.oncrpc.SimpleTcpServer: > Started listening to TCP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,993 TRACE org.apache.hadoop.oncrpc.RpcCall: > Xid:692394656, messageType:RPC_CALL, rpcVersion:2, program:10, version:2, > procedure:1, credential:(AuthFlavor:AUTH_NONE), > verifier:(AuthFlavor:AUTH_NONE) > 2018-03-05 12:49:31,998 FATAL org.apache.hadoop.mount.MountdBase: Failed to > start the server. Cause: > java.lang.UnsupportedOperationException: Unsupported verifier flavorAUTH_SYS > at > org.apache.hadoop.oncrpc.security.Verifier.readFlavorAndVerifier(Verifier.java:45) > at org.apache.hadoop.oncrpc.RpcDeniedReply.read(RpcDeniedReply.java:50) > at org.apache.hadoop.oncrpc.RpcReply.read(RpcReply.java:67) > at org.apache.hadoop.oncrpc.SimpleUdpClient.run(SimpleUdpClient.java:71) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:130) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:101) > at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:83) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69) > at > org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:60) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243) > 2018-03-05 12:49:32,007 INFO org.apache.hadoop.util.ExitUtil: Exiting with > status 1{noformat} > Reading the code comment for class Verifier, I think this bug existed since > its inception > {code:java} > /** > * Base class for verifier. Currently our authentication only supports 3 types > * of auth flavors: {@link RpcAuthInfo.AuthFlavor#AUTH_NONE}, {@link > RpcAuthInfo.AuthFlavor#AUTH_SYS}, > * and {@link RpcAuthInfo.AuthFlavor#RPCSEC_GSS}. Thus for verifier we only > need to handle > * AUTH_NONE and RPCSEC_GSS > */ > public abstract class Verifier extends RpcAuthInfo {{code} > The verifier should also handle AUTH_SYS too. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Assigned] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou reassigned HADOOP-15407: Assignee: Da Zhou (was: Esfandiar Manii) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Da Zhou >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch, HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou updated HADOOP-15407: - Status: Open (was: Patch Available) Trying to get Jenkins to run > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Esfandiar Manii >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch, HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou updated HADOOP-15407: - Attachment: HADOOP-15407-HADOOP-15407.008.patch Status: Patch Available (was: Open) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Esfandiar Manii >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-HADOOP-15407.008.patch, HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15407) Support Windows Azure Storage - Blob file system in Hadoop
[ https://issues.apache.org/jira/browse/HADOOP-15407?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Da Zhou updated HADOOP-15407: - Attachment: (was: HADOOP-15407-HADOOP-15407.008.patch) > Support Windows Azure Storage - Blob file system in Hadoop > -- > > Key: HADOOP-15407 > URL: https://issues.apache.org/jira/browse/HADOOP-15407 > Project: Hadoop Common > Issue Type: New Feature > Components: fs/azure >Affects Versions: 3.2.0 >Reporter: Esfandiar Manii >Assignee: Esfandiar Manii >Priority: Major > Attachments: HADOOP-15407-001.patch, HADOOP-15407-002.patch, > HADOOP-15407-003.patch, HADOOP-15407-004.patch, > HADOOP-15407-HADOOP-15407.006.patch, HADOOP-15407-HADOOP-15407.007.patch, > HADOOP-15407-patch-atop-patch-007.patch > > > *{color:#212121}Description{color}* > This JIRA adds a new file system implementation, ABFS, for running Big Data > and Analytics workloads against Azure Storage. This is a complete rewrite of > the previous WASB driver with a heavy focus on optimizing both performance > and cost. > {color:#212121} {color} > *{color:#212121}High level design{color}* > At a high level, the code here extends the FileSystem class to provide an > implementation for accessing blobs in Azure Storage. The scheme abfs is used > for accessing it over HTTP, and abfss for accessing over HTTPS. The following > URI scheme is used to address individual paths: > {color:#212121} {color} > > {color:#212121}abfs[s]://@.dfs.core.windows.net/{color} > {color:#212121} {color} > {color:#212121}ABFS is intended as a replacement to WASB. WASB is not > deprecated but is in pure maintenance mode and customers should upgrade to > ABFS once it hits General Availability later in CY18.{color} > {color:#212121}Benefits of ABFS include:{color} > {color:#212121}· Higher scale (capacity, throughput, and IOPS) Big > Data and Analytics workloads by allowing higher limits on storage > accounts{color} > {color:#212121}· Removing any ramp up time with Storage backend > partitioning; blocks are now automatically sharded across partitions in the > Storage backend{color} > {color:#212121} . This avoids the need for using > temporary/intermediate files, increasing the cost (and framework complexity > around committing jobs/tasks){color} > {color:#212121}· Enabling much higher read and write throughput on > single files (tens of Gbps by default){color} > {color:#212121}· Still retaining all of the Azure Blob features > customers are familiar with and expect, and gaining the benefits of future > Blob features as well{color} > {color:#212121}ABFS incorporates Hadoop Filesystem metrics to monitor the > file system throughput and operations. Ambari metrics are not currently > implemented for ABFS, but will be available soon.{color} > {color:#212121} {color} > *{color:#212121}Credits and history{color}* > Credit for this work goes to (hope I don't forget anyone): Shane Mainali, > {color:#212121}Thomas Marquardt, Zichen Sun, Georgi Chalakov, Esfandiar > Manii, Amit Singh, Dana Kaban, Da Zhou, Junhua Gu, Saher Ahwal, Saurabh Pant, > and James Baker. {color} > {color:#212121} {color} > *Test* > ABFS has gone through many test procedures including Hadoop file system > contract tests, unit testing, functional testing, and manual testing. All the > Junit tests provided with the driver are capable of running in both > sequential/parallel fashion in order to reduce the testing time. > {color:#212121}Besides unit tests, we have used ABFS as the default file > system in Azure HDInsight. Azure HDInsight will very soon offer ABFS as a > storage option. (HDFS is also used but not as default file system.) Various > different customer and test workloads have been run against clusters with > such configurations for quite some time. Benchmarks such as Tera*, TPC-DS, > Spark Streaming and Spark SQL, and others have been run to do scenario, > performance, and functional testing. Third parties and customers have also > done various testing of ABFS.{color} > {color:#212121}The current version reflects to the version of the code > tested and used in our production environment.{color} -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15307) Improve NFS error handling: Unsupported verifier flavorAUTH_SYS
[ https://issues.apache.org/jira/browse/HADOOP-15307?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509965#comment-16509965 ] Xiao Chen commented on HADOOP-15307: +1, based on the verbal understanding of the HDFS-5085 comment. (A test doesn't seem necessary for this change which is just adding a if condition) > Improve NFS error handling: Unsupported verifier flavorAUTH_SYS > --- > > Key: HADOOP-15307 > URL: https://issues.apache.org/jira/browse/HADOOP-15307 > Project: Hadoop Common > Issue Type: Bug > Components: nfs > Environment: CentOS 7.4, CDH5.13.1, Kerberized Hadoop cluster >Reporter: Wei-Chiu Chuang >Assignee: Gabor Bota >Priority: Major > Attachments: HADOOP-15307.001.patch, HADOOP-15307.002.patch, > HADOOP-15307.003.patch, HADOOP-15307.004.patch, HADOOP-15307.005.patch > > > When NFS gateway starts and if the portmapper request is denied by rpcbind > for any reason (in our case, /etc/hosts.allow did not have the localhost), > NFS gateway fails with the following obscure exception: > {noformat} > 2018-03-05 12:49:31,976 INFO org.apache.hadoop.oncrpc.SimpleUdpServer: > Started listening to UDP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,988 INFO org.apache.hadoop.oncrpc.SimpleTcpServer: > Started listening to TCP requests at port 4242 for Rpc program: mountd at > localhost:4242 with workerCount 1 > 2018-03-05 12:49:31,993 TRACE org.apache.hadoop.oncrpc.RpcCall: > Xid:692394656, messageType:RPC_CALL, rpcVersion:2, program:10, version:2, > procedure:1, credential:(AuthFlavor:AUTH_NONE), > verifier:(AuthFlavor:AUTH_NONE) > 2018-03-05 12:49:31,998 FATAL org.apache.hadoop.mount.MountdBase: Failed to > start the server. Cause: > java.lang.UnsupportedOperationException: Unsupported verifier flavorAUTH_SYS > at > org.apache.hadoop.oncrpc.security.Verifier.readFlavorAndVerifier(Verifier.java:45) > at org.apache.hadoop.oncrpc.RpcDeniedReply.read(RpcDeniedReply.java:50) > at org.apache.hadoop.oncrpc.RpcReply.read(RpcReply.java:67) > at org.apache.hadoop.oncrpc.SimpleUdpClient.run(SimpleUdpClient.java:71) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:130) > at org.apache.hadoop.oncrpc.RpcProgram.register(RpcProgram.java:101) > at org.apache.hadoop.mount.MountdBase.start(MountdBase.java:83) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startServiceInternal(Nfs3.java:56) > at org.apache.hadoop.hdfs.nfs.nfs3.Nfs3.startService(Nfs3.java:69) > at > org.apache.hadoop.hdfs.nfs.nfs3.PrivilegedNfsGatewayStarter.start(PrivilegedNfsGatewayStarter.java:60) > at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method) > at > sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62) > at > sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43) > at java.lang.reflect.Method.invoke(Method.java:498) > at org.apache.commons.daemon.support.DaemonLoader.start(DaemonLoader.java:243) > 2018-03-05 12:49:32,007 INFO org.apache.hadoop.util.ExitUtil: Exiting with > status 1{noformat} > Reading the code comment for class Verifier, I think this bug existed since > its inception > {code:java} > /** > * Base class for verifier. Currently our authentication only supports 3 types > * of auth flavors: {@link RpcAuthInfo.AuthFlavor#AUTH_NONE}, {@link > RpcAuthInfo.AuthFlavor#AUTH_SYS}, > * and {@link RpcAuthInfo.AuthFlavor#RPCSEC_GSS}. Thus for verifier we only > need to handle > * AUTH_NONE and RPCSEC_GSS > */ > public abstract class Verifier extends RpcAuthInfo {{code} > The verifier should also handle AUTH_SYS too. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15529) ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in Windows
[ https://issues.apache.org/jira/browse/HADOOP-15529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509953#comment-16509953 ] Hudson commented on HADOOP-15529: - SUCCESS: Integrated in Jenkins build Hadoop-trunk-Commit #14410 (See [https://builds.apache.org/job/Hadoop-trunk-Commit/14410/]) HADOOP-15529. ContainerLaunch#testInvalidEnvVariableSubstitutionType is (inigoiri: rev 6e756e8a620e4d6dc3192986679060c52063489b) * (edit) hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/java/org/apache/hadoop/yarn/server/nodemanager/containermanager/launcher/TestContainerLaunch.java > ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in > Windows > -- > > Key: HADOOP-15529 > URL: https://issues.apache.org/jira/browse/HADOOP-15529 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Fix For: 2.10.0, 3.2.0, 3.1.1, 2.9.2, 3.0.4 > > Attachments: HADOOP-15529.v1.patch, HADOOP-15529.v2.patch > > > YARN-5219 introduced 2 unit tests designed for Unix. They currently failing > in Windows. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15483) Upgrade jquery to version 3.3.1
[ https://issues.apache.org/jira/browse/HADOOP-15483?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Sunil Govindan updated HADOOP-15483: Attachment: HADOOP-15483.008.patch > Upgrade jquery to version 3.3.1 > --- > > Key: HADOOP-15483 > URL: https://issues.apache.org/jira/browse/HADOOP-15483 > Project: Hadoop Common > Issue Type: Task >Reporter: Lokesh Jain >Assignee: Lokesh Jain >Priority: Major > Attachments: HADOOP-15483.001.patch, HADOOP-15483.002.patch, > HADOOP-15483.003.patch, HADOOP-15483.004.patch, HADOOP-15483.005.patch, > HADOOP-15483.006.patch, HADOOP-15483.007.patch, HADOOP-15483.008.patch > > > This Jira aims to upgrade jquery to version 3.3.1. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15483) Upgrade jquery to version 3.3.1
[ https://issues.apache.org/jira/browse/HADOOP-15483?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509916#comment-16509916 ] Sunil Govindan commented on HADOOP-15483: - Attaching latest patch. Earlier patch seems missed some files. [~msingh] could you please confirm whether this latest patch is fine? > Upgrade jquery to version 3.3.1 > --- > > Key: HADOOP-15483 > URL: https://issues.apache.org/jira/browse/HADOOP-15483 > Project: Hadoop Common > Issue Type: Task >Reporter: Lokesh Jain >Assignee: Lokesh Jain >Priority: Major > Attachments: HADOOP-15483.001.patch, HADOOP-15483.002.patch, > HADOOP-15483.003.patch, HADOOP-15483.004.patch, HADOOP-15483.005.patch, > HADOOP-15483.006.patch, HADOOP-15483.007.patch, HADOOP-15483.008.patch > > > This Jira aims to upgrade jquery to version 3.3.1. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15529) ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in Windows
[ https://issues.apache.org/jira/browse/HADOOP-15529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HADOOP-15529: - Resolution: Fixed Hadoop Flags: Reviewed Fix Version/s: 3.0.4 2.9.2 3.1.1 3.2.0 2.10.0 Status: Resolved (was: Patch Available) Thanks [~giovanni.fumarola] for the fix. Committed to trunk, branch-3.1, branch-3.0, branch-2, and branch-2.9. > ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in > Windows > -- > > Key: HADOOP-15529 > URL: https://issues.apache.org/jira/browse/HADOOP-15529 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Fix For: 2.10.0, 3.2.0, 3.1.1, 2.9.2, 3.0.4 > > Attachments: HADOOP-15529.v1.patch, HADOOP-15529.v2.patch > > > YARN-5219 introduced 2 unit tests designed for Unix. They currently failing > in Windows. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509906#comment-16509906 ] Giovanni Matteo Fumarola commented on HADOOP-15532: --- Thanks [~elgoiri] for opening it. Attached [^HADOOP-15532.v1.patch] with the fix. Before the fix: !BeforeFix.png! After the fix: !AfterFix.png! > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: AfterFix.png, BeforeFix.png, HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Attachment: BeforeFix.png > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: AfterFix.png, BeforeFix.png, HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Attachment: AfterFix.png > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: AfterFix.png, BeforeFix.png, HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15529) ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in Windows
[ https://issues.apache.org/jira/browse/HADOOP-15529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509903#comment-16509903 ] Íñigo Goiri commented on HADOOP-15529: -- +1 on [^HADOOP-15529.v2.patch]. Committing all the way to branch-2.9. > ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in > Windows > -- > > Key: HADOOP-15529 > URL: https://issues.apache.org/jira/browse/HADOOP-15529 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15529.v1.patch, HADOOP-15529.v2.patch > > > YARN-5219 introduced 2 unit tests designed for Unix. They currently failing > in Windows. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Issue Type: Test (was: Bug) > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Test >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Issue Type: Bug (was: Sub-task) Parent: (was: HADOOP-15475) > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Bug >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Status: Patch Available (was: Open) > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Attachment: HADOOP-15532.v1.patch > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15532.v1.patch > > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Íñigo Goiri updated HADOOP-15532: - Priority: Minor (was: Major) > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Minor > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509901#comment-16509901 ] Íñigo Goiri commented on HADOOP-15532: -- HADOOP-13380 changed the path location for the tests. > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Major > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
[ https://issues.apache.org/jira/browse/HADOOP-15532?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15532: -- Issue Type: Sub-task (was: Test) Parent: HADOOP-15475 > TestBasicDiskValidator fails with NoSuchFileException > - > > Key: HADOOP-15532 > URL: https://issues.apache.org/jira/browse/HADOOP-15532 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Íñigo Goiri >Assignee: Giovanni Matteo Fumarola >Priority: Major > > TestBasicDiskValidator is failing with NoSuchFileException once in a while. > The daily Linux build shows the error > [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Created] (HADOOP-15532) TestBasicDiskValidator fails with NoSuchFileException
Íñigo Goiri created HADOOP-15532: Summary: TestBasicDiskValidator fails with NoSuchFileException Key: HADOOP-15532 URL: https://issues.apache.org/jira/browse/HADOOP-15532 Project: Hadoop Common Issue Type: Test Reporter: Íñigo Goiri Assignee: Giovanni Matteo Fumarola TestBasicDiskValidator is failing with NoSuchFileException once in a while. The daily Linux build shows the error [here|https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/809/testReport/org.apache.hadoop.util/TestBasicDiskValidator/]. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15529) ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in Windows
[ https://issues.apache.org/jira/browse/HADOOP-15529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509890#comment-16509890 ] genericqa commented on HADOOP-15529: | (/) *{color:green}+1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 30s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 25m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 56s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 12s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 36s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 11m 36s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 51s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 23s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 33s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 53s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 9s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 31s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 6s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 22s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 19m 32s{color} | {color:green} hadoop-yarn-server-nodemanager in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 21s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 76m 21s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15529 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927488/HADOOP-15529.v2.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux 9582c664ef9f 3.13.0-137-generic #186-Ubuntu SMP Mon Dec 4 19:09:19 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 652bcbb | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_171 | | findbugs | v3.1.0-RC1 | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14759/testReport/ | | Max. process+thread count | 302 (vs. ulimit of 1) | | modules | C: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager U: hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14759/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. >
[jira] [Commented] (HADOOP-15475) Fix broken unit tests on Windows
[ https://issues.apache.org/jira/browse/HADOOP-15475?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509849#comment-16509849 ] Íñigo Goiri commented on HADOOP-15475: -- I committed YARN-8359 and we cut the failures a bunch. There are also a couple smaller JIRAs that went in. ||Run||Date||Total||Skipped||Failed||Passed|| |[459|https://builds.apache.org/job/hadoop-trunk-win/459/testReport/]|2018/05/06|19386|1465|846|17075| |[460|https://builds.apache.org/job/hadoop-trunk-win/460/testReport/]|2018/05/07|16906|1418|754|14734| |[463|https://builds.apache.org/job/hadoop-trunk-win/463/testReport/]|2018/05/10|19252|1472|540|17240| |[467|https://builds.apache.org/job/hadoop-trunk-win/467/testReport/]|2018/05/14|19317|1472|573|17272| |[469|https://builds.apache.org/job/hadoop-trunk-win/469/testReport/]|2018/05/16|19326|1472|516|17388| |[476|https://builds.apache.org/job/hadoop-trunk-win/476/testReport/]|2018/05/23|19355|1473|446|17436| |[479|https://builds.apache.org/job/hadoop-trunk-win/479/testReport/]|2018/05/26|19384|1473|375|17536| |[481|https://builds.apache.org/job/hadoop-trunk-win/481/testReport/]|2018/05/23|19409|1472|405|17532| |[483|https://builds.apache.org/job/hadoop-trunk-win/483/testReport/]|2018/05/30|19390|1473|348|17569| |[484|https://builds.apache.org/job/hadoop-trunk-win/484/testReport/]|2018/05/31|19390|1473|288|17629| |[485|https://builds.apache.org/job/hadoop-trunk-win/484/testReport/]|2018/06/01|19173|1473|270|17430| |[489|https://builds.apache.org/job/hadoop-trunk-win/489/testReport/]|2018/06/05|19394|1482|242|17670| |[495|https://builds.apache.org/job/hadoop-trunk-win/495/testReport/]|2018/06/11|19332|1482|213|17637| Right now we are at 213 failures but I've seen it as small as 202 in build [494|https://builds.apache.org/job/hadoop-trunk-win/494/testReport/]. Today we should be committing MAPREDUCE-7108 which would take us <200. > Fix broken unit tests on Windows > > > Key: HADOOP-15475 > URL: https://issues.apache.org/jira/browse/HADOOP-15475 > Project: Hadoop Common > Issue Type: Test >Affects Versions: 3.1.0, 2.9.1 >Reporter: Anbang Hu >Assignee: Anbang Hu >Priority: Minor > Labels: Windows > > There are hundreds of unit tests that fail on Windows. This JIRA tracks the > effort to fix them. > The main reasons for unit test failures on Windows are: > * Windows/Linux path formats (e.g., HDFS-10256). > * Line separator. > * Locked files: Windows locks files when opening them. > ** The typical trigger is not cleaning MiniDFSCluster leaves files locked > when a test times out; they need to be cleaned using After. > * Memory lock size. > * Slow DNS resolution (e.g., HDFS-13569). > * Locked ports (e.g., HDFS-11700) -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-14178) Move Mockito up to version 2.x
[ https://issues.apache.org/jira/browse/HADOOP-14178?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509841#comment-16509841 ] Akira Ajisaka commented on HADOOP-14178: 019 patch: Fixed TestDockerContainerRuntime This patch is ready for review. > Move Mockito up to version 2.x > -- > > Key: HADOOP-14178 > URL: https://issues.apache.org/jira/browse/HADOOP-14178 > Project: Hadoop Common > Issue Type: Sub-task > Components: test >Affects Versions: 2.9.0 >Reporter: Steve Loughran >Assignee: Akira Ajisaka >Priority: Major > Attachments: HADOOP-14178.001.patch, HADOOP-14178.002.patch, > HADOOP-14178.003.patch, HADOOP-14178.004.patch, HADOOP-14178.005-wip.patch, > HADOOP-14178.005-wip2.patch, HADOOP-14178.005-wip3.patch, > HADOOP-14178.005-wip4.patch, HADOOP-14178.005-wip5.patch, > HADOOP-14178.005-wip6.patch, HADOOP-14178.005.patch, HADOOP-14178.006.patch, > HADOOP-14178.007.patch, HADOOP-14178.008.patch, HADOOP-14178.009.patch, > HADOOP-14178.010.patch, HADOOP-14178.011.patch, HADOOP-14178.012.patch, > HADOOP-14178.013.patch, HADOOP-14178.014.patch, HADOOP-14178.015.patch, > HADOOP-14178.016.patch, HADOOP-14178.017.patch, HADOOP-14178.018.patch, > HADOOP-14178.019.patch > > > I don't know when Hadoop picked up Mockito, but it has been frozen at 1.8.5 > since the switch to maven in 2011. > Mockito is now at version 2.1, [with lots of Java 8 > support|https://github.com/mockito/mockito/wiki/What%27s-new-in-Mockito-2]. > That' s not just defining actions as closures, but in supporting Optional > types, mocking methods in interfaces, etc. > It's only used for testing, and, *provided there aren't regressions*, cost of > upgrade is low. The good news: test tools usually come with good test > coverage. The bad: mockito does go deep into java bytecodes. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-14178) Move Mockito up to version 2.x
[ https://issues.apache.org/jira/browse/HADOOP-14178?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Akira Ajisaka updated HADOOP-14178: --- Attachment: HADOOP-14178.019.patch > Move Mockito up to version 2.x > -- > > Key: HADOOP-14178 > URL: https://issues.apache.org/jira/browse/HADOOP-14178 > Project: Hadoop Common > Issue Type: Sub-task > Components: test >Affects Versions: 2.9.0 >Reporter: Steve Loughran >Assignee: Akira Ajisaka >Priority: Major > Attachments: HADOOP-14178.001.patch, HADOOP-14178.002.patch, > HADOOP-14178.003.patch, HADOOP-14178.004.patch, HADOOP-14178.005-wip.patch, > HADOOP-14178.005-wip2.patch, HADOOP-14178.005-wip3.patch, > HADOOP-14178.005-wip4.patch, HADOOP-14178.005-wip5.patch, > HADOOP-14178.005-wip6.patch, HADOOP-14178.005.patch, HADOOP-14178.006.patch, > HADOOP-14178.007.patch, HADOOP-14178.008.patch, HADOOP-14178.009.patch, > HADOOP-14178.010.patch, HADOOP-14178.011.patch, HADOOP-14178.012.patch, > HADOOP-14178.013.patch, HADOOP-14178.014.patch, HADOOP-14178.015.patch, > HADOOP-14178.016.patch, HADOOP-14178.017.patch, HADOOP-14178.018.patch, > HADOOP-14178.019.patch > > > I don't know when Hadoop picked up Mockito, but it has been frozen at 1.8.5 > since the switch to maven in 2011. > Mockito is now at version 2.1, [with lots of Java 8 > support|https://github.com/mockito/mockito/wiki/What%27s-new-in-Mockito-2]. > That' s not just defining actions as closures, but in supporting Optional > types, mocking methods in interfaces, etc. > It's only used for testing, and, *provided there aren't regressions*, cost of > upgrade is low. The good news: test tools usually come with good test > coverage. The bad: mockito does go deep into java bytecodes. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15529) ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in Windows
[ https://issues.apache.org/jira/browse/HADOOP-15529?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15529: -- Attachment: HADOOP-15529.v2.patch > ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in > Windows > -- > > Key: HADOOP-15529 > URL: https://issues.apache.org/jira/browse/HADOOP-15529 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15529.v1.patch, HADOOP-15529.v2.patch > > > YARN-5219 introduced 2 unit tests designed for Unix. They currently failing > in Windows. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15528: -- Attachment: (was: HADOOP-15529.v2.patch) > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15529) ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in Windows
[ https://issues.apache.org/jira/browse/HADOOP-15529?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509780#comment-16509780 ] Giovanni Matteo Fumarola commented on HADOOP-15529: --- Attached v2 with the ternary operator change. > ContainerLaunch#testInvalidEnvVariableSubstitutionType is not supported in > Windows > -- > > Key: HADOOP-15529 > URL: https://issues.apache.org/jira/browse/HADOOP-15529 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Minor > Attachments: HADOOP-15529.v1.patch, HADOOP-15529.v2.patch > > > YARN-5219 introduced 2 unit tests designed for Unix. They currently failing > in Windows. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15528) Deprecate ContainerLaunch#link by using FileUtil#SymLink
[ https://issues.apache.org/jira/browse/HADOOP-15528?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Giovanni Matteo Fumarola updated HADOOP-15528: -- Attachment: HADOOP-15529.v2.patch > Deprecate ContainerLaunch#link by using FileUtil#SymLink > > > Key: HADOOP-15528 > URL: https://issues.apache.org/jira/browse/HADOOP-15528 > Project: Hadoop Common > Issue Type: Sub-task >Reporter: Giovanni Matteo Fumarola >Assignee: Giovanni Matteo Fumarola >Priority: Major > Attachments: HADOOP-15528-HADOOP-15461.v1.patch, HADOOP-15529.v2.patch > > > {{ContainerLaunch}} currently uses its own utility to create links (including > winutils). > This should be deprecated and rely on {{FileUtil#SymLink}} which is already > multi-platform and pure Java. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509764#comment-16509764 ] Eric Yang commented on HADOOP-15527: In the batch file, hadoop_stop_daemon is renamed to hadoop_stop_daemon_changing_pid. Is this change necessary? > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15523) Shell command timeout given is in seconds whereas it is taken as millisec while scheduling
[ https://issues.apache.org/jira/browse/HADOOP-15523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509503#comment-16509503 ] genericqa commented on HADOOP-15523: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 14s{color} | {color:blue} Docker mode activated. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 1 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 30m 18s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 27m 44s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 22s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 10s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 37s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 33s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 58s{color} | {color:green} trunk passed {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 46s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} compile {color} | {color:green} 26m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javac {color} | {color:green} 26m 36s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 20s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} whitespace {color} | {color:red} 0m 0s{color} | {color:red} The patch has 1 line(s) that end in whitespace. Use git apply --whitespace=fix <>. Refer https://git-scm.com/docs/git-apply {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 50s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | | {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 1m 40s{color} | {color:green} the patch passed {color} | | {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 55s{color} | {color:green} the patch passed {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 8m 18s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 39s{color} | {color:green} The patch does not generate ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black}126m 16s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15523 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927440/HADOOP-15523-002.patch | | Optional Tests | asflicense compile javac javadoc mvninstall mvnsite unit shadedclient findbugs checkstyle | | uname | Linux c12468e57e40 3.13.0-139-generic #188-Ubuntu SMP Tue Jan 9 14:43:09 UTC 2018 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 2b2f672 | | maven | version: Apache Maven 3.3.9 | | Default Java | 1.8.0_171 | | findbugs | v3.1.0-RC1 | | whitespace | https://builds.apache.org/job/PreCommit-HADOOP-Build/14758/artifact/out/whitespace-eol.txt | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14758/testReport/ | | Max. process+thread count | 1489 (vs. ulimit of 1) | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14758/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This
[jira] [Commented] (HADOOP-15523) Shell command timeout given is in seconds whereas it is taken as millisec while scheduling
[ https://issues.apache.org/jira/browse/HADOOP-15523?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509379#comment-16509379 ] Bilwa S T commented on HADOOP-15523: Thanks [~nandakumar131] for looking into the issue. I have done modifications specified by you. > Shell command timeout given is in seconds whereas it is taken as millisec > while scheduling > -- > > Key: HADOOP-15523 > URL: https://issues.apache.org/jira/browse/HADOOP-15523 > Project: Hadoop Common > Issue Type: Bug >Reporter: Bilwa S T >Assignee: Bilwa S T >Priority: Major > Attachments: HADOOP-15523-001.patch, HADOOP-15523-002.patch > > > ShellBasedUnixGroupsMapping has a property > {{hadoop.security.groups.shell.command.timeout}} to control how long to wait > for the fetch groups command which can be configured in seconds. but while > scheduling the time taken is millisecs. so currently if u give value as 60s, > it is taken as 60ms. > {code:java} > timeout = conf.getTimeDuration( > CommonConfigurationKeys. > HADOOP_SECURITY_GROUP_SHELL_COMMAND_TIMEOUT_SECS, > CommonConfigurationKeys. > HADOOP_SECURITY_GROUP_SHELL_COMMAND_TIMEOUT_SECS_DEFAULT, > TimeUnit.SECONDS);{code} > Time unit given is in seconds but it should be millisecs -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15523) Shell command timeout given is in seconds whereas it is taken as millisec while scheduling
[ https://issues.apache.org/jira/browse/HADOOP-15523?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Bilwa S T updated HADOOP-15523: --- Attachment: HADOOP-15523-002.patch > Shell command timeout given is in seconds whereas it is taken as millisec > while scheduling > -- > > Key: HADOOP-15523 > URL: https://issues.apache.org/jira/browse/HADOOP-15523 > Project: Hadoop Common > Issue Type: Bug >Reporter: Bilwa S T >Assignee: Bilwa S T >Priority: Major > Attachments: HADOOP-15523-001.patch, HADOOP-15523-002.patch > > > ShellBasedUnixGroupsMapping has a property > {{hadoop.security.groups.shell.command.timeout}} to control how long to wait > for the fetch groups command which can be configured in seconds. but while > scheduling the time taken is millisecs. so currently if u give value as 60s, > it is taken as 60ms. > {code:java} > timeout = conf.getTimeDuration( > CommonConfigurationKeys. > HADOOP_SECURITY_GROUP_SHELL_COMMAND_TIMEOUT_SECS, > CommonConfigurationKeys. > HADOOP_SECURITY_GROUP_SHELL_COMMAND_TIMEOUT_SECS_DEFAULT, > TimeUnit.SECONDS);{code} > Time unit given is in seconds but it should be millisecs -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509318#comment-16509318 ] genericqa commented on HADOOP-15527: | (x) *{color:red}-1 overall{color}* | \\ \\ || Vote || Subsystem || Runtime || Comment || | {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 21s{color} | {color:blue} Docker mode activated. {color} | | {color:blue}0{color} | {color:blue} patch {color} | {color:blue} 0m 3s{color} | {color:blue} The patch file was not named according to hadoop's naming conventions. Please see https://wiki.apache.org/hadoop/HowToContribute for instructions. {color} | || || || || {color:brown} Prechecks {color} || | {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s{color} | {color:green} The patch does not contain any @author tags. {color} | | {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 0s{color} | {color:green} The patch appears to include 2 new or modified test files. {color} | || || || || {color:brown} trunk Compile Tests {color} || | {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 27m 5s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 4s{color} | {color:green} trunk passed {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 10m 58s{color} | {color:green} branch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Patch Compile Tests {color} || | {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 1m 5s{color} | {color:green} the patch passed {color} | | {color:red}-1{color} | {color:red} shellcheck {color} | {color:red} 0m 4s{color} | {color:red} The patch generated 1 new + 20 unchanged - 0 fixed = 21 total (was 20) {color} | | {color:green}+1{color} | {color:green} shelldocs {color} | {color:green} 0m 11s{color} | {color:green} There were no new shelldocs issues. {color} | | {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 0s{color} | {color:green} The patch has no whitespace issues. {color} | | {color:green}+1{color} | {color:green} shadedclient {color} | {color:green} 12m 19s{color} | {color:green} patch has no errors when building and testing our client artifacts. {color} | || || || || {color:brown} Other Tests {color} || | {color:green}+1{color} | {color:green} unit {color} | {color:green} 2m 15s{color} | {color:green} hadoop-common in the patch passed. {color} | | {color:red}-1{color} | {color:red} asflicense {color} | {color:red} 0m 24s{color} | {color:red} The patch generated 1 ASF License warnings. {color} | | {color:black}{color} | {color:black} {color} | {color:black} 56m 13s{color} | {color:black} {color} | \\ \\ || Subsystem || Report/Notes || | Docker | Client=17.05.0-ce Server=17.05.0-ce Image:yetus/hadoop:abb62dd | | JIRA Issue | HADOOP-15527 | | JIRA Patch URL | https://issues.apache.org/jira/secure/attachment/12927431/HADOOP-15527.txt | | Optional Tests | asflicense mvnsite unit shellcheck shelldocs | | uname | Linux f8d160a6775c 3.13.0-137-generic #186-Ubuntu SMP Mon Dec 4 19:09:19 UTC 2017 x86_64 x86_64 x86_64 GNU/Linux | | Build tool | maven | | Personality | /testptch/patchprocess/precommit/personality/provided.sh | | git revision | trunk / 2b2f672 | | maven | version: Apache Maven 3.3.9 | | shellcheck | v0.4.6 | | shellcheck | https://builds.apache.org/job/PreCommit-HADOOP-Build/14757/artifact/out/diff-patch-shellcheck.txt | | Test Results | https://builds.apache.org/job/PreCommit-HADOOP-Build/14757/testReport/ | | asflicense | https://builds.apache.org/job/PreCommit-HADOOP-Build/14757/artifact/out/patch-asflicense-problems.txt | | Max. process+thread count | 334 (vs. ulimit of 1) | | modules | C: hadoop-common-project/hadoop-common U: hadoop-common-project/hadoop-common | | Console output | https://builds.apache.org/job/PreCommit-HADOOP-Build/14757/console | | Powered by | Apache Yetus 0.8.0-SNAPSHOT http://yetus.apache.org | This message was automatically generated. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager &
[jira] [Commented] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=16509272#comment-16509272 ] Vinod Kumar Vavilapalli commented on HADOOP-15527: -- Uploaded a patch that loops around for HADOOP_STOP_TIMEOUT after kill -9. It also optimizes the wait before kill -9. Earlier, we would always unnecessarily wait for HADOOP_STOP_TIMEOUT after SIGTERM, even if the process disappeared in the mean while. > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Status: Patch Available (was: Open) > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org
[jira] [Updated] (HADOOP-15527) Sometimes daemons keep running even after "kill -9" from daemon-stop script
[ https://issues.apache.org/jira/browse/HADOOP-15527?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Vinod Kumar Vavilapalli updated HADOOP-15527: - Attachment: HADOOP-15527.txt > Sometimes daemons keep running even after "kill -9" from daemon-stop script > --- > > Key: HADOOP-15527 > URL: https://issues.apache.org/jira/browse/HADOOP-15527 > Project: Hadoop Common > Issue Type: Bug >Reporter: Vinod Kumar Vavilapalli >Assignee: Vinod Kumar Vavilapalli >Priority: Major > Attachments: HADOOP-15527.txt > > > I'm seeing that sometimes daemons keep running for a little while even after > "kill -9" from daemon-stop scripts. > Debugging more, I see several instances of "ERROR: Unable to kill ${pid}". > Saw this specifically with ResourceManager & NodeManager - {{yarn --daemon > stop nodemanager}}. Though it is possible that other daemons may run into > this too. > Saw this on both Centos as well as Ubuntu. -- This message was sent by Atlassian JIRA (v7.6.3#76005) - To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org For additional commands, e-mail: common-issues-h...@hadoop.apache.org