Hadoop QA commented on YARN-3350:

{color:red}-1 overall{color}.  Here are the results of testing the latest 
  against trunk revision 3ff1ba2.

    {color:green}+1 @author{color}.  The patch does not contain any @author 

    {color:red}-1 tests included{color}.  The patch doesn't appear to include 
any new or modified tests.
                        Please justify why no new tests are needed for this 
                        Also please list what manual steps were performed to 
verify this patch.

    {color:green}+1 javac{color}.  The applied patch does not increase the 
total number of javac compiler warnings.

    {color:green}+1 javadoc{color}.  There were no new javadoc warning messages.

    {color:green}+1 eclipse:eclipse{color}.  The patch built with 

    {color:green}+1 findbugs{color}.  The patch does not introduce any new 
Findbugs (version 2.0.3) warnings.

    {color:green}+1 release audit{color}.  The applied patch does not increase 
the total number of release audit warnings.

    {color:green}+1 core tests{color}.  The patch passed unit tests in 

Test results: 
Console output: https://builds.apache.org/job/PreCommit-YARN-Build/6970//console

This message is automatically generated.

> YARN RackResolver spams logs with messages at info level
> --------------------------------------------------------
>                 Key: YARN-3350
>                 URL: https://issues.apache.org/jira/browse/YARN-3350
>             Project: Hadoop YARN
>          Issue Type: Improvement
>    Affects Versions: 2.6.0
>            Reporter: Wilfred Spiegelenburg
>            Assignee: Wilfred Spiegelenburg
>         Attachments: YARN-3350.patch, yarn-RackResolver-log.txt
> When you run an application the container logs shows a lot of messages for 
> the RackResolver:
> 2015-03-10 00:58:30,483 INFO [RMCommunicator Allocator] 
> org.apache.hadoop.yarn.util.RackResolver: Resolved node175.example.com to 
> /rack15
> A real world example for a large job was generating 20+ messages in 2 
> milliseconds during a sustained period of time flooding the logs causing the 
> node to run out of disk space.

This message was sent by Atlassian JIRA

Reply via email to