[ 
https://issues.apache.org/jira/browse/HADOOP-953?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#action_12470476
 ] 

Owen O'Malley commented on HADOOP-953:
--------------------------------------

This is already possible. Just change the HADOOP_ROOT_LOGGER variable in 
bin/hadoop-daemon.sh to WARN,DRFA or even ERROR,DRFA. (We really should make it 
a default rather than hard coding it in the script.) It will dramatically cut 
down the messages.

> huge log files
> --------------
>
>                 Key: HADOOP-953
>                 URL: https://issues.apache.org/jira/browse/HADOOP-953
>             Project: Hadoop
>          Issue Type: Improvement
>    Affects Versions: 0.10.1
>         Environment: N/A
>            Reporter: Andrew McNabb
>
> On our system, it's not uncommon to get 20 MB of logs with each MapReduce 
> job.  It would be very helpful if it were possible to configure Hadoop 
> daemons to write logs only when major things happen, but the only conf 
> options I could find are for increasing the amount of output.  The disk is 
> really a bottleneck for us, and I believe that short jobs would run much more 
> quickly with less disk usage.  We also believe that the high disk usage might 
> be triggering a kernel bug on some of our machines, causing them to crash.  
> If the 20 MB of logs went down to 20 KB, we would probably still have all of 
> the information we needed.
> Thanks!

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to