[
https://issues.apache.org/jira/browse/HADOOP-953?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12632523#action_12632523
]
Edward J. Yoon commented on HADOOP-953:
---------------------------------------
I would recommend the HADOOP-3210 idea for an huge log files.
> huge log files
> --------------
>
> Key: HADOOP-953
> URL: https://issues.apache.org/jira/browse/HADOOP-953
> Project: Hadoop Core
> Issue Type: Improvement
> Affects Versions: 0.10.1
> Environment: N/A
> Reporter: Andrew McNabb
>
> On our system, it's not uncommon to get 20 MB of logs with each MapReduce
> job. It would be very helpful if it were possible to configure Hadoop
> daemons to write logs only when major things happen, but the only conf
> options I could find are for increasing the amount of output. The disk is
> really a bottleneck for us, and I believe that short jobs would run much more
> quickly with less disk usage. We also believe that the high disk usage might
> be triggering a kernel bug on some of our machines, causing them to crash.
> If the 20 MB of logs went down to 20 KB, we would probably still have all of
> the information we needed.
> Thanks!
--
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.