Arv Mistry wrote:
I'm using hadoop 0.17.0. Unfortunately I cant upgrade to 0.19.0 just
yet.

I'm trying to control the amount of extraneous files. I noticed there
are the following log files produced by hadoop;

        On Slave
                - userlogs (for each map/reduce job)
                        - stderr
                        - stdout
                        - syslog
                - datanode .log file
                - datanode .out file
                - tasktracker .log file
                - tasktracker .out file

        On Master
                - jobtracker .log file
                - jobtracker .out file
                - namenode   .log file
                - namenode   .out file
                - secondarynamenode .log file
                - secondarynamenode .out file           
                - job .xml file
                - history
                        - xml file for job

        
Does any body know of how to configure hadoop so I don't have to delete
these files manually? Or just so that they don't get created at all.

For the history files, I set hadoop.job.history.user.location to none in
the hadoop-site.xml file but I still get the history files created.
Setting hadoop.job.history.user.location to "none", makes only history location specified for user. JT still has history location. History will be cleanup after a month.

Userlogs will be cleaned up after "mapred.userlog.retain.hours", by default , 24hrs.

Thanks
Amareshwari
Also I set in the log4j.properties the hadoop.root.logger=WARN but I
still see INFO messages in datanode,jobtracker etc logs

Thanks, in advance

Cheers Arv

Reply via email to