Ah, that is Hadoop logging, which is everywhere. Check where you can configure 
Hadoop's logging.
M.

 
 
-----Original message-----
> From:shubham.gupta <[email protected]>
> Sent: Thursday 25th August 2016 6:27
> To: [email protected]
> Subject: Re: Application creating huge amount of logs : Nutch 2.3.1 + Hadoop 
> 2.7.1
> 
> Hey
> 
> Logs are created when spills of map job are created during the FETCH job 
> and are stored in /home/hadoop/nodelogs/usercache/root/appcache. The 
> total size of logs sums up to over 13GB which occupies a lot of disk 
> space of the datanode and I have to delete those logs for smooth 
> functioning of nutch.
> 
> Also, I am unclear as to which parameter should be changed in the 
> log4j.properties to reduce this size.
> 
> Shubham Gupta
> 
> On 08/24/2016 05:20 PM, Markus Jelsma wrote:
> > If it is Nutch logging, change its level in conf/log4j.properties. It can 
> > also be Hadoop logging.
> > M.
> >   
> > -----Original message-----
> >> From:shubham.gupta <[email protected]>
> >> Sent: Tuesday 23rd August 2016 8:15
> >> To: [email protected]
> >> Subject: Application creating huge amount of logs : Nutch 2.3.1 + Hadoop 
> >> 2.7.1
> >>
> >> Hey
> >>
> >> I have integrated Nutch 2.3.1 with Hadoop 2.7.1, and the fetcher.parse
> >> property is set TRUE and the database used is MongoDB. While the map job
> >> of nutch runs, it creates a huge size of nodelogs over 13GB in size. And
> >> the cause of such huge amount of files in unknown. Any suggestion would
> >> help.
> >>
> >> Thanks in advance.
> >>
> >> Shubham Gupta
> >>
> 
> 

Reply via email to