Hi again
> i know the process is not stuck.. and the process is running because i > turned on the hadoop logs and i can see logs being written to it... > I'm not sure how to check if the task is completely stuck or not... > run jps to identify the process id then *jstack id* several times to see if it is blocked at the same place how much space do you have left on the partition where /tmp is mounted? J. > Below is the sample log as i'm sending this email.... Its been on the > updatedb process for the last 19 days and the it has been generating > debug logs similar to this........ > > Has anyone else has this same issue before... > > > 2009-11-02 13:34:21,112 DEBUG mapred.Counters - Creating group > org.apache.hadoop.mapred.Task$FileSystemCounter with bundle > 2009-11-02 13:34:21,112 DEBUG mapred.Counters - Adding LOCAL_READ > 2009-11-02 13:34:21,112 DEBUG mapred.Counters - Adding LOCAL_WRITE > 2009-11-02 13:34:21,112 DEBUG mapred.Counters - Creating group > org.apache.hadoop.mapred.Task$Counter with bundle > 2009-11-02 13:34:21,112 DEBUG mapred.Counters - Adding > COMBINE_OUTPUT_RECORDS > 2009-11-02 13:34:21,112 DEBUG mapred.Counters - Adding MAP_INPUT_RECORDS > 2009-11-02 13:34:21,113 DEBUG mapred.Counters - Adding MAP_OUTPUT_BYTES > 2009-11-02 13:34:21,113 DEBUG mapred.Counters - Adding MAP_INPUT_BYTES > 2009-11-02 13:34:21,113 DEBUG mapred.Counters - Adding MAP_OUTPUT_RECORDS > 2009-11-02 13:34:21,113 DEBUG mapred.Counters - Adding > COMBINE_INPUT_RECORDS > 2009-11-02 13:34:21,643 INFO mapred.JobClient - map 93% reduce 0% > 2009-11-02 13:34:22,121 INFO mapred.MapTask - Spilling map output: > record full = true > 2009-11-02 13:34:22,121 INFO mapred.MapTask - bufstart = 10420198; > bufend = 13893589; bufvoid = 99614720 > 2009-11-02 13:34:22,121 INFO mapred.MapTask - kvstart = 131070; kvend > = 65533; length = 327680 > 2009-11-02 13:34:22,427 INFO mapred.MapTask - Finished spill 3 > 2009-11-02 13:34:23,301 INFO mapred.MapTask - Starting flush of map output > 2009-11-02 13:34:23,384 INFO mapred.MapTask - Finished spill 4 > 2009-11-02 13:34:23,390 DEBUG mapred.MapTask - > MapId=attempt_local_0001_m_000003_0 Reducer=0Spill =0(0,224, 228) > 2009-11-02 13:34:23,390 DEBUG mapred.MapTask - > MapId=attempt_local_0001_m_000003_0 Reducer=0Spill =1(0,242, 246) > 2009-11-02 13:34:23,390 DEBUG mapred.MapTask - > MapId=attempt_local_0001_m_000003_0 Reducer=0Spill =2(0,242, 246) > 2009-11-02 13:34:23,390 DEBUG mapred.MapTask - > MapId=attempt_local_0001_m_000003_0 Reducer=0Spill =3(0,242, 246) > 2009-11-02 13:34:23,390 DEBUG mapred.MapTask - > MapId=attempt_local_0001_m_000003_0 Reducer=0Spill =4(0,242, 246) > 2009-11-02 13:34:23,390 INFO mapred.Merger - Merging 5 sorted segments > 2009-11-02 13:34:23,392 INFO mapred.Merger - Down to the last > merge-pass, with 5 segments left of total size: 1192 bytes > 2009-11-02 13:34:23,393 INFO mapred.MapTask - Index: (0, 354, 358) > 2009-11-02 13:34:23,394 INFO mapred.TaskRunner - > Task:attempt_local_0001_m_000003_0 is done. And is in the process of > commiting > 2009-11-02 13:34:23,395 DEBUG mapred.TaskRunner - > attempt_local_0001_m_000003_0 Progress/ping thread exiting since it > got interrupted > 2009-11-02 13:34:23,395 INFO mapred.LocalJobRunner - > > file:/opt/tsweb/nutch-1.0/newHyperseekCrawl/db/current/part-00000/data:100663296+33554432 > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Creating group > org.apache.hadoop.mapred.Task$FileSystemCounter with bundle > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding LOCAL_READ > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding LOCAL_WRITE > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Creating group > org.apache.hadoop.mapred.Task$Counter with bundle > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding > COMBINE_OUTPUT_RECORDS > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding MAP_INPUT_RECORDS > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding MAP_OUTPUT_BYTES > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding MAP_INPUT_BYTES > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding MAP_OUTPUT_RECORDS > 2009-11-02 13:34:23,396 DEBUG mapred.Counters - Adding > COMBINE_INPUT_RECORDS > 2009-11-02 13:34:23,397 INFO mapred.TaskRunner - Task > 'attempt_local_0001_m_000003_0' done. > 2009-11-02 13:34:23,397 DEBUG mapred.SortedRanges - currentIndex 0 0:0 > 2009-11-02 13:34:23,397 DEBUG conf.Configuration - > java.io.IOException: config(config) > at > org.apache.hadoop.conf.Configuration.<init>(Configuration.java:192) > at org.apache.hadoop.mapred.JobConf.<init>(JobConf.java:139) > at > org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:132) > > 2009-11-02 13:34:23,398 DEBUG mapred.MapTask - Writing local split to > /tmp/hadoop-root/mapred/local/localRunner/split.dta > 2009-11-02 13:34:23,451 DEBUG mapred.TaskRunner - > attempt_local_0001_m_000004_0 Progress/ping thread started > 2009-11-02 13:34:23,452 INFO mapred.MapTask - numReduceTasks: 1 > 2009-11-02 13:34:23,453 INFO mapred.MapTask - io.sort.mb = 100 > 2009-11-02 13:34:23,644 INFO mapred.JobClient - map 100% reduce 0% > > Mathan > On Mon, Nov 2, 2009 at 4:11 AM, Andrzej Bialecki <a...@getopt.org> wrote: > > Kalaimathan Mahenthiran wrote: > >> > >> I forgot to add the detail... > >> > >> The segment i'm trying to do updatedb on has 1.3 millions urls fetched > >> and 1.08 million urls parsed.. > >> > >> Any help related to this would be appreciated... > >> > >> > >> On Sun, Nov 1, 2009 at 11:53 PM, Kalaimathan Mahenthiran > >> <matha...@gmail.com> wrote: > >>> > >>> hi everyone > >>> > >>> I'm using nutch 1.0. I have fetched successfully and currently on the > >>> updatedb process. I'm doing updatedb and its taking so long. I don't > >>> know why its taking this long. I have a new machine with quad core > >>> processor and 8 gb of ram. > >>> > >>> I believe this system is really good in terms of processing power. I > >>> don't think processing power is the problem here. I noticed that all > >>> the ram is getting using up. close to 7.7gb by the updatedb process. > >>> The computer is becoming is really slow. > >>> > >>> The updatedb process has been running for the last 19 days continually > >>> with the message merging segment data into db.. Does anyone know why > >>> its taking so long... Is there any configuration setting i can do to > >>> increase the speed of the updatedb process... > > > > First, this process normally takes just a few minutes, depending on the > > hardware, and not several days - so something is wrong. > > > > * do you run this in "local" or pseudo-distributed mode (i.e. running a > real > > jobtracker and tasktracker?) Try the pseudo-distributed mode, because > then > > you can monitor the progress in the web UI. > > > > * how many reduce tasks do you have? with large updates it helps if you > run > >> 1 reducer, to split the final sorting. > > > > * if the task appears to be completely stuck, please generate a thread > dump > > (kill -SIGQUIT) and see where it's stuck. This could be related to > > urlfilter-regex or urlnormalizer-regex - you can identify if these are > > problematic by removing them from the config and re-running the > operation. > > > > * minor issue - when specifying the path names of segments and crawldb, > do > > NOT append the trailing slash - it's not harmful in this particular case, > > but you could have a nasty surprise when doing e.g. copy / mv operations > ... > > > > -- > > Best regards, > > Andrzej Bialecki <>< > > ___. ___ ___ ___ _ _ __________________________________ > > [__ || __|__/|__||\/| Information Retrieval, Semantic Web > > ___|||__|| \| || | Embedded Unix, System Integration > > http://www.sigram.com Contact: info at sigram dot com > > > > > -- DigitalPebble Ltd http://www.digitalpebble.com