Hi Amna, The logs usually go to hadoop.log. If you set a text direct to some text file while running the crawl you would see it there as well.
Just do a file search for hadoop.log in your nutch folder, you should be able to find it. I think another best way to make sure the plugin works is by using Junit test cases. Cheers, Abi On Fri, Feb 11, 2011 at 5:10 PM, Amna Waqar <[email protected]> wrote: > hello everybody, > I write a plugin for checking the meta tags http-equiv and do some > processing based on its content values like content="text/html ; > charset=UTF-8" > Properties HttpMetaTags = metaTags.getHttpEquivTags() ; > > for (Enumeration tagNames = HttpMetaTags.propertyNames(); > tagNames.hasMoreElements(); ) { > if (tagNames.nextElement().equals("http-equiv")) { > desc = HttpMetaTags.getProperty("http-equiv"); > if(desc == null) > LOG.info("No http-equiv tag for this page"); > else if(desc.equals("")) { > LOG.info("Found an empty http-equiv tag"); > } else { > LOG.info("Found an http-equiv tag; contents: " + desc); > } > } > } > > if((desc != null) && !(desc.equals("")) && (desc.equals("text/html; > charset=UTF-8")) ) { > LOG.info("Adding http-equiv; contents: " + desc); > parse.getData().getContentMeta().set(META_DESCRIPTION_NAME, desc); > > } > how can i see at which point this plugin is used and the location where > log,info values are output > i want to see the desc string value. > please help me. >

