Thanks for your help.

I am copying here the content.

# Licensed to the Apache Software Foundation (ASF) under one or more
# contributor license agreements.  See the NOTICE file distributed with
# this work for additional information regarding copyright ownership.
# The ASF licenses this file to You under the Apache License, Version 2.0
# (the "License"); you may not use this file except in compliance with
# the License.  You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.


# The default url filter.
# Better for whole-internet crawling.

# Each non-comment, non-blank line contains a regular expression
# prefixed by '+' or '-'.  The first matching pattern in the file
# determines whether a URL is included or ignored.  If no pattern
# matches, the URL is ignored.

# skip file: ftp: and mailto: urls
-^(ftp|mailto):

# skip image and other suffixes we can't yet parse
# for a more extensive coverage use the urlfilter-suffix plugin
-\.(gif|GIF|jpg|JPG|png|PNG|ico|ICO|css|CSS|sit|SIT|eps|EPS|wmf|WMF|zip|ZIP|ppt|PPT|mpg|MPG|xls|XLS|gz|GZ|rpm|RPM|tgz|TGZ|mov|MOV|exe|EXE|jpeg|JPEG|bmp|BMP|js|JS)$

# skip URLs containing certain characters as probable queries, etc.
#-[?*!@=]

# skip URLs with slash-delimited segment that repeats 3+ times, to break
loops
#-.*(/[^/]+)/[^/]+\1/[^/]+\1/

# accept anything else
+.



On Sun, Jun 30, 2013 at 6:38 PM, h b <[email protected]> wrote:

> What does your conf/regex_urlfilters
> file contain?
> Did you change this file?
> On Jun 30, 2013 5:10 AM, "Sznajder ForMailingList" <
> [email protected]>
> wrote:
>
> > Thanks a lot for your help
> >
> > however, I still did not resovle this issue...
> >
> >
> > I attach there the logs after 2 rounds of
> > "generate/fetch/parse/updatedb"
> >
> > the DB still contains only the seed url , not more...
> >
> >
> >
> >
> > On Thu, Jun 27, 2013 at 12:37 AM, Lewis John Mcgibbney <
> > [email protected]> wrote:
> >
> >> Try each step with a crawlId and see if this provides you with better
> >> results.
> >>
> >> Unless you truncated all data between Nutch tasks then you should be
> >> seeing
> >> more data in HBase.
> >> As Tejas asked... what do the logs say?
> >>
> >>
> >> On Wed, Jun 26, 2013 at 3:40 AM, Sznajder ForMailingList <
> >> [email protected]> wrote:
> >>
> >> > Hi Lewis,
> >> >
> >> > Thanks for your reply
> >> >
> >> > I just set the values:
> >> >
> >> >  gora.datastore.default=org.apache.gora.hbase.store.HBaseStore
> >> >
> >> >
> >> > I already removed the Hbase table in the past. Can it be a cause?
> >> >
> >> > Benjamin
> >> >
> >> >
> >> >
> >> >
> >> > On Tue, Jun 25, 2013 at 7:34 PM, Lewis John Mcgibbney <
> >> > [email protected]> wrote:
> >> >
> >> > > Have you changed from the default MemStore gora storage to something
> >> > else?
> >> > >
> >> > > On Tuesday, June 25, 2013, Sznajder ForMailingList <
> >> > > [email protected]>
> >> > > wrote:
> >> > > > thanks Tejas
> >> > > >
> >> > > > Yes, I cheecked the logs and  no Error appears in them
> >> > > >
> >> > > > I let the http.content.limit and parser.html.impl with their
> default
> >> > > > value...
> >> > > >
> >> > > > Benajmin
> >> > > >
> >> > > >
> >> > > > On Tue, Jun 25, 2013 at 6:14 PM, Tejas Patil <
> >> [email protected]
> >> > > >wrote:
> >> > > >
> >> > > >> Did you check the logs (NUTCH_HOME/logs/hadoop.log) for any
> >> exception
> >> > or
> >> > > >> error messages ?
> >> > > >> Also you might have a look at these configs in nutch-site.xml
> >> (default
> >> > > >> values are in nutch-default.xml):
> >> > > >> http.content.limit and parser.html.impl
> >> > > >>
> >> > > >>
> >> > > >> On Tue, Jun 25, 2013 at 7:04 AM, Sznajder ForMailingList <
> >> > > >> [email protected]> wrote:
> >> > > >>
> >> > > >> > Hello
> >> > > >> >
> >> > > >> > I installed Nutch 2.2 on my linux machine.
> >> > > >> >
> >> > > >> > I defined the seed directory with one file containing:
> >> > > >> > http://en.wikipedia.org/
> >> > > >> > http://edition.cnn.com/
> >> > > >> >
> >> > > >> >
> >> > > >> > I ran the following:
> >> > > >> > sh bin/nutch inject ~/DataExplorerCrawl_gpfs/seed/
> >> > > >> >
> >> > > >> > After this step:
> >> > > >> > the call
> >> > > >> > -bash-4.1$ sh bin/nutch readdb -stats
> >> > > >> >
> >> > > >> > returns
> >> > > >> > TOTAL urls:     2
> >> > > >> > status 0 (null):        2
> >> > > >> > avg score:      1.0
> >> > > >> >
> >> > > >> >
> >> > > >> > Then, I ran the following:
> >> > > >> > bin/nutch generate -topN 10
> >> > > >> > bin/nutch fetch -all
> >> > > >> > bin/nutch parse -all
> >> > > >> > bin/nutch updatedb
> >> > > >> > bin/nutch generate -topN 1000
> >> > > >> > bin/nutch fetch -all
> >> > > >> > bin/nutch parse -all
> >> > > >> > bin/nutch updatedb
> >> > > >> >
> >> > > >> >
> >> > > >> > However, the stats call after these steps is still:
> >> > > >> > the call
> >> > > >> > -bash-4.1$ sh bin/nutch readdb -stats
> >> > > >> > status 5 (status_redir_perm):   1
> >> > > >> > max score:      2.0
> >> > > >> > TOTAL urls:     3
> >> > > >> > avg score:      1.3333334
> >> > > >> >
> >> > > >> >
> >> > > >> >
> >> > > >> > Only 3 urls?!
> >> > > >> > What do I miss?
> >> > > >> >
> >> > > >> > thanks
> >> > > >> >
> >> > > >> > Benjamin
> >> > > >> >
> >> > > >>
> >> > > >
> >> > >
> >> > > --
> >> > > *Lewis*
> >> > >
> >> >
> >>
> >>
> >>
> >> --
> >> *Lewis*
> >>
> >
> >
>

Reply via email to