Hi Robert,

You could look at pageDoc & pageScore to improve things for deep paging (
http://wiki.apache.org/solr/CommonQueryParameters#pageDoc_and_pageScore).

Regards,
Aloke

On Sat, Dec 8, 2012 at 8:08 AM, Upayavira <u...@odoko.co.uk> wrote:

> Yes, expected.
>
> When it does a search for the first, say, 10 results, it must scan
> through all docs, recording just the highest ten scoring ones.
>
> To find documents 1000 to 1010, it must scan through all docs, recording
> the best scoring 1010 documents, and then discard the first 1000. This
> is much more expensive. Try it on google, they won't let you go beyond
> around 900 pages or such (or is it 900 results?)
>
> Upayavira
>
> On Sat, Dec 8, 2012, at 01:10 AM, Petersen, Robert wrote:
> > Hi guys,
> >
> >
> > Sometimes we get a bot crawling our search function on our retail web
> > site.  The ebay crawler loves to do this (Request.UserAgent:
> > Terapeakbot).  They just do a star search and then iterate through page
> > after page.  I've noticed that when they get to higher page numbers like
> > page 9000, the searches are taking more than 20 seconds.  Is this
> > expected behavior?  We're requesting standard facets with the search as
> > well as incorporating boosting by function query.  Our index is almost 15
> > million docs now and we're on Solr 3.6.1, this isn't causing any errors
> > to occur at the solr layer but our web layer times out the search after
> > 20 seconds and logs the exception.
> >
> >
> >
> > Thanks
> >
> > Robi
> >
>

Reply via email to