The problem is that it works differently than fetch or parse.  If I say
fetch -all I get every url that has been tagged for fetch, regardless of
the batch id.  Same with parse -all.

However, when I do solrindex -all, I get the same thing as solrindex
-reindex.  What I'm looking for is the same as the fetch and parse all
where it will index all of the items that have just been parsed.

As for passing in a batchid, how can you get the batch id from a CLI
generate call?  That is why I was using -all in the first place.


On Wed, May 15, 2013 at 3:47 PM, Lewis John Mcgibbney <
[email protected]> wrote:

> Is this not desired/intended behaviour for the -all switch?
> If not then we should be able to pass an argument to only index a batchId
> or crawlId
>
>
> On Mon, May 13, 2013 at 11:06 AM, Bai Shen <[email protected]>
> wrote:
>
> > I'm using 2.x HEAD now and I'm still seeing the same problem.  When I
> call
> > solrindex -all it still indexes everything, not just the newly parsed
> > items.
> >
> >
> > On Wed, May 1, 2013 at 2:13 PM, Lewis John Mcgibbney <
> > [email protected]> wrote:
> >
> > > What version are you using?
> > > If you can I would advise you to upgrade to 2.x HEAD.
> > >
> > >
> > > On Wed, May 1, 2013 at 4:32 AM, Bai Shen <[email protected]>
> > wrote:
> > >
> > > > My crawl loop consists of the following.
> > > >
> > > > generate -topN
> > > > fetch -all
> > > > parse -all
> > > > updatedb
> > > > solrindex -all
> > > >
> > > > With the fetch and parse the -all only pulls the batch that was
> > > generated,
> > > > skipping all of the other urls.  However, the solrindex seems to be
> > > > equivalent to -reindex, commiting everything not just what hasn't
> been
> > > > sent.
> > > >
> > > > Anyone else run into this issue?
> > > >
> > > > Thanks.
> > > >
> > >
> > >
> > >
> > > --
> > > *Lewis*
> > >
> >
>
>
>
> --
> *Lewis*
>

Reply via email to