Thanks for the clarification.
--
Patrick Mézard

On 23/07/2020 19:23, Shashanka Balakuntala wrote:
Hi Patrick,
Yes, I did want to mention that it will not affect previous fetch lists. Sorry 
for the confusion.

Thanks,
Shashanka Balakuntala


On Thu, 23 Jul 2020, 22:40 Patrick Mézard, <[email protected] 
<mailto:[email protected]>> wrote:

    Hello,

    On 23/07/2020 14:37, Shashanka Balakuntala wrote:
     > Hi Patrick,
     >
     > Yes, the idea that you have suggested would work, but i do have to 
mention
     > that it might just affect the next iteration. So you can just clean the
     > last parse segment and parse again and updatedb with the plugins 
activated
     > and that would do.

    I do not follow you. How could the similarity scores of all documents be 
collected and used by updatedb without reparsing all content? From what I see, 
the similarity scorer operates during the parse phase and the score should be 
recorded in crawl_parse.

     > Deleting all the the parsed segments might not work because, because a 
url
     > with score less than threshold will not be generated or fetched, so none 
of
     > its outlinks will be fetched as well. So if you just delete parse segment
     > and do the process, it would mean the all the already fetched segments 
will
     > not be impacted. So it will update the scoring, if you just need the 
score
     > for something else, please do go ahead with this.

    Again, I am confused. My mental model is:

    - Delete and reparse everything. I means similarity scores are taken in 
account and included all segments crawl_parse.
    - Run updatedb on all segments. CrawlDatum entries will be gathered by 
"url" and some final score will be generated in the reduce phase, probably 
favoring the more recent score.

    Now, maybe the existing crawldb might interfere during the final merge and 
I should clear it somehow, but otherwise, once the similarity scores are 
reflected in the updated crawldb, the next generate phase will take them in 
account.

    Obviously, they will not retroactively affect the previous fetch lists. Is 
it what you tried to tell me?

    Thanks for your comments,
    --
    Patrick Mézard

     > Lets see if anyone has any other items to add or clear here.
     >
     > *Regards*
     >    Shashanka Balakuntala Srinivasa
     >
     >
     >
     > On Thu, Jul 23, 2020 at 2:40 PM Patrick Mézard <[email protected] 
<mailto:[email protected]>> wrote:
     >
     >> Hello,
     >>
     >> I have crawled a first document set using a combination of depth and 
opic
     >> scoring plugins. I would like to add the similarity scoring plugin but
     >> obviously the crawldb scores should be updated for it and following
     >> "generate" phases to be effective. Is there a recommended approach to
     >> achieve this?
     >>
     >> My current understanding is since the similarity plugin operates in 
parse
     >> phase, I would have to remove all parsed data from segments, re-parse 
them
     >> and updatedb? Would that work? Is there anything smarter?
     >>
     >> Thanks,
     >> --
     >> Patrick Mézard
     >>
     >


Reply via email to