You can use NUTCH-1526 to dump segment contents and index them then to whatever 
you want, or use NUTCH-1785 to directly index a document's raw binary content 
to whatever configured or custom back-end plugin. Both will help to do what you 
need.

-----Original message-----
From: Xavier Morera<[email protected]>
Sent: Tuesday 9th December 2014 22:37
To: dev <[email protected]>
Subject: Re: Crawling a site and saving the page html exactly as is in a 
database

Hi Chris Mattmann,

We will soon test it out. Is it ok if I let you know if I have questions or 
comments?

Thanks,

Xavier

On Fri, Sep 19, 2014 at 12:31 AM, Mattmann, Chris A (3980) 
<[email protected] <mailto:[email protected]>> wrote:
Please check out NUTCH-1526 [1] which I am currently targeting for

contribution to 1.10-trunk and the 2.x branch. Id be happy to

discuss. Thank you!

Please try the patch out - it will dump out the web pages, images,

etc. all content that is stored in the segments as the original

files that were crawled.

There is a review board link here:

https://reviews.apache.org/r/9119/ <https://reviews.apache.org/r/9119/>

Cheers,

Chris

[1] https://issues.apache.org/jira/browse/NUTCH-1526 
<https://issues.apache.org/jira/browse/NUTCH-1526>

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Chris Mattmann, Ph.D.

Chief Architect

Instrument Software and Science Data Systems Section (398)

NASA Jet Propulsion Laboratory Pasadena, CA 91109 USA

Office: 168-519, Mailstop: 168-527

Email: [email protected] <mailto:[email protected]>

WWW:  http://sunset.usc.edu/~mattmann/ <http://sunset.usc.edu/~mattmann/>

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

Adjunct Associate Professor, Computer Science Department

University of Southern California, Los Angeles, CA 90089 USA

++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++++

-----Original Message-----

From: Xavier Morera <[email protected] <mailto:[email protected]>>

Reply-To: "[email protected] <mailto:[email protected]>" 
<[email protected] <mailto:[email protected]>>

Date: Thursday, September 18, 2014 3:21 PM

To: dev <[email protected] <mailto:[email protected]>>

Subject: Crawling a site and saving the page html exactly as is in a

database

>Hi,

>

>

>I have a requirement to crawl a site and save the crawled html pages into

>a database exactly as is. How complicated can this be? I need for it to

>keep all html tags.

>

>

>Also, are there any examples available that I could use as a base?

>

>

>Regards,

>Xavier

>

>

>--

>Xavier Morera

>email: [email protected] <mailto:[email protected]>

>CR: +(506) 8849 8866 <tel:%2B%28506%29%208849%208866>

>US: +1 (305) 600 4919skype: xmorera

>

>

>

>

>

--

Xavier Morera

Entrepreneur | Author & Trainer | Consultant | Developer & Scrum Master

www.xaviermorera.com <http://www.xaviermorera.com/>

office:  (305)
600-4919

cel:     +506 8849-8866

skype: xmorera

Twitter <https://twitter.com/xmorera> | LinkedIn 
<https://www.linkedin.com/in/xmorera> | Pluralsight Author 
<http://www.pluralsight.com/author/xavier-morera>


Reply via email to