RE: Downloading Full Copies of Web Pages

2004-10-20 Thread Karthik N S
Hi


Try

nutch   [ http://www.nutch.org/docs/en/about.html ]  underneath it uses
Lucene  :)





-Original Message-
From: Luciano Barbosa [mailto:[EMAIL PROTECTED]
Sent: Wednesday, October 20, 2004 3:06 AM
To: [EMAIL PROTECTED]
Subject: Downloading Full Copies of Web Pages


Hi folks,
I want to download full copies of web pages and storage them locally as
well the hyperlink structures as local directories. I tried to use
Lucene, but I've realized that  it doesn't have a crawler.
Does anyone know a software that make this?
Thanks,

-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]


-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]



Re: Downloading Full Copies of Web Pages

2004-10-20 Thread John Moylan
wget does this. Little point in reinventing the wheel.
Luciano Barbosa wrote:
Hi folks,
I want to download full copies of web pages and storage them locally as 
well the hyperlink structures as local directories. I tried to use 
Lucene, but I've realized that  it doesn't have a crawler.
Does anyone know a software that make this?
Thanks,

-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]
**
The information in this e-mail is confidential and may be legally privileged.
It is intended solely for the addressee. Access to this e-mail by anyone else
is unauthorised. If you are not the intended recipient, any disclosure,
copying, distribution, or any action taken or omitted to be taken in reliance
on it, is prohibited and may be unlawful.
Please note that emails to, from and within RTÉ may be subject to the Freedom
of Information Act 1997 and may be liable to disclosure.
**
-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]


Downloading Full Copies of Web Pages

2004-10-19 Thread Luciano Barbosa
Hi folks,
I want to download full copies of web pages and storage them locally as 
well the hyperlink structures as local directories. I tried to use 
Lucene, but I've realized that  it doesn't have a crawler.
Does anyone know a software that make this?
Thanks,

-
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]