Once more, this time to list http://aspn.activestate.com/ASPN/docs/ActivePython/2.5/diveintopython/html/html_processing/extracting_data.html has some sample recipes that should give you a good starting point.
On 22/07/07, Sean Murphy <[EMAIL PROTECTED]> wrote:
All. I wish to extract specific links from a web page. A part of the requirement is to be able to drill down about three to four levels to extract the information. The first page shall have 26 to 30 links I want to extract. The levels beneath the first page is a lot higher. I only want the text, not the underlying HTML code, but if I get the URL path I can deal. Wget grabs to much information for my use. I only know the higher levels of Perl and I am starting to learn Ruby. So my coding under Linux is not very advance at all. Sean Murphy Skype: smurf20005 Life is a challenge, treat it that way. -- SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/ Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html
-- There is nothing more worthy of contempt than a man who quotes himself - Zhasper, 2004 -- SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/ Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html