Hi,

> thanks. I think you might have anwsered my next question already:

> what I'm doing is like: wget url > html; lynx html > text

Ah, skip the lynx step. Just work with the html directly.

All the tools (sed, awk, grep) can work directly with html.

> even it's somewhat outside of my abilities, I guess I could try Perl?
> (if I can find some sample code)

To some extent it depends on how variable the original page is.

Once you skip the lynx step, you might even find that it's trivial. For
instance, the values you want might all be in a tag with a single, fixed
"class" or "id"; you can just grep for that, then strip off the HTML
with sed.

This is especially likely if they're all from the same website,
generated from a single template.


Jiri
-- 
Jiří Baum <[email protected]>
Sabik Software Solutions Pty Ltd
0413 183 117
http://www.baum.com.au/sabik
-- 
SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/
Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html

Reply via email to