Hi, > thanks. I think you might have anwsered my next question already:
> what I'm doing is like: wget url > html; lynx html > text Ah, skip the lynx step. Just work with the html directly. All the tools (sed, awk, grep) can work directly with html. > even it's somewhat outside of my abilities, I guess I could try Perl? > (if I can find some sample code) To some extent it depends on how variable the original page is. Once you skip the lynx step, you might even find that it's trivial. For instance, the values you want might all be in a tag with a single, fixed "class" or "id"; you can just grep for that, then strip off the HTML with sed. This is especially likely if they're all from the same website, generated from a single template. Jiri -- Jiří Baum <[email protected]> Sabik Software Solutions Pty Ltd 0413 183 117 http://www.baum.com.au/sabik -- SLUG - Sydney Linux User's Group Mailing List - http://slug.org.au/ Subscription info and FAQs: http://slug.org.au/faq/mailinglists.html
