Please direct your question to the list, no to me personally. When I have
time on the list, I answer what I can... if I start answering questions off
a) others don't benefit
b) others can't join in
c) I consider it a billable, personalised service to you
Parsing HTML is pretty complex stuff, and I haven't finished anything yet.
For me, I wanted to get all <A ...> tags, then get all the elements out of
it, change some, and glue it all back together, AND insert it back into the
HTML page in the right spot. Not easy! However it does help if you know
exactly what you want, and can limit the requirements.
HTML is a bitch to parse. There are tonnes options for attributes of a
href = "http://foo.com"
href = 'http://foo.com'
etc. Then there's the fact that you may only want to search for links
within your own domain, or only those outside your domain, and will you
want to keep query strings, port numbers, etc etc intact?
In your case, you might only want an array with a list of all the http://
URLs found in the page... eg:
$urls => 'http://www.indent.com.au/home.php';
$urls => 'http://www.indent.com.au/about.php';
$urls => 'http://google.com';
OR, you might also want the
...and whatever else in there as well.
OR you might want the whole hyperlink '<A' through to '>'.
It's not up to us to *guess* what you want to achieve. It's up to you to
CLEARLY SPECIFY the requirements.
1. what is your source?
a HTML file on disk?
a http:// URL on the web?
2. what specifically do you want to search for?
3. what do you want to do with the matches?
You need to be a LOT more specific.
on 19/08/02 5:49 AM, Roman Be¯lonohy´ ([EMAIL PROTECTED]) wrote:
> Hi man,
> I'd need to do a script which return all hyperlinks found on some webpage.
> You wrote about the similar script you're writing.
> Could you show me it or advice some approach ?
> thank you,
PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php