Gohaku wrote:
There are some pages I would like to extract links from but I can't access the page using the following:
ini_set('php.user_agent', 'Mozilla/5.0');
$url = "http://www.google.com/search?q=php";;
$fp = fopen($url,"r");
$buffer = fread($fp,1000000);
echo $buffer;

1) The php.ini option is user_agent, not php.user_agent. Changing that has the desired effect.


2) If you want to query google you're better off using the google SOAP API - it's legal for a start. See the following for more info...

http://www.google.com/apis/
http://www.devshed.com/Server_Side/PHP/GoogleAPI/page1.html

Also, what is the best number to use when allocating Buffer space for a URL?
Is 1000000 too much?
Thanks.
-Gohaku

3) You're better off looping to get the full contents than trying to 'guess' how big the file will be. Try the following...


<?php
    ini_set('user_agent', 'Mozilla/5.0');
    $url = 'http://www.google.com/search?q=php';
    $fp = fopen($url, "r");
    while (!feof($fp))
    {
        $buffer = fread($fp, 1024);
        echo $buffer;
    }
?>

--
Stuart

--
PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php



Reply via email to