On my site I paginate query results by limiting rows output to a
value, say LIMIT, and then the 2nd, 3rd pages run the same query with
$skip=LIMIT, $skip=(LIMIT*2) value posted back.  I use the following
code to "skip" these result rows, which is just fetching the next row
to an unused array.

//if there are rows to skip
if ($result_rows > $rows_to_skip) {
   while ( $rows_to_skip ) {
      // eat a row

Can I make this more efficient?  Is there a way to eliminate this data
before it leaves the mySQL server (and would it be faster)?

If it makes any difference the average result row is probably around 40-50
bytes but some of these queries can return up to 850 rows..

[EMAIL PROTECTED] ** http://mrclay.org

PHP General Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php

Reply via email to