El Jue 25 Sep 2003 13:08, Dave [Hawk-Systems] escribió:
> we have a number of csv dumps that occasionally have to be used to update
> tables in a postgres database...
> normally this is done by uploading the file and whomever running a php
> based parser which opens the file into an array, does a split on the comma,
> then executes an insert or update to the database with the split array
> values.
> we have a new upload that involves a 143mb file, compared to previous
> upload sizes of 10k to 150k.  Is there a more efficient way to handle this
> rather than having PHP load the entire file into an array (which would have
> to be in memory during the operation, correct?).

I guess you are using file() to read the entire file. Try using fopen() and 
read line by line with fread().

Porqué usar una base de datos relacional cualquiera,
si podés usar PostgreSQL?
Martín Marqués                  |        [EMAIL PROTECTED]
Programador, Administrador, DBA |       Centro de Telematica
                       Universidad Nacional
                            del Litoral

PHP Database Mailing List (http://www.php.net/)
To unsubscribe, visit: http://www.php.net/unsub.php

Reply via email to