Can you elaborate? On May 21, 1:34 am, Galou <gael.duc...@gmail.com> wrote: > It could be due to the php xdebug module. > > Galou > > On 20 mai, 09:14, Mohd Shakir Zakaria <mohdsha...@gmail.com> wrote: > > > > > > > > > I keep on getting this error when running one of my scripts; > > > PHP Fatal error: Allowed memory size of 1073741824 bytes exhausted > > (tried to allocate 71 bytes) in ... > > lib/symfony-1.4.11/lib/plugins/sfDoctrinePlugin/lib/vendor/doctrine/Doctrin > > e/Connection/Statement.php > > on line 246, ... > > > The following is the stripped-down version of the script that's > > triggering the error; > > > public function executeImportFile(sfWebRequest $request) > > { > > ini_set('memory_limit', '1024M'); > > set_time_limit ( 0 ); > > > //more codes here... > > > $files = scandir($workspace.'/'.$directory); > > > foreach ($files as $file) { > > $path = $workspace.'/'.$directory.'/'.$file; > > > if ($file != "." && $file != "..") { > > $this->importfile($path); > > } > > } > > > } > > > protected function importfile($path){ > > ini_set('memory_limit', '1024M'); > > set_time_limit ( 0 ); > > > $connection = > > sfContext::getInstance()->getDatabaseManager()->getDatabase('doctrine')->ge > > tDoctrineConnection(); > > $connection->beginTransaction(); > > try { > > > //more codes here... > > > while ($data = $reader->read()) //reads each line of a csv file > > { > > // send the line to another private function to be processed > > // and then write to database > > $this->writewave($data); > > } > > > $connection->commit(); > > > } catch (Exception $e) { > > $connection->rollback(); > > } > > > } > > > What the script does is basically to read all the csv files in a > > folder (which contains tens of thousands of lines each), process it, > > and the write it to the database using Doctrine's transaction. > > > While I don't think I need to set the memory limit and the time limit > > in both functions, the script quits as Doctrine uses up all the the > > allocated 1GB of memory. > > > It will normally stop after processing 10 files, and allocating more > > memory will allow it to process a bit more files, and will still > > crash. > > > Is there anything that I'm missing here that's causing the memory not > > to free up after processing each files? > > > Mohd Shakir Zakariahttp://www.mohdshakir.net
-- If you want to report a vulnerability issue on symfony, please send it to security at symfony-project.com You received this message because you are subscribed to the Google Groups "symfony users" group. To post to this group, send email to symfony-users@googlegroups.com To unsubscribe from this group, send email to symfony-users+unsubscr...@googlegroups.com For more options, visit this group at http://groups.google.com/group/symfony-users?hl=en