Dan Friedman wrote:
I have an standalone application that I am importing data into. The data is
being imported into an array. The data is just text - formatted as XML.
Each record being imported is pretty small - about 20 lines containing [on
average] about 365 characters.
The application processes the data [creating a new key in the array] VERY
fast until it hits about the 415th record. Then, it begins to slow down.
With each new record it gets slower and slower. It will eventually finish
without error.
This is happening on Windows only. On a Mac, it whizzes through all the
records with no problem.
Is this memory? Is there something that I can do? Anyone got any
suggestions?
Performance degradation so noticeable isn't likely to be caused by a
shortage of memory, esp. with such small data sets.
You could test it by rebooting, opening your System Monitor to show
available RAM, run the routine a few times and see if memory stays
allocated beyond what would seem normal.
My hunch is that there's something about the loop that's causing the
slowdown, perhaps a "repeat with i=1 to the number of lines" or
something like that which doesn't scale well.
Can you post the portion of your code that seems to be affected?
--
Richard Gaskin
Managing Editor, revJournal
_______________________________________________________
Rev tips, tutorials and more: http://www.revJournal.com
_______________________________________________
use-revolution mailing list
[email protected]
Please visit this url to subscribe, unsubscribe and manage your subscription
preferences:
http://lists.runrev.com/mailman/listinfo/use-revolution