Kay- Wednesday, March 24, 2010, 11:01:15 PM, you wrote:
> If the numbers were remotely close, ie if I wanted to prove it was faster to > create 99% than delete 1%, then yes, I totally agree, but the variation in > my random 10% is insignificant compared to the slowness of using 'delete > line x', especially on really large data sets. This is getting a bit off topic, but you can't prove the above. If you used a counter instead of a random number, I'd say you were accurate. But you're deliberately calling the random number generator, which has a finite chance of returning all ones. > Yes I could empty these but I don't see they match your statement 'the > variable will be continuing to increase'? Memory, now that's another issue, <g> that's because I missed the fact that your revised code uses different variables for the three tests... > What I'm after is the fastest way to take a HUGE amount of data and reduce > it by roughly 5-10%. The repeat for each code I supplied seems to do that, > if anyone has any other code that is faster PLEASE provide. As I said at the > beginning of this thread I'm dealing with two nested repeat loops each > dealing with 1.4 million cycles!! Having something tenfold slower is NOT > what I'm after but is what I was seeing because I was using delete. Is it possible to refactor your code to use something other than nested repeat loops? My guess from what you've posted so far is that there isn't, but I'm grasping at straws... I don't know what you're aiming at for an end result - could a filter command help out? -- -Mark Wieder [email protected] _______________________________________________ use-revolution mailing list [email protected] Please visit this url to subscribe, unsubscribe and manage your subscription preferences: http://lists.runrev.com/mailman/listinfo/use-revolution
