Hello List, I am working on a script to parse large files, by large I mean 4 million line+ in length and when splitting on the delimiter ( ; ) there are close to 300 fields per record, but I am only interested in the first 44.
I have begin testing to see how fast the file can be read in a few different scenarios: while( <> ) { } It only takes about 6 seconds to read 4,112,220 lines. But when I introduce split as such: while (<>) { chomp($_); my @tokens = split( ";", $_ ); } It takes around 7 minutes to reach eof. I also tried using a LIMIT on split as shown below: It helped greatly by only taking a little over 1 minute but I am curious if there is a way to still improve the time to read in the file or is this a reasonable time. while (<>) { chomp($_); my @tokens = split( ";", $_, 44 ); } Thank you, Chris