antonv wrote:
> I have a bit of experience programming and I am pretty sure I get my parts of
> the code pretty well optimized. I made sure that in the loop I have only the
> stuff needed and I'm loading all the stuff before.
> 
> The biggest bottleneck is happening because I'm unpacking grib files to csv
> files using Degrib in command line. That operation is usually around half an

Instead of going to csv files--which are *very* inefficient to write, 
store, and then read in again--why not convert directly to netcdf, and 
then read your data in from netcdf as needed for plotting?  I suspect 
this will speed things up quite a bit.  Numpy support for netcdf is very 
good.  Of course, direct numpy-enabled access to the grib files might be 
even better, eliminating the translation phase entirely.  Have you 
looked into http://www.pyngl.ucar.edu/Nio.shtml?

Eric


> hour using no more than 50% of the processor but it maxes out the memory
> usage and it definitely is hard drive intensive as it ends up writing over 4
> GB of data. I have noticed also that on a lower spec AMD desktop this runs
> faster than on my P4 Intel Laptop, my guess being that the laptop hdd is
> 5400 rpm and the desktop is 7200 rpm.
> 
> Next step is to take all those csv files and make images from them. For this
> one I haven't dug too deep to see what is happening but it seems to be the
> other way, using the cpu a lot more while keeping the memory usage high too.
> 
> Thanks,
> Anton


------------------------------------------------------------------------------
This SF.net email is sponsored by:
High Quality Requirements in a Collaborative Environment.
Download a free trial of Rational Requirements Composer Now!
http://p.sf.net/sfu/www-ibm-com
_______________________________________________
Matplotlib-users mailing list
Matplotlib-users@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/matplotlib-users

Reply via email to