I have a situation in which I need to load a large number of separate pdb files into a single pymol session. In this case, the number is ~150, but it could potentially be more. However, the amount of time required to load a file appears to be strongly dependent on the number of files already loaded. For example:

# of structures loaded  time to load all structures (seconds)
5       0.82
10      2.49
20      11.05
30      29.85
40      62.48
50      115.25
60      189.79
70      302.67
80      432.82
90      589.23

unfortunately, this means that to load 150 structures takes over an hour. I observe this behavior whether I am loading the structures all at once using a python script, or one at a time. In both cases, I am using the cmd.load() api function, but the built-in load command gives similar results. The structures I am loading are (nearly) identical: each has 263 residues (in a single chain); each individual pdb file is about 215KB.

I am running this on a dual 2.0 GHz G5 system with 1.5 GB memory. The long loading times are consistent between the two versions of pymol I have installed: OSX/X11 hybrid version 0.97 and MacPyMol version 0.95. During the long loading times, there is plenty of memory available, but the processor load stays at 50% (i.e. one processor on my machine is fully loaded throughout).

My gut feeling is that this situation should not be, but I don't yet understand the structure of the code well enough to debug it. Can anyone shed light on this issue?

Thanks in advance,
Ben Allen



Reply via email to