Hi Edward,
Thanks for the response. So, with 5 relaxation data sets, only tm8
should be removed -- no need to remove m8 as well? Also, if only 4
relaxation data sets were available, could {tm6-8 and m8} be removed
to use the full_analysis.py protocol?
Thanks,
Doug
On Jan 10, 2008, at 1:31 PM, Edward d'Auvergne wrote:
> Hi,
>
> If you have 5 relaxation data sets, you can use the full_analysis.py
> script but you will need to remove model tm8. This is the only model
> with 6 parameters and doing the analysis without it might just work
> (the other tm0 to tm9 models may compensate adequately).
>
> I've looked at the script and it seems fine. I think the issue is
> that the model-free problem is not simply an optimisation issue. It
> is the simultaneous combination of global optimisation (mathematics)
> with model selection (statistics). You are not searching for the
> global minimum in one space, as in a normal optimisation problem, but
> for the global minimum across and enormous number of spaces
> simultaneously. I formulated the totality of this problem using set
> theory here http://www.rsc.org/Publishing/Journals/MB/article.asp?
> doi=b702202f
> or in my PhD thesis at
> http://eprints.infodiv.unimelb.edu.au/archive/00002799/. In your
> script, the CONV_LOOP flag allows you to automatically loop over many
> global optimisations. Each iteration of the loop is the mathematical
> optimisation part. But the entire loop itself allows for the sliding
> between these different spaces. Note that this is a very, very
> complex problem involving huge numbers spaces or universes, each of
> which consists of a large number of dimensions. There was a mistake
> in my Molecular BioSystems paper in that the number of spaces is
> really equal to n*m^l where n is the number of diffusion models, m is
> the number of model-free models (10 if you use m0 to m9), and l is the
> number of spin systems. So if you have 200 residues, the number of
> spaces is on the order of 10 to the power of 200. The number of
> dimensions for this system is on the order of 10^2 to 10^3. So the
> problem is to find the 'best' minimum in 10^200 spaces, each
> consisting of 10^2 to 10^3 dimensions (the universal solution or the
> solution in the universal set). The problem is just a little more
> complex than most people think!!!
>
> So, my opinion of the problem is that the starting position of one of
> the 2 solutions is not good. In one (or maybe both) you are stuck in
> the wrong universe (out of billions of billions of billions of
> billions....). And you can't slide out of that universe using the
> looping procedure in your script. That's why I designed the new
> model-free analysis protocol used by the full_analysis.py script
> (http://www.springerlink.com/content/u170k174t805r344/?
> p=23cf5337c42e457abe3e5a1aeb38c520&pi=3
> or the thesis again). The aim of this new protocol is so that you
> start in a universe much closer to the one with the universal solution
> that you can ever get with the initial diffusion tensor estimate.
> Then you can easily slide, in less than 20 iterations, to the
> universal solution using the looping procedure. For a published
> example of this type of failure, see the section titled "Failure of
> the diffusion seeded paradigm" in the previous link to the
> "Optimisation of NMR dynamic models II" paper.
>
> Does this description make sense? Does it answer all your questions?
>
> Regards,
>
> Edward
>
>
>
> On Jan 10, 2008 5:49 PM, Douglas Kojetin
> <[EMAIL PROTECTED]> wrote:
>> Hi All,
>>
>> I am working with five relaxation data sets (r1, r2 and noe at 400
>> MHz; r1 and r2 and 600 MHz), and therefore cannot use the
>> full_analysis.py protocol. I have obtained estimates for tm,
>> Dratio, theta and phi using Art Palmer's quadric_diffusion program.
>> I modified the full_analysis.py protocol to optimize a prolate tensor
>> using these estimates (attached file: mod.py). I have performed the
>> optimization of the prolate tensor using either (1) my original
>> structure or (2) the same structure rotated and translated by the
>> quadric_diffusion program. It seems that relax does not converge to
>> a single global optimum, as different values of tm, Da, theta and phi
>> are reported.
>>
>> Using my original structure:
>> #tm = 6.00721299718e-09
>> #Da = 14256303.3975
>> #theta = 11.127323614211441
>> #phi = 62.250251959733312
>>
>> Using the rotated/translated structure by the quadric_diffusion
>> program:
>> #tm = 5.84350638161e-09
>> #Da = 11626835.475
>> #theta = 8.4006873071400197
>> #phi = 113.6068898953142
>>
>> The only difference between the two calculations is the orientation
>> of the input PDB structure file. For another set of five rates
>> (different protein), there is a >0.3 ns difference in the converged
>> tm values.
>>
>> Is my modified protocol (in mod.py) setup properly? Or is this a
>> more complex issue in the global optimization? In previous attempts,
>> I've also noticed that separate runs with differences in the
>> estimates for Dratio, theta and phi also converge to different
>> optimized diffusion tensor variables.
>>
>> Doug
>>
>>
>> _______________________________________________
>> relax (http://nmr-relax.com)
>>
>> This is the relax-users mailing list
>> [email protected]
>>
>> To unsubscribe from this list, get a password
>> reminder, or change your subscription options,
>> visit the list information page at
>> https://mail.gna.org/listinfo/relax-users
>>
>>
_______________________________________________
relax (http://nmr-relax.com)
This is the relax-users mailing list
[email protected]
To unsubscribe from this list, get a password
reminder, or change your subscription options,
visit the list information page at
https://mail.gna.org/listinfo/relax-users