Hi Edward,
For the first run listed below (tensor, # runs to convergence):
sphere, 4
prolate, 7
oblate, 9
ellipsoid, 9
For the second run listed below (tensor, # runs to convergence):
sphere, 4
prolate, 5
oblate, 9
ellipsoid, 9
Doug
On Feb 7, 2008, at 4:03 AM, Edward d'Auvergne wrote:
> Hi,
>
> Something strange is happening in your analysis. Unfortunately with
> the limited information in your post, I really cannot start to track
> it down. The fact that the local_tm and sphere runs are the same in
> both is a good sign. Do you have the number of iterations required
> for the convergence of each tensor?
>
> The full_analysis.py script should be insensitive to the input
> structure, but there is one point at which this might not be exactly
> true. The grid search for the initial tensor parameter values prior
> to minimisation is dependent. You can think of the grid search as a
> cage which stays fixed in space while the molecule spins around inside
> it. But the subsequent Newton optimisation should easily recover from
> the small differences of the increments between grid points. That
> might be a place to start looking though.
>
> The only point in relax where a random number generator is utilised is
> in the Monte Carlo simulations. It might also be worth looking at the
> Dr value in the ellipsoid optimisations. If this value is close to
> zero, then the results from the prolate and ellipsoid diffusion
> tensors may actually be almost the same. In which case, the
> differences don't matter. For understanding molecule motions, the
> model itself is of no interest - this is, of course, a simple exercise
> in the field of mathematical modelling (a mathematics library will
> show you the extent of this field). Note that it is what the model
> says about the dynamics which is of interest, not the details of the
> model itself. So two completely different models may actually say the
> same thing, but maybe with small statistically insignificant
> difference.
>
> That being said, this problem should not occur. But as I cannot do
> anything with the limited information, you may need to hunt down where
> the problem lies yourself (whether in relax, the operation of relax,
> the use of the quadric_diffusion program, or elsewhere).
>
> Regards,
>
> Edward
>
>
>
> On Jan 30, 2008 3:24 PM, Douglas Kojetin
> <[EMAIL PROTECTED]> wrote:
>> Hi Edward,
>>
>> As a followup to this, I performed two relax runs using six datasets
>> (r1, r2 and noe at two fields) with two identical structures, but one
>> had been rotated/translated using the quadric_diffusion program
>> provided by the Palmer lab. For one structure, a prolate tensor is
>> chosen, whereas an elliposid tensor is chosen for the rotated/
>> translated structure:
>>
>> ## ORIGINAL PDB
>> Run Chi2 Criterion
>> local_tm 102.67810 870.67810
>> sphere 177.96407 807.96407
>> prolate 152.70721 796.70721
>> oblate 178.61058 810.61058
>> ellipsoid 155.78475 801.78475
>>
>> The model from the run 'prolate' has been selected.
>>
>> ## ROTATED/TRANSLATED PDB
>> Run Chi2 Criterion
>> local_tm 102.67810 870.67810
>> sphere 177.96407 807.96407
>> prolate 175.13432 803.13432
>> oblate 178.61979 810.61979
>> ellipsoid 155.82168 801.82168
>>
>> The model from the run 'ellipsoid' has been selected.
>>
>>
>> There are no differences in the models selected for two of the three
>> structure-dependent runs (oblate and ellipsoid tensor runs) , but
>> there are a handful of differences in the models selected for the
>> prolate tensor runs. Is the full_analysis protocol sensitive to the
>> orientation of the input structure, or could this be a result of
>> different runs using something equivalent to different random number
>> seeds?
>>
>> Doug
>>
>>
>>
>>
>> On Jan 10, 2008, at 2:36 PM, Edward d'Auvergne wrote:
>>
>>> Yes, with 4 data sets you could remove tm6 to tm8. You would also
>>> need to remove m8. But in this situation, you will be significantly
>>> biasing the initial position (the starting universe will be further
>>> away from that of the universal solution). I don't know how well
>>> this
>>> new protocol will perform 4 data sets, i.e. this is untested, but I
>>> would be highly reluctant to trust it. The relaxation data type and
>>> field strength will be very important. I would even be wary using 5
>>> data sets, especially if the missing data set is the higher-field
>>> NOE.
>>> So I would never recommend using 4 data sets.
>>>
>>> Regards,
>>>
>>> Edward
>>>
>>>
>>> On Jan 10, 2008 8:12 PM, Douglas Kojetin
>>> <[EMAIL PROTECTED]> wrote:
>>>> Hi Edward,
>>>>
>>>> Thanks for the response. So, with 5 relaxation data sets, only tm8
>>>> should be removed -- no need to remove m8 as well? Also, if only 4
>>>> relaxation data sets were available, could {tm6-8 and m8} be
>>>> removed
>>>> to use the full_analysis.py protocol?
>>>>
>>>> Thanks,
>>>> Doug
>>>>
>>>>
>>>>
>>>> On Jan 10, 2008, at 1:31 PM, Edward d'Auvergne wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> If you have 5 relaxation data sets, you can use the
>>>>> full_analysis.py
>>>>> script but you will need to remove model tm8. This is the only
>>>>> model
>>>>> with 6 parameters and doing the analysis without it might just
>>>>> work
>>>>> (the other tm0 to tm9 models may compensate adequately).
>>>>>
>>>>> I've looked at the script and it seems fine. I think the issue is
>>>>> that the model-free problem is not simply an optimisation
>>>>> issue. It
>>>>> is the simultaneous combination of global optimisation
>>>>> (mathematics)
>>>>> with model selection (statistics). You are not searching for the
>>>>> global minimum in one space, as in a normal optimisation problem,
>>>>> but
>>>>> for the global minimum across and enormous number of spaces
>>>>> simultaneously. I formulated the totality of this problem
>>>>> using set
>>>>> theory here http://www.rsc.org/Publishing/Journals/MB/article.asp?
>>>>> doi=b702202f
>>>>> or in my PhD thesis at
>>>>> http://eprints.infodiv.unimelb.edu.au/archive/00002799/. In your
>>>>> script, the CONV_LOOP flag allows you to automatically loop over
>>>>> many
>>>>> global optimisations. Each iteration of the loop is the
>>>>> mathematical
>>>>> optimisation part. But the entire loop itself allows for the
>>>>> sliding
>>>>> between these different spaces. Note that this is a very, very
>>>>> complex problem involving huge numbers spaces or universes,
>>>>> each of
>>>>> which consists of a large number of dimensions. There was a
>>>>> mistake
>>>>> in my Molecular BioSystems paper in that the number of spaces is
>>>>> really equal to n*m^l where n is the number of diffusion models,
>>>>> m is
>>>>> the number of model-free models (10 if you use m0 to m9), and l
>>>>> is the
>>>>> number of spin systems. So if you have 200 residues, the
>>>>> number of
>>>>> spaces is on the order of 10 to the power of 200. The number of
>>>>> dimensions for this system is on the order of 10^2 to 10^3. So
>>>>> the
>>>>> problem is to find the 'best' minimum in 10^200 spaces, each
>>>>> consisting of 10^2 to 10^3 dimensions (the universal solution
>>>>> or the
>>>>> solution in the universal set). The problem is just a little more
>>>>> complex than most people think!!!
>>>>>
>>>>> So, my opinion of the problem is that the starting position of
>>>>> one of
>>>>> the 2 solutions is not good. In one (or maybe both) you are
>>>>> stuck in
>>>>> the wrong universe (out of billions of billions of billions of
>>>>> billions....). And you can't slide out of that universe using the
>>>>> looping procedure in your script. That's why I designed the new
>>>>> model-free analysis protocol used by the full_analysis.py script
>>>>> (http://www.springerlink.com/content/u170k174t805r344/?
>>>>> p=23cf5337c42e457abe3e5a1aeb38c520&pi=3
>>>>> or the thesis again). The aim of this new protocol is so that you
>>>>> start in a universe much closer to the one with the universal
>>>>> solution
>>>>> that you can ever get with the initial diffusion tensor estimate.
>>>>> Then you can easily slide, in less than 20 iterations, to the
>>>>> universal solution using the looping procedure. For a published
>>>>> example of this type of failure, see the section titled
>>>>> "Failure of
>>>>> the diffusion seeded paradigm" in the previous link to the
>>>>> "Optimisation of NMR dynamic models II" paper.
>>>>>
>>>>> Does this description make sense? Does it answer all your
>>>>> questions?
>>>>>
>>>>> Regards,
>>>>>
>>>>> Edward
>>>>>
>>>>>
>>>>>
>>>>> On Jan 10, 2008 5:49 PM, Douglas Kojetin
>>>>> <[EMAIL PROTECTED]> wrote:
>>>>>> Hi All,
>>>>>>
>>>>>> I am working with five relaxation data sets (r1, r2 and noe at
>>>>>> 400
>>>>>> MHz; r1 and r2 and 600 MHz), and therefore cannot use the
>>>>>> full_analysis.py protocol. I have obtained estimates for tm,
>>>>>> Dratio, theta and phi using Art Palmer's quadric_diffusion
>>>>>> program.
>>>>>> I modified the full_analysis.py protocol to optimize a prolate
>>>>>> tensor
>>>>>> using these estimates (attached file: mod.py). I have performed
>>>>>> the
>>>>>> optimization of the prolate tensor using either (1) my original
>>>>>> structure or (2) the same structure rotated and translated by the
>>>>>> quadric_diffusion program. It seems that relax does not
>>>>>> converge to
>>>>>> a single global optimum, as different values of tm, Da, theta
>>>>>> and phi
>>>>>> are reported.
>>>>>>
>>>>>> Using my original structure:
>>>>>> #tm = 6.00721299718e-09
>>>>>> #Da = 14256303.3975
>>>>>> #theta = 11.127323614211441
>>>>>> #phi = 62.250251959733312
>>>>>>
>>>>>> Using the rotated/translated structure by the quadric_diffusion
>>>>>> program:
>>>>>> #tm = 5.84350638161e-09
>>>>>> #Da = 11626835.475
>>>>>> #theta = 8.4006873071400197
>>>>>> #phi = 113.6068898953142
>>>>>>
>>>>>> The only difference between the two calculations is the
>>>>>> orientation
>>>>>> of the input PDB structure file. For another set of five rates
>>>>>> (different protein), there is a >0.3 ns difference in the
>>>>>> converged
>>>>>> tm values.
>>>>>>
>>>>>> Is my modified protocol (in mod.py) setup properly? Or is this a
>>>>>> more complex issue in the global optimization? In previous
>>>>>> attempts,
>>>>>> I've also noticed that separate runs with differences in the
>>>>>> estimates for Dratio, theta and phi also converge to different
>>>>>> optimized diffusion tensor variables.
>>>>>>
>>>>>> Doug
>>>>>>
>>>>>>
>>>>>> _______________________________________________
>>>>>> relax (http://nmr-relax.com)
>>>>>>
>>>>>> This is the relax-users mailing list
>>>>>> [email protected]
>>>>>>
>>>>>> To unsubscribe from this list, get a password
>>>>>> reminder, or change your subscription options,
>>>>>> visit the list information page at
>>>>>> https://mail.gna.org/listinfo/relax-users
>>>>>>
>>>>>>
>>>>
>>>>
>>
>>
_______________________________________________
relax (http://nmr-relax.com)
This is the relax-users mailing list
[email protected]
To unsubscribe from this list, get a password
reminder, or change your subscription options,
visit the list information page at
https://mail.gna.org/listinfo/relax-users