Re: [ccp4bb] Drawing reaction mechanism

2012-06-01 Thread Daniel Fernandez
Hi,

this one is free:
http://www.acdlabs.com/resources/freeware/chemsketch/

Cheers,

Daniel


--
On Fri, Jun 1, 2012 12:01 AM ART Appu kumar wrote:

Dear ccp4 user,
  Sorry for offset question, Anyone please tell me
the name of programme which can be used to draw enzyme reaction mechanism (
like Sn1 and Sn2). Your kind help will be much appreciated
Thank you
Appu


Re: [ccp4bb] Calculation of volume/size of cavity

2012-06-01 Thread Daniel Fernandez
You could use Laskowski's surfnet,

http://www.biochem.ucl.ac.uk/~roman/surfnet/surfnet.html

HTH,

Daniel

--
On Thu, May 31, 2012 3:59 PM ART Allan Pang wrote:

A list that you might want to check if it is useful for you.

http://www.caver.cz/index.php?sid=133

Enjoy.

Allan

Quoting Zhou, Tongqing (NIH/VRC) [E] tz...@mail.nih.gov:

 Dear All,
 
 I am looking for a way to calculate the size of a protein cavity  which is 
 occupied by a loop from a ligand protein. The goal is to  see what's the 
 maximum length of peptide allowed in this cavity.
 
 Thanks,
 
 Tongqing
 
 Tongqing Zhou, Ph.D.
 Staff Scientist
 Structural Biology Section
 Vaccine Research Center, NIAID/NIH
 Building 40, Room 4609B
 40 Convent Drive, MSC3027
 Bethesda, MD 20892
 (301) 594-8710 (Tel)
 (301) 793-0794 (Cell)
 (301) 480-2658 (Fax)
 **
 The information in this e-mail and any of its attachments is  confidential 
 and may contain sensitive information. It should not be  used by anyone who 
 is not the original intended recipient. If you  have received this e-mail in 
 error please inform the sender and  delete it from your mailbox or any other 
 storage devices. National  Institute of Allergy and Infectious Diseases 
 shall not accept  liability for any statements made that are sender's own 
 and not  expressly made on behalf of the NIAID by one of its representatives.
 **
 
 



--Allan Pang

PhD Student

G35 Joseph Priestley Building
Queen Mary University of London
London
E1 4NS

Phone number: 02078828480

Twitter: @xerophytes


Re: [ccp4bb] Drawing reaction mechanism

2012-06-01 Thread Robbert Kim
Hi Appu,

I use ChemDoodle ( http://www.chemdoodle.com ) which works quite well, but
unfortunately is not free.
You can try the trial version though.

Robbert

On Fri, Jun 1, 2012 at 4:01 AM, Appu kumar appu.kum...@gmail.com wrote:

 Dear ccp4 user,
   Sorry for offset question, Anyone please tell me
 the name of programme which can be used to draw enzyme reaction mechanism (
 like Sn1 and Sn2). Your kind help will be much appreciated
 Thank you
 Appu



Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Ian Tickle
On 1 June 2012 03:22, Edward A. Berry ber...@upstate.edu wrote:
 Leo will probably answer better than I can, but I would say I/SigI counts
 only
 the present reflection, so eliminating noise by anisotropic truncation
 should
 improve it, raising the average I/SigI in the last shell.

We always include unmeasured reflections with I/sigma(I) = 0 in the
calculation of the mean I/sigma(I) (i.e. we divide the sum of
I/sigma(I) for measureds by the predicted total no of reflections incl
unmeasureds), since for unmeasureds I is (almost) completely unknown
and therefore sigma(I) is effectively infinite (or at least finite but
large since you do have some idea of what range I must fall in).  A
shell with I/sigma(I) = 2 and 50% completeness clearly doesn't carry
the same information content as one with the same I/sigma(I) and
100% complete; therefore IMO it's very misleading to quote
I/sigma(I) including only the measured reflections.  This also means
we can use a single cut-off criterion (we use mean I/sigma(I)  1),
and we don't need another arbitrary cut-off criterion for
completeness.  As many others seem to be doing now, we don't use
Rmerge, Rpim etc as criteria to estimate resolution, they're just too
unreliable - Rmerge is indeed dead and buried!

Actually a mean value of I/sigma(I) of 2 is highly statistically
significant, i.e. very unlikely to have arisen by chance variations,
and the significance threshold for the mean must be much closer to 1
than to 2.  Taking an average always increases the statistical
significance, therefore it's not valid to compare an _average_ value
of I/sigma(I) = 2 with a _single_ value of I/sigma(I) = 3 (taking 3
sigma as the threshold of statistical significance of an individual
measurement): that's a case of comparing apples with pears.  In
other words in the outer shell you would need a lot of highly
significant individual values  3 to attain an overall average of 2
since the majority of individual values will be  1.

 F/sigF is expected to be better than I/sigI because dx^2 = 2Xdx,
 dx^2/x^2 = 2dx/x, dI/I = 2* dF/F  (or approaches that in the limit . . .)

That depends on what you mean by 'better': every metric must be
compared with a criterion appropriate to that metric. So if we are
comparing I/sigma(I) with a criterion value = 3, then we must compare
F/sigma(F) with criterion value = 6 ('in the limit' of zero I), in
which case the comparison is no 'better' (in terms of information
content) with I than with F: they are entirely equivalent.  It's
meaningless to compare F/sigma(F) with the criterion value appropriate
to I/sigma(I): again that's comparing apples and pears!

Cheers

-- Ian


Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread aaleshin
Please excuse my ignorance, but I cannot understand why Rmerge is unreliable 
for estimation of the resolution?
I mean, from a theoretical point of view, 1/sigma is indeed a better 
criterion, but it is not obvious from a practical point of view.

1/sigma depends on a method for sigma estimation, and so same data processed 
by different programs may have different 1/sigma. Moreover, HKL2000 allows 
users to adjust sigmas manually. Rmerge estimates sigmas from differences 
between measurements of same structural factor, and hence is independent of our 
preferences.  But, it also has a very important ability to validate consistency 
of the merged data. If my crystal changed during the data collection, or 
something went wrong with the diffractometer, Rmerge will show it immediately, 
but 1/sigma  will not.

So, please explain why should we stop using Rmerge as a criterion of data 
resolution? 

Alex
Sanford-Burnham Medical Research Institute
10901 North Torrey Pines Road
La Jolla, California 92037



On Jun 1, 2012, at 5:07 AM, Ian Tickle wrote:

 On 1 June 2012 03:22, Edward A. Berry ber...@upstate.edu wrote:
 Leo will probably answer better than I can, but I would say I/SigI counts
 only
 the present reflection, so eliminating noise by anisotropic truncation
 should
 improve it, raising the average I/SigI in the last shell.
 
 We always include unmeasured reflections with I/sigma(I) = 0 in the
 calculation of the mean I/sigma(I) (i.e. we divide the sum of
 I/sigma(I) for measureds by the predicted total no of reflections incl
 unmeasureds), since for unmeasureds I is (almost) completely unknown
 and therefore sigma(I) is effectively infinite (or at least finite but
 large since you do have some idea of what range I must fall in).  A
 shell with I/sigma(I) = 2 and 50% completeness clearly doesn't carry
 the same information content as one with the same I/sigma(I) and
 100% complete; therefore IMO it's very misleading to quote
 I/sigma(I) including only the measured reflections.  This also means
 we can use a single cut-off criterion (we use mean I/sigma(I)  1),
 and we don't need another arbitrary cut-off criterion for
 completeness.  As many others seem to be doing now, we don't use
 Rmerge, Rpim etc as criteria to estimate resolution, they're just too
 unreliable - Rmerge is indeed dead and buried!
 
 Actually a mean value of I/sigma(I) of 2 is highly statistically
 significant, i.e. very unlikely to have arisen by chance variations,
 and the significance threshold for the mean must be much closer to 1
 than to 2.  Taking an average always increases the statistical
 significance, therefore it's not valid to compare an _average_ value
 of I/sigma(I) = 2 with a _single_ value of I/sigma(I) = 3 (taking 3
 sigma as the threshold of statistical significance of an individual
 measurement): that's a case of comparing apples with pears.  In
 other words in the outer shell you would need a lot of highly
 significant individual values  3 to attain an overall average of 2
 since the majority of individual values will be  1.
 
 F/sigF is expected to be better than I/sigI because dx^2 = 2Xdx,
 dx^2/x^2 = 2dx/x, dI/I = 2* dF/F  (or approaches that in the limit . . .)
 
 That depends on what you mean by 'better': every metric must be
 compared with a criterion appropriate to that metric. So if we are
 comparing I/sigma(I) with a criterion value = 3, then we must compare
 F/sigma(F) with criterion value = 6 ('in the limit' of zero I), in
 which case the comparison is no 'better' (in terms of information
 content) with I than with F: they are entirely equivalent.  It's
 meaningless to compare F/sigma(F) with the criterion value appropriate
 to I/sigma(I): again that's comparing apples and pears!
 
 Cheers
 
 -- Ian


Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Phil Evans
As the K  D paper points out, as the signal/noise declines at higher 
resolution, Rmerge goes up to infinity, so there is no sensible way to set a 
limiting value to determine resolution.

That is not to say that Rmerge has no use: as you say it's a reasonably good 
metric to plot against image number to detect a problem. It just not a suitable 
metric for deciding resolution

I/sigI is pretty good for this, even though the sigma estimates are not very 
reliable. CC1/2 is probably better since it is independent of sigmas and has 
defined values from 1.0 down to 0.0 as signal/noise decreases. But we should be 
careful of any dogma which says what data we should discard, and what the 
cutoff limits should be: I/sigI  3,2, or 1? CC1/2  0.2, 0.3, 0.5 ...? Usually 
it does not make a huge difference, but why discard useful data? Provided the 
data are properly weighted in refinement by weights incorporating observed 
sigmas (true in  Refmac, not true in phenix.refine at present I believe), 
adding extra weak data should do no harm, at least out to some point. Program 
algorithms are improving in their treatment of weak data, but are by no means 
perfect.

One problem as discussed earlier in this thread is that we have got used to the 
idea that nominal resolution is a single number indicating the quality of a 
structure, but this has never been true, irrespective of the cutoff method. 
Apart from the considerable problem of anisotropy, we all need to note the 
wisdom of Ethan Merritt

 We should also encourage people not to confuse the quality of 
 the data with the quality of the model.

Phil



On 1 Jun 2012, at 18:59, aaleshin wrote:

 Please excuse my ignorance, but I cannot understand why Rmerge is unreliable 
 for estimation of the resolution?
 I mean, from a theoretical point of view, 1/sigma is indeed a better 
 criterion, but it is not obvious from a practical point of view.
 
 1/sigma depends on a method for sigma estimation, and so same data 
 processed by different programs may have different 1/sigma. Moreover, 
 HKL2000 allows users to adjust sigmas manually. Rmerge estimates sigmas from 
 differences between measurements of same structural factor, and hence is 
 independent of our preferences.  But, it also has a very important ability to 
 validate consistency of the merged data. If my crystal changed during the 
 data collection, or something went wrong with the diffractometer, Rmerge will 
 show it immediately, but 1/sigma  will not.
 
 So, please explain why should we stop using Rmerge as a criterion of data 
 resolution? 
 
 Alex
 Sanford-Burnham Medical Research Institute
 10901 North Torrey Pines Road
 La Jolla, California 92037
 
 
 
 On Jun 1, 2012, at 5:07 AM, Ian Tickle wrote:
 
 On 1 June 2012 03:22, Edward A. Berry ber...@upstate.edu wrote:
 Leo will probably answer better than I can, but I would say I/SigI counts
 only
 the present reflection, so eliminating noise by anisotropic truncation
 should
 improve it, raising the average I/SigI in the last shell.
 
 We always include unmeasured reflections with I/sigma(I) = 0 in the
 calculation of the mean I/sigma(I) (i.e. we divide the sum of
 I/sigma(I) for measureds by the predicted total no of reflections incl
 unmeasureds), since for unmeasureds I is (almost) completely unknown
 and therefore sigma(I) is effectively infinite (or at least finite but
 large since you do have some idea of what range I must fall in).  A
 shell with I/sigma(I) = 2 and 50% completeness clearly doesn't carry
 the same information content as one with the same I/sigma(I) and
 100% complete; therefore IMO it's very misleading to quote
 I/sigma(I) including only the measured reflections.  This also means
 we can use a single cut-off criterion (we use mean I/sigma(I)  1),
 and we don't need another arbitrary cut-off criterion for
 completeness.  As many others seem to be doing now, we don't use
 Rmerge, Rpim etc as criteria to estimate resolution, they're just too
 unreliable - Rmerge is indeed dead and buried!
 
 Actually a mean value of I/sigma(I) of 2 is highly statistically
 significant, i.e. very unlikely to have arisen by chance variations,
 and the significance threshold for the mean must be much closer to 1
 than to 2.  Taking an average always increases the statistical
 significance, therefore it's not valid to compare an _average_ value
 of I/sigma(I) = 2 with a _single_ value of I/sigma(I) = 3 (taking 3
 sigma as the threshold of statistical significance of an individual
 measurement): that's a case of comparing apples with pears.  In
 other words in the outer shell you would need a lot of highly
 significant individual values  3 to attain an overall average of 2
 since the majority of individual values will be  1.
 
 F/sigF is expected to be better than I/sigI because dx^2 = 2Xdx,
 dx^2/x^2 = 2dx/x, dI/I = 2* dF/F  (or approaches that in the limit . . .)
 
 That depends on what you mean by 'better': every metric must be
 compared with a criterion 

Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Ed Pozharski
http://www.nature.com/nsmb/journal/v4/n4/abs/nsb0497-269.html
http://scripts.iucr.org/cgi-bin/paper?S0021889800018227

Just collect 360 sweep instead of 180 on a non-decaying crystal and see
Rmerge go up due to increase in multiplicity (and enough with redundancy
term - the extra data is not really *redundant*).  Is your resolution
worse or better?

This has been argued over before.  Rmerge has some value in comparing
two datasets collected in perfectly identical conditions to see which
crystal is better and it may predict to some extent what R-values you
might expect.  Otherwise, it's unreliable.

Given that it's been 15 years since this was pointed out in no less than
Nature group magazine, and we still hear that Rmerge should decide
resolution cutoff, chances are increasingly slim that I will personally
see the dethroning of that other major oppressor, R-value.

On Fri, 2012-06-01 at 10:59 -0700, aaleshin wrote:
 Please excuse my ignorance, but I cannot understand why Rmerge is unreliable 
 for estimation of the resolution?
 I mean, from a theoretical point of view, 1/sigma is indeed a better 
 criterion, but it is not obvious from a practical point of view.
 
 1/sigma depends on a method for sigma estimation, and so same data 
 processed by different programs may have different 1/sigma. Moreover, 
 HKL2000 allows users to adjust sigmas manually. Rmerge estimates sigmas from 
 differences between measurements of same structural factor, and hence is 
 independent of our preferences.  But, it also has a very important ability to 
 validate consistency of the merged data. If my crystal changed during the 
 data collection, or something went wrong with the diffractometer, Rmerge will 
 show it immediately, but 1/sigma  will not.
 
 So, please explain why should we stop using Rmerge as a criterion of data 
 resolution? 
 
 Alex
 Sanford-Burnham Medical Research Institute
 10901 North Torrey Pines Road
 La Jolla, California 92037
 
 
 
 On Jun 1, 2012, at 5:07 AM, Ian Tickle wrote:
 
  On 1 June 2012 03:22, Edward A. Berry ber...@upstate.edu wrote:
  Leo will probably answer better than I can, but I would say I/SigI counts
  only
  the present reflection, so eliminating noise by anisotropic truncation
  should
  improve it, raising the average I/SigI in the last shell.
  
  We always include unmeasured reflections with I/sigma(I) = 0 in the
  calculation of the mean I/sigma(I) (i.e. we divide the sum of
  I/sigma(I) for measureds by the predicted total no of reflections incl
  unmeasureds), since for unmeasureds I is (almost) completely unknown
  and therefore sigma(I) is effectively infinite (or at least finite but
  large since you do have some idea of what range I must fall in).  A
  shell with I/sigma(I) = 2 and 50% completeness clearly doesn't carry
  the same information content as one with the same I/sigma(I) and
  100% complete; therefore IMO it's very misleading to quote
  I/sigma(I) including only the measured reflections.  This also means
  we can use a single cut-off criterion (we use mean I/sigma(I)  1),
  and we don't need another arbitrary cut-off criterion for
  completeness.  As many others seem to be doing now, we don't use
  Rmerge, Rpim etc as criteria to estimate resolution, they're just too
  unreliable - Rmerge is indeed dead and buried!
  
  Actually a mean value of I/sigma(I) of 2 is highly statistically
  significant, i.e. very unlikely to have arisen by chance variations,
  and the significance threshold for the mean must be much closer to 1
  than to 2.  Taking an average always increases the statistical
  significance, therefore it's not valid to compare an _average_ value
  of I/sigma(I) = 2 with a _single_ value of I/sigma(I) = 3 (taking 3
  sigma as the threshold of statistical significance of an individual
  measurement): that's a case of comparing apples with pears.  In
  other words in the outer shell you would need a lot of highly
  significant individual values  3 to attain an overall average of 2
  since the majority of individual values will be  1.
  
  F/sigF is expected to be better than I/sigI because dx^2 = 2Xdx,
  dx^2/x^2 = 2dx/x, dI/I = 2* dF/F  (or approaches that in the limit . . .)
  
  That depends on what you mean by 'better': every metric must be
  compared with a criterion appropriate to that metric. So if we are
  comparing I/sigma(I) with a criterion value = 3, then we must compare
  F/sigma(F) with criterion value = 6 ('in the limit' of zero I), in
  which case the comparison is no 'better' (in terms of information
  content) with I than with F: they are entirely equivalent.  It's
  meaningless to compare F/sigma(F) with the criterion value appropriate
  to I/sigma(I): again that's comparing apples and pears!
  
  Cheers
  
  -- Ian

-- 
Edwin Pozharski, PhD, Assistant Professor
University of Maryland, Baltimore
--
When the Way is forgotten duty and justice appear;
Then knowledge and wisdom are born along with 

Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Jacob Keller
I don't think any data should be discarded, and I think that although
we are not there yet, refinement should work directly with the images,
iterating back and forth through all the various levels of data
processing. As I think was pointed out by Wang, even an intensity of 0
provides information placing limits on the possible true values of
that reflection. It seems that the main reason data were discarded
historically was because of the limitations of (under)grad students
going through multiple layers of films, evaluating intensities for
each spot, or other similar processing limits, most of which are not
really applicable today. A whole iterated refinement protocol now
takes, what, 15 minutes?

Jacob



On Fri, Jun 1, 2012 at 1:29 PM, Ed Pozharski epozh...@umaryland.edu wrote:
 http://www.nature.com/nsmb/journal/v4/n4/abs/nsb0497-269.html
 http://scripts.iucr.org/cgi-bin/paper?S0021889800018227

 Just collect 360 sweep instead of 180 on a non-decaying crystal and see
 Rmerge go up due to increase in multiplicity (and enough with redundancy
 term - the extra data is not really *redundant*).  Is your resolution
 worse or better?

 This has been argued over before.  Rmerge has some value in comparing
 two datasets collected in perfectly identical conditions to see which
 crystal is better and it may predict to some extent what R-values you
 might expect.  Otherwise, it's unreliable.

 Given that it's been 15 years since this was pointed out in no less than
 Nature group magazine, and we still hear that Rmerge should decide
 resolution cutoff, chances are increasingly slim that I will personally
 see the dethroning of that other major oppressor, R-value.

 On Fri, 2012-06-01 at 10:59 -0700, aaleshin wrote:
 Please excuse my ignorance, but I cannot understand why Rmerge is unreliable 
 for estimation of the resolution?
 I mean, from a theoretical point of view, 1/sigma is indeed a better 
 criterion, but it is not obvious from a practical point of view.

 1/sigma depends on a method for sigma estimation, and so same data 
 processed by different programs may have different 1/sigma. Moreover, 
 HKL2000 allows users to adjust sigmas manually. Rmerge estimates sigmas from 
 differences between measurements of same structural factor, and hence is 
 independent of our preferences.  But, it also has a very important ability 
 to validate consistency of the merged data. If my crystal changed during the 
 data collection, or something went wrong with the diffractometer, Rmerge 
 will show it immediately, but 1/sigma  will not.

 So, please explain why should we stop using Rmerge as a criterion of data 
 resolution?

 Alex
 Sanford-Burnham Medical Research Institute
 10901 North Torrey Pines Road
 La Jolla, California 92037



 On Jun 1, 2012, at 5:07 AM, Ian Tickle wrote:

  On 1 June 2012 03:22, Edward A. Berry ber...@upstate.edu wrote:
  Leo will probably answer better than I can, but I would say I/SigI counts
  only
  the present reflection, so eliminating noise by anisotropic truncation
  should
  improve it, raising the average I/SigI in the last shell.
 
  We always include unmeasured reflections with I/sigma(I) = 0 in the
  calculation of the mean I/sigma(I) (i.e. we divide the sum of
  I/sigma(I) for measureds by the predicted total no of reflections incl
  unmeasureds), since for unmeasureds I is (almost) completely unknown
  and therefore sigma(I) is effectively infinite (or at least finite but
  large since you do have some idea of what range I must fall in).  A
  shell with I/sigma(I) = 2 and 50% completeness clearly doesn't carry
  the same information content as one with the same I/sigma(I) and
  100% complete; therefore IMO it's very misleading to quote
  I/sigma(I) including only the measured reflections.  This also means
  we can use a single cut-off criterion (we use mean I/sigma(I)  1),
  and we don't need another arbitrary cut-off criterion for
  completeness.  As many others seem to be doing now, we don't use
  Rmerge, Rpim etc as criteria to estimate resolution, they're just too
  unreliable - Rmerge is indeed dead and buried!
 
  Actually a mean value of I/sigma(I) of 2 is highly statistically
  significant, i.e. very unlikely to have arisen by chance variations,
  and the significance threshold for the mean must be much closer to 1
  than to 2.  Taking an average always increases the statistical
  significance, therefore it's not valid to compare an _average_ value
  of I/sigma(I) = 2 with a _single_ value of I/sigma(I) = 3 (taking 3
  sigma as the threshold of statistical significance of an individual
  measurement): that's a case of comparing apples with pears.  In
  other words in the outer shell you would need a lot of highly
  significant individual values  3 to attain an overall average of 2
  since the majority of individual values will be  1.
 
  F/sigF is expected to be better than I/sigI because dx^2 = 2Xdx,
  dx^2/x^2 = 2dx/x, dI/I = 2* dF/F  (or approaches that in the limit . . .)

Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Ian Tickle
Let's say you collect data (or rather indices) to 1.4 Ang but the real
resolution is 2.8 Ang and you use all the data in refinement with no
resolution cut-off, so there are 8 times as many data.  Then your 15
mins becomes 2 hours - is that still acceptable?  It's unlikely that
you'll see any difference in the results so was all that extra
computing worth the effort?

Now work out the total number of pixels in one of your datasets (i.e.
no of pixels per image times no of images).  Divide that by the no of
reflections in the a.u. and multiply by 15 mins (it's probably in the
region of 400 days!): still acceptable?  Again it's unlikely you'll
see any significant difference in the results (assuming you only use
the Bragg spots), so again was it worth it?

What matters in terms of information content is not the absolute
intensity but the ratio intensity / (expected intensity).  As the data
get weaker at higher d* I falls off, but so does I and the ratio I /
I becomes progressively more unreliable at determining the
information content.  So a zero I when the other intensities in the
same d* shell are strong is indeed a powerful constraint (this I
suspect is what Wang meant), however if the other intensities in the
shell are also all zero it tells you next to nothing.

-- Ian

On 1 June 2012 20:03, Jacob Keller j-kell...@fsm.northwestern.edu wrote:
 I don't think any data should be discarded, and I think that although
 we are not there yet, refinement should work directly with the images,
 iterating back and forth through all the various levels of data
 processing. As I think was pointed out by Wang, even an intensity of 0
 provides information placing limits on the possible true values of
 that reflection. It seems that the main reason data were discarded
 historically was because of the limitations of (under)grad students
 going through multiple layers of films, evaluating intensities for
 each spot, or other similar processing limits, most of which are not
 really applicable today. A whole iterated refinement protocol now
 takes, what, 15 minutes?

 Jacob



 On Fri, Jun 1, 2012 at 1:29 PM, Ed Pozharski epozh...@umaryland.edu wrote:
 http://www.nature.com/nsmb/journal/v4/n4/abs/nsb0497-269.html
 http://scripts.iucr.org/cgi-bin/paper?S0021889800018227

 Just collect 360 sweep instead of 180 on a non-decaying crystal and see
 Rmerge go up due to increase in multiplicity (and enough with redundancy
 term - the extra data is not really *redundant*).  Is your resolution
 worse or better?

 This has been argued over before.  Rmerge has some value in comparing
 two datasets collected in perfectly identical conditions to see which
 crystal is better and it may predict to some extent what R-values you
 might expect.  Otherwise, it's unreliable.

 Given that it's been 15 years since this was pointed out in no less than
 Nature group magazine, and we still hear that Rmerge should decide
 resolution cutoff, chances are increasingly slim that I will personally
 see the dethroning of that other major oppressor, R-value.

 On Fri, 2012-06-01 at 10:59 -0700, aaleshin wrote:
 Please excuse my ignorance, but I cannot understand why Rmerge is 
 unreliable for estimation of the resolution?
 I mean, from a theoretical point of view, 1/sigma is indeed a better 
 criterion, but it is not obvious from a practical point of view.

 1/sigma depends on a method for sigma estimation, and so same data 
 processed by different programs may have different 1/sigma. Moreover, 
 HKL2000 allows users to adjust sigmas manually. Rmerge estimates sigmas 
 from differences between measurements of same structural factor, and hence 
 is independent of our preferences.  But, it also has a very important 
 ability to validate consistency of the merged data. If my crystal changed 
 during the data collection, or something went wrong with the 
 diffractometer, Rmerge will show it immediately, but 1/sigma  will not.

 So, please explain why should we stop using Rmerge as a criterion of data 
 resolution?

 Alex
 Sanford-Burnham Medical Research Institute
 10901 North Torrey Pines Road
 La Jolla, California 92037



 On Jun 1, 2012, at 5:07 AM, Ian Tickle wrote:

  On 1 June 2012 03:22, Edward A. Berry ber...@upstate.edu wrote:
  Leo will probably answer better than I can, but I would say I/SigI counts
  only
  the present reflection, so eliminating noise by anisotropic truncation
  should
  improve it, raising the average I/SigI in the last shell.
 
  We always include unmeasured reflections with I/sigma(I) = 0 in the
  calculation of the mean I/sigma(I) (i.e. we divide the sum of
  I/sigma(I) for measureds by the predicted total no of reflections incl
  unmeasureds), since for unmeasureds I is (almost) completely unknown
  and therefore sigma(I) is effectively infinite (or at least finite but
  large since you do have some idea of what range I must fall in).  A
  shell with I/sigma(I) = 2 and 50% completeness clearly doesn't carry
  the same information 

Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Leonid Sazanov
Hi, as we reported in our paper in Table 1 (actually Supplementary Table 1), at 
the end of Scaling 2, completeness in the outer shell after aniso truncation 
was 54%. Whilst 96% completeness and I/sigma 0.8 is of course before aniso 
truncation. I/sigma after truncation would be higher, but it is not clear to me 
how to calculate that number exactly, since aniso truncation is done post data 
scaling. One could of course re-process images in Mosflm with applied aniso 
limits and then scale data, but that would not be exactly the same.

From many trials with strongly anisotropic data we found that for map 
calculation and refinement it is best to cut data anisotropically where F/sigma 
is approaching 2.5-2.7 in each direction, as long as completeness in the outer 
shell remains above 50% or so. Usually the highest useful resolution is also 
where the correlation coefficient between random half-data-set estimates of 
intensities in SCALA falls below about 0.5 (as advocated by Phil Evans, I 
think). CC seems to be less affected by anisotropy (in this case it reached 0.5 
at 3.0 angstrom, which was another criterion to cut data at 3.0).

HTH.
Leo




I am little curious about the anisotropically truncated data for 3RKO:

Percent Possible(All)   96.0
Mean I Over Sigma(Observed) 0.8

In the supplementary table of the nature paper it was made clear that this 
3.16-3.0A, I/sigmaI=0.8 and Rmerge=1.216 shell was the outer shell of the 
anisotropically truncated data. The authors did also report the 
isotropically truncated resolution to be 3.2A with I/sigmaI=1.3 and 
Rmerge=73%.

The authors also stated in the main text that

the best native data set was anisotropically scaled and truncated to 3.4 Å, 
3.0 Å and 3.0 Å resolution, where the F/σ ratio drops to ~2.6–2.8 along 
the a*, b* and c* axes, respectively (scaling 2, Supplementary Table 1)

My question is, is the I/sigmaI=0.8 a consequence of many reflections with 
nearly 0 I/sigmaI being included in the calculation? Then what does the 96% 
completeness mean? Does it mean that 96% completeness in the spherical shell 
of 3.16-3.0A was achieved, by including a great number of I=0 reflections?


Zhijie


Re: [ccp4bb] Death of Rmerge

2012-06-01 Thread Jacob Keller
 Let's say you collect data (or rather indices) to 1.4 Ang but the real
 resolution is 2.8 Ang and you use all the data in refinement with no
 resolution cut-off, so there are 8 times as many data.  Then your 15
 mins becomes 2 hours - is that still acceptable?  It's unlikely that
 you'll see any difference in the results so was all that extra
 computing worth the effort?

 Now work out the total number of pixels in one of your datasets (i.e.
 no of pixels per image times no of images).  Divide that by the no of
 reflections in the a.u. and multiply by 15 mins (it's probably in the
 region of 400 days!): still acceptable?  Again it's unlikely you'll
 see any significant difference in the results (assuming you only use
 the Bragg spots), so again was it worth it?

 What matters in terms of information content is not the absolute
 intensity but the ratio intensity / (expected intensity).  As the data
 get weaker at higher d* I falls off, but so does I and the ratio I /
 I becomes progressively more unreliable at determining the
 information content.  So a zero I when the other intensities in the
 same d* shell are strong is indeed a powerful constraint (this I
 suspect is what Wang meant), however if the other intensities in the
 shell are also all zero it tells you next to nothing.

 -- Ian


I envisioned a process of iteration through the various stages of
processing, so still using integration, scaling, etc. to reduce data
before refinement, but maybe feeding back model-based information to
inform the processing of the images. Something like, Refmac says to
Mosflm: kill frames 1100-1200: they're too radiation-damaged. But I
like your idea of using all the pixels--that would be the ultimate,
wouldn't it! Actually, the best would be to have the refinement
already going when collecting data, and informing which frames to
take, and for how long! In a couple years that too will take no time
at all, but then again, we'll probably have atomic-precision real-time
in vivo microscopes by then anyway, and crystallography will have
become an (interesting!) historical curiosity...

JPK

***
Jacob Pearson Keller
Northwestern University
Medical Scientist Training Program
email: j-kell...@northwestern.edu
***