Well, in the example I made the individual atomic B factors, the Wilson B, and "true B-factor" were all the same thing.  Keeps it simple.

But to be clear, yes: the B factors at the end of refinement are not the "true B-factors", they are our best estimation of them.  Given that no protein model has ever explained the data to within experimental error, there is clearly a difference between these two things.

I suppose suspicions of B factors arise historically because of how much they have been abused.  In all our efforts to fit Gaussian pegs into weird-shaped holes we have done a lot of crazy things with B factors. Observations/parameters and all that.  You can usually tell something is amiss when you see physically unreasonable things, like covalently bonded neighbors having drastically different Bs.  And of course, like all good validation methods these eventually turned into restraints.  Thru-bond and thru-space B factor restraints are the default now.

That said, I do expect that the "average" B factor is accurate for most structures.  This is because even a small error in overall B is easily corrected in scaling.  The distribution of atomic B factors, however, is harder.  The connection to the Wilson B depends on which part of the Wilson plot you look at.  Any atoms with very large B factors (such as the waters in the bulk solvent) will not contribute significantly to high-angle structure factors.  So, if you are looking at the slope of the Wilson plot at high (-ish) angle you will not see anything coming from these atoms.  Similar is true for disordered side chains, etc.

A common way a refined B factor can go wrong is if the atom in question really should be split into two conformers.  In those cases the refined B factor will be too high, and too anisotropic.  You can also fill in little bits of noise or systematic error with low-occupancy water atoms that have B factors so sharp as to only contribute to a few map grid points.  These are probably not real. On the other hand, anything built into a wrong place or just general nothingness will often refine to a very large B factor.

So yes, all these are caveats in the model-vs-data difference. However, if you have a data set that extends to 1.8 A resolution those outermost spots will be due to the atoms in the structure with the lowest "true" B factors.  Atoms with high B factors don't really play a role in determining the outer resolution limit.

-James Holton
MAD Scientist

On 3/9/2020 1:07 AM, Dale Tronrud wrote:
    Just a note: James Holton said "true B-factor" not "true B-factors".
  I believe he was talking about the overall B not the individual B's.

Dale Tronrud

On 3/8/2020 3:25 PM, Rangana Warshamanage wrote:
Sorry for not being clear enough.
If B-factors at the end of refinement are the "true B-factors" then they
represent a true property of data. They should be good enough to assess
the model quality directly. This is what I meant by B factor validation.
However, how far are the final B-factors similar to true B-factors is
another question.

Rangana


On Sun, Mar 8, 2020 at 7:06 PM Ethan A Merritt <[email protected]
<mailto:[email protected]>> wrote:

     On Sunday, 8 March 2020 01:08:32 PDT Rangana Warshamanage wrote:
     > "The best estimate we have of the "true" B factor is the model B
     factors
     > we get at the end of refinement, once everything is converged,
     after we
     > have done all the building we can.  It is this "true B factor"
     that is a
     > property of the data, not the model, "
     >
     > If this is the case, why can't we use model B factors to validate our
     > structure? I know some people are skeptical about this approach
     because B
     > factors are refinable parameters.
     >
     > Rangana

     It is not clear to me exactly what you are asking.

     B factors _should_ be validated, precisely because they are refined
     parameters
     that are part of your model.   Where have you seen skepticism?

     Maybe you thinking of the frequent question "should the averaged
     refined B
     equal the Wilson B reported by data processing?".  That discussion usual
     wanders off into explanations of why the Wilson B estimate is or is not
     reliable, what "average B" actually means, and so on.  For me the bottom
     line is that comparison of Bavg to the estimated Wilson B is an
     extremely
     weak validation test.  There are many better tests for model quality.

             Ethan





------------------------------------------------------------------------

To unsubscribe from the CCP4BB list, click the following link:
https://www.jiscmail.ac.uk/cgi-bin/webadmin?SUBED1=CCP4BB&A=1

########################################################################

To unsubscribe from the CCP4BB list, click the following link:
https://www.jiscmail.ac.uk/cgi-bin/webadmin?SUBED1=CCP4BB&A=1

########################################################################

To unsubscribe from the CCP4BB list, click the following link:
https://www.jiscmail.ac.uk/cgi-bin/webadmin?SUBED1=CCP4BB&A=1

Reply via email to