Re: [ccp4bb] SeMet data

2019-08-28 Thread Stefan Gajewski
Hi Lindsey,

I do agree with the others on improving your experimental strategy to enhance 
the anomalous signal in your data, but did you try to solve the structure with 
the data you already have? I have seen cases where there is only a pathetic 
anomalous signal from fully Se-Met incorporated crystals in any experimental 
setting and those could eventually be solved. You probably have to deal with a 
significant number of false positives in your substructure but if the anomalous 
signal is observed accurately you are quite likely able to solve the structure 
with weak data alone.  

How much anomalous signal did you get out of those frames and how much did you 
expect to get? Have you tried merging data from several crystals collected at 
the same wavelength ?

Best,
Stefan



To unsubscribe from the CCP4BB list, click the following link:
https://www.jiscmail.ac.uk/cgi-bin/webadmin?SUBED1=CCP4BB=1


Re: [ccp4bb] align structure in Pymol based on dsDNA helix

2017-05-15 Thread Stefan Gajewski
Dear Xiao Lei,

You might want to consider UCSF Chimera for this task

https://www.cgl.ucsf.edu/chimera/docs/ContributedSoftware/matchalign/matchalign.html

Best,
Stefan


Re: [ccp4bb] Interesting DNA contamination

2015-06-26 Thread Stefan Gajewski
Pramod,

You already got good suggestions on how to handle DNA contamination in protein 
preparations.

Let me point out briefly that you haven't demonstrated yet that your 
contamination is DNA. 

I had the same observation when purifying UvsX. A very persistent and strong 
contamination in all my preps at ~500kb. To test weather it was DNA or RNA I 
boiled the protein 30 minutes and incubated it with DNAase and RNAse and result 
was the same. I concluded it was neither RNA nor DNA and continued as if 
nothing had happened. 

This publication is reporting the same observation:

Formosa and Alberts (1986) Purification and characterization of the T4 
bacteriophage uvsX protein. J Biol Chem. 1986 May 5;261(13):6107-18.

If you ever find out what it is that runs like 500kb DNA on Agarose, please let 
me know. 
S.


Re: [ccp4bb] Interesting DNA contamination

2015-06-26 Thread Stefan Gajewski
Correction,

I meant to say 0.5kb, not 500kb

sorry for that.
S.


Re: [ccp4bb] experimental phasing at low resolution

2015-04-28 Thread Stefan Gajewski
Hi,

I doubt you have a TaBr cluster bound in your soaks. 

You'd expect a strong anomalous signal from cluster compounds (that's what they 
are made for) and you can easily get a low resolution anomalous signal from a 
few dislocated HA in the solvent channels, especially if you don't counter soak 
the crystals (expose them to HA free buffer briefly prior to freezing to leech 
the solvent channels from excess HA).

I wouldn't collect MAD data on cluster compound soaks anyway. If they are 
present, you should be able to solve it with SAD.  Alternatively, try the LIII 
and LII edges at the respective peak wavelength instead of collecting data at 
wavelength around the same edge. I don't think that's your problem because I 
don't think the TaBr bound at all in your crystals. In your case, I'd recommend 
trying a different HA soak first. If you cannot express your protein with 
selenium but you can get your hands on xenon derivatization equipment, try 
xenon!! Mercury is always worth a try IMHO. 

Non-isomorphous data from various soaks is not very helpful, that single SAD 
dataset with that little bit of extra signal IS very helpful. Optimize your 
cryo protectant and data collection strategy (inverse wedges, kappa goni, 
etc.). 

If you want to try MAD, try collecting all wavelength data from the same 
crystal. Don't do MAD if you have radiation sensitive non-isomorphous crystals. 
  

An non interpretable MR solution will help you locate the HA sites as long it 
is correct. So, if you have one, try to use it to phase an anomalous difference 
map. 

You indicated that you have several datasets from two compounds. Try merging 
the 'identical' datasets and see if that works better. 

SHARP is really good for MAD data, but you need to collect and process the data 
right, which is difficult.

SHELX is great for low resolution/low signal SAD/SIR. I'd always use SHELX 
first because it was the only thing working for me below 5A so far (sample 
size=1, anecdotal evidence, handle with care!). 

good luck,
S.


[ccp4bb] A crystal symmetry broken beyond repair?

2014-02-22 Thread Stefan Gajewski
I have three brief questions (number two should be the easy one).

1. Where could an extra electron density of an (physically impossible)
additional molecule in an otherwise perfect map come from? (even in a wrong
solution in a {probably} too high space group?)

2. I would like to see how my data looks like in a RvR plot. How can I make
one?  (Acta Cryst.D62, 83-95.)

3. Is it possible to have an order/disorder crystal defect with a pure
rotational disorder, e.g. the unit cell translations are not affected by
the disorder, only the orientations of the molecules in separate layers?

Stefan


[ccp4bb] Triclinic solution for a dataset with pseudo-translational symmetry, possible pseudo-centering and possible pseudo-merohedral twinning. How to proceed?

2013-10-18 Thread Stefan Gajewski
Dear experts,

This is a messy question concerning a messy dataset and I try to put it
correctly.

I have a severely anisotropic dataset up to 2.3A that merges quite fine in
all 4/mmm. Automatic indexing is primitive tetragonal but I can push it to
a centered tetragonal cell if I want to. The crystals and the diffraction
images are highly reproducible.

There is one non-origin peak of ~45% origin peak height with vector 0.5 0.5
0.46. Intensity statistics do not indicate twinning.

Now I have a good MR solution with a 99% model in P1 that reveals four
proteins in the cell. Dimer A,C and Dimer B,D with pseudo-translation A to
B and C to D and a twofold between A,C and B,D. The twofold lies almost
perfectly within a=b and B,C mimic a center at half c. Each protein
(A,B,C,D) is a pseudo-octameric homo-oligomer.

It is also still possible that the orientation of the multimers is off by
one rotation in my MR solution and there actually exists a higher symmetry,
although there are no convincing MR solutions in C2, P2, P2(1), yet.

Initial refinement in P1 stalled at R=0.35 and R-free 0.40. There are seven
possible pseudo-merohedral twin laws and all of them decrease the R-factors
to R~0.30 and R-free ~0.35 in refinement, so I can't tell which one is the
correct one or if it is twinned at all.

What should I try next?

Is there a good advice how to refine such a type of structure? How can I
determine if it is twinned or not ? Should I omit anisotropic scaling in
this case and cut back the resolution to preserve the weak reflections on
the original scale?


feeling a little lost here with all the pseudo symmetries ,

Stefan Gajewski


Re: [ccp4bb] Resolution, R factors and data quality

2013-08-28 Thread Stefan Gajewski
Jim,

This is coming from someone who just got enlightened a few weeks ago on 
resolution cut-offs.

I am asked often:  What value of CC1/2 should I cut my resolution at? 

The KD paper mentioned that the CC(1/2) criterion loses its significance at ~9 
according to student test.

I doubt that this can be a generally true guideline for a resolution cut-off. 
The structures I am doing right now were cut off at ~20 to ~80 CC(1/2)

You probably do not want to do the same mistake again, we all made before, when 
cutting resolution based on Rmerge/Rmeas, do you?


 What should I tell my students?  I've got a course coming up and I am sure 
 they will ask me again.

This is actually the more valuable insight I got from the KD paper. You don't 
use the CC(1/2) as an absolute indicator but rather as an suggestion. The 
resolution limit is determined by the refinement, not by the data processing.

I think I will handle my data in future as follows:

Bins with CC(1/2) less than 9 should be initially excluded.

The structure is then refined against all reflections in the file and only 
those bins that add information to the map/structure are kept in the final 
rounds. In most cases this will probably be more than CC(1/2) 25. If the last 
shell (CC~9) still adds information to the model, process the images again, 
e.g. till CC(1/2) drops to 0, and see if some more useful information is in 
there. You could also go ahead and use CC(1/2) 0 as initial cut-off, but I 
think that will rather increase computation time than help your structure in 
most cases.


So yes, I would feel comfortable with giving true resolution limits based on 
the refinement of the model, and not based on any number derived from data 
processing. In the end, you can always say  I tried it and this was the 
highest resolution I could model vs. I cut at _numerical value X of this 
parameter_ because everybody else does so.


Re: [ccp4bb] Stuck rfree - possible non merohedral twinning ?

2013-08-26 Thread Stefan Gajewski
Hi Mahesh,

First of all, I risk going out on a limb here since I have no demonstrable 
experience concerning your topic. I was trying to solve a non-merohedral 
dataset for years, and I failed (like failure as defined in the dictionary). 
That is hardly a reference, but I got to read a little on the topic and maybe 
I got some comments that can be useful for you.

So, if you data would be twinned by non-merohedry, you would not have a valid 
solution by now. You would have only a fraction of the reflections indexed and 
a significant fraction in that index would be overlaps. The others would be 
partial overlaps/splits and the rest fully split reflections (singlets). 
Solving a non-merohedrally twinned structure requires developing an independent 
crystal rotation matrix for each twin domain, integrating in the correct point 
group, leaving all related reflections unmerged and with the original index, 
solving the structure and identifying the twin law (the one that describes the 
superposition of one twin domain on top of the other one) during structure 
solution/refinement. According to my knowledge, this has not been done with a 
novel structure as of now. (As for anything I say here, someone please educate 
me if I am wrong) The really mean part with non-merohedral data is, that you 
cannot even index the data properly, although a variety of indexing choices 
will give you reasonably well integration/merging statistics and you might not 
notice it in the beginning. It is very hard to identify the correct cell. A 
rough guideline is, that in many cases, crystals twinned by non-merohedry 
display a relationship of a~2b (2a~b, depending on choice) so you can try to 
double the axis for either a or b in a tetragonal cell and see if your 
predictions for a primitive cell with those dimensions overlap better with your 
reflections. The true space group of the lattice will emerge at a later point 
of time. The best way to do the indexing is supposedly using programs for 
small-molecule crystallography. When (=if) you have the two matrices, you can 
forward them to your protein x-ray crystallography software and integrate twice 
using one matrix at a time. A much better way to deal with non-merohedry is to 
make different crystals. Or to optimize the cryo. 

If your data is twinned by pseudo-merohedry, most of the reflections will be 
overlaps with a very few splits, rather in higher resolution bins than in lower 
ones. In case your data is really twined, I suspect that is what you are 
dealing with here by looking at the diffraction patterns you showed before. 
This means you can solve your structure by integrating your data with only one 
crystal rotation matrix. However, your data might appear to be not only in a 
higher point group but also in a higher lattice system than it actually is. 
Your data then emulates a P422 but your structure does not generate the crystal 
by that symmetry relationship. Since there is no twin law for P422, 
phenix.xtriage can not help you there with merged data. Treating twinning by 
pseudo-merohedry is pretty much straight forward. You must process your data in 
the correct point group and later identify and apply the twin law. Start with 
P1 and work up if you have enough frames.

Also, if you have more than one molecule in the ASU, you might have simply 
overestimated the symmetry. I have seen this once in a monoclinic cell with two 
copies in the ASU where one angle was ~90.08, but not 90.00. It still merged 
reasonably well as an orthorombic cell, but it was not twined.

In either case, try processing your images in a lower point group first, and 
look carefully at your data.

Did your predicted reflections cover all the observed reflections? Did you 
predict reflections (entire lunes?) that are not there? How does your native 
Patterson map look like if you process your data in P1? 

good luck,
s.


Re: [ccp4bb] Where to cut the data in this medium resolution dataset

2013-07-23 Thread Stefan Gajewski
Nat,

What do correct B-factors look like?  What refinement strategy did you
 use for them?


1) If I see strong positive density in the Fo-Fc map along the backbone of
two turns of an correctly placed alpha helix, therefore the B-factors are
too high in that region. The model after refinement suggest less scattering
in that region than is observed which is most likely explained by
incorrectly high B-factors.

2) xyz, TLS and individual B-factors (no grouping). SS and ramachandran
restraints on coordinates and NCS restraints on coordinates and B-factors.




 Note that the R-free value in the 3.4A shell is lower than the R-work
 (and also the Rpim in that shell!) which clearly indicates this refinement
 was not stable.


 I don't think it indicates anything about the stability of refinement -
 my guess would be that the NCS is biasing R-free.  I suppose it could also
 indicate that the data in the 3.6-3.4 range are basically noise, although
 if the maps look better then that would suggest the opposite.


 I think the refinement is not parametrized correctly.

Thank you,

S.




[ccp4bb] Where to cut the data in this medium resolution dataset

2013-07-22 Thread Stefan Gajewski
Hey!

I was reading a lot lately on data processing and the ongoing debate in the
community on how to submit Table 1.

Here is an example of medium resolution data integrated with XDS, merged in
SCALA and preliminarily refined in phenix.


Overall  InnerShell  OuterShell
  Low resolution limit   49.06 49.06  3.58
  High resolution limit   3.40 10.74  3.40

  Rmerge 0.224 0.050 1.324
  Rmerge in top intensity bin0.049- -
  Rmeas (within I+/I-)   0.235 0.052 1.391
  Rmeas (all I+  I-)0.235 0.052 1.391
  Rpim (within I+/I-)0.067 0.014 0.407
  Rpim (all I+  I-) 0.067 0.014 0.407
  Fractional partial bias0.000 0.000 0.000
  Total number of observations  275312  8491 39783
  Total number unique21137   765  3016
  Mean((I)/sd(I)) 10.2  31.2   2.3
  Completeness98.4  95.7  98.4
  Multiplicity13.0  11.1  13.2

r_work=0.2461 (0.3998)  r_free=  0.2697 (0.3592)
SigmaA highest shell = 0.78
scale factor highest shell (phenix.refine) = 0.87


   Overall  InnerShell  OuterShell
  Low resolution limit   49.59 49.59  3.80
  High resolution limit   3.60 11.39  3.60

  Rmerge 0.231 0.048 0.819
  Rmerge in top intensity bin0.045- -
  Rmeas (within I+/I-)   0.242 0.050 0.860
  Rmeas (all I+  I-)0.242 0.050 0.860
  Rpim (within I+/I-)0.069 0.014 0.249
  Rpim (all I+  I-) 0.069 0.014 0.249
  Fractional partial bias0.000 0.000 0.000
  Total number of observations  230945  6997 33402
  Total number unique17794   646  2531
  Mean((I)/sd(I)) 11.5  30.7   3.7
  Completeness98.3  95.5  98.7
  Multiplicity13.0  10.8  13.2

r_work= 0.2372 (0.3585)  r_free= 0.2663  (0.3770)
SigmaA highest shell = 0.79
scale factor highest shell (phenix.refine) = 0.95

XSCALE gives significantly lower average Rrim and Rmerge for both
integrations (~18%) and CC(1/2) is above 0.7 in all bins

The diffraction pattern looks great, the 3.4A reflections are visible by
eye and the edge of the detector is about 2.8A. The crystals were 10x20x50
um in size and spacegroup is P6522.

The maps shows signs of over fitting, the B-factors do not look correct in
my opinion. Note that the R-free value in the 3.4A shell is lower than the
R-work (and also the Rpim in that shell!) which clearly indicates this
refinement was not stable.

The structure contains no beta sheets and refinement also profits greatly
from very rigid high-order NCS. The maps are very detailed, in fact better
than some 2.8A maps I've seen before.  The 0.2A in question here are
actually quite helpful to increase the map quality, so I keep wondering if
I should deposit the structure with them or keep them only for my own
interpretation.

Before I continue optimizing the integration/refinement I would like to
hear suggestions from the experts where to make the resolution cut-off in
this case?
Do I have all information I need to make that decision?
What arguments should I present when dealing with the reviewers? I mean,
the Rrim/Rmerge values are really very high.

Thank you for your input,

Stefan Gajewski


[ccp4bb] Fun Question - Is multiple isomorphous replacement an obsolete technique?

2012-06-05 Thread Stefan Gajewski
Hey!

I was just wondering, do you know of any recent (~10y) publication that
presented a structure solution solely based on MIR? Without the use of any
anomalous signal of some sort?

When was the last time you saw a structure that was solved without the use
of anomalous signal or homology model? Is there a way to look up the answer
(e.g. filter settings in the RCSB) I am not aware of?

Thanks,
S.

(Disclaimer: I am aware that isomorpous data is a valuable source of
information)


Re: [ccp4bb] Expression of Viral proteins for crystallography

2012-01-24 Thread Stefan Gajewski
Rubén,

the previous answer probably addresses your problem accurately. 

However. If your protein of interest modifies DNA/RNA, it is quite common that 
your pET constructs will mutate rapidly in E.coli. Lac operons tend to leak 
quite a bit, which is not enough to detect the protein prior to induction but 
can definitely drive the selection for mutated plasmids in E.coli. We saw that 
behavior in a nuclease cloned from homogeneous genomic DNA. Every colony had a 
different mutant when grown on LB.  Also mutants that do not appear to affect 
the enzyme activity judged by the structure. We assumed, that any mutation that 
slightly reduces the toxicity for E.coli will be selected for. So, if you 
experience the same issue again with synthetic DNA, or have the patience to 
repeat, try adding ~1% glucose to all media, liquid and solid. This will 
tighten up the Lac operon and worked well for our nuclease project.  You still 
have to sequence every batch you purify. pLysS cells can also work, but we 
still ended up adding glucose because it is cheap and doesn't cause trouble 
downstream. 

Stefan


Re: [ccp4bb] very strange lattice: high anisotropy, 78% solvent content and maybe merohedral twinning

2011-12-14 Thread Stefan Gajewski
Jürgen,

Have you checked a simple selfrotation function in your currently favored
 space group ?


Yes, both selfrotation function and patterson map do not look suspicious in
I422.


 Do you have sufficient data collected to start out in P1 or C2 ? Then I
 would start there and systematically look at selfrotation functions in
 those space groups. Also check the native Patterson for translational NCS.


I got ~220 degrees before the radiation damage became significant, so there
should be enough data and I will look into it.

4 A is not great for stable refinement of cell parameters, which program
 did you use and which parameters did you fix?


HKL2000 without fixing parameters. mosflm can't hold on the lattice and I
haven't tried d*trek, yet.


 Did you use main.ncs=true in the SA approach ?


yes


 Pointless or xtriage ?


xtriage



 Why does it take a year to grow those crystals ?


Well, other crystals don't diffract and the protein is quite stable in
solution.
That's how it is, I guess?


 Out of curiosity, how did you collect on this crystal without overlapping
 reflections ?


I got away with 1 second exposure and 0.75 degree oscillation at 650mm
detector distance. there are some predicted overlaps but they are in those
regions that are empty due to anisotropy. Our cryo condition gives well
separated small spots of nice, round shape, mosaicity is ~0.6. The pattern
itself looks great, although the beamstop shadow is quite big on the
frames.

Thank You,
Stefan


[ccp4bb] very strange lattice: high anisotropy, 78% solvent content and maybe merohedral twinning

2011-12-13 Thread Stefan Gajewski
I am looking at a highly unusual crystal lattice right now and can't figure out 
what is going on, so I decided to ask the experts.

I recently got data on a oligomeric protein with many highly correlated NCS 
units (4.0A resolution, linear R-sym is 0.16-0.21 in I4, I422, F222, C2 and 
0.12 in P1) with severe anisotropic diffraction (according to diffraction 
anisotropy server, the F/sigma drops below 3 at a=6.1 b=6.1 c=never, suggested 
isotropic B-sharpening -125A^2) This lattice has a problem. The apparent unit 
cell is rather huge (roughly 180 180 620 / 90 90 90)

The unit cell dimensions are almost perfectly I4 and the presence of systematic 
absent reflections 50 I/s in I41 and I4122 suggest no screw axis. I used a 
very closely related structure solved at 4.2A as molecular replacement model 
and got a solution from the anisotropy corrected data in I422 space group with 
two oligomers in the asymmetric unit cell. 

Confidence of the MR solution is quite high since (a)the MR replacement put 
one model one NCS raster off the true position resulting in a clash with the 
second one in an empty region of the map and additional electron density on the 
other side which corresponds perfectly to the wrongly positioned monomer, and 
(b) after rotating the model in the right position I could refine the 
structure to R-work=0.31. R-free=0.35 in one run of rigid body refinement 
followed by NCS restrained simulated annealing refinement (phenix.refine), 
which is in my opinion really good at such an early stage of refinement given 
the low overall resolution and even lower completeness of strong reflections in 
a and b due to high anisotropy (observables to atoms ratio is about 3:1) . I 
can even see clear density for some of the bulky sidechains which were not 
included in the model. 

Now here is the baffling thing. The unit cell is almost empty with an apparent 
solvent content of 78%. The molecules cluster around the c-axis and at the 
origin with an empty gap in a and b of at least 15A and up to 165A(!) in the 
longest dimension. There is no sign of electron density that would indicate  a 
missing protein in that region and ~98% of my model is already accounted for by 
the density in the 2Fo-Fc map, making a contact of disordered protein regions 
across the ASUs unlikely. In fact, the protein density is well defined at the 
closest gap and no mainchain atom is unaccounted for in that region. The 
oligomer has a magnitude of ~105A x 70A.  I heavily doubt that a crystal 
lattice with such little contacts and holes as huge as these can exist and 
therefore think that: 

(a) the R-factors are misleading me to think the solution is correct and 
complete
(b) I must have been doing something really wrong

Since proteins from this family have a well established history of producing 
twinned crystals I had a look at that possibility. Analyzing the anisotropy 
corrected I4 data for twinning (Padilla  Yeates method) revealed a 2-fold twin 
law with a twin fraction of 0.42 which would make the discrimination between an 
almost perfectly merohedral twin in I4 and a (non twinned ?) I422 extremely 
difficult (to me).  MR with anisotropy corrected I4 data gave the same crystal 
packing and hence the same void solvent region. MR in lower point groups was 
not successful so far although I haven't pursued that idea vigorously. The same 
data in I422 has no indication for twinning and in C2 three 2-fold twin laws.

Anomalous data is not easily available since those crystals grow in about one 
year and getting another crystal is also not very likely because this IS the 
other crystal.

I am clueless now on how to proceed here and would appreciate advice from 
experienced crystallographers on what to try first.

Am I worrying too much about the packing? 
Is it even possible to have such an enormously huge solvent region in a protein 
crystal? 
What is the recommended protocol when dealing with many and very strongly 
correlated NCS units, putative twinning and severe anisotropy all at the same 
time?

Stefan Gajewski