Re: [ccp4bb] Copy NCS chain in coot

2011-03-31 Thread Javier Garcia

Dear Zhang yu,

I also found that same problem recently. I think is a small bug from Coot.
It will only copy NCS if the molecule's name you want to copy is molecule A.
So maybe will work if you rename your built chain as A.

Good luck!

Javier

On 31/03/11 00:16, zhang yu wrote:

Dear all,

I have two fold NCS in ASU.  I could find phase by molecular 
replacement, but I have to build other chains by myself, since the 
model is not complete. During model building in coot, I built one 
chain in one NCS, and I merged this chain into the previous molecule. 
Now it is the chain M'.


 When I tried to directly copy this chain to its NCS mate, It didn't work.

I followed the instruction posted 
beforehttp://www.mail-archive.com/coot@jiscmail.ac.uk/msg01984.html;.


What I did is
1. Change the master chain to chain M Draw-NCS ghost 
control-change to chain M
(The terminal said 'there are no ghosts when I change the master 
chain to chain M)
2. go to Extension-NCS-copy chain-, coot ask to fill the master 
chain ID, and I typed M


Nothing happened.  Could someone tell me what is the correct procedure 
to do it ?


Bye the way, for those chains present in both NCS maps, I modified one 
chain and applied the change to its NCS mate by using the same above 
procedure, and it worked.


Thanks

--
Yu Zhang
HHMI associate
Waksman Institute, Rutgers University
190 Frelinghuysen Rd.
Piscataway, NJ, 08904




Re: [ccp4bb] [phenixbb] what to do with disordered side chains

2011-03-31 Thread Huanwang Yang
If a part of sequence has no density, this part will be cut from 
coordinates.
If the side chain of a residue is  lack of density, the opinion is not 
converged.

What about the ligand, if no density is observed?

Huanwang

Ed Pozharski wrote:

The results of the online survey on what to do with disordered side
chains (from total of 240 responses):

Delete the atoms 43%
Let refinement take care of it by inflating B-factors41%
Set occupancy to zero12%
Other 4%

Other suggestions were:

- Place atoms in most likely spot based on rotomer and contacts and
indicate high positional sigmas on ATMSIG records
- To invent refinement that will spread this residues over many rotamers
as this is what actually happened
- Delet the atoms but retain the original amino acid name
- choose the most common rotamer (B-factors don't inflate, they just
rise slightly)
- Depends. if the disordered region is unteresting, delete atoms.
Otherwise, try to model it in one or more disordered model (and then
state it clearly in the pdb file)
- In case that no density is in the map, model several conformations of
the missing segment and insert it into the PDB file with zero
occupancies. It is equivalent what the NMR people do. 
- Model it in and compare the MD simulations with SAXS

- I would assumne Dale Tronrod suggestion the best. Sigatm labels.
- Let the refinement inflate B-factors, then set occupancy to zero in
the last round.

Thanks to all for participation,

Ed.

  


Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Bertrand, Jay Aaron [Nervianoms]
I checked with someone in our protein production group and got the following 
response:

We also stopped doing the virus titration with the plaque assay and instead are 
performing expression test with different concentration of virus from the 3rd 
amplification. But for some viruses we still have doubts concerning the 
amplification success, so we are now evaluating a new technology using qPCR 
with the following kit (http://oetltd.com/products/category/baculoquant/). So 
you might have a look and see if it could be useful for your group. We would 
also be curious to hear if anyone else has experience with this approach.

I hope this helps.
Jay

-Original Message-
From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of Nathaniel 
Clark
Sent: Wednesday, March 30, 2011 11:38 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] titering baculovirus ?

We don't have a problem getting them to stick to the plates in serum-free 
media, or in 5% FBS media.  The more challenging part is getting the plating 
density just right, too low and the plaques are too big, to high and they are 
too small.  Or if the cells dry out, or if your agarose overlay is too hot, 
etc...

However, we have actually stopped titering all together.  We find early stocks 
(from co-transfection, or plaque purification) are 'low', but after ~2 rounds 
of amplification in adherent culture of the 'low'
titer stock(using a large volume of low-titer virus in a t25 flask), we can add 
~0.5 ml to a 30-100 ml shake flask of Sf9's (serum free
media) and get a high titer stock( ie. 10^8 pfu/ml).  From there we amplify 
with ~ 10 mls of virus in a 1 L shaker culture, and we have our large volume 
high-titer stock.  Sometimes we will incubate the cells in pure virus stock in 
a t25 flask for 1 hour, take the virus off, and add fresh media, as a way to 
rescue low-titer stocks.

If you are just trying to titer (not plaque-purify), you can just take 10 fold 
dilutions of your virus, and do several small scale infections
in 6 well plates, 10 ml shaker cultures, whatever you prefer.   At the
lowest virus concentration where you see a synchronous infection (judged by 
protein expression levels, or cell-diameter if you have a cell counter, or by 
viewing with a trained eye), you call that an MOI=1.  From there you know the 
number of cells in the plate, and the volume of virus you added, so you can 
calculate an effective titer.
Plaque assays are really difficult and slow, and if you are just trying to make 
protein, an effective titer is fine, the absolute number isn't that helpful, Nat

Nat Clark
Graduate Student
Garman Lab
Biochemistry and Molecular Biology Dept.
UMass Amherst


On Wed, Mar 30, 2011 at 5:14 PM, Gloria Borgstahl gborgst...@gmail.com wrote:
 Hi Guys,
 we are learning to work with Sf9 cells and Carol in my lab wanted me 
 to ask you the following question.  Many thanks for any help, G

 I need to titer a baculovirus stock in my suspension-adapted Sf9 
 cells.   I know that these can be encouraged to attach better to 
 tissue culture plastic if they have added FBS (about 10%), but am not 
 sure that they will not be migrating and hiding plaques.  Does anyone 
 have suggestions about how to keep them more firmly anchored during 
 the baculovirus titration, or about another cell line that we could use 
 instead?


This message has been scanned for malware by Websense. www.websense.com


Re: [ccp4bb] [phenixbb] what to do with disordered side chains

2011-03-31 Thread Randy Read
Judging from some examples in the literature, if there's no density for the 
ligand, you publish in Nature!

On 31 Mar 2011, at 14:32, Huanwang Yang wrote:

 If a part of sequence has no density, this part will be cut from coordinates.
 If the side chain of a residue is  lack of density, the opinion is not 
 converged.
 What about the ligand, if no density is observed?
 
 Huanwang
 
 Ed Pozharski wrote:
 The results of the online survey on what to do with disordered side
 chains (from total of 240 responses):
 
 Delete the atoms 43%
 Let refinement take care of it by inflating B-factors41%
 Set occupancy to zero12%
 Other 4%
 
 Other suggestions were:
 
 - Place atoms in most likely spot based on rotomer and contacts and
 indicate high positional sigmas on ATMSIG records
 - To invent refinement that will spread this residues over many rotamers
 as this is what actually happened
 - Delet the atoms but retain the original amino acid name
 - choose the most common rotamer (B-factors don't inflate, they just
 rise slightly)
 - Depends. if the disordered region is unteresting, delete atoms.
 Otherwise, try to model it in one or more disordered model (and then
 state it clearly in the pdb file)
 - In case that no density is in the map, model several conformations of
 the missing segment and insert it into the PDB file with zero
 occupancies. It is equivalent what the NMR people do. - Model it in and 
 compare the MD simulations with SAXS
 - I would assumne Dale Tronrod suggestion the best. Sigatm labels.
 - Let the refinement inflate B-factors, then set occupancy to zero in
 the last round.
 
 Thanks to all for participation,
 
 Ed.
 
  

--
Randy J. Read
Department of Haematology, University of Cambridge
Cambridge Institute for Medical Research  Tel: + 44 1223 336500
Wellcome Trust/MRC Building   Fax: + 44 1223 336827
Hills RoadE-mail: rj...@cam.ac.uk
Cambridge CB2 0XY, U.K.   www-structmed.cimr.cam.ac.uk


[ccp4bb] Question about TEV cleavage

2011-03-31 Thread Prince, D Bryan
Hello all:

I am curious about what level of recovery is reasonable when performing
a TEV cleavage to remove 6HIS tags (N-terminal) from a protein.
Currently we are experiencing 50% loss of soluble protein after TEV
cleavage, and feel that this is too much. Are there better systems for
his tag cleavage that have better soluble protein recoveries than TEV?

Many thanks (in advance)

Bryan

--
Confidentiality Notice: This message is private and may contain confidential 
and proprietary information. If you have received this message in error, please 
notify us and remove it from your system and note that you must not copy, 
distribute or take any action in reliance on it. Any unauthorized use or 
disclosure of the contents of this message is not permitted and may be unlawful.
 


[ccp4bb] off topic: protease identification

2011-03-31 Thread Brett, Thomas
Hi all:
I was wondering if anyone had any tips on identifying proteases. I have a 
protein for which I know the proteolytic cleavage site. What are the best ways 
to identify the protease that does the cutting either:
1) bioinformatically (i.e., a good database to search using the cleavage site 
or a consensus)
2) experimentally (some engineered substrate to trap/identify the substrate or 
any other method?)
Thanks in advance 
-Tom

Tom J. Brett, PhD
Assistant Professor of Medicine
Division of Pulmonary and Critical Care
Washington University School of Medicine
Campus Box 8052, 660 S. Euclid
Saint Louis, MO 63110
http://brettlab.dom.wustl.edu/

Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Herman . Schreuder
Dear Quyen,
I am afraid you won't get any better answers than you got so far. There is no 
holy bible telling you what to do with disordered side chains. I fully agree 
with James that you should try to get the best possible model, which best 
explains your data and that will be your decision. Here are my 2 cents:
 
-If you see alternative positions, you have to build them. 
-If you do not see alternative positions, I would not replace one fantasy (some 
call it most likely) orientation with 2 or 3 fantasy orientations.
-I personally belong to the let the B-factors take care of it camp, but that 
is my personal opinion. Leaving side chains out could lead to 
misinterpretations by slightly less savy users of our data, especially when 
charge distributions are being studied. Besides, we know (almost) for sure that 
the side chain is there, it is only disordered and as we just learned, even 
slightly less savy users know what flaming red side chains mean. Even if they 
may not be mathematically entirely correct, huge B-factors clearly indicate 
that there is disorder involved.
-I would not let occupancies take up the slack since even very savy users have 
never heard of them and again, the side chain is fully occupied, only 
disordered. Of course if you build alternate positions, you have to divede the 
occupancies amongst them.
 
Best,
Herman




From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
Quyen Hoang
Sent: Thursday, March 31, 2011 3:55 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] what to do with disordered side chains



We are getting off topic a little bit.

Original topic: is it better to not build disordered sidechains or 
build them and let B-factors take care of it?
Ed's poll got almost a 50:50 split.
Question still unanswered.

Second topic introduced by Pavel: Your B-factors are valid within a 
harmonic (small) approximation of atomic vibrations. Larger scale motions you 
are talking about go beyond the harmonic approximation, and using the B-factor 
to model them is abusing the corresponding mathematical model. 
And that these large scale motions (disorders) are better represented 
by alternative conformations and associated with them occupancies.

My question is, how many people here do this?
If you're currently doing what Pavel suggested here, how do you decide 
where to keep the upper limit of B-factors and what the occupancies are for 
each atom (data with resolution of 2.0A or worse)? I mean, do you cap the 
B-factor at a reasonable number to represent natural atomic vibrations (which 
is very small as Pavel pointed out) and then let the occupancies pick up the 
slack? More importantly, what is your reason for doing this?

Cheers and thanks for your contribution,
Quyen


On Mar 30, 2011, at 5:20 PM, Pavel Afonine wrote:


Mark, 
alternative conformations and associated with them occupancies 
are to describe the larger scale disorder (the one that goes beyond the 
B-factor's capability to cope with). 
Multi-model PDB files is another option.
Best,
Pavel.



On Wed, Mar 30, 2011 at 2:15 PM, VAN RAAIJ , MARK JOHAN 
mjvanra...@cnb.csic.es wrote:


yet, apart from (and additionally to) modelling two 
conformations of the side-chain, the B-factor is the only tool we have (now). 

Quoting Pavel Afonine:

 Hi  Quyen,


 (...) And if B-factor is an estimate of thermo-motion 
(or static disorder),
 then would it not be reasonable to accept that 
building the side-chain and
 let B-factor sky rocket might reflect reality more 
so than not building it?


 NO.  Your B-factors are valid within a harmonic 
(small) approximation of
 atomic vibrations. Larger scale motions you are 
talking about go beyond the
 harmonic approximation, and using the B-factor to 
model them is abusing the
 corresponding mathematical model.
 
http://www.phenix-online.org/newsletter/CCN_2010_07.pdf

 Pavel.



Mark J van Raaij
Laboratorio M-4

Dpto de Estructura de Macromoléculas
Centro Nacional de Biotecnología - CSIC 

c/Darwin 3, Campus Cantoblanco
   

Re: [ccp4bb] Question about TEV cleavage

2011-03-31 Thread Xiaopeng Hu
Our experience is do not shake the tube during TEV cleavage,I dont know why, 
but it does help.

xiaopeng


Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Brad Bennett
Though I'm not meaning to turn this into a plaque assay burn session, for
many reasons we have also abandoned it and instead titer our baculovirus
stocks (P3 and P4) using flow cytometry. We use an antibody against viral
gp64 that has been labeled with PE and stain infected cells in 96 well
plates with the PE-antibody. We then measure PE fluorescence on a flow
cytometer (you can also do cell counts, viability determinations, etc.). We
equate 1 fluorescent event as 1 infected cell. Since we know how many cells
have been plated in each well, we can determine the percentage infected in
each well. We calculate a non-normalized titer from this data alone or we
compare this data to a standardized virus and determine a normalized titer
using a standard curve. From infection to having a titer in hand takes about
24 hours. Of course, the potential bottleneck is access to a flow cytometer!
I can give more experimental details off-board.

I should say that for getting an idea of relative titers and to test protein
expression on a small scale, we also do the effective titer tests as
suggested by Nat, with cell morphology and immunoblots as our read-out of
virus potency and recombinant protein expression, respectively. No doubt,
this will get you a long way but at some point, I argue, you need to
determine an actual, absolute titer for duplication of results,
troubleshooting, monitoring virus health over time, publications, etc.

HTH-
Brad

On Thu, Mar 31, 2011 at 9:23 AM, Bertrand, Jay Aaron [Nervianoms] 
jay.bertr...@nervianoms.com wrote:

 I checked with someone in our protein production group and got the
 following response:

 We also stopped doing the virus titration with the plaque assay and instead
 are performing expression test with different concentration of virus from
 the 3rd amplification. But for some viruses we still have doubts concerning
 the amplification success, so we are now evaluating a new technology using
 qPCR with the following kit (
 http://oetltd.com/products/category/baculoquant/). So you might have a
 look and see if it could be useful for your group. We would also be curious
 to hear if anyone else has experience with this approach.

 I hope this helps.
 Jay

 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 Nathaniel Clark
 Sent: Wednesday, March 30, 2011 11:38 PM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] titering baculovirus ?

 We don't have a problem getting them to stick to the plates in serum-free
 media, or in 5% FBS media.  The more challenging part is getting the plating
 density just right, too low and the plaques are too big, to high and they
 are too small.  Or if the cells dry out, or if your agarose overlay is too
 hot, etc...

 However, we have actually stopped titering all together.  We find early
 stocks (from co-transfection, or plaque purification) are 'low', but after
 ~2 rounds of amplification in adherent culture of the 'low'
 titer stock(using a large volume of low-titer virus in a t25 flask), we can
 add ~0.5 ml to a 30-100 ml shake flask of Sf9's (serum free
 media) and get a high titer stock( ie. 10^8 pfu/ml).  From there we
 amplify with ~ 10 mls of virus in a 1 L shaker culture, and we have our
 large volume high-titer stock.  Sometimes we will incubate the cells in pure
 virus stock in a t25 flask for 1 hour, take the virus off, and add fresh
 media, as a way to rescue low-titer stocks.

 If you are just trying to titer (not plaque-purify), you can just take 10
 fold dilutions of your virus, and do several small scale infections
 in 6 well plates, 10 ml shaker cultures, whatever you prefer.   At the
 lowest virus concentration where you see a synchronous infection (judged by
 protein expression levels, or cell-diameter if you have a cell counter, or
 by viewing with a trained eye), you call that an MOI=1.  From there you know
 the number of cells in the plate, and the volume of virus you added, so you
 can calculate an effective titer.
 Plaque assays are really difficult and slow, and if you are just trying to
 make protein, an effective titer is fine, the absolute number isn't that
 helpful, Nat

 Nat Clark
 Graduate Student
 Garman Lab
 Biochemistry and Molecular Biology Dept.
 UMass Amherst


 On Wed, Mar 30, 2011 at 5:14 PM, Gloria Borgstahl gborgst...@gmail.com
 wrote:
  Hi Guys,
  we are learning to work with Sf9 cells and Carol in my lab wanted me
  to ask you the following question.  Many thanks for any help, G
 
  I need to titer a baculovirus stock in my suspension-adapted Sf9
  cells.   I know that these can be encouraged to attach better to
  tissue culture plastic if they have added FBS (about 10%), but am not
  sure that they will not be migrating and hiding plaques.  Does anyone
  have suggestions about how to keep them more firmly anchored during
  the baculovirus titration, or about another cell line that we could use
 instead?


 This message has been scanned for malware by Websense. 

[ccp4bb] problem of conventions

2011-03-31 Thread Anita Lewit-Bentley

Dear all,

I would like to share my experiencde with a rather unexpected problem  
of indexing conventions. Perhaps I can save people some time


I have a crystal in the more unusual P21212 space-group (No 18). Its  
unit cell lengths are bac (please note). I systematically use XDS  
for data integration, since so far it was able to handle even the most  
horrible-looking spots.


Now XDS indexed my data in space-group 18, but with the axes order  
abc! It had, in fact, invented a space-group P22121, which does  
not exist. I did not realise this until I had spent a couple of weeks  
with beautiful peaks in rotation functions, but hopeless results in  
translation functions. It wasn't until I looked more closely into the  
definition of the screw axes that I realised the problem.


POINTLESS does not allow a reindexing of reflexions within the same  
space-group, but fortunately REINDEX did the trick at the level of  
intensities, because I like to use SCALA for careful scaling of my data.


So, basically, beyond just warning people who might encounter similar  
problems, I was wo,dering if XDS could perhaps reindex reflexions  
according to Int. Table conventions once the screw axes of a crystal  
system have been identified?


With best wishes,

Anita


Anita Lewit-Bentley
Unité d'Immunologie Structurale
CNRS URA 2185
Département de Biologie Structurale  Chimie
Institut Pasteur
25 rue du Dr. Roux
75724 Paris cedex 15
FRANCE

Tel: 33- (0)1 45 68 88 95
FAX: 33-(0)1 40 61 30 74
email: ale...@pasteur.fr



Re: [ccp4bb] problem of conventions

2011-03-31 Thread Santarsiero, Bernard D.
If you are using CCP4, it can accomodate P22121. However, just reindex in
CCP4 to the correct setting with P21212.

Bernie Santarsiero


On Thu, March 31, 2011 9:28 am, Anita Lewit-Bentley wrote:
 Dear all,

 I would like to share my experiencde with a rather unexpected problem
 of indexing conventions. Perhaps I can save people some time

 I have a crystal in the more unusual P21212 space-group (No 18). Its
 unit cell lengths are bac (please note). I systematically use XDS
 for data integration, since so far it was able to handle even the most
 horrible-looking spots.

 Now XDS indexed my data in space-group 18, but with the axes order
 abc! It had, in fact, invented a space-group P22121, which does
 not exist. I did not realise this until I had spent a couple of weeks
 with beautiful peaks in rotation functions, but hopeless results in
 translation functions. It wasn't until I looked more closely into the
 definition of the screw axes that I realised the problem.

 POINTLESS does not allow a reindexing of reflexions within the same
 space-group, but fortunately REINDEX did the trick at the level of
 intensities, because I like to use SCALA for careful scaling of my data.

 So, basically, beyond just warning people who might encounter similar
 problems, I was wo,dering if XDS could perhaps reindex reflexions
 according to Int. Table conventions once the screw axes of a crystal
 system have been identified?

 With best wishes,

 Anita


 Anita Lewit-Bentley
 Unité d'Immunologie Structurale
 CNRS URA 2185
 Département de Biologie Structurale  Chimie
 Institut Pasteur
 25 rue du Dr. Roux
 75724 Paris cedex 15
 FRANCE

 Tel: 33- (0)1 45 68 88 95
 FAX: 33-(0)1 40 61 30 74
 email: ale...@pasteur.fr




Re: [ccp4bb] Copy NCS chain in coot

2011-03-31 Thread zhang yu
Dear Javier,

Thanks for your reply. I am using coot 0.6.1, and I could copy molecules
with names besides A only if chains are already present in both NCS.

For chains which are only present in one NCS , I did find out a way to copy
to the other NCS by using 'Find NCS ligand' , although it is not a ligand.



For example, there are already chain A - chain D in the first NCS, and
chain E-chain H in the second NCS. I build chain 'M' from the scratch in
the first NCS, and I want to copy it to its NCS mate.


1. Change the master chain to chain A
  Draw-NCS ghost control-change to chain A

2. go to Extension-NCS-NCS ligands-,
In upper dialog box, Protein with NCS, Fill the master chain ID to A.
In the lower dialogue box Molecules contain ligands, choose the right
molecule and Fill 'M' and 'residue ranges'.

Coot will give a few positions to fit the chain, just choose the right one.

Yu

2011/3/31 Javier Garcia gar...@ysbl.york.ac.uk

 Dear Zhang yu,

 I also found that same problem recently. I think is a small bug from Coot.
 It will only copy NCS if the molecule's name you want to copy is molecule
 A.
 So maybe will work if you rename your built chain as A.

 Good luck!

 Javier


 On 31/03/11 00:16, zhang yu wrote:

 Dear all,

 I have two fold NCS in ASU.  I could find phase by molecular replacement,
 but I have to build other chains by myself, since the model is not complete.
 During model building in coot, I built one chain in one NCS, and I merged
 this chain into the previous molecule. Now it is the chain M'.

  When I tried to directly copy this chain to its NCS mate, It didn't work.

 I followed the instruction posted before
 http://www.mail-archive.com/coot@jiscmail.ac.uk/msg01984.html;.

 What I did is
 1. Change the master chain to chain M Draw-NCS ghost
 control-change to chain M
 (The terminal said 'there are no ghosts when I change the master chain to
 chain M)
 2. go to Extension-NCS-copy chain-, coot ask to fill the master chain
 ID, and I typed M

 Nothing happened.  Could someone tell me what is the correct procedure to
 do it ?

 Bye the way, for those chains present in both NCS maps, I modified one
 chain and applied the change to its NCS mate by using the same above
 procedure, and it worked.

 Thanks

 --
 Yu Zhang
 HHMI associate
 Waksman Institute, Rutgers University
 190 Frelinghuysen Rd.
 Piscataway, NJ, 08904






-- 
Yu Zhang
HHMI associate
Waksman Institute, Rutgers University
190 Frelinghuysen Rd.
Piscataway, NJ, 08904


Re: [ccp4bb] Question about TEV cleavage

2011-03-31 Thread Charles Allerston
I echo this.  

Some time ago I was working on a target which seemed to precipitate when 
cleaving overnight with TEV. I wasted a fair bit of time trying to optimise 
cleavage conditions with a myriad of buffers.  In the end, just by not 
agitating my solution, there was no precipitation and recovery was extremely 
good.  

I never agitate my solutions when cleaving now and have not had any problems 
with cleavage (to do with this issue, anyway) or recovery.


cheers
charlie



Dr. Charles Allerston
Genome Integrity Group
Structural Genomics Consortium
Nuffield Department of Medicine
Old Road Campus
University of Oxford
OX3 7DQ
http://www.sgc.ox.ac.uk/





-Original Message-
From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of Xiaopeng 
Hu
Sent: 31 March 2011 15:23
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] Question about TEV cleavage

Our experience is do not shake the tube during TEV cleavage,I dont know why, 
but it does help.

xiaopeng


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Jacob Keller
Why not have the b-factors take care of it until some magic cutoff
number? When they reach the cutoff, two things happen:

1. Occupancies are set to zero for those side chains, to represent our
lack of ability to model the region,

2. B-factors are set to exactly 500, as a flag allowing casual
b-factor-savvy users to identify suspicious regions, since they will
probably not see occupancies, but *will* see b-factors. Therefore, all
0-occupancy atoms will automatically have b-factors = 500. I believe
it is true that if the occupancies are zero, the b-factors are totally
irrelevant for all calculations?

Doesn't this satisfy both parties?

Jacob





On Thu, Mar 31, 2011 at 9:22 AM,  herman.schreu...@sanofi-aventis.com wrote:
 Dear Quyen,
 I am afraid you won't get any better answers than you got so far. There is
 no holy bible telling you what to do with disordered side chains. I fully
 agree with James that you should try to get the best possible model, which
 best explains your data and that will be your decision. Here are my 2 cents:

 -If you see alternative positions, you have to build them.
 -If you do not see alternative positions, I would not replace one fantasy
 (some call it most likely) orientation with 2 or 3 fantasy orientations.
 -I personally belong to the let the B-factors take care of it camp, but
 that is my personal opinion. Leaving side chains out could lead to
 misinterpretations by slightly less savy users of our data, especially when
 charge distributions are being studied. Besides, we know (almost) for sure
 that the side chain is there, it is only disordered and as we just learned,
 even slightly less savy users know what flaming red side chains mean. Even
 if they may not be mathematically entirely correct, huge B-factors clearly
 indicate that there is disorder involved.
 -I would not let occupancies take up the slack since even very savy users
 have never heard of them and again, the side chain is fully occupied, only
 disordered. Of course if you build alternate positions, you have to divede
 the occupancies amongst them.

 Best,
 Herman

 
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of Quyen
 Hoang
 Sent: Thursday, March 31, 2011 3:55 PM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] what to do with disordered side chains

 We are getting off topic a little bit.
 Original topic: is it better to not build disordered sidechains or build
 them and let B-factors take care of it?
 Ed's poll got almost a 50:50 split.
 Question still unanswered.
 Second topic introduced by Pavel: Your B-factors are valid within a
 harmonic (small) approximation of atomic vibrations. Larger scale motions
 you are talking about go beyond the harmonic approximation, and using the
 B-factor to model them is abusing the corresponding mathematical model.
 And that these large scale motions (disorders) are better represented by
 alternative conformations and associated with them occupancies.
 My question is, how many people here do this?
 If you're currently doing what Pavel suggested here, how do you decide where
 to keep the upper limit of B-factors and what the occupancies are for each
 atom (data with resolution of 2.0A or worse)? I mean, do you cap the
 B-factor at a reasonable number to represent natural atomic vibrations
 (which is very small as Pavel pointed out) and then let the occupancies pick
 up the slack? More importantly, what is your reason for doing this?
 Cheers and thanks for your contribution,
 Quyen

 On Mar 30, 2011, at 5:20 PM, Pavel Afonine wrote:

 Mark,
 alternative conformations and associated with them occupancies are to
 describe the larger scale disorder (the one that goes beyond the B-factor's
 capability to cope with).
 Multi-model PDB files is another option.
 Best,
 Pavel.


 On Wed, Mar 30, 2011 at 2:15 PM, VAN RAAIJ , MARK JOHAN
 mjvanra...@cnb.csic.es wrote:

 yet, apart from (and additionally to) modelling two conformations of the
 side-chain, the B-factor is the only tool we have (now).
 Quoting Pavel Afonine:

  Hi  Quyen,
 
 
  (...) And if B-factor is an estimate of thermo-motion (or static
  disorder),
  then would it not be reasonable to accept that building the side-chain
  and
  let B-factor sky rocket might reflect reality more so than not building
  it?
 
 
  NO.  Your B-factors are valid within a harmonic (small) approximation of
  atomic vibrations. Larger scale motions you are talking about go beyond
  the
  harmonic approximation, and using the B-factor to model them is abusing
  the
  corresponding mathematical model.
  http://www.phenix-online.org/newsletter/CCN_2010_07.pdf
 
  Pavel.
 

 Mark J van Raaij
 Laboratorio M-4
 Dpto de Estructura de Macromoléculas
 Centro Nacional de Biotecnología - CSIC
 c/Darwin 3, Campus Cantoblanco
 28049 Madrid
 tel. 91 585 4616
 email: mjvanra...@cnb.csic.es






-- 
***
Jacob Pearson Keller
Northwestern University
Medical 

Re: [ccp4bb] problem of conventions

2011-03-31 Thread Phil Evans
The IUCr standard is to make abc, see

http://nvl.nist.gov/pub/nistpubs/jres/107/4/j74mig.pdf
http://nvl.nist.gov/pub/nistpubs/jres/106/6/j66mig.pdf

and P 2 21 21 is a perfectly valid space group

Pointless will reindex within the same point group, or you can choose the abc 
convention (SETTING CELL_BASED) or the reference setting P 21 21 2 (SETTING 
SYMMETRY_BASED), so you have a choice

Most programs are perfectly happy with space group P 2 21 21 (I'm not sure 
about [auto]Sharp)

Phil

On 31 Mar 2011, at 15:28, Anita Lewit-Bentley wrote:

 Dear all,
 
 I would like to share my experiencde with a rather unexpected problem of 
 indexing conventions. Perhaps I can save people some time
 
 I have a crystal in the more unusual P21212 space-group (No 18). Its unit 
 cell lengths are bac (please note). I systematically use XDS for data 
 integration, since so far it was able to handle even the most 
 horrible-looking spots.
 
 Now XDS indexed my data in space-group 18, but with the axes order abc! It 
 had, in fact, invented a space-group P22121, which does not exist. I did 
 not realise this until I had spent a couple of weeks with beautiful peaks in 
 rotation functions, but hopeless results in translation functions. It wasn't 
 until I looked more closely into the definition of the screw axes that I 
 realised the problem.
 
 POINTLESS does not allow a reindexing of reflexions within the same 
 space-group, but fortunately REINDEX did the trick at the level of 
 intensities, because I like to use SCALA for careful scaling of my data.
 
 So, basically, beyond just warning people who might encounter similar 
 problems, I was wo,dering if XDS could perhaps reindex reflexions according 
 to Int. Table conventions once the screw axes of a crystal system have been 
 identified?
 
 With best wishes,
 
 Anita
 
 
 Anita Lewit-Bentley
 Unité d'Immunologie Structurale
 CNRS URA 2185
 Département de Biologie Structurale  Chimie
 Institut Pasteur
 25 rue du Dr. Roux
 75724 Paris cedex 15
 FRANCE
 
 Tel: 33- (0)1 45 68 88 95
 FAX: 33-(0)1 40 61 30 74
 email: ale...@pasteur.fr
 


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Nat Echols
On Thu, Mar 31, 2011 at 7:42 AM, Jacob Keller 
j-kell...@fsm.northwestern.edu wrote:

 Why not have the b-factors take care of it until some magic cutoff
 number? When they reach the cutoff, two things happen:

 1. Occupancies are set to zero for those side chains, to represent our
 lack of ability to model the region,

 2. B-factors are set to exactly 500, as a flag allowing casual
 b-factor-savvy users to identify suspicious regions, since they will
 probably not see occupancies, but *will* see b-factors. Therefore, all
 0-occupancy atoms will automatically have b-factors = 500. I believe
 it is true that if the occupancies are zero, the b-factors are totally
 irrelevant for all calculations?

 Doesn't this satisfy both parties?


No, because now you're not only presenting the user with made-up
coordinates, you're giving them a made-up B-factor as well, so there is
effectively no property of those atoms that is based on experimental data
rather than subjective criteria.  Regardless of any  problems inherent in
letting the B-factors take care of all forms of disorder, they are
nonetheless a refined parameter.

-Nat


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Phoebe Rice
 - they all know what B is and how to look for regions of high B 
 (with, say, pymol) and they know not to make firm conclusions about H-bonds 
 to flaming red side chains.

But this knowledge may be quite wrong.  If the flaming red really indicates 
large vibrational motion then yes, one whould not bet on stable H-bonds.
But if the flaming red indicates that a well-ordered sidechain was incorrectly
modeled at full occupancy when in fact it is only present at half-occupancy
then no, the H-bond could be strong but only present in that half-occupancy
conformation.  One presumes that the other half-occupancy location (perhaps
missing from the model) would have its own H-bonding network.


I beg to differ.  If a side chain has 2 or more positions, one should be a bit 
careful about making firm conclusions based on only one of those, even if it 
isn't clear exactly why one should use caution.  Also, isn't the isotropic B we 
fit at medium resolution more of a spherical cow approximation to physical 
reality anyway?

  Phoebe

  


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Jacob Keller
Well, I guess I was thinking to make the b-factor such a preposterous
value that no one would possibly believe it. Setting occupancies to
zero effectively places a stumbling block, because people see the
residues and think they are actually supported by data. So to
counter-balance this, I thought putting up a high-b-factor flag would
prevent people from tripping over the stumbling block. Look, you could
even set the b-factor to 1 if you want--just something so people
totally discount those coordinates.

Jacob

On Thu, Mar 31, 2011 at 9:55 AM, Nat Echols nathaniel.ech...@gmail.com wrote:
 On Thu, Mar 31, 2011 at 7:42 AM, Jacob Keller
 j-kell...@fsm.northwestern.edu wrote:

 Why not have the b-factors take care of it until some magic cutoff
 number? When they reach the cutoff, two things happen:

 1. Occupancies are set to zero for those side chains, to represent our
 lack of ability to model the region,

 2. B-factors are set to exactly 500, as a flag allowing casual
 b-factor-savvy users to identify suspicious regions, since they will
 probably not see occupancies, but *will* see b-factors. Therefore, all
 0-occupancy atoms will automatically have b-factors = 500. I believe
 it is true that if the occupancies are zero, the b-factors are totally
 irrelevant for all calculations?

 Doesn't this satisfy both parties?

 No, because now you're not only presenting the user with made-up
 coordinates, you're giving them a made-up B-factor as well, so there is
 effectively no property of those atoms that is based on experimental data
 rather than subjective criteria.  Regardless of any  problems inherent in
 letting the B-factors take care of all forms of disorder, they are
 nonetheless a refined parameter.
 -Nat



-- 
***
Jacob Pearson Keller
Northwestern University
Medical Scientist Training Program
cel: 773.608.9185
email: j-kell...@northwestern.edu
***


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Herman . Schreuder
Dear Jacob,

What do we gain? As Dale pointed out, we are already abusing either occupancy, 
B-factor or delete the side chain to compensate for our inability to tell the 
user that the side chain is disordered. With your proposal, we would fudge both 
occupancy and B-factor, which in my eyes is even worse as fudging just one of 
the two. 

Also, who should decide on the magic number: the all-knowing gurus at the 
protein data bank? Maybe we should really start using cif files, which allow to 
specify coordinate uncertainties.

Best regards,
Herman

 

-Original Message-
From: Jacob Keller [mailto:j-kell...@fsm.northwestern.edu] 
Sent: Thursday, March 31, 2011 4:43 PM
To: Schreuder, Herman RD/DE
Cc: CCP4BB@jiscmail.ac.uk
Subject: Re: [ccp4bb] what to do with disordered side chains

Why not have the b-factors take care of it until some magic cutoff number? 
When they reach the cutoff, two things happen:

1. Occupancies are set to zero for those side chains, to represent our lack of 
ability to model the region,

2. B-factors are set to exactly 500, as a flag allowing casual b-factor-savvy 
users to identify suspicious regions, since they will probably not see 
occupancies, but *will* see b-factors. Therefore, all 0-occupancy atoms will 
automatically have b-factors = 500. I believe it is true that if the 
occupancies are zero, the b-factors are totally irrelevant for all calculations?

Doesn't this satisfy both parties?

Jacob





On Thu, Mar 31, 2011 at 9:22 AM,  herman.schreu...@sanofi-aventis.com wrote:
 Dear Quyen,
 I am afraid you won't get any better answers than you got so far. 
 There is no holy bible telling you what to do with disordered side 
 chains. I fully agree with James that you should try to get the best 
 possible model, which best explains your data and that will be your decision. 
 Here are my 2 cents:

 -If you see alternative positions, you have to build them.
 -If you do not see alternative positions, I would not replace one 
 fantasy (some call it most likely) orientation with 2 or 3 fantasy 
 orientations.
 -I personally belong to the let the B-factors take care of it camp, 
 but that is my personal opinion. Leaving side chains out could lead to 
 misinterpretations by slightly less savy users of our data, especially 
 when charge distributions are being studied. Besides, we know (almost) 
 for sure that the side chain is there, it is only disordered and as we 
 just learned, even slightly less savy users know what flaming red side 
 chains mean. Even if they may not be mathematically entirely correct, 
 huge B-factors clearly indicate that there is disorder involved.
 -I would not let occupancies take up the slack since even very savy 
 users have never heard of them and again, the side chain is fully 
 occupied, only disordered. Of course if you build alternate positions, 
 you have to divede the occupancies amongst them.

 Best,
 Herman

 
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
 Quyen Hoang
 Sent: Thursday, March 31, 2011 3:55 PM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] what to do with disordered side chains

 We are getting off topic a little bit.
 Original topic: is it better to not build disordered sidechains or 
 build them and let B-factors take care of it?
 Ed's poll got almost a 50:50 split.
 Question still unanswered.
 Second topic introduced by Pavel: Your B-factors are valid within a 
 harmonic (small) approximation of atomic vibrations. Larger scale 
 motions you are talking about go beyond the harmonic approximation, 
 and using the B-factor to model them is abusing the corresponding 
 mathematical model.
 And that these large scale motions (disorders) are better represented 
 by alternative conformations and associated with them occupancies.
 My question is, how many people here do this?
 If you're currently doing what Pavel suggested here, how do you decide 
 where to keep the upper limit of B-factors and what the occupancies 
 are for each atom (data with resolution of 2.0A or worse)? I mean, do 
 you cap the B-factor at a reasonable number to represent natural 
 atomic vibrations (which is very small as Pavel pointed out) and then 
 let the occupancies pick up the slack? More importantly, what is your reason 
 for doing this?
 Cheers and thanks for your contribution, Quyen

 On Mar 30, 2011, at 5:20 PM, Pavel Afonine wrote:

 Mark,
 alternative conformations and associated with them occupancies are to 
 describe the larger scale disorder (the one that goes beyond the 
 B-factor's capability to cope with).
 Multi-model PDB files is another option.
 Best,
 Pavel.


 On Wed, Mar 30, 2011 at 2:15 PM, VAN RAAIJ , MARK JOHAN 
 mjvanra...@cnb.csic.es wrote:

 yet, apart from (and additionally to) modelling two conformations of 
 the side-chain, the B-factor is the only tool we have (now).
 Quoting Pavel Afonine:

  Hi  Quyen,
 
 
  (...) And if B-factor is an estimate of thermo-motion 

Re: [ccp4bb] Question about TEV cleavage

2011-03-31 Thread Laurie Betts
We have the same experience with the DO NOT AGITATE.  We purify our own
His-tagged TEV protease, and flash-freeze it IMMEDIATELY after IMAC prep at
about 1 mg/mL.

Laurie Betts
UNC



On Thu, Mar 31, 2011 at 10:37 AM, Charles Allerston 
charles.allers...@sgc.ox.ac.uk wrote:

 I echo this.

 Some time ago I was working on a target which seemed to precipitate when
 cleaving overnight with TEV. I wasted a fair bit of time trying to optimise
 cleavage conditions with a myriad of buffers.  In the end, just by not
 agitating my solution, there was no precipitation and recovery was extremely
 good.

 I never agitate my solutions when cleaving now and have not had any
 problems with cleavage (to do with this issue, anyway) or recovery.


 cheers
 charlie



 Dr. Charles Allerston
 Genome Integrity Group
 Structural Genomics Consortium
 Nuffield Department of Medicine
 Old Road Campus
 University of Oxford
 OX3 7DQ
 http://www.sgc.ox.ac.uk/





 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 Xiaopeng Hu
 Sent: 31 March 2011 15:23
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] Question about TEV cleavage

 Our experience is do not shake the tube during TEV cleavage,I dont know
 why, but it does help.

 xiaopeng



Re: [ccp4bb] off topic: protease identification

2011-03-31 Thread S. Jack Lin
Hi Tom,

You can try the MEROPS database. There is a search feature called What
peptidases can cleave this bond.

http://merops.sanger.ac.uk/cgi-bin/specsearch.pl

Cheers,
Jack


On Thu, Mar 31, 2011 at 6:59 AM, Brett, Thomas tbr...@dom.wustl.edu wrote:

 Hi all:
 I was wondering if anyone had any tips on identifying proteases. I have a
 protein for which I know the proteolytic cleavage site. What are the best
 ways to identify the protease that does the cutting either:
 1) bioinformatically (i.e., a good database to search using the cleavage
 site or a consensus)
 2) experimentally (some engineered substrate to trap/identify the substrate
 or any other method?)
 Thanks in advance
 -Tom

 Tom J. Brett, PhD
 Assistant Professor of Medicine
 Division of Pulmonary and Critical Care
 Washington University School of Medicine
 Campus Box 8052, 660 S. Euclid
 Saint Louis, MO 63110
 http://brettlab.dom.wustl.edu/


Re: [ccp4bb] Question about TEV cleavage

2011-03-31 Thread Klaus Piontek

And not at full moon!

Klaus


Am 31.03.2011 um 16:23 schrieb Xiaopeng Hu:

Our experience is do not shake the tube during TEV cleavage,I dont  
know why, but it does help.


xiaopeng


Dr. Klaus Piontek
Albert-Ludwigs-University Freiburg
Institute of Organic Chemistry and Biochemistry, Room 401 H  
Albertstrasse 21

D-79104 Freiburg Germany
Phone: ++49-761-203-6036
Fax: ++49-761-203-8714
Email: klaus.pion...@ocbc.uni-freiburg.de
Web: http://www.chemie.uni-freiburg.de/orgbio/w3platt/



Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Quyen Hoang

Thank you for your post, Herman.
Since there is no holy bible to provide guidance, perhaps we should  
hold off the idea of electing a powerful dictator to enforce this?
- at least until we all can come to a consensus on how the dictator  
should dictate...


Cheers,
Quyen


On Mar 31, 2011, at 10:22 AM, herman.schreu...@sanofi-aventis.com wrote:


Dear Quyen,
I am afraid you won't get any better answers than you got so far.  
There is no holy bible telling you what to do with disordered side  
chains. I fully agree with James that you should try to get the best  
possible model, which best explains your data and that will be your  
decision. Here are my 2 cents:


-If you see alternative positions, you have to build them.
-If you do not see alternative positions, I would not replace one  
fantasy (some call it most likely) orientation with 2 or 3 fantasy  
orientations.
-I personally belong to the let the B-factors take care of it  
camp, but that is my personal opinion. Leaving side chains out could  
lead to misinterpretations by slightly less savy users of our data,  
especially when charge distributions are being studied. Besides, we  
know (almost) for sure that the side chain is there, it is only  
disordered and as we just learned, even slightly less savy users  
know what flaming red side chains mean. Even if they may not be  
mathematically entirely correct, huge B-factors clearly indicate  
that there is disorder involved.
-I would not let occupancies take up the slack since even very savy  
users have never heard of them and again, the side chain is fully  
occupied, only disordered. Of course if you build alternate  
positions, you have to divede the occupancies amongst them.


Best,
Herman

From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf  
Of Quyen Hoang

Sent: Thursday, March 31, 2011 3:55 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] what to do with disordered side chains

We are getting off topic a little bit.

Original topic: is it better to not build disordered sidechains or  
build them and let B-factors take care of it?

Ed's poll got almost a 50:50 split.
Question still unanswered.

Second topic introduced by Pavel: Your B-factors are valid within a  
harmonic (small) approximation of atomic vibrations. Larger scale  
motions you are talking about go beyond the harmonic approximation,  
and using the B-factor to model them is abusing the corresponding  
mathematical model.
And that these large scale motions (disorders) are better  
represented by alternative conformations and associated with them  
occupancies.


My question is, how many people here do this?
If you're currently doing what Pavel suggested here, how do you  
decide where to keep the upper limit of B-factors and what the  
occupancies are for each atom (data with resolution of 2.0A or  
worse)? I mean, do you cap the B-factor at a reasonable number to  
represent natural atomic vibrations (which is very small as Pavel  
pointed out) and then let the occupancies pick up the slack? More  
importantly, what is your reason for doing this?


Cheers and thanks for your contribution,
Quyen


On Mar 30, 2011, at 5:20 PM, Pavel Afonine wrote:


Mark,
alternative conformations and associated with them occupancies are  
to describe the larger scale disorder (the one that goes beyond the  
B-factor's capability to cope with).

Multi-model PDB files is another option.
Best,
Pavel.


On Wed, Mar 30, 2011 at 2:15 PM, VAN RAAIJ , MARK JOHAN mjvanra...@cnb.csic.es 
 wrote:
yet, apart from (and additionally to) modelling two conformations  
of the side-chain, the B-factor is the only tool we have (now).


Quoting Pavel Afonine:

 Hi  Quyen,


 (...) And if B-factor is an estimate of thermo-motion (or static  
disorder),
 then would it not be reasonable to accept that building the side- 
chain and
 let B-factor sky rocket might reflect reality more so than not  
building it?



 NO.  Your B-factors are valid within a harmonic (small)  
approximation of
 atomic vibrations. Larger scale motions you are talking about go  
beyond the
 harmonic approximation, and using the B-factor to model them is  
abusing the

 corresponding mathematical model.
 http://www.phenix-online.org/newsletter/CCN_2010_07.pdf

 Pavel.


Mark J van Raaij
Laboratorio M-4
Dpto de Estructura de Macromoléculas
Centro Nacional de Biotecnología - CSIC

c/Darwin 3, Campus Cantoblanco
28049 Madrid
tel. 91 585 4616
email: mjvanra...@cnb.csic.es








Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Ed Pozharski
On Thu, 2011-03-31 at 17:04 +0200, herman.schreu...@sanofi-aventis.com
wrote:
 Maybe we should really start using cif files, which allow to specify
 coordinate uncertainties.

PDB has SIGATM record for that purpose

-- 
I'd jump in myself, if I weren't so good at whistling.
   Julian, King of Lemurs


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Peter Keller
On Thu, 2011-03-31 at 11:27 -0400, Quyen Hoang wrote:
 Thank you for your post, Herman.
 Since there is no holy bible to provide guidance, perhaps we should
 hold off the idea of electing a powerful dictator to enforce this?
 - at least until we all can come to a consensus on how the dictator
 should dictate...

Well, that is partly what we have the IUCr for, isn't it? A couple of
people have referred to the wwPDB in this context, but IMHO the IUCr is
a much better forum to try to reach some kind of decision about these
issues. If the IUCr has a clear policy, the wwPDB can enforce it (like
they did with the deposition of structure factors). If the wwPDB takes a
lead a lot of people will get annoyed at them, when the real problem is
that crystallographic practitioners haven't come to any agreement
amongst themselves.

And yes, I know that balancing questions of scientific correctness with
the needs of more or less naive consumers of the data isn't
straightforward :-)

Regards,
Peter.

-- 
Peter Keller Tel.: +44 (0)1223 353033
Global Phasing Ltd., Fax.: +44 (0)1223 366889
Sheraton House,
Castle Park,
Cambridge CB3 0AX
United Kingdom


[ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Zbyszek Otwinowski
The B-factor in crystallography represents the convolution (sum) of two
types of uncertainties about the atom (electron cloud) position:

1) dispersion of atom positions in crystal lattice
2) uncertainty of the experimenter's knowledge  about the atom position.

In general, uncertainty needs not to be described by Gaussian function.
However, communicating uncertainty using the second moment of its
distribution is a widely accepted practice, with frequently implied
meaning that it corresponds to a Gaussian probability function. B-factor
is simply a scaled (by 8 times pi squared) second moment of uncertainty
distribution.

In the previous, long thread, confusion was generated by the additional
assumption that B-factor also corresponds to a Gaussian probability
distribution and not just to a second moment of any probability
distribution. Crystallographic literature often implies the Gaussian
shape, so there is some justification for such an interpretation, where
the more complex probability distribution is represented by the sum of
displaced Gaussians, where the area under each Gaussian component
corresponds to the occupancy of an alternative conformation.

For data with a typical resolution for macromolecular crystallography,
such multi-Gaussian description of the atom position's uncertainty is not
practical, as it would lead to instability in the refinement and/or
overfitting. Due to this, a simplified description of the atom's position
uncertainty by just the second moment of probability distribution is the
right approach. For this reason, the PDB format is highly suitable for the
description of positional uncertainties,  the only difference with other
fields being the unusual form of squaring and then scaling up the standard
uncertainty. As this calculation can be easily inverted, there is no loss
of information. However, in teaching one should probably stress more this
unusual form of presenting the standard deviation.

A separate issue is the use of restraints on B-factor values, a subject
that probably needs a longer discussion.

With respect to the previous thread, representing poorly-ordered (so
called 'disordered') side chains by the most likely conformer with
appropriately high B-factors is fully justifiable, and currently is
probably the best solution to a difficult problem.

Zbyszek Otwinowski



 - they all know what B is and how to look for regions of high B
 (with, say, pymol) and they know not to make firm conclusions about
 H-bonds
 to flaming red side chains.

But this knowledge may be quite wrong.  If the flaming red really
 indicates
large vibrational motion then yes, one whould not bet on stable H-bonds.
But if the flaming red indicates that a well-ordered sidechain was
 incorrectly
modeled at full occupancy when in fact it is only present at
 half-occupancy
then no, the H-bond could be strong but only present in that
 half-occupancy
conformation.  One presumes that the other half-occupancy location
 (perhaps
missing from the model) would have its own H-bonding network.


 I beg to differ.  If a side chain has 2 or more positions, one should be a
 bit careful about making firm conclusions based on only one of those, even
 if it isn't clear exactly why one should use caution.  Also, isn't the
 isotropic B we fit at medium resolution more of a spherical cow
 approximation to physical reality anyway?

   Phoebe





Zbyszek Otwinowski
UT Southwestern Medical Center at Dallas
5323 Harry Hines Blvd.
Dallas, TX 75390-8816
Tel. 214-645-6385
Fax. 214-645-6353


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Roberts, Sue A - (suer)
Regarding suggestions that the pdb or the IUCR to tell us what to do:

IMO -

Neither of the usual solutions - (a) deleting side chains when there is no 
density or (b) letting B factors go where they will - is without problems (this 
is clear from the ongoing discussion).  I would be really unhappy if some 
authority unilaterally imposed either of these solutions on the protein 
crystallographic community.

Sue


Dr. Sue A. Roberts
Dept. of Chemistry and Biochemistry
University of Arizona
1041 E. Lowell St.,  Tucson, AZ 85721
Phone: 520 621 8171
s...@email.arizona.edu
http://www.biochem.arizona.edu/xray











Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Gerard Bricogne
Dear Quyen,

On Thu, Mar 31, 2011 at 11:27:58AM -0400, Quyen Hoang wrote:
 Thank you for your post, Herman.
 Since there is no holy bible to provide guidance, perhaps we should hold 
 off the idea of electing a powerful dictator to enforce this?
 - at least until we all can come to a consensus on how the dictator 
 should dictate...


 ... but that might well be even harder than to decide what to do with
disordered side chains ... .


 With best wishes,
 
  Gerard.

--

 ===
 * *
 * Gerard Bricogne g...@globalphasing.com  *
 * *
 * Global Phasing Ltd. *
 * Sheraton House, Castle Park Tel: +44-(0)1223-353033 *
 * Cambridge CB3 0AX, UK   Fax: +44-(0)1223-366889 *
 * *
 ===


 On Mar 31, 2011, at 10:22 AM, herman.schreu...@sanofi-aventis.com wrote:

 Dear Quyen,
 I am afraid you won't get any better answers than you got so far. There is 
 no holy bible telling you what to do with disordered side chains. I fully 
 agree with James that you should try to get the best possible model, which 
 best explains your data and that will be your decision. Here are my 2 
 cents:

 -If you see alternative positions, you have to build them.
 -If you do not see alternative positions, I would not replace one fantasy 
 (some call it most likely) orientation with 2 or 3 fantasy orientations.
 -I personally belong to the let the B-factors take care of it camp, but 
 that is my personal opinion. Leaving side chains out could lead to 
 misinterpretations by slightly less savy users of our data, especially 
 when charge distributions are being studied. Besides, we know (almost) for 
 sure that the side chain is there, it is only disordered and as we just 
 learned, even slightly less savy users know what flaming red side chains 
 mean. Even if they may not be mathematically entirely correct, huge 
 B-factors clearly indicate that there is disorder involved.
 -I would not let occupancies take up the slack since even very savy users 
 have never heard of them and again, the side chain is fully occupied, only 
 disordered. Of course if you build alternate positions, you have to divede 
 the occupancies amongst them.

 Best,
 Herman

 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
 Quyen Hoang
 Sent: Thursday, March 31, 2011 3:55 PM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] what to do with disordered side chains

 We are getting off topic a little bit.

 Original topic: is it better to not build disordered sidechains or build 
 them and let B-factors take care of it?
 Ed's poll got almost a 50:50 split.
 Question still unanswered.

 Second topic introduced by Pavel: Your B-factors are valid within a 
 harmonic (small) approximation of atomic vibrations. Larger scale motions 
 you are talking about go beyond the harmonic approximation, and using the 
 B-factor to model them is abusing the corresponding mathematical model.
 And that these large scale motions (disorders) are better represented by 
 alternative conformations and associated with them occupancies.

 My question is, how many people here do this?
 If you're currently doing what Pavel suggested here, how do you decide 
 where to keep the upper limit of B-factors and what the occupancies are 
 for each atom (data with resolution of 2.0A or worse)? I mean, do you cap 
 the B-factor at a reasonable number to represent natural atomic vibrations 
 (which is very small as Pavel pointed out) and then let the occupancies 
 pick up the slack? More importantly, what is your reason for doing this?

 Cheers and thanks for your contribution,
 Quyen


 On Mar 30, 2011, at 5:20 PM, Pavel Afonine wrote:

 Mark,
 alternative conformations and associated with them occupancies are to 
 describe the larger scale disorder (the one that goes beyond the 
 B-factor's capability to cope with).
 Multi-model PDB files is another option.
 Best,
 Pavel.


 On Wed, Mar 30, 2011 at 2:15 PM, VAN RAAIJ , MARK JOHAN 
 mjvanra...@cnb.csic.es wrote:
 yet, apart from (and additionally to) modelling two conformations of the 
 side-chain, the B-factor is the only tool we have (now).

 Quoting Pavel Afonine:

  Hi  Quyen,
 
 
  (...) And if B-factor is an estimate of thermo-motion (or static 
 disorder),
  then would it not be reasonable to accept that building the side-chain 
 and
  let B-factor sky rocket might reflect reality more so than not 
 building it?
 
 
  NO.  Your B-factors are valid within a harmonic (small) approximation 
 of
  atomic vibrations. Larger scale motions you are talking about go beyond 
 the
  harmonic approximation, and using the 

Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Katya Heldwein
A related question: how do most people amplify their baculovirus
stocks? Adherent cultures vs suspension? Fold dilution at each stage
(P1 to P2, P2 to P3)? Duration of each amplification stage?

We have some viral stocks that go off rather quickly (1-2 months)
despite being stored with FBS in a cool, dark place.


Katya


On Thu, Mar 31, 2011 at 10:23 AM, Brad Bennett bradbennet...@gmail.com wrote:
 Though I'm not meaning to turn this into a plaque assay burn session, for
 many reasons we have also abandoned it and instead titer our baculovirus
 stocks (P3 and P4) using flow cytometry. We use an antibody against viral
 gp64 that has been labeled with PE and stain infected cells in 96 well
 plates with the PE-antibody. We then measure PE fluorescence on a flow
 cytometer (you can also do cell counts, viability determinations, etc.). We
 equate 1 fluorescent event as 1 infected cell. Since we know how many cells
 have been plated in each well, we can determine the percentage infected in
 each well. We calculate a non-normalized titer from this data alone or we
 compare this data to a standardized virus and determine a normalized titer
 using a standard curve. From infection to having a titer in hand takes about
 24 hours. Of course, the potential bottleneck is access to a flow cytometer!
 I can give more experimental details off-board.
 I should say that for getting an idea of relative titers and to test protein
 expression on a small scale, we also do the effective titer tests as
 suggested by Nat, with cell morphology and immunoblots as our read-out of
 virus potency and recombinant protein expression, respectively. No doubt,
 this will get you a long way but at some point, I argue, you need to
 determine an actual, absolute titer for duplication of results,
 troubleshooting, monitoring virus health over time, publications, etc.

 HTH-
 Brad

 On Thu, Mar 31, 2011 at 9:23 AM, Bertrand, Jay Aaron [Nervianoms]
 jay.bertr...@nervianoms.com wrote:

 I checked with someone in our protein production group and got the
 following response:

 We also stopped doing the virus titration with the plaque assay and
 instead are performing expression test with different concentration of virus
 from the 3rd amplification. But for some viruses we still have doubts
 concerning the amplification success, so we are now evaluating a new
 technology using qPCR with the following kit
 (http://oetltd.com/products/category/baculoquant/). So you might have a look
 and see if it could be useful for your group. We would also be curious to
 hear if anyone else has experience with this approach.

 I hope this helps.
 Jay

 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 Nathaniel Clark
 Sent: Wednesday, March 30, 2011 11:38 PM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] titering baculovirus ?

 We don't have a problem getting them to stick to the plates in serum-free
 media, or in 5% FBS media.  The more challenging part is getting the plating
 density just right, too low and the plaques are too big, to high and they
 are too small.  Or if the cells dry out, or if your agarose overlay is too
 hot, etc...

 However, we have actually stopped titering all together.  We find early
 stocks (from co-transfection, or plaque purification) are 'low', but after
 ~2 rounds of amplification in adherent culture of the 'low'
 titer stock(using a large volume of low-titer virus in a t25 flask), we
 can add ~0.5 ml to a 30-100 ml shake flask of Sf9's (serum free
 media) and get a high titer stock( ie. 10^8 pfu/ml).  From there we
 amplify with ~ 10 mls of virus in a 1 L shaker culture, and we have our
 large volume high-titer stock.  Sometimes we will incubate the cells in pure
 virus stock in a t25 flask for 1 hour, take the virus off, and add fresh
 media, as a way to rescue low-titer stocks.

 If you are just trying to titer (not plaque-purify), you can just take 10
 fold dilutions of your virus, and do several small scale infections
 in 6 well plates, 10 ml shaker cultures, whatever you prefer.   At the
 lowest virus concentration where you see a synchronous infection (judged
 by protein expression levels, or cell-diameter if you have a cell counter,
 or by viewing with a trained eye), you call that an MOI=1.  From there you
 know the number of cells in the plate, and the volume of virus you added, so
 you can calculate an effective titer.
 Plaque assays are really difficult and slow, and if you are just trying to
 make protein, an effective titer is fine, the absolute number isn't that
 helpful, Nat

 Nat Clark
 Graduate Student
 Garman Lab
 Biochemistry and Molecular Biology Dept.
 UMass Amherst


 On Wed, Mar 30, 2011 at 5:14 PM, Gloria Borgstahl gborgst...@gmail.com
 wrote:
  Hi Guys,
  we are learning to work with Sf9 cells and Carol in my lab wanted me
  to ask you the following question.  Many thanks for any help, G
 
  I need to titer a baculovirus stock in my suspension-adapted Sf9
 

Re: [ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Frank von Delft
This is a lovely summary, and we should make our students read it. - But 
I'm afraid I do not see how it supports the closing statement in the 
last paragraph... phx.



On 31/03/2011 17:06, Zbyszek Otwinowski wrote:

The B-factor in crystallography represents the convolution (sum) of two
types of uncertainties about the atom (electron cloud) position:

1) dispersion of atom positions in crystal lattice
2) uncertainty of the experimenter's knowledge  about the atom position.

In general, uncertainty needs not to be described by Gaussian function.
However, communicating uncertainty using the second moment of its
distribution is a widely accepted practice, with frequently implied
meaning that it corresponds to a Gaussian probability function. B-factor
is simply a scaled (by 8 times pi squared) second moment of uncertainty
distribution.

In the previous, long thread, confusion was generated by the additional
assumption that B-factor also corresponds to a Gaussian probability
distribution and not just to a second moment of any probability
distribution. Crystallographic literature often implies the Gaussian
shape, so there is some justification for such an interpretation, where
the more complex probability distribution is represented by the sum of
displaced Gaussians, where the area under each Gaussian component
corresponds to the occupancy of an alternative conformation.

For data with a typical resolution for macromolecular crystallography,
such multi-Gaussian description of the atom position's uncertainty is not
practical, as it would lead to instability in the refinement and/or
overfitting. Due to this, a simplified description of the atom's position
uncertainty by just the second moment of probability distribution is the
right approach. For this reason, the PDB format is highly suitable for the
description of positional uncertainties,  the only difference with other
fields being the unusual form of squaring and then scaling up the standard
uncertainty. As this calculation can be easily inverted, there is no loss
of information. However, in teaching one should probably stress more this
unusual form of presenting the standard deviation.

A separate issue is the use of restraints on B-factor values, a subject
that probably needs a longer discussion.

With respect to the previous thread, representing poorly-ordered (so
called 'disordered') side chains by the most likely conformer with
appropriately high B-factors is fully justifiable, and currently is
probably the best solution to a difficult problem.

Zbyszek Otwinowski




- they all know what B is and how to look for regions of high B
(with, say, pymol) and they know not to make firm conclusions about
H-bonds
to flaming red side chains.

But this knowledge may be quite wrong.  If the flaming red really
indicates
large vibrational motion then yes, one whould not bet on stable H-bonds.
But if the flaming red indicates that a well-ordered sidechain was
incorrectly
modeled at full occupancy when in fact it is only present at
half-occupancy
then no, the H-bond could be strong but only present in that
half-occupancy
conformation.  One presumes that the other half-occupancy location
(perhaps
missing from the model) would have its own H-bonding network.


I beg to differ.  If a side chain has 2 or more positions, one should be a
bit careful about making firm conclusions based on only one of those, even
if it isn't clear exactly why one should use caution.  Also, isn't the
isotropic B we fit at medium resolution more of a spherical cow
approximation to physical reality anyway?

   Phoebe





Zbyszek Otwinowski
UT Southwestern Medical Center at Dallas
5323 Harry Hines Blvd.
Dallas, TX 75390-8816
Tel. 214-645-6385
Fax. 214-645-6353


Re: [ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Hailiang Zhang
Dear Zbyszek:

Thanks a lot for your good summary. It is very interesting but, do you
think there are some references for more detailed description, especially
from mathematics point of view about correlating B-factor to the Gaussian
probability distribution (the B-factor unit of A^2 is my first doubt as
for the probability distribution description)? Thanks again for your
efforts!

Best Regards, Hailiang


 The B-factor in crystallography represents the convolution (sum) of two
 types of uncertainties about the atom (electron cloud) position:

 1) dispersion of atom positions in crystal lattice
 2) uncertainty of the experimenter's knowledge  about the atom position.

 In general, uncertainty needs not to be described by Gaussian function.
 However, communicating uncertainty using the second moment of its
 distribution is a widely accepted practice, with frequently implied
 meaning that it corresponds to a Gaussian probability function. B-factor
 is simply a scaled (by 8 times pi squared) second moment of uncertainty
 distribution.

 In the previous, long thread, confusion was generated by the additional
 assumption that B-factor also corresponds to a Gaussian probability
 distribution and not just to a second moment of any probability
 distribution. Crystallographic literature often implies the Gaussian
 shape, so there is some justification for such an interpretation, where
 the more complex probability distribution is represented by the sum of
 displaced Gaussians, where the area under each Gaussian component
 corresponds to the occupancy of an alternative conformation.

 For data with a typical resolution for macromolecular crystallography,
 such multi-Gaussian description of the atom position's uncertainty is not
 practical, as it would lead to instability in the refinement and/or
 overfitting. Due to this, a simplified description of the atom's position
 uncertainty by just the second moment of probability distribution is the
 right approach. For this reason, the PDB format is highly suitable for the
 description of positional uncertainties,  the only difference with other
 fields being the unusual form of squaring and then scaling up the standard
 uncertainty. As this calculation can be easily inverted, there is no loss
 of information. However, in teaching one should probably stress more this
 unusual form of presenting the standard deviation.

 A separate issue is the use of restraints on B-factor values, a subject
 that probably needs a longer discussion.

 With respect to the previous thread, representing poorly-ordered (so
 called 'disordered') side chains by the most likely conformer with
 appropriately high B-factors is fully justifiable, and currently is
 probably the best solution to a difficult problem.

 Zbyszek Otwinowski



 - they all know what B is and how to look for regions of high B
 (with, say, pymol) and they know not to make firm conclusions about
 H-bonds
 to flaming red side chains.

But this knowledge may be quite wrong.  If the flaming red really
 indicates
large vibrational motion then yes, one whould not bet on stable H-bonds.
But if the flaming red indicates that a well-ordered sidechain was
 incorrectly
modeled at full occupancy when in fact it is only present at
 half-occupancy
then no, the H-bond could be strong but only present in that
 half-occupancy
conformation.  One presumes that the other half-occupancy location
 (perhaps
missing from the model) would have its own H-bonding network.


 I beg to differ.  If a side chain has 2 or more positions, one should be
 a
 bit careful about making firm conclusions based on only one of those,
 even
 if it isn't clear exactly why one should use caution.  Also, isn't the
 isotropic B we fit at medium resolution more of a spherical cow
 approximation to physical reality anyway?

   Phoebe





 Zbyszek Otwinowski
 UT Southwestern Medical Center at Dallas
 5323 Harry Hines Blvd.
 Dallas, TX 75390-8816
 Tel. 214-645-6385
 Fax. 214-645-6353




Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Nathaniel Clark
We do adherent if we have a small volume of low titer virus, but as
soon as we have a decent titer we will start with a 30 ml suspension
flask.  Normally I add something like 0.5 ml, which is probably more
then I need.  We only use serum free media right now.  As I mentioned,
if you need to boost the titer of a low titer stock, we plate a t25
flask with Sf9s, take off the media, and incubate it in 1 ml of pure
virus (or two-fold diluted) for an hour or so, then remove (and save)
that virus stock and give the cells fresh media.

For fold dilution, just do a rough calculation of titer.  Once you are
at a high titer assume you  have 10^8-10^9 pfu/ml, so you can estimate
the ballpark volume to infect a given number of cells at an MOI of 1.
From early stocks you might have a titer of 10^6-10^7 pfu/ml.  If you
do an amplification that doesn't work, decide if you were to high
(MOI=10 not great for amplifying) or too low (MOI=0.1) and try
again.  Better yet, just set up 3 amplifications at the same time,
whichever worked the best, use for another round of amplification, or
for infections.

For duration, we do 3 days amplification (post-infection), but
sometimes we let them go a few days longer, which allows a secondary
infection of the cells that weren't infected by the viral innoculum.
There doesn't seem to be any harm to letting infections go longer; at
some point all the cells die, but since the virus is pretty stable,
it's no problem.  We have a CEDEX cell counter which tells you average
cell diameter, which is a great read out for infection.  Also looking
and density and viability versus an uninfected control will tell you
if the infection worked.
Nat

Nat Clark
Graduate Student
Garman Lab
Biochemistry and Molecular Biology Dept.
UMass Amherst

On Thu, Mar 31, 2011 at 12:36 PM, Katya Heldwein
katya.heldw...@gmail.com wrote:
 A related question: how do most people amplify their baculovirus
 stocks? Adherent cultures vs suspension? Fold dilution at each stage
 (P1 to P2, P2 to P3)? Duration of each amplification stage?

 We have some viral stocks that go off rather quickly (1-2 months)
 despite being stored with FBS in a cool, dark place.


 Katya


 On Thu, Mar 31, 2011 at 10:23 AM, Brad Bennett bradbennet...@gmail.com 
 wrote:
 Though I'm not meaning to turn this into a plaque assay burn session, for
 many reasons we have also abandoned it and instead titer our baculovirus
 stocks (P3 and P4) using flow cytometry. We use an antibody against viral
 gp64 that has been labeled with PE and stain infected cells in 96 well
 plates with the PE-antibody. We then measure PE fluorescence on a flow
 cytometer (you can also do cell counts, viability determinations, etc.). We
 equate 1 fluorescent event as 1 infected cell. Since we know how many cells
 have been plated in each well, we can determine the percentage infected in
 each well. We calculate a non-normalized titer from this data alone or we
 compare this data to a standardized virus and determine a normalized titer
 using a standard curve. From infection to having a titer in hand takes about
 24 hours. Of course, the potential bottleneck is access to a flow cytometer!
 I can give more experimental details off-board.
 I should say that for getting an idea of relative titers and to test protein
 expression on a small scale, we also do the effective titer tests as
 suggested by Nat, with cell morphology and immunoblots as our read-out of
 virus potency and recombinant protein expression, respectively. No doubt,
 this will get you a long way but at some point, I argue, you need to
 determine an actual, absolute titer for duplication of results,
 troubleshooting, monitoring virus health over time, publications, etc.

 HTH-
 Brad

 On Thu, Mar 31, 2011 at 9:23 AM, Bertrand, Jay Aaron [Nervianoms]
 jay.bertr...@nervianoms.com wrote:

 I checked with someone in our protein production group and got the
 following response:

 We also stopped doing the virus titration with the plaque assay and
 instead are performing expression test with different concentration of virus
 from the 3rd amplification. But for some viruses we still have doubts
 concerning the amplification success, so we are now evaluating a new
 technology using qPCR with the following kit
 (http://oetltd.com/products/category/baculoquant/). So you might have a look
 and see if it could be useful for your group. We would also be curious to
 hear if anyone else has experience with this approach.

 I hope this helps.
 Jay

 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 Nathaniel Clark
 Sent: Wednesday, March 30, 2011 11:38 PM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] titering baculovirus ?

 We don't have a problem getting them to stick to the plates in serum-free
 media, or in 5% FBS media.  The more challenging part is getting the plating
 density just right, too low and the plaques are too big, to high and they
 are too small.  Or if the cells dry 

Re: [ccp4bb] step refine speed of wincoot - a little drawback of the new interruptible_fit_protein() function in wincoot

2011-03-31 Thread Bernhard C. Lohkamp
Thank you for the comment. You are correct that there are being backups
made in the newer interruptable fit function (they were't previously in
the uninterruptable fitting). I wasnt thinking about that (*). This may
(especially on windows) slow down things a great deal as we write and
gzip these files as well. To note here, gzip may be slow on windows as
is writing files. In the pre-release versions I disabled the gzip on
windows by default (and in general made the compression of backup files
optional). This may speed up things (in certain circumstances).
I hope this clarifies things a bit...

B

(*) It's debatable if we want backups here or not. If you can and want
to interrupt the fitting you may want to go back one step, which
requires a backup. On the other side this will write a lot of backup
files which is slowing thing down by writing and compressing files, as
well as producing a lot of backup files. Discuss... (or Paul and I just
decide).

 Hi Xiaopeng, and those who are using the new interruptible
 fit_protein... or stepped_refine... in wincoot 0.6.1:


 I just took a look at wincoot 0.6.1's fitting.py file
 (WinCoot\share\coot\python\fitting.py). It seems that in the old
 fit_protein() and stepped_refine_protein() functions, the backup
 was programmed to be turned off during the cycles when it steps
 through the protein.

 However when using the Stepped refine... from the Extensions menu,
 the program starts to spit out tons of backup files. I then checked
 the extension.py. It turns out that these menu items are calling the
 new interruptible_fit_protein() function instead of the old
 uninterruptible ones. Some further experiment showed that this
 behavior was dependent on the backup state variable of the model
 being refined. When the backup state is 1 (the default value?),
 these new interruptible fitting functions will save a compressed
 backup file on each step.

 So, in order to save time when using the new interruptible protein
 fitting functions, one can type the following command in the python
 script window to tweak the backup state to 0:

 make_backup(imol) #to save a backup
 turn_off_backup(imol)
 where imol is the number associated with the pdb file being fitted.

 Do not forget to turn it back on after the fitting:

 turn_on_backup(imol)

 To check the current model's backup state:

 backup_state(imol)


 Zhijie


 --
 From: Xiaopeng Hu huxp...@mail.sysu.edu.cn
 Sent: Tuesday, March 29, 2011 9:02 AM
 To: Zhijie Li zhijie...@utoronto.ca
 Subject: Re: [ccp4bb] step refine speed of wincoot

 Dear Zhijie,

 Thanks for there suggestions.

 I tried to turn off backup (with script command) and it did help
 although not much. Is there any way to add this perference to setting
 of coot/wincoot?


 best

 xiaopeng

 - 原始邮件 -
 发件人: Zhijie Li zhijie...@utoronto.ca
 收件人: Xiaopeng Hu huxp...@mail.sysu.edu.cn
 发送时间: 星期二, 2011年 3 月 29日 下午 7:46:05
 主题: Re: [ccp4bb] step refine speed of wincoot

 Hi,

 If you feel the refinement to be too slow, you may turn off the smooth
 centering (in preferences) or change the centering steps to a smaller
 number
 to save unnecessary graphical calculations. To go extreme, you may even
 remove the real time display commands in the scripts - also a way to
 test if
 the difference observed is due to different graphical efficiency.
 Reducing
 the size of the displayed map also helps.

 The other thing you may need to consider is that coot/wincoot will
 save a
 backup file each time it updates the model, which means on each step
 of the
 refinement you have a new file generated (take a look at the coot-backup
 directory when doing stepped refinement). If your windows disk is
 terribly
 fragmented then sure you will spend a lot of time on writing these
 files.
 The other thing is, windows has a different file caching mechanism than
 linux, this can also cause a huge difference when a few hundred small
 temporary files are queued for writing. My impression is that both the
 ext2/3 file system and the way linux handles caching are more
 efficient for
 this kind of situations. You may try deleting the files in
 wincoot\coot-backup periodically and defragmenting that partition.
 Making a
 virtual disk in the RAM to put your backup directory there could be
 something to experiment on too.

 Zhijie
 --
 From: Xiaopeng Hu huxp...@mail.sysu.edu.cn
 Sent: Sunday, March 20, 2011 9:22 AM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: [ccp4bb] step refine speed of wincoot

 Dear all,

 I found the step refine speed of wincoot is much slower than that of
 linux
 coot (with the same pc). Is it normal or I need to configure something
 with the wincoot?

 best,

 Xiaopeng Hu



Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Jacob Keller
 What do we gain? As Dale pointed out, we are already abusing either 
 occupancy, B-factor or delete the side chain to compensate for our inability 
 to tell the user that the side chain is disordered. With your proposal, we 
 would fudge both occupancy and B-factor, which in my eyes is even worse as 
 fudging just one of the two.


We gain clarity to the non-crystallographer user: a b-factor of 278.9
sounds like possibly something real. A b-factor of exactly 1000 does
not. Both probably have the same believability, viz., ~zero. Also,
setting occupancy = zero is not fudging but rather respectfully
declining to comment based on lack of data. I think it is exactly the
same as omitting residues one cannot see in the density.


 Also, who should decide on the magic number: the all-knowing gurus at the 
 protein data bank? Maybe we should really start using cif files, which allow 
 to specify coordinate uncertainties.


I think a reasonable number could be derived and agreed upon, and
would not be surprised if there is such a derivation or analysis in
the literature answering the question:

At what b-factor does modelling an atom become insignificant with
respect to explaining/predicting/fitting the data?

That point would be the b-factor/occupancy cutoff.

JPK


Re: [ccp4bb] problem of conventions

2011-03-31 Thread Tim Gruene
Dear Anita,

I happen to have a very similar problem today.

Does XDS use the desired setting if you provide it with the correct cell and
space group during the IDXREF step? You can otherwise re-index in CORRECT.


To comment on Phil:
I fed the mtz-file from pointless into ctruncate (or maybe it was scala) which
left the space group string (P2 21 21) but turned the space group number 18 into
3018 - this does screw up autosharp and maybe also other programs which use the
space group number/ symbol and not the symmetry operators.

Tim

On Thu, Mar 31, 2011 at 04:28:18PM +0200, Anita Lewit-Bentley wrote:
 Dear all,
 
 I would like to share my experiencde with a rather unexpected
 problem of indexing conventions. Perhaps I can save people some
 time
 
 I have a crystal in the more unusual P21212 space-group (No 18). Its
 unit cell lengths are bac (please note). I systematically use XDS
 for data integration, since so far it was able to handle even the
 most horrible-looking spots.
 
 Now XDS indexed my data in space-group 18, but with the axes order
 abc! It had, in fact, invented a space-group P22121, which does
 not exist. I did not realise this until I had spent a couple of
 weeks with beautiful peaks in rotation functions, but hopeless
 results in translation functions. It wasn't until I looked more
 closely into the definition of the screw axes that I realised the
 problem.
 
 POINTLESS does not allow a reindexing of reflexions within the same
 space-group, but fortunately REINDEX did the trick at the level of
 intensities, because I like to use SCALA for careful scaling of my
 data.
 
 So, basically, beyond just warning people who might encounter
 similar problems, I was wo,dering if XDS could perhaps reindex
 reflexions according to Int. Table conventions once the screw axes
 of a crystal system have been identified?
 
 With best wishes,
 
 Anita
 
 
 Anita Lewit-Bentley
 Unité d'Immunologie Structurale
 CNRS URA 2185
 Département de Biologie Structurale  Chimie
 Institut Pasteur
 25 rue du Dr. Roux
 75724 Paris cedex 15
 FRANCE
 
 Tel: 33- (0)1 45 68 88 95
 FAX: 33-(0)1 40 61 30 74
 email: ale...@pasteur.fr
 

-- 
--
Tim Gruene
Institut fuer anorganische Chemie
Tammannstr. 4
D-37077 Goettingen

phone: +49 (0)551 39 22149

GPG Key ID = A46BEE1A



signature.asc
Description: Digital signature


Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Nathaniel Clark
We don't do anything special to store the virus, just a dark fridge.
You do need to mix well as the virus can settle.

You might want to try this 'TIPS' approach.  Basically you freeze
baculovirus infected cells, and use them as your virus innoculum.  I
think I tried it once and it didn't really work, but I didn't really
try too hard...
There is a research paper on the subject that I can't find, but here
is the first thing I found on google:
http://www.abrf.org/ResearchGroups/ProteinExpressionResearchGroup/Activities/Wasilko-PERG2007.pdf

Nat

On Thu, Mar 31, 2011 at 1:12 PM, chitra shintre
chitra.shin...@gmail.com wrote:
 Hi,
 We have observed this virus going off within 2 months as well. ...any idea
 why this happens?
 Chitra

 On 31 March 2011 18:07, Nathaniel Clark nathanielcl...@gmail.com wrote:

 We do adherent if we have a small volume of low titer virus, but as
 soon as we have a decent titer we will start with a 30 ml suspension
 flask.  Normally I add something like 0.5 ml, which is probably more
 then I need.  We only use serum free media right now.  As I mentioned,
 if you need to boost the titer of a low titer stock, we plate a t25
 flask with Sf9s, take off the media, and incubate it in 1 ml of pure
 virus (or two-fold diluted) for an hour or so, then remove (and save)
 that virus stock and give the cells fresh media.

 For fold dilution, just do a rough calculation of titer.  Once you are
 at a high titer assume you  have 10^8-10^9 pfu/ml, so you can estimate
 the ballpark volume to infect a given number of cells at an MOI of 1.
 From early stocks you might have a titer of 10^6-10^7 pfu/ml.  If you
 do an amplification that doesn't work, decide if you were to high
 (MOI=10 not great for amplifying) or too low (MOI=0.1) and try
 again.  Better yet, just set up 3 amplifications at the same time,
 whichever worked the best, use for another round of amplification, or
 for infections.

 For duration, we do 3 days amplification (post-infection), but
 sometimes we let them go a few days longer, which allows a secondary
 infection of the cells that weren't infected by the viral innoculum.
 There doesn't seem to be any harm to letting infections go longer; at
 some point all the cells die, but since the virus is pretty stable,
 it's no problem.  We have a CEDEX cell counter which tells you average
 cell diameter, which is a great read out for infection.  Also looking
 and density and viability versus an uninfected control will tell you
 if the infection worked.
 Nat

 Nat Clark
 Graduate Student
 Garman Lab
 Biochemistry and Molecular Biology Dept.
 UMass Amherst

 On Thu, Mar 31, 2011 at 12:36 PM, Katya Heldwein
 katya.heldw...@gmail.com wrote:
  A related question: how do most people amplify their baculovirus
  stocks? Adherent cultures vs suspension? Fold dilution at each stage
  (P1 to P2, P2 to P3)? Duration of each amplification stage?
 
  We have some viral stocks that go off rather quickly (1-2 months)
  despite being stored with FBS in a cool, dark place.
 
 
  Katya
 
 
  On Thu, Mar 31, 2011 at 10:23 AM, Brad Bennett bradbennet...@gmail.com
  wrote:
  Though I'm not meaning to turn this into a plaque assay burn session,
  for
  many reasons we have also abandoned it and instead titer our
  baculovirus
  stocks (P3 and P4) using flow cytometry. We use an antibody against
  viral
  gp64 that has been labeled with PE and stain infected cells in 96 well
  plates with the PE-antibody. We then measure PE fluorescence on a flow
  cytometer (you can also do cell counts, viability determinations,
  etc.). We
  equate 1 fluorescent event as 1 infected cell. Since we know how many
  cells
  have been plated in each well, we can determine the percentage infected
  in
  each well. We calculate a non-normalized titer from this data alone
  or we
  compare this data to a standardized virus and determine a normalized
  titer
  using a standard curve. From infection to having a titer in hand takes
  about
  24 hours. Of course, the potential bottleneck is access to a flow
  cytometer!
  I can give more experimental details off-board.
  I should say that for getting an idea of relative titers and to test
  protein
  expression on a small scale, we also do the effective titer tests as
  suggested by Nat, with cell morphology and immunoblots as our read-out
  of
  virus potency and recombinant protein expression, respectively. No
  doubt,
  this will get you a long way but at some point, I argue, you need to
  determine an actual, absolute titer for duplication of results,
  troubleshooting, monitoring virus health over time, publications, etc.
 
  HTH-
  Brad
 
  On Thu, Mar 31, 2011 at 9:23 AM, Bertrand, Jay Aaron [Nervianoms]
  jay.bertr...@nervianoms.com wrote:
 
  I checked with someone in our protein production group and got the
  following response:
 
  We also stopped doing the virus titration with the plaque assay and
  instead are performing expression test with different concentration of
  virus
  

Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread chitra shintre
Hi,
We have observed this virus going off within 2 months as well. ...any idea
why this happens?

Chitra

On 31 March 2011 18:07, Nathaniel Clark nathanielcl...@gmail.com wrote:

 We do adherent if we have a small volume of low titer virus, but as
 soon as we have a decent titer we will start with a 30 ml suspension
 flask.  Normally I add something like 0.5 ml, which is probably more
 then I need.  We only use serum free media right now.  As I mentioned,
 if you need to boost the titer of a low titer stock, we plate a t25
 flask with Sf9s, take off the media, and incubate it in 1 ml of pure
 virus (or two-fold diluted) for an hour or so, then remove (and save)
 that virus stock and give the cells fresh media.

 For fold dilution, just do a rough calculation of titer.  Once you are
 at a high titer assume you  have 10^8-10^9 pfu/ml, so you can estimate
 the ballpark volume to infect a given number of cells at an MOI of 1.
 From early stocks you might have a titer of 10^6-10^7 pfu/ml.  If you
 do an amplification that doesn't work, decide if you were to high
 (MOI=10 not great for amplifying) or too low (MOI=0.1) and try
 again.  Better yet, just set up 3 amplifications at the same time,
 whichever worked the best, use for another round of amplification, or
 for infections.

 For duration, we do 3 days amplification (post-infection), but
 sometimes we let them go a few days longer, which allows a secondary
 infection of the cells that weren't infected by the viral innoculum.
 There doesn't seem to be any harm to letting infections go longer; at
 some point all the cells die, but since the virus is pretty stable,
 it's no problem.  We have a CEDEX cell counter which tells you average
 cell diameter, which is a great read out for infection.  Also looking
 and density and viability versus an uninfected control will tell you
 if the infection worked.
 Nat

 Nat Clark
 Graduate Student
 Garman Lab
 Biochemistry and Molecular Biology Dept.
 UMass Amherst

 On Thu, Mar 31, 2011 at 12:36 PM, Katya Heldwein
 katya.heldw...@gmail.com wrote:
  A related question: how do most people amplify their baculovirus
  stocks? Adherent cultures vs suspension? Fold dilution at each stage
  (P1 to P2, P2 to P3)? Duration of each amplification stage?
 
  We have some viral stocks that go off rather quickly (1-2 months)
  despite being stored with FBS in a cool, dark place.
 
 
  Katya
 
 
  On Thu, Mar 31, 2011 at 10:23 AM, Brad Bennett bradbennet...@gmail.com
 wrote:
  Though I'm not meaning to turn this into a plaque assay burn session,
 for
  many reasons we have also abandoned it and instead titer our baculovirus
  stocks (P3 and P4) using flow cytometry. We use an antibody against
 viral
  gp64 that has been labeled with PE and stain infected cells in 96 well
  plates with the PE-antibody. We then measure PE fluorescence on a flow
  cytometer (you can also do cell counts, viability determinations, etc.).
 We
  equate 1 fluorescent event as 1 infected cell. Since we know how many
 cells
  have been plated in each well, we can determine the percentage infected
 in
  each well. We calculate a non-normalized titer from this data alone or
 we
  compare this data to a standardized virus and determine a normalized
 titer
  using a standard curve. From infection to having a titer in hand takes
 about
  24 hours. Of course, the potential bottleneck is access to a flow
 cytometer!
  I can give more experimental details off-board.
  I should say that for getting an idea of relative titers and to test
 protein
  expression on a small scale, we also do the effective titer tests as
  suggested by Nat, with cell morphology and immunoblots as our read-out
 of
  virus potency and recombinant protein expression, respectively. No
 doubt,
  this will get you a long way but at some point, I argue, you need to
  determine an actual, absolute titer for duplication of results,
  troubleshooting, monitoring virus health over time, publications, etc.
 
  HTH-
  Brad
 
  On Thu, Mar 31, 2011 at 9:23 AM, Bertrand, Jay Aaron [Nervianoms]
  jay.bertr...@nervianoms.com wrote:
 
  I checked with someone in our protein production group and got the
  following response:
 
  We also stopped doing the virus titration with the plaque assay and
  instead are performing expression test with different concentration of
 virus
  from the 3rd amplification. But for some viruses we still have doubts
  concerning the amplification success, so we are now evaluating a new
  technology using qPCR with the following kit
  (http://oetltd.com/products/category/baculoquant/). So you might have
 a look
  and see if it could be useful for your group. We would also be curious
 to
  hear if anyone else has experience with this approach.
 
  I hope this helps.
  Jay
 
  -Original Message-
  From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
  Nathaniel Clark
  Sent: Wednesday, March 30, 2011 11:38 PM
  To: CCP4BB@JISCMAIL.AC.UK
  Subject: Re: [ccp4bb] 

Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Nat Echols
On Thu, Mar 31, 2011 at 10:14 AM, Jacob Keller 
j-kell...@fsm.northwestern.edu wrote:

  Also,
 setting occupancy = zero is not fudging but rather respectfully
 declining to comment based on lack of data. I think it is exactly the
 same as omitting residues one cannot see in the density.


No, it's not the same.  If you have placed any atoms, even with zero
occupancy, you have said something about where you expect the atoms to be,
or at least where the refinement program thinks they should be.  Declining
to comment would be deleting them, not guessing.

I think a reasonable number could be derived and agreed upon, and
 would not be surprised if there is such a derivation or analysis in
 the literature answering the question:

 At what b-factor does modelling an atom become insignificant with
 respect to explaining/predicting/fitting the data?

 That point would be the b-factor/occupancy cutoff.


Although atoms with very high B-factors may have almost no impact on
F(calc), if the occupancy is non-zero they will still be driven by gradients
with respect to X-ray data, and their positions (or changes thereof) will in
turn affect other atoms, through geometry restraints if not F(calc).  So
there is no point at which these atoms cease to be relevant to the task of
fitting.

-Nat


Re: [ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Ethan Merritt
On Thursday, March 31, 2011 10:05:22 am Hailiang Zhang wrote:
 Dear Zbyszek:
 
 Thanks a lot for your good summary. It is very interesting but, do you
 think there are some references for more detailed description, especially
 from mathematics point of view about correlating B-factor to the Gaussian
 probability distribution (the B-factor unit of A^2 is my first doubt as
 for the probability distribution description)? Thanks again for your
 efforts!
 
 Best Regards, Hailiang

I already cited the IUCr standard once, but here it is again:
 Trueblood, et al, 1996; Acta Cryst. A52, 770-781 
 http://dx.doi.org/10.1107/S0108767396005697



-- 
Ethan A Merritt
Biomolecular Structure Center,  K-428 Health Sciences Bldg
University of Washington, Seattle 98195-7742


Re: [ccp4bb] problem of conventions

2011-03-31 Thread Ian Tickle
 To comment on Phil:
 I fed the mtz-file from pointless into ctruncate (or maybe it was scala) which
 left the space group string (P2 21 21) but turned the space group number 18 
 into
 3018 - this does screw up autosharp and maybe also other programs which use 
 the
 space group number/ symbol and not the symmetry operators.

If that's the case autosharp should be fixed so it recognises the
correct space group! (in this P22121 or #3018 in CCP4-ese).  It has
been fixed in autoBuster so maybe you're using an old version of
autosharp?

-- Ian


Re: [ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Dale Tronrud

   While what you say here is quite true and is useful for us to
remember, your list is quite short.  I can add another

3) The systematic error introduced by assuming full occupancy for all sites.

There are, of course, many other factors that we don't account for
that our refinement programs tend to dump into the B factors.

   The definition of that number in the PDB file, as listed in the mmCIF
dictionary, only includes your first factor --

http://mmcif.rcsb.org/dictionaries/mmcif_std.dic/Items/_atom_site.B_iso_or_equiv.html

and these numbers are routinely interpreted as though that definition is
the law.  Certainly the whole basis of TLS refinement is that the B factors
are really Atomic Displacement Parameters.   In addition the stereochemical
restraints on B factors are derived from the assumption that these parameters
are ADPs.  Convoluting all these other factors with the ADPs causes serious
problems for those who analyze B factors as measures of motion.

   The fact that current refinement programs mix all these factors with the
ADP for an atom to produce a vaguely defined B factor should be considered
a flaw to be corrected and not an opportunity to pile even more factors into
this field in the PDB file.

Dale Tronrud


On 3/31/2011 9:06 AM, Zbyszek Otwinowski wrote:

The B-factor in crystallography represents the convolution (sum) of two
types of uncertainties about the atom (electron cloud) position:

1) dispersion of atom positions in crystal lattice
2) uncertainty of the experimenter's knowledge  about the atom position.

In general, uncertainty needs not to be described by Gaussian function.
However, communicating uncertainty using the second moment of its
distribution is a widely accepted practice, with frequently implied
meaning that it corresponds to a Gaussian probability function. B-factor
is simply a scaled (by 8 times pi squared) second moment of uncertainty
distribution.

In the previous, long thread, confusion was generated by the additional
assumption that B-factor also corresponds to a Gaussian probability
distribution and not just to a second moment of any probability
distribution. Crystallographic literature often implies the Gaussian
shape, so there is some justification for such an interpretation, where
the more complex probability distribution is represented by the sum of
displaced Gaussians, where the area under each Gaussian component
corresponds to the occupancy of an alternative conformation.

For data with a typical resolution for macromolecular crystallography,
such multi-Gaussian description of the atom position's uncertainty is not
practical, as it would lead to instability in the refinement and/or
overfitting. Due to this, a simplified description of the atom's position
uncertainty by just the second moment of probability distribution is the
right approach. For this reason, the PDB format is highly suitable for the
description of positional uncertainties,  the only difference with other
fields being the unusual form of squaring and then scaling up the standard
uncertainty. As this calculation can be easily inverted, there is no loss
of information. However, in teaching one should probably stress more this
unusual form of presenting the standard deviation.

A separate issue is the use of restraints on B-factor values, a subject
that probably needs a longer discussion.

With respect to the previous thread, representing poorly-ordered (so
called 'disordered') side chains by the most likely conformer with
appropriately high B-factors is fully justifiable, and currently is
probably the best solution to a difficult problem.

Zbyszek Otwinowski




- they all know what B is and how to look for regions of high B
(with, say, pymol) and they know not to make firm conclusions about
H-bonds
to flaming red side chains.


But this knowledge may be quite wrong.  If the flaming red really
indicates
large vibrational motion then yes, one whould not bet on stable H-bonds.
But if the flaming red indicates that a well-ordered sidechain was
incorrectly
modeled at full occupancy when in fact it is only present at
half-occupancy
then no, the H-bond could be strong but only present in that
half-occupancy
conformation.  One presumes that the other half-occupancy location
(perhaps
missing from the model) would have its own H-bonding network.



I beg to differ.  If a side chain has 2 or more positions, one should be a
bit careful about making firm conclusions based on only one of those, even
if it isn't clear exactly why one should use caution.  Also, isn't the
isotropic B we fit at medium resolution more of a spherical cow
approximation to physical reality anyway?

   Phoebe






Zbyszek Otwinowski
UT Southwestern Medical Center at Dallas
5323 Harry Hines Blvd.
Dallas, TX 75390-8816
Tel. 214-645-6385
Fax. 214-645-6353


Re: [ccp4bb] problem of conventions

2011-03-31 Thread Ian Tickle
 I would like to share my experiencde with a rather unexpected problem of 
 indexing conventions. Perhaps I can save people some
 time

 I have a crystal in the more unusual P21212 space-group (No 18). Its unit 
 cell lengths are bac (please note). I systematically
 use XDS for data integration, since so far it was able to handle even the 
 most horrible-looking spots.

 Now XDS indexed my data in space-group 18, but with the axes order abc! It 
 had, in fact, invented a space-group P22121,
 which does not exist. I did not realise this until I had spent a couple of 
 weeks with beautiful peaks in rotation functions, but
 hopeless results in translation functions. It wasn't until I looked more 
 closely into the definition of the screw axes that I realised the
 problem.

 POINTLESS does not allow a reindexing of reflexions within the same 
 space-group, but fortunately REINDEX did the trick at the
 level of intensities, because I like to use SCALA for careful scaling of my 
 data.

I was wo,dering if XDS could perhaps reindex reflexions according
 to Int. Table conventions once the screw axes of a crystal system have been
 identified?

The International Tables / IUCr / NIST convention _is_  a=b=c for
orthorhombic so no re-indexing is necessary or desirable.  See IT vol.
A 5th ed. (2002), table 9.3.4.1 (p. 758 in my edition) for all the
conventional cells.  The problem may be that some programs are not
sticking to the agreed convention - but then the obvious solution is
to fix the program (or use a different one).  Is the problem that XDS
is indexing it correctly as P22121 but calling it SG #18 (i.e. instead
of the correct #3018).  That would certainly confuse all CCP4 programs
which generally tend to use the space-group number first if it's
available.

I'm not clear what you mean when you say P22121 doesn't exist?  It's
clearly shown in my edition of IT (p. 202).  Maybe your lab needs to
invest in the most recent edition of IT?

Cheers

-- Ian


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Dale Tronrud

On 3/31/2011 10:14 AM, Jacob Keller wrote:
 What do we gain? As Dale pointed out, we are already abusing either occupancy, B-factor or delete the side chain to compensate 
for our inability to tell the user that the side chain is disordered. With your proposal, we would fudge both occupancy and 
B-factor, which in my eyes is even worse as fudging just one of the two.



 We gain clarity to the non-crystallographer user: a b-factor of 278.9
 sounds like possibly something real. A b-factor of exactly 1000 does
 not. Both probably have the same believability, viz., ~zero. Also,
 setting occupancy = zero is not fudging but rather respectfully
 declining to comment based on lack of data. I think it is exactly the
 same as omitting residues one cannot see in the density.


   These things are never clear unless there is a solid definition of
the terms you are using.  I don't think you can come up with an out of
band value for the B factor that doesn't have a legitimate meaning as
an atomic displacement parameter for someone.  How large a B factor you
can meaningfully define depends on your lower resolution limit.  People
working with electron microscopy or small angle X-ray scattering could
easily build models with ADPs far larger than anything we normally
encounter.

   In addition, you can't define 1000 as a magic value since the PDB
format will only allow values up to 999.99, and I presume maintaining
the PDB format is one of your goals.  Of course, you could choose -99.99
as the magic value but that would break all of our existing software
and I presume you don't want that either.  Actually defining any value
for the B factor as the magic value would break all of our software.
The only advantage of a large, positive, number is that it would create
bugs that are more subtle.

   The fundamental problem with your solution is that you are trying to
cram two pieces of information into a single number.  Such density always
causes problems.  Each concept needs its own value.

   You could implement your solution easily in mmCIF.  Just create a new
tag, say _atom_site.imaginary_site, which is either true or false
for every atom.  Then everyone would be able to either filter out the fake
atoms or leave them in, without ambiguity or confusion.

   If you object that the naive user of structural models wouldn't know
to check this tag - they aren't going to know about your magic B factor
either.  You can't out-think someone who's not paying attention.  At
some point you have to assume that people being paid to perform research
will learn the basics of the data they are using, even if you know that
assumption is not 100% true.

Dale Tronrud


Re: [ccp4bb] problem of conventions

2011-03-31 Thread Santarsiero, Bernard D.
Interesting. My IT, both volume I and volume A (1983) only have P21212 for
space group #18. Do I have to purchase a new volume A every year to keep
up with the new conventions?

Cheers,

Bernie


On Thu, March 31, 2011 12:57 pm, Ian Tickle wrote:
 I would like to share my experiencde with a rather unexpected problem of
 indexing conventions. Perhaps I can save people some
 time

 I have a crystal in the more unusual P21212 space-group (No 18). Its
 unit cell lengths are bac (please note). I systematically
 use XDS for data integration, since so far it was able to handle even
 the most horrible-looking spots.

 Now XDS indexed my data in space-group 18, but with the axes order
 abc! It had, in fact, invented a space-group P22121,
 which does not exist. I did not realise this until I had spent a couple
 of weeks with beautiful peaks in rotation functions, but
 hopeless results in translation functions. It wasn't until I looked more
 closely into the definition of the screw axes that I realised the
 problem.

 POINTLESS does not allow a reindexing of reflexions within the same
 space-group, but fortunately REINDEX did the trick at the
 level of intensities, because I like to use SCALA for careful scaling of
 my data.

I was wo,dering if XDS could perhaps reindex reflexions according
 to Int. Table conventions once the screw axes of a crystal system have
 been
 identified?

 The International Tables / IUCr / NIST convention _is_  a=b=c for
 orthorhombic so no re-indexing is necessary or desirable.  See IT vol.
 A 5th ed. (2002), table 9.3.4.1 (p. 758 in my edition) for all the
 conventional cells.  The problem may be that some programs are not
 sticking to the agreed convention - but then the obvious solution is
 to fix the program (or use a different one).  Is the problem that XDS
 is indexing it correctly as P22121 but calling it SG #18 (i.e. instead
 of the correct #3018).  That would certainly confuse all CCP4 programs
 which generally tend to use the space-group number first if it's
 available.

 I'm not clear what you mean when you say P22121 doesn't exist?  It's
 clearly shown in my edition of IT (p. 202).  Maybe your lab needs to
 invest in the most recent edition of IT?

 Cheers

 -- Ian



[ccp4bb] Question about GST cleavage

2011-03-31 Thread gauri misra
Just an offshoot of the same Question..
I would like to ask whether the same applies for GST-tag digestion using
thrombin..
No agitation gives better results in the above case too...
Any personal experiences

On Thu, Mar 31, 2011 at 11:29 AM, Klaus Piontek 
klaus.pion...@ocbc.uni-freiburg.de wrote:

 And not at full moon!

 Klaus


 Am 31.03.2011 um 16:23 schrieb Xiaopeng Hu:

 Our experience is do not shake the tube during TEV cleavage,I dont know
 why, but it does help.

 xiaopeng


 Dr. Klaus Piontek
 Albert-Ludwigs-University Freiburg
 Institute of Organic Chemistry and Biochemistry, Room 401 H Albertstrasse
 21
 D-79104 Freiburg Germany
 Phone: ++49-761-203-6036
 Fax: ++49-761-203-8714
 Email: klaus.pion...@ocbc.uni-freiburg.de
 Web: http://www.chemie.uni-freiburg.de/orgbio/w3platt/




Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Chun Luo
As many have mentioned, there is no need to titer baculoviruses for protein 
expression purpose. Plaque assay or other titering methods can give 10-fold 
variation between two operators.  Cells can be different from time to time as 
well. Sf9 cells in different labs are quite different. MOI reported in 
literature is pretty much meaningless. I think most authors and especially 
their labmates agree on this point. The important thing is how insect cells are 
infected. In this aspect, industry is at least 10 year in advance than 
academia. We have developed methods to quantify the infection and ensure high 
quality virus production and protein expression. Nat has described some steps 
how to do it.
When properly made and stored, recombinant baculoviruses are stable for at 
least a year without the need of adding FBS or other stabilizer. We have used 
several 2 year old viruses and got the exact same results as when they were 
freshly made. TIPS is one great way for long term virus storage, but reqires 
expensive equipment and know-how. Bacmids can be stored cheaply (or plasmids if 
not using Bac-to-Bac system). It takes only a week to generate enough high 
tiger viruses sufficient for 5-10L expression. For most academic research, 
there may be no need to do any viral amplification.
Cheers,
Chun
Accelagen

-Original Message-
From: Nathaniel Clark nathanielcl...@gmail.com
Sender:   CCP4 bulletin board CCP4BB@JISCMAIL.AC.UK
Date: Thu, 31 Mar 2011 13:20:14 
To: CCP4BB@JISCMAIL.AC.UK
Reply-To: Nathaniel Clark nathanielcl...@gmail.com
Subject: Re: [ccp4bb] titering baculovirus ?

We don't do anything special to store the virus, just a dark fridge.
You do need to mix well as the virus can settle.

You might want to try this 'TIPS' approach.  Basically you freeze
baculovirus infected cells, and use them as your virus innoculum.  I
think I tried it once and it didn't really work, but I didn't really
try too hard...
There is a research paper on the subject that I can't find, but here
is the first thing I found on google:
http://www.abrf.org/ResearchGroups/ProteinExpressionResearchGroup/Activities/Wasilko-PERG2007.pdf

Nat

On Thu, Mar 31, 2011 at 1:12 PM, chitra shintre
chitra.shin...@gmail.com wrote:
 Hi,
 We have observed this virus going off within 2 months as well. ...any idea
 why this happens?
 Chitra

 On 31 March 2011 18:07, Nathaniel Clark nathanielcl...@gmail.com wrote:

 We do adherent if we have a small volume of low titer virus, but as
 soon as we have a decent titer we will start with a 30 ml suspension
 flask.  Normally I add something like 0.5 ml, which is probably more
 then I need.  We only use serum free media right now.  As I mentioned,
 if you need to boost the titer of a low titer stock, we plate a t25
 flask with Sf9s, take off the media, and incubate it in 1 ml of pure
 virus (or two-fold diluted) for an hour or so, then remove (and save)
 that virus stock and give the cells fresh media.

 For fold dilution, just do a rough calculation of titer.  Once you are
 at a high titer assume you  have 10^8-10^9 pfu/ml, so you can estimate
 the ballpark volume to infect a given number of cells at an MOI of 1.
 From early stocks you might have a titer of 10^6-10^7 pfu/ml.  If you
 do an amplification that doesn't work, decide if you were to high
 (MOI=10 not great for amplifying) or too low (MOI=0.1) and try
 again.  Better yet, just set up 3 amplifications at the same time,
 whichever worked the best, use for another round of amplification, or
 for infections.

 For duration, we do 3 days amplification (post-infection), but
 sometimes we let them go a few days longer, which allows a secondary
 infection of the cells that weren't infected by the viral innoculum.
 There doesn't seem to be any harm to letting infections go longer; at
 some point all the cells die, but since the virus is pretty stable,
 it's no problem.  We have a CEDEX cell counter which tells you average
 cell diameter, which is a great read out for infection.  Also looking
 and density and viability versus an uninfected control will tell you
 if the infection worked.
 Nat

 Nat Clark
 Graduate Student
 Garman Lab
 Biochemistry and Molecular Biology Dept.
 UMass Amherst

 On Thu, Mar 31, 2011 at 12:36 PM, Katya Heldwein
 katya.heldw...@gmail.com wrote:
  A related question: how do most people amplify their baculovirus
  stocks? Adherent cultures vs suspension? Fold dilution at each stage
  (P1 to P2, P2 to P3)? Duration of each amplification stage?
 
  We have some viral stocks that go off rather quickly (1-2 months)
  despite being stored with FBS in a cool, dark place.
 
 
  Katya
 
 
  On Thu, Mar 31, 2011 at 10:23 AM, Brad Bennett bradbennet...@gmail.com
  wrote:
  Though I'm not meaning to turn this into a plaque assay burn session,
  for
  many reasons we have also abandoned it and instead titer our
  baculovirus
  stocks (P3 and P4) using flow cytometry. We use an antibody against
  viral
  gp64 that 

Re: [ccp4bb] Question about TEV cleavage

2011-03-31 Thread Chun Luo
Properly made TEV protease should work with or without rocking. We have done QC 
of our TurboTEV, a dual GST- and His-tagged TEV, under many conditions and used 
on hundreds of proteins. The most common cause of incomplete digestion is the 
insolubility of target protein or bad construct design. Shaking or agitation 
may oxidize the protein. TEV is a Cys protease that can be oxidized as well. 
Adding reducing agents or EDTA may well prevent some of the effects of 
agitation. 
Unlike plasmid DNA, protein solutions should be mixed GENTLY.
Regarding the original question, 50% loss may not be a bad thing if it is about 
total protein. In our experience, it is not uncommn to have only 50% recovery 
after tag removal of a Ni pool. Some proteins aggregate after tag removal even 
it's His-tag. Again the loss can be a good thing.
Chun
Accelagen 
-Original Message-
From: Laurie Betts laurie.betts0...@gmail.com
Sender:   CCP4 bulletin board CCP4BB@JISCMAIL.AC.UK
Date: Thu, 31 Mar 2011 11:11:33 
To: CCP4BB@JISCMAIL.AC.UK
Reply-To: Laurie Betts laurie.betts0...@gmail.com
Subject: Re: [ccp4bb] Question about TEV cleavage

We have the same experience with the DO NOT AGITATE.  We purify our own
His-tagged TEV protease, and flash-freeze it IMMEDIATELY after IMAC prep at
about 1 mg/mL.

Laurie Betts
UNC



On Thu, Mar 31, 2011 at 10:37 AM, Charles Allerston 
charles.allers...@sgc.ox.ac.uk wrote:

 I echo this.

 Some time ago I was working on a target which seemed to precipitate when
 cleaving overnight with TEV. I wasted a fair bit of time trying to optimise
 cleavage conditions with a myriad of buffers.  In the end, just by not
 agitating my solution, there was no precipitation and recovery was extremely
 good.

 I never agitate my solutions when cleaving now and have not had any
 problems with cleavage (to do with this issue, anyway) or recovery.


 cheers
 charlie



 Dr. Charles Allerston
 Genome Integrity Group
 Structural Genomics Consortium
 Nuffield Department of Medicine
 Old Road Campus
 University of Oxford
 OX3 7DQ
 http://www.sgc.ox.ac.uk/





 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 Xiaopeng Hu
 Sent: 31 March 2011 15:23
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] Question about TEV cleavage

 Our experience is do not shake the tube during TEV cleavage,I dont know
 why, but it does help.

 xiaopeng




Re: [ccp4bb] titering baculovirus ?

2011-03-31 Thread Nathaniel Clark
On another note, we have moved from baculovirus to stable poly or
monoclonal Tn5 cell lines using the pIB/V5-HIS vector.  It removes all
concerns about titering and viral propagation, and for most of our
proteins, the expression level is the same or better.  We secrete all
our proteins, so the stable cell line is great, you just grow several
liters of the cells to a high density, harvest the supernatant, and
purify.  It's much simpler.
Nat

On Thu, Mar 31, 2011 at 3:52 PM, Chun Luo c...@accelagen.com wrote:
 As many have mentioned, there is no need to titer baculoviruses for protein 
 expression purpose. Plaque assay or other titering methods can give 10-fold 
 variation between two operators.  Cells can be different from time to time as 
 well. Sf9 cells in different labs are quite different. MOI reported in 
 literature is pretty much meaningless. I think most authors and especially 
 their labmates agree on this point. The important thing is how insect cells 
 are infected. In this aspect, industry is at least 10 year in advance than 
 academia. We have developed methods to quantify the infection and ensure high 
 quality virus production and protein expression. Nat has described some steps 
 how to do it.
 When properly made and stored, recombinant baculoviruses are stable for at 
 least a year without the need of adding FBS or other stabilizer. We have used 
 several 2 year old viruses and got the exact same results as when they were 
 freshly made. TIPS is one great way for long term virus storage, but reqires 
 expensive equipment and know-how. Bacmids can be stored cheaply (or plasmids 
 if not using Bac-to-Bac system). It takes only a week to generate enough 
 high tiger viruses sufficient for 5-10L expression. For most academic 
 research, there may be no need to do any viral amplification.
 Cheers,
 Chun
 Accelagen

 -Original Message-
 From:         Nathaniel Clark nathanielcl...@gmail.com
 Sender:       CCP4 bulletin board CCP4BB@JISCMAIL.AC.UK
 Date:         Thu, 31 Mar 2011 13:20:14
 To: CCP4BB@JISCMAIL.AC.UK
 Reply-To:     Nathaniel Clark nathanielcl...@gmail.com
 Subject: Re: [ccp4bb] titering baculovirus ?

 We don't do anything special to store the virus, just a dark fridge.
 You do need to mix well as the virus can settle.

 You might want to try this 'TIPS' approach.  Basically you freeze
 baculovirus infected cells, and use them as your virus innoculum.  I
 think I tried it once and it didn't really work, but I didn't really
 try too hard...
 There is a research paper on the subject that I can't find, but here
 is the first thing I found on google:
 http://www.abrf.org/ResearchGroups/ProteinExpressionResearchGroup/Activities/Wasilko-PERG2007.pdf

 Nat

 On Thu, Mar 31, 2011 at 1:12 PM, chitra shintre
 chitra.shin...@gmail.com wrote:
 Hi,
 We have observed this virus going off within 2 months as well. ...any idea
 why this happens?
 Chitra

 On 31 March 2011 18:07, Nathaniel Clark nathanielcl...@gmail.com wrote:

 We do adherent if we have a small volume of low titer virus, but as
 soon as we have a decent titer we will start with a 30 ml suspension
 flask.  Normally I add something like 0.5 ml, which is probably more
 then I need.  We only use serum free media right now.  As I mentioned,
 if you need to boost the titer of a low titer stock, we plate a t25
 flask with Sf9s, take off the media, and incubate it in 1 ml of pure
 virus (or two-fold diluted) for an hour or so, then remove (and save)
 that virus stock and give the cells fresh media.

 For fold dilution, just do a rough calculation of titer.  Once you are
 at a high titer assume you  have 10^8-10^9 pfu/ml, so you can estimate
 the ballpark volume to infect a given number of cells at an MOI of 1.
 From early stocks you might have a titer of 10^6-10^7 pfu/ml.  If you
 do an amplification that doesn't work, decide if you were to high
 (MOI=10 not great for amplifying) or too low (MOI=0.1) and try
 again.  Better yet, just set up 3 amplifications at the same time,
 whichever worked the best, use for another round of amplification, or
 for infections.

 For duration, we do 3 days amplification (post-infection), but
 sometimes we let them go a few days longer, which allows a secondary
 infection of the cells that weren't infected by the viral innoculum.
 There doesn't seem to be any harm to letting infections go longer; at
 some point all the cells die, but since the virus is pretty stable,
 it's no problem.  We have a CEDEX cell counter which tells you average
 cell diameter, which is a great read out for infection.  Also looking
 and density and viability versus an uninfected control will tell you
 if the infection worked.
 Nat

 Nat Clark
 Graduate Student
 Garman Lab
 Biochemistry and Molecular Biology Dept.
 UMass Amherst

 On Thu, Mar 31, 2011 at 12:36 PM, Katya Heldwein
 katya.heldw...@gmail.com wrote:
  A related question: how do most people amplify their baculovirus
  stocks? Adherent cultures vs suspension? Fold 

Re: [ccp4bb] problem of conventions

2011-03-31 Thread Ian Tickle
There are no 'new' conventions to keep up with: recent editions of the
old volume 1 or new A do not disagree on the question of the unit cell
conventions (except for minor details which don't affect the majority
of the common space groups), where by recent I mean going back ~ 70
years.  So it's certainly not the case that the conventions are
changing every year (that would be silly!) - they have been defined
exactly once in the last 100 years!  I believe the unit cell
conventions currently in use were actually first defined by the 1952
edition of International Tables, so both the 1969 edition (volume '1')
and the 1983 edition (1st of volume 'A') will certainly describe them.
 I have only the 2002 edition (the 5th) so I can't tell you exactly
where to find the relevant info in the older editions.  The very first
edition of IT (1935 I believe) did not define the unit cell
conventions, only the space groups, so I wouldn't recommend that!

The older editions did not include information on alternate
space-group settings simply in order to save paper: the 1952 edition
was published in the years following WW2 when there was a paper
shortage, so this was an important consideration!  Only one setting
(the 'standard setting') of each space group, chosen arbitrarily, was
described and the crystallographer was expected to permute it to get
the desired setting.  If you need to see all the alternate settings
laid out explicitly then you need to get hold of a recent (e.g. the
5th printed or 1st online) edition; failing that you have to work them
out yourself!  I thought the alternate settings were first described
(though possibly without the diagrams) in the 1st (1983) edition of
volume A, but I'm relying on memory and could well be wrong.  The
setting was often chosen to be consistent with a pre-existing
isomorphous structure (i.e. generally isomorphism overrides
convention); if there was none either the setting was defined by the
unit cell convention, or often it was simply easiest to use the
standard setting.  Of course not everyone followed the conventions: it
was common to write programs that could handle only the standard
settings (and it still is!).  Wiser programmers allowed space-groups
to be defined arbitrarily by the equivalent positions instead of the
number or symbol, so then it was straightforward to select any desired
alternate setting.

Note that the convention describes the unit cells, from which the
space-group symbols are then derived, not the other way around.  The
ratiionale behind this is simple: there was a time not so long ago
(which I remember!) when data collection and structure solution for
even routine structures was actually non-trivial (I'm not implying
that it's always trivial even nowadays!).  However it was possible
relatively straightforwardly to obtain the unit cell from precession
photos (i.e. the indexing).  It used to be common practice to publish
an initial communication giving the unit cell and possibly a tentative
space group; this would be followed up (often several years later!) by
structures determined to successively higher resolution as more data
was collected.  Of course it was not possible to be 100% certain of
the space-group assignment from the precession photos (and for several
space-groups there is of course no unique space-group determinable
from the systematic absences alone); final space-group assignment
often had to wait several years for the structure determination.
Hence it made sense to define the setting from the unit cell, not the
space group.

I recommend the 2 papers from the US National Institute of Standards 
Technology (see Phil's posting) for more on this: the NIST conventions
are the same as the IUCr ones, i.e. based on the unit cell (in fact
Alan Mighell when he was at NIST wrote much of unit-cell convention
material in IT).

-- Ian

On Thu, Mar 31, 2011 at 7:36 PM, Santarsiero, Bernard D. b...@uic.edu wrote:
 Interesting. My IT, both volume I and volume A (1983) only have P21212 for
 space group #18. Do I have to purchase a new volume A every year to keep
 up with the new conventions?

 Cheers,

 Bernie


 On Thu, March 31, 2011 12:57 pm, Ian Tickle wrote:
 I would like to share my experiencde with a rather unexpected problem of
 indexing conventions. Perhaps I can save people some
 time

 I have a crystal in the more unusual P21212 space-group (No 18). Its
 unit cell lengths are bac (please note). I systematically
 use XDS for data integration, since so far it was able to handle even
 the most horrible-looking spots.

 Now XDS indexed my data in space-group 18, but with the axes order
 abc! It had, in fact, invented a space-group P22121,
 which does not exist. I did not realise this until I had spent a couple
 of weeks with beautiful peaks in rotation functions, but
 hopeless results in translation functions. It wasn't until I looked more
 closely into the definition of the screw axes that I realised the
 problem.

 POINTLESS does not allow a 

Re: [ccp4bb] Question about TEV cleavage

2011-03-31 Thread Zhijie Li
Totally agree with Chun. 

We are using a His tagged S219V construct that's very similary to the one 
described in the Waugh paper. To my experience, agitation, 37C incubation, low 
salt buffer, (etc.?), should all be avoided when using TEV. When using 
relatively large amount of enzyme(0.1~0.2mg/mL final,  in total about 1/50 
-1/10 of the total substrate by mol), we can often cleave our ProteinA fusions 
to nearly 90% complete on IgG beads. We do it at 4C overnight, elute, check the 
cleavage by SDS gel, if not complete, we add enzyme, cut O/N again, elute, 
check on SDS again. Then that's the 90% cleavage I am refering to. I also did 
O/N cleavage on column at 22C, and the cleave was quite good. In solution, my 
experience was positive too. In some cases even added at less than 1:200 molar 
ratio, the enzyme can still achieve a nearly complete cleavage at 4C in one or 
two days. 

One thing that could greatly affect the cleavage is the accessibility of the 
cleavage site. Looking from the crystal structure of TEV protease, the Q in the 
ENLYFQ-G/S has better be at least 2-3 aa away from the folded protein domain 
coming up. I had one construct made by LIC, which had only one G outside the 
folded domain (revealed by later crystal structure), then it could not be 
cleaved at all. When two more amino acids were added, it cleaved fine.

Zhijie


From: Chun Luo 
Sent: Thursday, March 31, 2011 4:42 PM
To: CCP4BB@JISCMAIL.AC.UK 
Subject: Re: [ccp4bb] Question about TEV cleavage


Properly made TEV protease should work with or without rocking. We have done QC 
of our TurboTEV, a dual GST- and His-tagged TEV, under many conditions and used 
on hundreds of proteins. The most common cause of incomplete digestion is the 
insolubility of target protein or bad construct design. Shaking or agitation 
may oxidize the protein. TEV is a Cys protease that can be oxidized as well. 
Adding reducing agents or EDTA may well prevent some of the effects of 
agitation. 
Unlike plasmid DNA, protein solutions should be mixed GENTLY.
Regarding the original question, 50% loss may not be a bad thing if it is about 
total protein. In our experience, it is not uncommn to have only 50% recovery 
after tag removal of a Ni pool. Some proteins aggregate after tag removal even 
it's His-tag. Again the loss can be a good thing.
Chun
Accelagen 


From: Laurie Betts laurie.betts0...@gmail.com 
Sender: CCP4 bulletin board CCP4BB@JISCMAIL.AC.UK 
Date: Thu, 31 Mar 2011 11:11:33 -0400
To: CCP4BB@JISCMAIL.AC.UK
ReplyTo: Laurie Betts laurie.betts0...@gmail.com 
Subject: Re: [ccp4bb] Question about TEV cleavage


We have the same experience with the DO NOT AGITATE.  We purify our own 
His-tagged TEV protease, and flash-freeze it IMMEDIATELY after IMAC prep at 
about 1 mg/mL.  

Laurie Betts
UNC 




On Thu, Mar 31, 2011 at 10:37 AM, Charles Allerston 
charles.allers...@sgc.ox.ac.uk wrote:

  I echo this.

  Some time ago I was working on a target which seemed to precipitate when 
cleaving overnight with TEV. I wasted a fair bit of time trying to optimise 
cleavage conditions with a myriad of buffers.  In the end, just by not 
agitating my solution, there was no precipitation and recovery was extremely 
good.

  I never agitate my solutions when cleaving now and have not had any problems 
with cleavage (to do with this issue, anyway) or recovery.


  cheers
  charlie



  Dr. Charles Allerston
  Genome Integrity Group
  Structural Genomics Consortium
  Nuffield Department of Medicine
  Old Road Campus
  University of Oxford
  OX3 7DQ
  http://www.sgc.ox.ac.uk/






  -Original Message-
  From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
Xiaopeng Hu
  Sent: 31 March 2011 15:23
  To: CCP4BB@JISCMAIL.AC.UK
  Subject: Re: [ccp4bb] Question about TEV cleavage

  Our experience is do not shake the tube during TEV cleavage,I dont know why, 
but it does help.

  xiaopeng



Re: [ccp4bb] problem of conventions

2011-03-31 Thread James Holton
I have the 2002 edition, and indeed it only contains space group
numbers up to 230.  The page numbers quoted by Ian contain space group
numbers 17 and 18.

Although I am all for program authors building in support for the
screwy orthorhombics (as I call them), I should admit that my
fuddy-duddy strategy for dealing with them remains simply to use space
groups 17 and 18, and permute the cell edges around with REINDEX to
put the unique (screw or non-screw) axis on the c position.  I have
yet to encounter a program that gets broken when presented with data
that doesn't have abc, but there are many non-CCP4 programs out
there that still don't seem to understand P22121, P21221, P2122 and
P2212.

This is not the only space group convention issue out there!  The
R3x vs H3x business continues to be annoying to this day!

-James Holton
MAD Scientist

On Thu, Mar 31, 2011 at 11:36 AM, Santarsiero, Bernard D. b...@uic.edu wrote:
 Interesting. My IT, both volume I and volume A (1983) only have P21212 for
 space group #18. Do I have to purchase a new volume A every year to keep
 up with the new conventions?

 Cheers,

 Bernie


 On Thu, March 31, 2011 12:57 pm, Ian Tickle wrote:
 I would like to share my experiencde with a rather unexpected problem of
 indexing conventions. Perhaps I can save people some
 time

 I have a crystal in the more unusual P21212 space-group (No 18). Its
 unit cell lengths are bac (please note). I systematically
 use XDS for data integration, since so far it was able to handle even
 the most horrible-looking spots.

 Now XDS indexed my data in space-group 18, but with the axes order
 abc! It had, in fact, invented a space-group P22121,
 which does not exist. I did not realise this until I had spent a couple
 of weeks with beautiful peaks in rotation functions, but
 hopeless results in translation functions. It wasn't until I looked more
 closely into the definition of the screw axes that I realised the
 problem.

 POINTLESS does not allow a reindexing of reflexions within the same
 space-group, but fortunately REINDEX did the trick at the
 level of intensities, because I like to use SCALA for careful scaling of
 my data.

I was wo,dering if XDS could perhaps reindex reflexions according
 to Int. Table conventions once the screw axes of a crystal system have
 been
 identified?

 The International Tables / IUCr / NIST convention _is_  a=b=c for
 orthorhombic so no re-indexing is necessary or desirable.  See IT vol.
 A 5th ed. (2002), table 9.3.4.1 (p. 758 in my edition) for all the
 conventional cells.  The problem may be that some programs are not
 sticking to the agreed convention - but then the obvious solution is
 to fix the program (or use a different one).  Is the problem that XDS
 is indexing it correctly as P22121 but calling it SG #18 (i.e. instead
 of the correct #3018).  That would certainly confuse all CCP4 programs
 which generally tend to use the space-group number first if it's
 available.

 I'm not clear what you mean when you say P22121 doesn't exist?  It's
 clearly shown in my edition of IT (p. 202).  Maybe your lab needs to
 invest in the most recent edition of IT?

 Cheers

 -- Ian




Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Dale Tronrud

On 3/31/2011 12:52 PM, Jacob Keller wrote:

The only advantage of a large, positive, number is that it would create
bugs that are more subtle.


Although most of the users on this BB probably know more about the
software coding, I am surprised that bugs--even subtle ones--would be
introduced by residues flagged with 0 occupancy and b-factor = 500.
Can you elaborate/enumerate?


   The principle problems with defining a particular value of the B factor
as magical have already been listed in this thread.  B factors are usually
restrained to the values of the atoms their atom is bonded to and
sometimes to other atoms they pack against.  You may set the B factor
equal to 500.00 but it will not stick.  At worst its presence will pull
up the B factors of nearby atoms that do have density.

   In addition, the only refinement program I know of that takes occupancy
into account when restraining bond lengths and angles is CNX.  The presence
of atoms with occ=0 will affect the location of atoms they share geometry
restraints with.

   Of course you could modify the refinement programs, and every other
program that reads crystallographic models, to deal with your redefinition
of _atom_site.B_iso_or_equiv.  In fact you would have to, just as you would
have to when you change the definition of any of the parameters in our
models.  If we have to modify code, why not create a solution that is
explicit, clear, and as consistent with previous practices as possible?


I think that the worst that could happen is that the unexperienced yet
b-factor-savvy user would be astonished by the huge b-factors, even if
he did not realize they were flags. At best, being surprised at the
precise number 500, he would look into the pdb file and see occupancy
= zero, google it, and learn something new about crystallography.


   How about positive difference map peaks on neighboring atoms?  How
about values for B factors that don't relate to the mean square motion
of the atom, despite that being the direct definition of the B factor?

   The concept of an unexperienced yet b-factor-savvy user is amusing.
I'm not b-factor-savvy.  Atomic displacement values are easy, but I'm
learning new subtleties about B factors all the time.




   The fundamental problem with your solution is that you are trying to
cram two pieces of information into a single number.  Such density always
causes problems.  Each concept needs its own value.


What two pieces of information into what single number? Occupancy = 0
tells you that the atom cannot be modelled, and B=500 is merely a flag
for same, and always goes with occ=0. What is so dense? On the
contrary, I think the info is redundant if anything...


   To be honest I had forgotten that you were proposing that the occupancy
be set to zero at the same time.  Besides putting two pieces of information
in the B factor column (The B factor's value and a flag for imaginaryness.)
You do the same for occupancy (the occupancy's value and a flag for
imaginaryness.)  This violates another rule of data structures - that
each concept be stored in one, and only one, place.  How do you interpret
an atom with an occupancy of zero but a B factor of 250?  How about an
atom with a B factor of 500.00 and an occupancy of 1.00?  Now we have the
confusing situation that the B factor can only be interpreted in the
context of the the value of the occupancy and vice versa.  Database-savvy
people (and I'm not one of them either) are not going to like this.

   If you want to calculate the average B factor for a model, certainly
those atoms with their B factor = 500 should not be included.  However,
I gather we do need to include those equal to 500 if their occupancy is
not equal to 0.0.  This is a mess.  In a database application we can't
simply SELECT the row with the B factors and average them.  We have to
SELECT both the B factor and occupancy rows and perform some really
weird if statements element by element - just to calculate an average!
What should be a simple task becomes very complex.  Will a graduate
student code the calculation correctly?  Probably not.  They will likely
not recall all the complicated interpretations of special values your
convention would require.

   Now consider this.  Refinement is running along and the occupancy for
an atom happens to overshoot and, in the middle of refinement, assumes
a value of 0.00.  There is positive difference density the next cycle.
(I did say that it overshot.)  Should the refinement program interpret
that Occ=0.00 to mean that the atom is imaginary and should not be
considered as part of the crystallographic model?  Wouldn't it be bad
if the atom suddenly disappeared because of a fluctuation?  Or should
the refinement program use one definition of occupancy during
refinement, but write a PDB file occupancy that has a different definition?
(It might be relevant to this line of thought to recall that the TNT
refinement package writes each intermediate coordinate file to disk 

Re: [ccp4bb] Question about GST cleavage

2011-03-31 Thread Nian Huang
I once cleaved a GST tag on the resin using TEV by rocking overnight at 4 C.
I would say it is 100% cutting judging from the gel. One thing to add is
that the protein bound so tight to the beads that cutting tag is the only
way to elute it except by SDS. I haven't had any trouble with TEV and
thrombin with rocking or shaking. The important part is that you have to
have good protease to start with.

Nian Huang, Ph.D.
UT Southwestern Medical Center

On Thu, Mar 31, 2011 at 2:41 PM, gauri misra kamga...@gmail.com wrote:

 Just an offshoot of the same Question..
 I would like to ask whether the same applies for GST-tag digestion using
 thrombin..
 No agitation gives better results in the above case too...
 Any personal experiences

 On Thu, Mar 31, 2011 at 11:29 AM, Klaus Piontek 
 klaus.pion...@ocbc.uni-freiburg.de wrote:

 And not at full moon!

 Klaus


 Am 31.03.2011 um 16:23 schrieb Xiaopeng Hu:

 Our experience is do not shake the tube during TEV cleavage,I dont know
 why, but it does help.

 xiaopeng


  Dr. Klaus Piontek
 Albert-Ludwigs-University Freiburg
 Institute of Organic Chemistry and Biochemistry, Room 401 H Albertstrasse
 21
 D-79104 Freiburg Germany
 Phone: ++49-761-203-6036
 Fax: ++49-761-203-8714
 Email: klaus.pion...@ocbc.uni-freiburg.de
 Web: http://www.chemie.uni-freiburg.de/orgbio/w3platt/





[ccp4bb] xds question: inverse beam, lots of wedges

2011-03-31 Thread Patrick Loll
We've just collected a number of inverse beam data sets. It turns out the 
crystals showed little radiation damage, so we have a lot of data: 2 x 360 deg 
for each crystal, broken up into 30 deg wedges. The collection order went like 
this: 0-30 deg, 180-210, 30-60, 210-240, etc.

Now, assuming no slippage, I could simply integrate the first set of data 
(non-inverse?) in one run: 0-360 deg. However, since the 12 individual wedges 
making up this 360 deg sweep were not collected  immediately one after the 
other, I don't expect the scale factors for individual images to vary smoothly 
(there should be discontinuities at the boundaries between wedges). If I do 
integrate the data in one fell swoop, am I in danger of introducing errors? For 
example, I seem to recall that denzo had built-in restraints to ensure that 
scale factors for adjacent images didn't vary by too much. Is there a similar 
restraint that in XDS that I might run afoul of?

The alternative is to integrate each each wedge separately, but with 24 wedges 
per xtal, this is starting to look a little tedious.

Cheers,
Pat


Re: [ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Zbyszek Otwinowski

Dale Tronrud wrote:

   While what you say here is quite true and is useful for us to
remember, your list is quite short.  I can add another

3) The systematic error introduced by assuming full occupancy for all 
sites.


You are right that structural heterogeneity is an additional factor.
Se-Met expression is one of the examples where the Se-Met residue is 
often not fully incorporated, and therefore its side chains have mixed 
with Met composition.

Obviously, solvent molecules may have partial occupancies.
Also, in heavily exposed crystals chemical reactions result in loss of 
the functional groups (e.g. by decarboxylation).
However, in most cases even if side chains have multiple conformations 
their total occupancy is 1.0.




There are, of course, many other factors that we don't account for
that our refinement programs tend to dump into the B factors.

   The definition of that number in the PDB file, as listed in the mmCIF
dictionary, only includes your first factor --

http://mmcif.rcsb.org/dictionaries/mmcif_std.dic/Items/_atom_site.B_iso_or_equiv.html 



and these numbers are routinely interpreted as though that definition is
the law.  Certainly the whole basis of TLS refinement is that the B factors
are really Atomic Displacement Parameters.   In addition the stereochemical
restraints on B factors are derived from the assumption that these 
parameters

are ADPs.  Convoluting all these other factors with the ADPs causes serious
problems for those who analyze B factors as measures of motion.

   The fact that current refinement programs mix all these factors with the
ADP for an atom to produce a vaguely defined B factor should be 
considered
a flaw to be corrected and not an opportunity to pile even more factors 
into

this field in the PDB file.



B-factors describe overall uncertainty of the current model. Refinement 
programs, which do not introduce or remove parts of the model (e.g. are 
not able to add additional conformations) intrinsically pile up all 
uncertainties into B-factors. Solutions, which you would like to see 
implemented, require a model-building like approach. The test of the 
success of such approach would be a substantial decrease of R-free 
values. If anybody can show it, it would be great.


Zbyszek


Dale Tronrud






On 3/31/2011 9:06 AM, Zbyszek Otwinowski wrote:

The B-factor in crystallography represents the convolution (sum) of two
types of uncertainties about the atom (electron cloud) position:

1) dispersion of atom positions in crystal lattice
2) uncertainty of the experimenter's knowledge  about the atom position.

In general, uncertainty needs not to be described by Gaussian function.
However, communicating uncertainty using the second moment of its
distribution is a widely accepted practice, with frequently implied
meaning that it corresponds to a Gaussian probability function. B-factor
is simply a scaled (by 8 times pi squared) second moment of uncertainty
distribution.

In the previous, long thread, confusion was generated by the additional
assumption that B-factor also corresponds to a Gaussian probability
distribution and not just to a second moment of any probability
distribution. Crystallographic literature often implies the Gaussian
shape, so there is some justification for such an interpretation, where
the more complex probability distribution is represented by the sum of
displaced Gaussians, where the area under each Gaussian component
corresponds to the occupancy of an alternative conformation.

For data with a typical resolution for macromolecular crystallography,
such multi-Gaussian description of the atom position's uncertainty is not
practical, as it would lead to instability in the refinement and/or
overfitting. Due to this, a simplified description of the atom's position
uncertainty by just the second moment of probability distribution is the
right approach. For this reason, the PDB format is highly suitable for 
the

description of positional uncertainties,  the only difference with other
fields being the unusual form of squaring and then scaling up the 
standard

uncertainty. As this calculation can be easily inverted, there is no loss
of information. However, in teaching one should probably stress more this
unusual form of presenting the standard deviation.

A separate issue is the use of restraints on B-factor values, a subject
that probably needs a longer discussion.

With respect to the previous thread, representing poorly-ordered (so
called 'disordered') side chains by the most likely conformer with
appropriately high B-factors is fully justifiable, and currently is
probably the best solution to a difficult problem.

Zbyszek Otwinowski




- they all know what B is and how to look for regions of high B
(with, say, pymol) and they know not to make firm conclusions about
H-bonds
to flaming red side chains.


But this knowledge may be quite wrong.  If the flaming red really
indicates
large vibrational motion then yes, one whould not bet 

Re: [ccp4bb] xds question: inverse beam, lots of wedges

2011-03-31 Thread Jan Abendroth
Pat,
at least give it a try with the one sweep approach.
We have collected plenty of 360deg data sets on a Rigaku system which requires 
two omega sweeps at phi 0 and 180 deg. These data sets are for in-house 
phasing. We haven't seen big issues with running XDS over these images as one 
continuous sweep. Monitoring scalefactors might be a good indicator.

Good luck
Jan

On Mar 31, 2011, at 3:08 PM, Patrick Loll wrote:

 We've just collected a number of inverse beam data sets. It turns out the 
 crystals showed little radiation damage, so we have a lot of data: 2 x 360 
 deg for each crystal, broken up into 30 deg wedges. The collection order went 
 like this: 0-30 deg, 180-210, 30-60, 210-240, etc.
 
 Now, assuming no slippage, I could simply integrate the first set of data 
 (non-inverse?) in one run: 0-360 deg. However, since the 12 individual wedges 
 making up this 360 deg sweep were not collected  immediately one after the 
 other, I don't expect the scale factors for individual images to vary 
 smoothly (there should be discontinuities at the boundaries between wedges). 
 If I do integrate the data in one fell swoop, am I in danger of introducing 
 errors? For example, I seem to recall that denzo had built-in restraints to 
 ensure that scale factors for adjacent images didn't vary by too much. Is 
 there a similar restraint that in XDS that I might run afoul of?
 
 The alternative is to integrate each each wedge separately, but with 24 
 wedges per xtal, this is starting to look a little tedious.
 
 Cheers,
 Pat

--
Jan Abendroth
Emerald BioStructures
Seattle / Bainbridge Island WA, USA
home: Jan.Abendroth_at_gmail.com
work: JAbendroth_at_embios.com
http://www.emeraldbiostructures.com


Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Sanishvili, Ruslan
Hi All,
Notwithstanding the stimulating discussion about the B-factor, I'd like
to chime in with my $0.02 on the original question of to build or not to
build and what are the rules and standards... and sorry for the lengthy
e-mail - I was trying to respond to several comments at once.

I thought there was a very well-defined rule: models based on
experimental data should represent the experimental data correctly. If a
model has parts that are not substantiated by experimental data and are
based only on assumptions, it's no longer an experimental model. Based
on this, one should leave out the atoms for which there is no observable
electron density. And one need not say that we were unable to build a
model of a missing side chains (or any other segments of the structure).
There is also no need to guess or fake most probable conformations of
the unobserved parts. Instead, it should be reported that the segment in
question was so flexible that it could not be described by just one or
two (and may be three) conformers. As such, this observation stands on
its own feet just like any other observation of visible segments and
there is no need to fake a model. If the work was done properly, a model
with missing parts is not intrinsically inferior to other, more complete
model. The fact that a side chain displays flexibility may be
biologically much more relevant than some well-defined Ile in the core
of the molecule. Omitting unobserved side chains from the model would
also help avoid assumptions as if we know for sure that the side chain
is there. Given side chain actually may not be there for some reason or
another. Sequence errors and radiation-induced damage come to mind, for
example. The latter is also often the reason that the side chain may not
be fully occupied in the structure derived from a specific data set
(i.e. the sum of occupancies of all its existing conformations may not
be 1, contrary to earlier suggestions in the thread). Back in the day I
personally spent large amounts of time and effort constructing and
refining multi-conformational models of some side chains because I was
sure they were there somewhere. Later on, as we learned more, I realized
that some of them have been sheered by radiation damage and actually
were not there. As knowledge advances, many of our assumptions may
crumble and that's why we ought to keep experimentally visible models
apart from those with assumed parts.

As for the downstream consumers of our models, we may not need to
confuse them with strange B factors or occupancies. We just need to give
them correct information. Namely, that the given part(s) of the molecule
could not be seen experimentally due to its flexibility (or, in some
cases, to radiation damage).  There was an interesting suggestion of two
models - one accurately describing the experimental observations and the
other for the downstream users. It would be a good way to separate Sci
from Fi but there may be a problem. When theories are derived further
downstream, it'll be impossible to keep track of what came from Sci and
what came from Fi versions.

Best regards,
N.


Ruslan Sanishvili (Nukri), Ph.D.

GM/CA-CAT
Biosciences Division, ANL
9700 S. Cass Ave.
Argonne, IL 60439

Tel: (630)252-0665
Fax: (630)252-0667
rsanishv...@anl.gov

-Original Message-
From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
Dale Tronrud
Sent: Thursday, March 31, 2011 4:51 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] what to do with disordered side chains

On 3/31/2011 12:52 PM, Jacob Keller wrote:
 The only advantage of a large, positive, number is that it would
create
 bugs that are more subtle.

 Although most of the users on this BB probably know more about the
 software coding, I am surprised that bugs--even subtle ones--would be
 introduced by residues flagged with 0 occupancy and b-factor = 500.
 Can you elaborate/enumerate?

The principle problems with defining a particular value of the B
factor
as magical have already been listed in this thread.  B factors are
usually
restrained to the values of the atoms their atom is bonded to and
sometimes to other atoms they pack against.  You may set the B factor
equal to 500.00 but it will not stick.  At worst its presence will pull
up the B factors of nearby atoms that do have density.

In addition, the only refinement program I know of that takes
occupancy
into account when restraining bond lengths and angles is CNX.  The
presence
of atoms with occ=0 will affect the location of atoms they share
geometry
restraints with.

Of course you could modify the refinement programs, and every other
program that reads crystallographic models, to deal with your
redefinition
of _atom_site.B_iso_or_equiv.  In fact you would have to, just as you
would
have to when you change the definition of any of the parameters in our
models.  If we have to modify code, why not create a solution that is
explicit, clear, and as consistent with previous practices as 

[ccp4bb] Jrh input Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Jrh
Dear Ed,
Thankyou for this and apologies for late reply.
If one has chemical evidence for the presence of residues but these residues 
are disordered I find the delete atoms option disagreeable. Such a static 
disorder situation should be described by a high atomic displacement parameter, 
in my view. (nb the use of ADP is better than B factor terminology). 
Yours sincerely,
John
Prof John R Helliwell DSc


On 29 Mar 2011, at 22:43, Ed Pozharski epozh...@umaryland.edu wrote:

 The results of the online survey on what to do with disordered side
 chains (from total of 240 responses):
 
 Delete the atoms 43%
 Let refinement take care of it by inflating B-factors41%
 Set occupancy to zero12%
 Other 4%
 
 Other suggestions were:
 
 - Place atoms in most likely spot based on rotomer and contacts and
 indicate high positional sigmas on ATMSIG records
 - To invent refinement that will spread this residues over many rotamers
 as this is what actually happened
 - Delet the atoms but retain the original amino acid name
 - choose the most common rotamer (B-factors don't inflate, they just
 rise slightly)
 - Depends. if the disordered region is unteresting, delete atoms.
 Otherwise, try to model it in one or more disordered model (and then
 state it clearly in the pdb file)
 - In case that no density is in the map, model several conformations of
 the missing segment and insert it into the PDB file with zero
 occupancies. It is equivalent what the NMR people do. 
 - Model it in and compare the MD simulations with SAXS
 - I would assumne Dale Tronrod suggestion the best. Sigatm labels.
 - Let the refinement inflate B-factors, then set occupancy to zero in
 the last round.
 
 Thanks to all for participation,
 
 Ed.
 
 -- 
 I'd jump in myself, if I weren't so good at whistling.
   Julian, King of Lemurs


Re: [ccp4bb] The meaning of B-factor, was Re: [ccp4bb] what to do with disordered side chains

2011-03-31 Thread Zbyszek Otwinowski
Regarding the closing statement about the best solution to poorly 
ordered side chains:


I described in the previous e-mail the probabilistic interpretation of 
B-factors. In the case of very high uncertainty = poorly ordered side 
chains, I prefer to deposit the conformer representing maximum a 
posteriori, even if it does not represent all possible conformations.
Maximum a posteriori will have significant contribution from the most 
probable conformation of side chain (prior knowledge) and should not 
conflict with likelihood (electron density map).
Thus, in practice I model the most probable conformation as long as it 
it in even very weak electron density, does not overlap significantly 
with negative difference electron density and do not clash with other 
residues.


As a user of PDB files I much prefer the simplest and the most 
informative representation of the result. Removing parts of side chains 
that carry charges, as already mentioned, is not particularly helpful 
for the downstream uses. NMR-like deposits are not among my favorites, 
either. Having multiple conformations with low occupancies increases 
potential for a confusion, while benefits are not clear to me.


Zbyszek

Frank von Delft wrote:
This is a lovely summary, and we should make our students read it. - But 
I'm afraid I do not see how it supports the closing statement in the 
last paragraph... phx.



On 31/03/2011 17:06, Zbyszek Otwinowski wrote:

The B-factor in crystallography represents the convolution (sum) of two
types of uncertainties about the atom (electron cloud) position:

1) dispersion of atom positions in crystal lattice
2) uncertainty of the experimenter's knowledge  about the atom position.

In general, uncertainty needs not to be described by Gaussian function.
However, communicating uncertainty using the second moment of its
distribution is a widely accepted practice, with frequently implied
meaning that it corresponds to a Gaussian probability function. B-factor
is simply a scaled (by 8 times pi squared) second moment of uncertainty
distribution.

In the previous, long thread, confusion was generated by the additional
assumption that B-factor also corresponds to a Gaussian probability
distribution and not just to a second moment of any probability
distribution. Crystallographic literature often implies the Gaussian
shape, so there is some justification for such an interpretation, where
the more complex probability distribution is represented by the sum of
displaced Gaussians, where the area under each Gaussian component
corresponds to the occupancy of an alternative conformation.

For data with a typical resolution for macromolecular crystallography,
such multi-Gaussian description of the atom position's uncertainty is not
practical, as it would lead to instability in the refinement and/or
overfitting. Due to this, a simplified description of the atom's position
uncertainty by just the second moment of probability distribution is the
right approach. For this reason, the PDB format is highly suitable for 
the

description of positional uncertainties,  the only difference with other
fields being the unusual form of squaring and then scaling up the 
standard

uncertainty. As this calculation can be easily inverted, there is no loss
of information. However, in teaching one should probably stress more this
unusual form of presenting the standard deviation.

A separate issue is the use of restraints on B-factor values, a subject
that probably needs a longer discussion.

With respect to the previous thread, representing poorly-ordered (so
called 'disordered') side chains by the most likely conformer with
appropriately high B-factors is fully justifiable, and currently is
probably the best solution to a difficult problem.

Zbyszek Otwinowski




- they all know what B is and how to look for regions of high B
(with, say, pymol) and they know not to make firm conclusions about
H-bonds
to flaming red side chains.

But this knowledge may be quite wrong.  If the flaming red really
indicates
large vibrational motion then yes, one whould not bet on stable 
H-bonds.

But if the flaming red indicates that a well-ordered sidechain was
incorrectly
modeled at full occupancy when in fact it is only present at
half-occupancy
then no, the H-bond could be strong but only present in that
half-occupancy
conformation.  One presumes that the other half-occupancy location
(perhaps
missing from the model) would have its own H-bonding network.

I beg to differ.  If a side chain has 2 or more positions, one should 
be a
bit careful about making firm conclusions based on only one of those, 
even

if it isn't clear exactly why one should use caution.  Also, isn't the
isotropic B we fit at medium resolution more of a spherical cow
approximation to physical reality anyway?

   Phoebe





Zbyszek Otwinowski
UT Southwestern Medical Center at Dallas
5323 Harry Hines Blvd.
Dallas, TX 75390-8816
Tel. 214-645-6385
Fax. 214-645-6353





--
Zbyszek 

Re: [ccp4bb] problem of conventions

2011-03-31 Thread Ian Tickle
On Thu, Mar 31, 2011 at 10:43 PM, James Holton jmhol...@lbl.gov wrote:
 I have the 2002 edition, and indeed it only contains space group
 numbers up to 230.  The page numbers quoted by Ian contain space group
 numbers 17 and 18.

You need to distinguish the 'IT space group number' which indeed goes
up to 230 (i.e. the number of unique settings), from the 'CCP4 space
group number' which, peculiar to CCP4 (which is why I called it
'CCP4-ese'), adds a multiple of 1000 to get a unique number for the
alternate settings as used in the API.  The page I mentioned show the
diagrams for IT SG #18 P22121 (CCP4 #3018), P21221 (CCP4 #2018) and
P21212 (CCP4 #18), so they certainly are all there!

 Although I am all for program authors building in support for the
 screwy orthorhombics (as I call them), I should admit that my
 fuddy-duddy strategy for dealing with them remains simply to use space
 groups 17 and 18, and permute the cell edges around with REINDEX to
 put the unique (screw or non-screw) axis on the c position.

Re-indexing is not an option for us (indeed if there were no
alternative, it would be a major undertaking), because the integrity
of our LIMS database requires that all protein-ligand structures from
the same target  crystal form are indexed with the same (or nearly
the same) cell and space group (and it makes life so much easier!).
With space-groups such as P22121 it can happen (indeed it has
happened) that it was not possible to define the space group correctly
at the processing stage due to ambiguous absences; indeed it was only
after using the SGALternative ALL option in Phaser and refining each
TF solution that we identified the space group correctly as P22121.

Having learnt the lesson the hard way, we routinely use P222 for all
processing of orthorhombics, which of course always gives the
conventional abc setting, and only assign the space group well down
the pipeline and only when we are 100% confident; by that time it's
too late to re-index (indeed why on earth would we want to give
ourselves all that trouble?).  This is therefore totally analogous to
the scenario of yesteryear that I described where it was common to see
a 'unit cell' communication followed some years later by the structure
paper (though we have compressed the gap somewhat!), and we base the
setting on the unit cell convention for exactly the same reason.

It's only if you're doing 1 structure at a time that you can afford
the luxury of re-indexing - and also the pain: many times I've seen
even experienced people getting their files mixed up and trying to
refine with differently indexed MTZ  PDB files (why is my R factor so
high?)!  My advice would be - _never_ re-index!

-- Ian


  I have
 yet to encounter a program that gets broken when presented with data
 that doesn't have abc, but there are many non-CCP4 programs out
 there that still don't seem to understand P22121, P21221, P2122 and
 P2212.

I find that surprising!  Exactly which 'many' programs are those?  You
really should report them to CCP4 (or to me if it's one of mine) so
they can be fixed!  We've been using CCP4 programs as integral
components of our processing pipeline (from data processing through to
validation) for the last 10 years and I've never come across one
that's broken in the way you describe (I've found many broken for
other reasons and either fixed it myself or reported it - you should
do the same!).  Any program which uses csymlib with syminfo.lib can
automatically handle all space groups defined in syminfo, which
includes all the common alternates you mentioned (and others such as
I2).  The only program I'm aware of that's limited to the standard
settings is sftools (because it has its own internal space group table
- it would be nice to see it updated to use syminfo!).

 This is not the only space group convention issue out there!  The
 R3x vs H3x business continues to be annoying to this day!

Yeah to that!  H centring was defined in IT long ago (look it up) and
it has nothing to do with the R setting!

-- Ian


Re: [ccp4bb] unknown electron density~

2011-03-31 Thread Artem Evdokimov
Try fitting a xylitol molecule in it but watch out for the distortions
caused by proximity to symmetry axis.

Artem

On Thu, Mar 31, 2011 at 1:16 PM, Shu XU xushuh...@gmail.com wrote:

 Hi, there.
 I'm refining a 1.76 A structure. The r-work stuck at 21%, rfree is 24%
 after adding waters.
 But there is something between the interfaces of the dimers.
 The space group is I4122, so it looks like something small, five or
 six atoms all together (without hydrogens).
 I tried to fit Tris in, however, it didn't look good at all.
 The crystal came out from Tris-cl buffer, ammonium sulfate, and there
 is some NaCl in the protein buffer too.
 The crystals were dropped into 20% xylitol and flash frozen.
 Fo-Fc maps are attached at 3 sigma level.
 Let me know what you're thoughts are.
 Thank you for your help.

 Shu


 --
 Shu Xu
 Ph.D candidate in Chemistry
 Department of Chemistry
 The University of Toledo
 Toledo, OH 43606
 Tel 419-530-1524



Re: [ccp4bb] step refine speed of wincoot - a little drawback of the new interruptible_fit_protein() function in wincoot

2011-03-31 Thread Zhijie Li

Hi Bernhard,

I realized this when comparing the functions yesterday. Exactly like what 
you said, when a user interrupts the process, what he/she wants at the 
moment is most likely to be going back a few residues instead of starting 
all over again. I am quite happy to have both interruptible and 
uninterruptible functions to choose from. Thanks for the effort.
But maybe it would make the impatient ones happier by presenting both 
versions in the Extension menu?
Besides speed, one other advantage I can see for the old function is that by 
having a simpler structure, it is way easier for the users to make their own 
modifications. May I beg you not to remove it from future versions?


Regards,
Zhijie


--
From: Bernhard C. Lohkamp bernh...@chem.gla.ac.uk
Sent: Thursday, March 31, 2011 1:11 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] step refine speed of wincoot - a little drawback of 
the new interruptible_fit_protein() function in wincoot



Thank you for the comment. You are correct that there are being backups
made in the newer interruptable fit function (they were't previously in
the uninterruptable fitting). I wasnt thinking about that (*). This may
(especially on windows) slow down things a great deal as we write and
gzip these files as well. To note here, gzip may be slow on windows as
is writing files. In the pre-release versions I disabled the gzip on
windows by default (and in general made the compression of backup files
optional). This may speed up things (in certain circumstances).
I hope this clarifies things a bit...

B

(*) It's debatable if we want backups here or not. If you can and want
to interrupt the fitting you may want to go back one step, which
requires a backup. On the other side this will write a lot of backup
files which is slowing thing down by writing and compressing files, as
well as producing a lot of backup files. Discuss... (or Paul and I just
decide).


Hi Xiaopeng, and those who are using the new interruptible
fit_protein... or stepped_refine... in wincoot 0.6.1:


I just took a look at wincoot 0.6.1's fitting.py file
(WinCoot\share\coot\python\fitting.py). It seems that in the old
fit_protein() and stepped_refine_protein() functions, the backup
was programmed to be turned off during the cycles when it steps
through the protein.

However when using the Stepped refine... from the Extensions menu,
the program starts to spit out tons of backup files. I then checked
the extension.py. It turns out that these menu items are calling the
new interruptible_fit_protein() function instead of the old
uninterruptible ones. Some further experiment showed that this
behavior was dependent on the backup state variable of the model
being refined. When the backup state is 1 (the default value?),
these new interruptible fitting functions will save a compressed
backup file on each step.

So, in order to save time when using the new interruptible protein
fitting functions, one can type the following command in the python
script window to tweak the backup state to 0:


make_backup(imol) #to save a backup
turn_off_backup(imol)

where imol is the number associated with the pdb file being fitted.

Do not forget to turn it back on after the fitting:


turn_on_backup(imol)


To check the current model's backup state:


backup_state(imol)



Zhijie


--
From: Xiaopeng Hu huxp...@mail.sysu.edu.cn
Sent: Tuesday, March 29, 2011 9:02 AM
To: Zhijie Li zhijie...@utoronto.ca
Subject: Re: [ccp4bb] step refine speed of wincoot


Dear Zhijie,

Thanks for there suggestions.

I tried to turn off backup (with script command) and it did help
although not much. Is there any way to add this perference to setting
of coot/wincoot?


best

xiaopeng

- 原始邮件 -
发件人: Zhijie Li zhijie...@utoronto.ca
收件人: Xiaopeng Hu huxp...@mail.sysu.edu.cn
发送时间: 星期二, 2011年 3 月 29日 下午 7:46:05
主题: Re: [ccp4bb] step refine speed of wincoot

Hi,

If you feel the refinement to be too slow, you may turn off the smooth
centering (in preferences) or change the centering steps to a smaller
number
to save unnecessary graphical calculations. To go extreme, you may even
remove the real time display commands in the scripts - also a way to
test if
the difference observed is due to different graphical efficiency.
Reducing
the size of the displayed map also helps.

The other thing you may need to consider is that coot/wincoot will
save a
backup file each time it updates the model, which means on each step
of the
refinement you have a new file generated (take a look at the coot-backup
directory when doing stepped refinement). If your windows disk is
terribly
fragmented then sure you will spend a lot of time on writing these
files.
The other thing is, windows has a different file caching mechanism than
linux, this can also cause a huge difference when a few hundred small
temporary files are queued for writing. My impression is that both the

Re: [ccp4bb] off topic: protease identification

2011-03-31 Thread Artem Evdokimov
This can be very hard to do because quite a few proteases are promiscuous
and will cut substrates solely based on masking of the polypeptide within
the structure of the protein. Typically these proteases will not stop
cutting at a single nick - they often proceed until they can't 'dig into' a
buried or obstructed section. Often one protease nicks the chain and other
proteases (or amino/carboxy peptidases) extend the gap. I would base the
search on the observed phenomenon - i.e. is it a single nick, a few residues
missing, or a whole swath or domain?

Theoretical aspects (i.e. searching for sequences) have been mentioned
already - practical ones of course include fractionating the environmental
factors (i.e. cell juice) where the cutting takes place and exposing the
uncut substrate protein (if you can get it!) to fractions, then
sub-fractionating; you can use class-specific inhibitors to further narrow
down the selection of enzymes; in some cases a trap (crosslinking, suicide
substrate, etc.) can be used to identify the perpetrator. In some cases it
may be possible to identify the enzyme by separating all possibilities on
e.g. a 2D gel then exposing a fluorogenic peptide to the gel and trying to
find a glowing spot (or spots). But again, this is not a very clean solution
due to promiscuous nature of many proteases and the sad fact that post-gel
renaturation is not guarranteed (not to mention that the protease in
question may be multisubunit, or activated by cofactors and gel will destroy
these interactions).

What do you actually see?

Artem


On Thu, Mar 31, 2011 at 8:59 AM, Brett, Thomas tbr...@dom.wustl.edu wrote:

 Hi all:
 I was wondering if anyone had any tips on identifying proteases. I have a
 protein for which I know the proteolytic cleavage site. What are the best
 ways to identify the protease that does the cutting either:
 1) bioinformatically (i.e., a good database to search using the cleavage
 site or a consensus)
 2) experimentally (some engineered substrate to trap/identify the substrate
 or any other method?)
 Thanks in advance
 -Tom

 Tom J. Brett, PhD
 Assistant Professor of Medicine
 Division of Pulmonary and Critical Care
 Washington University School of Medicine
 Campus Box 8052, 660 S. Euclid
 Saint Louis, MO 63110
 http://brettlab.dom.wustl.edu/


Re: [ccp4bb] problem of conventions

2011-03-31 Thread Santarsiero, Bernard D.
Ian,

I think it's amazing that we can program computers to resolve a  b  c
but it would be a major undertaking to store the matrix transformations
for 22121 to 21212 and reindex a cell to a standard setting. I was also
told that I was lazy to not reindex to the standard setting when I was a
grad student. Now it takes less than a minute to enter a transformation
and re-index.

The orthorhombic rule of a  b  c makes sense in 222 or 212121, but when
there is a standard setting of the 2-fold along the c-axis, then why not
adopt that? Often we chose a non-setting when there was a historical
precedence, as in the comparison of one structure to another, e.g., P21/c
with beta greater than 120deg vs. P21/n, etc. That is no more difficult
with modern computing than dragging along three space groups for #18.
There was a compactness to 230, and only 230 space groups. (I cheat, since
I agree there is both the rhombohedral and hexagonal cell settings for
R3bar.)

Bernie



On Thu, March 31, 2011 5:48 pm, Ian Tickle wrote:
 On Thu, Mar 31, 2011 at 10:43 PM, James Holton jmhol...@lbl.gov wrote:
 I have the 2002 edition, and indeed it only contains space group
 numbers up to 230.  The page numbers quoted by Ian contain space group
 numbers 17 and 18.

 You need to distinguish the 'IT space group number' which indeed goes
 up to 230 (i.e. the number of unique settings), from the 'CCP4 space
 group number' which, peculiar to CCP4 (which is why I called it
 'CCP4-ese'), adds a multiple of 1000 to get a unique number for the
 alternate settings as used in the API.  The page I mentioned show the
 diagrams for IT SG #18 P22121 (CCP4 #3018), P21221 (CCP4 #2018) and
 P21212 (CCP4 #18), so they certainly are all there!

 Although I am all for program authors building in support for the
 screwy orthorhombics (as I call them), I should admit that my
 fuddy-duddy strategy for dealing with them remains simply to use space
 groups 17 and 18, and permute the cell edges around with REINDEX to
 put the unique (screw or non-screw) axis on the c position.

 Re-indexing is not an option for us (indeed if there were no
 alternative, it would be a major undertaking), because the integrity
 of our LIMS database requires that all protein-ligand structures from
 the same target  crystal form are indexed with the same (or nearly
 the same) cell and space group (and it makes life so much easier!).
 With space-groups such as P22121 it can happen (indeed it has
 happened) that it was not possible to define the space group correctly
 at the processing stage due to ambiguous absences; indeed it was only
 after using the SGALternative ALL option in Phaser and refining each
 TF solution that we identified the space group correctly as P22121.

 Having learnt the lesson the hard way, we routinely use P222 for all
 processing of orthorhombics, which of course always gives the
 conventional abc setting, and only assign the space group well down
 the pipeline and only when we are 100% confident; by that time it's
 too late to re-index (indeed why on earth would we want to give
 ourselves all that trouble?).  This is therefore totally analogous to
 the scenario of yesteryear that I described where it was common to see
 a 'unit cell' communication followed some years later by the structure
 paper (though we have compressed the gap somewhat!), and we base the
 setting on the unit cell convention for exactly the same reason.

 It's only if you're doing 1 structure at a time that you can afford
 the luxury of re-indexing - and also the pain: many times I've seen
 even experienced people getting their files mixed up and trying to
 refine with differently indexed MTZ  PDB files (why is my R factor so
 high?)!  My advice would be - _never_ re-index!

 -- Ian


  I have
 yet to encounter a program that gets broken when presented with data
 that doesn't have abc, but there are many non-CCP4 programs out
 there that still don't seem to understand P22121, P21221, P2122 and
 P2212.

 I find that surprising!  Exactly which 'many' programs are those?  You
 really should report them to CCP4 (or to me if it's one of mine) so
 they can be fixed!  We've been using CCP4 programs as integral
 components of our processing pipeline (from data processing through to
 validation) for the last 10 years and I've never come across one
 that's broken in the way you describe (I've found many broken for
 other reasons and either fixed it myself or reported it - you should
 do the same!).  Any program which uses csymlib with syminfo.lib can
 automatically handle all space groups defined in syminfo, which
 includes all the common alternates you mentioned (and others such as
 I2).  The only program I'm aware of that's limited to the standard
 settings is sftools (because it has its own internal space group table
 - it would be nice to see it updated to use syminfo!).

 This is not the only space group convention issue out there!  The
 R3x vs H3x business continues to be annoying to this day!