[ccp4bb] mar345 version of generate_XDS.INP

2013-06-21 Thread Jan Gebauer

Dear all,

sorry for the slightly off-topic question:

Does anyone have a modified version of generate_XDS.INP for the mar345  
image plate detector?


I know I could -relative easily- make it on my own, but why inventing the  
wheel twice?


Regards,
Jan
--
Dr. Jan Gebauer
AG Prof. Baumann
Institut für Biochemie / Uni-Köln
Otto-Fischer-Str. 12-14 / 50674 Köln
Fon: +49 (221) 470 3212 
Fax: +49 (221) 470 5066

http://px.uni-koeln.de/


Re: [ccp4bb] mar345 version of generate_XDS.INP

2013-06-21 Thread Jan Gebauer

Dear all,

thanks to all for your first responses; however, I probably wasn't precise  
enough in my original message.


I DO have an XDS.INP for the mar345 image plate (and i also now where to  
find the others).


What I would like to have is a modified version of Kay Diederichs  
generate_XDS.INP script, to generate XDS.INPs automatically... It's very  
handy if you have datasets from multiple detectors.
(here the wiki:  
http://strucbio.biologie.uni-konstanz.de/xdswiki/index.php/Generate_XDS.INP)


Best,
Jan

Am 21.06.2013, 10:23 Uhr, schrieb Jan Gebauer  
ccp...@science.jan-gebauer.de:



Dear all,

sorry for the slightly off-topic question:

Does anyone have a modified version of generate_XDS.INP for the mar345  
image plate detector?


I know I could -relative easily- make it on my own, but why inventing  
the wheel twice?


Regards,
Jan


--
Dr. Jan Gebauer
AG Prof. Baumann
Institut für Biochemie / Uni-Köln
Otto-Fischer-Str. 12-14 / 50674 Köln
Fon: +49 (221) 470 3212 
Fax: +49 (221) 470 5066

http://px.uni-koeln.de/


Re: [ccp4bb] NAG-NAG

2013-06-21 Thread Eugene Osipov
Hello, Monika,
1) Add NAGs
2) Remove their hydrogens and O1 atom and put them in density
3) In menu: Calculate-merge molecules, choose your pdb and both NAG
residues
4) save coordinate file


2013/6/20 Monika Pathak m.pat...@nottingham.ac.uk

 Hi

 Please can I ask about how can I put NAG to asparagine in coot (I think
 its 2NAG that I can put in density) and if possible  to refine it in
 refmac then.

 Thanks in advance for help
 Regards
 Monika

 Monika Pathak
 University of Nottingham
 NG7 2RD


 This message and any attachment are intended solely for the addressee and
 may contain confidential information. If you have received this message in
 error, please send it back to me, and immediately delete it.   Please do
 not use, copy or disclose the information contained in this message or in
 any attachment.  Any views or opinions expressed by the author of this
 email do not necessarily reflect the views of the University of Nottingham.

 This message has been checked for viruses but the contents of an
 attachment may still contain software viruses which could damage your
 computer system, you are advised to perform your own checks. Email
 communications with the University of Nottingham may be monitored as
 permitted by UK legislation.




-- 
Eugene Osipov
Junior Research Scientist
Laboratory of Enzyme Engineering
A.N. Bach Institute of Biochemistry
Russian Academy of Sciences
Leninsky pr. 33, 119071 Moscow, Russia
e-mail: e.m.osi...@gmail.com


Re: [ccp4bb] PIMS XtalView Help

2013-06-21 Thread McCully, Dwayne (NIH/NIAMS) [C]
Hello Everyone,

I'm looking for a group/person that has PIMS and Xtailpims installed which does 
not throw Java errors while using the program.
I've been struggling to get the program to work without error trying different 
versions of the below programs.
I believe PIMS and Xtalpims has potential for my group so I looking for outside 
help.

I need to know the following, with no errors, to make sure my setup is correct:

Java Version
Tomcat Version
Windows or Linux version
Apache Version

Thanks in advance.
Dwayne

Dwayne McCully
Structural Biology Section, NIAID, NIH
Contractor


[ccp4bb] Postdoctoral position at Humboldt University Berlin and Helmholtz Zentrum Berlin (BESSY II)

2013-06-21 Thread Müller , Uwe
The Helmholtz Zentrum Berlin für Materialien und Energie, the 
Max-Delbrück-Center, the Freie Universität Berlin and the Humboldt Universität 
zu Berlin jointly operate three experimental stations for bio-macromolecular 
crystallography at BESSY II, one of the world’s most modern synchrotron 
radiation sources for VUV and soft X-rays.

The Institute of Biology of the Humboldt-Universität zu Berlin is seeking a

Post Doctoral Research Assistant

to conduct research on enzymes involved in the energy metabolism at the 
Laboratory for Structural Biology / Biochemistry (headed by Prof. Dr. Holger 
Dobbek) of the Humboldt-Universität zu Berlin. The successful candidate will 
also be involved in the operation of the BESSY II bio-macromolecular 
crystallography beamlines (headed by Dr. Uwe Müller  Dr. Manfred Weiss) and 
will take part in graduate student education.
Initially, a three-year contract will be offered to the successful candidate 
with the possibility for an extension. The salary will be based on German 
Federal TVöD (i.d.F.d. Anw.-TV HUB). The position is available immediately and 
applications will be considered until the position is filled.
Applicants should hold a Ph.D. in the biological, chemical or physical sciences 
and have a background in bio-macromolecular crystallography, biochemistry or 
computational chemistry. The position also requires the documented ability to 
conduct independent research as well as excellent communication and 
interpersonal skills.

Please send your application (CV, list of publications, 2 referees, .pdf-file 
format) in electronic form until July 12, 2013, with the Code DR/075/13  to:

Dr. Uwe Müller
Helmholtz-Zentrum Berlin / BESSY II
Albert-Einstein-Str. 15, 12489 Berlin
Dr. Uwe Mueller

Fon: +49 30 8062 14974
Fax: +49 30 8062 14975
url: www.helmholtz-berlin.de/bessy-mx
email:u...@helmholtz-berlin.de




Helmholtz-Zentrum Berlin für Materialien und Energie GmbH

Mitglied der Hermann von Helmholtz-Gemeinschaft Deutscher Forschungszentren e.V.

Aufsichtsrat: Vorsitzender Prof. Dr. Dr. h.c. mult. Joachim Treusch, stv. 
Vorsitzende Dr. Beatrix Vierkorn-Rudolph
Geschäftsführung: Prof. Dr. Anke Rita Kaysser-Pyzalla, Thomas Frederking

Sitz Berlin, AG Charlottenburg, 89 HRB 5583

Postadresse:
Hahn-Meitner-Platz 1
D-14109 Berlin

http://www.helmholtz-berlin.de


[ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Herman . Schreuder
Dear Bulletin Board,

After some headbanging (Refmac5 had helpfully created gap records for all 
insertions and deletions present in the structure), I got refmac5 running with 
the TWIN option. Refmac5 also found the k,h,-l domain and rejected the other 
possible domains because they were too small. The Rfactor's are now extremely 
good: ~14% and the Rfree's are for me acceptable: ~24%. Since I found the 
difference between R and Rfree somewhat large, I have been playing with the 
weighting. By using a weight of 0.01, I can bring the Rfactor up to 18%, but 
the Rfree stays about the same or even gets a little worse.

My question: is there a way to bring R and Rfree closer together, or is it 
related to the twinned data and is it something we have to live with?

Best regards,
Herman
 

-Ursprüngliche Nachricht-
Von: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von Miller, 
Mitchell D.
Gesendet: Donnerstag, 20. Juni 2013 17:43
An: CCP4BB@JISCMAIL.AC.UK
Betreff: Re: [ccp4bb] Twinning problem

You are welcome.  Let me also for the benefit of others who may search the 
archives in the future, let me correct two errors below - (typo and a 
miss-recollection).  

Specially, I was thinking that phenix.refine was now able to refine multiple 
twin laws, but according to Nat Echols on the phenix mailing list 
http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html
phenix.refine only handles 1 twin law at this time. 
(My typo was that and our second structure was 3nuz with twin fractions 0.38, 
0.32, 0.16 and 0.14 -- not 2nuz).

A useful search for deposited structures mentioning tetartohedral 
http://www.ebi.ac.uk/pdbe-srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETARTOHEDRAL
 

Regards,
Mitch


-Original Message-
From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
herman.schreu...@sanofi.com
Sent: Thursday, June 20, 2013 8:04 AM
To: CCP4BB@JISCMAIL.AC.UK
Subject: [ccp4bb] AW: Twinning problem

Dear Mitch (and Philip and Phil),

It is clear that I should give refmac a go with the non-detwinned F's and just 
the TWIN command.

Thank you for your suggestions,
Herman

 

-Ursprüngliche Nachricht-
Von: Miller, Mitchell D. [mailto:mmil...@slac.stanford.edu]
Gesendet: Donnerstag, 20. Juni 2013 16:18
An: Schreuder, Herman RD/DE
Betreff: RE: Twinning problem

Hi Herman,
 Have you considered the possibility of your crystals being tetartohedral 
twinned.  That is more than one of the twin laws may apply to your crystals.
E.g. in P32 it is possible to have tetartohedral twinning which would have
4 twin domains - (h,k,l), (k,h,-l), (-h,-k,l) and (-k,-h,-l). Perfect 
tetartohedral twinning of P3 would merge in P622 and each twin domain would 
have a faction of 0.25.

  We have had 2 cases like this (the first 2PRX was before there was support 
for this type of twinning except for in shelxl and we ended up with refined 
twin fractions of 0.38, 0.28, 0.19, 0.15 for the deposited crystal and a 2nd 
crystal that we did not deposit had twin fractions of 0.25, 0.27, 0.17, 0.31).  
The 2nd case we had was after support for twining (including tetartohedral 
twinning) was added to refmac (and I think phenix.refine can also handle this). 
 For 2NUZ, it was P32 with refined twin fractions of 0.25, 0.27, 0.17, 0.31.

  Pietro Roversi wrote a review of tetartohedral twinning for the CCP4 
proceedings issues of acta D http://dx.doi.org/10.1107/S0907444912006737 

  I would try refinement with refmac using the original (non-detwinned F's) 
with just the TWIN command to see if it ends up keeping twin fractions for all 
3 operators (4 domains) -- especially with crystals 1 and 3 which appear to 
have the largest estimates of the other twin fractions.

Regards,
Mitch


==
Mitchell Miller, Ph.D.
Joint Center for Structural Genomics
Stanford Synchrotron Radiation Lightsource
2575 Sand Hill Rd  -- SLAC MS 99
Menlo Park, CA  94025
Phone: 1-650-926-5036
FAX: 1-650-926-3292


-Original Message-
From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
herman.schreu...@sanofi.com
Sent: Thursday, June 20, 2013 6:47 AM
To: CCP4BB@JISCMAIL.AC.UK
Subject: [ccp4bb] Twinning problem

Dear Bulletin Board,
 
Prodded by pdb annotators, which are very hesitant to accept coordinate files 
when their Rfactor does not correspond with our Rfactor, I had a look again 
into some old data sets, which I suspect are twinned. Below are the results of 
some twinning tests with the Detwin program (top value: all reflections, lower 
value: reflections  Nsig*obs (whatever that may mean). The space group is P32, 
the resolution is 2.3 - 2.6 Å and data are reasonable complete: 95 - 100%.
 
From the Detwin analysis, it seems that the crystals are twinned with twin 
operator k,h,-l with a twinning fraction of 0.3 for crystal 1, 0.15 for 
crystal 2 and 0.4 for crystal 3. Crystal 2 can be refined while ignoring 
twinning to get 

Re: [ccp4bb] NAG-NAG

2013-06-21 Thread Paul Emsley

On 06/20/2013 06:56 PM, Monika Pathak wrote:

Hi

Please can I ask about how can I put NAG to asparagine in coot (I 
think its 2NAG that I can put in density) and if possible  to refine 
it in refmac then.





Calculate - Scripting - Scheme:

(add-linked-residue imol chain-id res-no ins-code NAG  NAG-ASN)

for some value of imol, chain-id, res-no and ins-code of course.

HTH,

Paul.


Re: [ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Roger Rowlett
I have found PDB_REDO is an efficient way of tweaking structure solutions.
Among other things, it sorts through various MATRIX and BFAC weights and
tests effectiveness of TLS if not used already. It usually improves typical
final structure solutions for our group by about 1% or so Rfree and
reduces the Rfree-R spread, as well as improves a variety of other
structural measures.

You can of course adjust weights and TLS manually, but if you can get a
local version of PDB_REDO running it saves a lot of time, and has some
extra features. I've modified v. 5.0.9 for this purpose. Depending on the
packages you do or don't have, it is a little tricky to get running locally.

Roger Rowlett
On Jun 21, 2013 4:42 AM, herman.schreu...@sanofi.com wrote:

 Dear Bulletin Board,

 After some headbanging (Refmac5 had helpfully created gap records for all
 insertions and deletions present in the structure), I got refmac5 running
 with the TWIN option. Refmac5 also found the k,h,-l domain and rejected the
 other possible domains because they were too small. The Rfactor's are now
 extremely good: ~14% and the Rfree's are for me acceptable: ~24%. Since I
 found the difference between R and Rfree somewhat large, I have been
 playing with the weighting. By using a weight of 0.01, I can bring the
 Rfactor up to 18%, but the Rfree stays about the same or even gets a little
 worse.

 My question: is there a way to bring R and Rfree closer together, or is it
 related to the twinned data and is it something we have to live with?

 Best regards,
 Herman


 -Ursprüngliche Nachricht-
 Von: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von
 Miller, Mitchell D.
 Gesendet: Donnerstag, 20. Juni 2013 17:43
 An: CCP4BB@JISCMAIL.AC.UK
 Betreff: Re: [ccp4bb] Twinning problem

 You are welcome.  Let me also for the benefit of others who may search the
 archives in the future, let me correct two errors below - (typo and a
 miss-recollection).

 Specially, I was thinking that phenix.refine was now able to refine
 multiple twin laws, but according to Nat Echols on the phenix mailing list
 http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html
 phenix.refine only handles 1 twin law at this time.
 (My typo was that and our second structure was 3nuz with twin fractions
 0.38, 0.32, 0.16 and 0.14 -- not 2nuz).

 A useful search for deposited structures mentioning tetartohedral
 http://www.ebi.ac.uk/pdbe-srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETARTOHEDRAL

 Regards,
 Mitch


 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 herman.schreu...@sanofi.com
 Sent: Thursday, June 20, 2013 8:04 AM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: [ccp4bb] AW: Twinning problem

 Dear Mitch (and Philip and Phil),

 It is clear that I should give refmac a go with the non-detwinned F's and
 just the TWIN command.

 Thank you for your suggestions,
 Herman



 -Ursprüngliche Nachricht-
 Von: Miller, Mitchell D. [mailto:mmil...@slac.stanford.edu]
 Gesendet: Donnerstag, 20. Juni 2013 16:18
 An: Schreuder, Herman RD/DE
 Betreff: RE: Twinning problem

 Hi Herman,
  Have you considered the possibility of your crystals being tetartohedral
 twinned.  That is more than one of the twin laws may apply to your crystals.
 E.g. in P32 it is possible to have tetartohedral twinning which would have
 4 twin domains - (h,k,l), (k,h,-l), (-h,-k,l) and (-k,-h,-l). Perfect
 tetartohedral twinning of P3 would merge in P622 and each twin domain would
 have a faction of 0.25.

   We have had 2 cases like this (the first 2PRX was before there was
 support for this type of twinning except for in shelxl and we ended up with
 refined twin fractions of 0.38, 0.28, 0.19, 0.15 for the deposited crystal
 and a 2nd crystal that we did not deposit had twin fractions of 0.25, 0.27,
 0.17, 0.31).  The 2nd case we had was after support for twining (including
 tetartohedral twinning) was added to refmac (and I think phenix.refine can
 also handle this).  For 2NUZ, it was P32 with refined twin fractions of
 0.25, 0.27, 0.17, 0.31.

   Pietro Roversi wrote a review of tetartohedral twinning for the CCP4
 proceedings issues of acta D http://dx.doi.org/10.1107/S0907444912006737

   I would try refinement with refmac using the original (non-detwinned
 F's) with just the TWIN command to see if it ends up keeping twin fractions
 for all 3 operators (4 domains) -- especially with crystals 1 and 3 which
 appear to have the largest estimates of the other twin fractions.

 Regards,
 Mitch


 ==
 Mitchell Miller, Ph.D.
 Joint Center for Structural Genomics
 Stanford Synchrotron Radiation Lightsource
 2575 Sand Hill Rd  -- SLAC MS 99
 Menlo Park, CA  94025
 Phone: 1-650-926-5036
 FAX: 1-650-926-3292


 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 herman.schreu...@sanofi.com
 Sent: Thursday, June 20, 2013 6:47 AM
 To: 

Re: [ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Eleanor Dodson
At your resolution that seems to me a reasonable gap between R and Rfree?
 Eleanor

On 21 Jun 2013, at 12:28, herman.schreu...@sanofi.com wrote:

 Dear Bulletin Board,
 
 After some headbanging (Refmac5 had helpfully created gap records for all 
 insertions and deletions present in the structure), I got refmac5 running 
 with the TWIN option. Refmac5 also found the k,h,-l domain and rejected the 
 other possible domains because they were too small. The Rfactor's are now 
 extremely good: ~14% and the Rfree's are for me acceptable: ~24%. Since I 
 found the difference between R and Rfree somewhat large, I have been playing 
 with the weighting. By using a weight of 0.01, I can bring the Rfactor up to 
 18%, but the Rfree stays about the same or even gets a little worse.
 
 My question: is there a way to bring R and Rfree closer together, or is it 
 related to the twinned data and is it something we have to live with?
 
 Best regards,
 Herman
 
 
 -Ursprüngliche Nachricht-
 Von: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von 
 Miller, Mitchell D.
 Gesendet: Donnerstag, 20. Juni 2013 17:43
 An: CCP4BB@JISCMAIL.AC.UK
 Betreff: Re: [ccp4bb] Twinning problem
 
 You are welcome.  Let me also for the benefit of others who may search the 
 archives in the future, let me correct two errors below - (typo and a 
 miss-recollection).  
 
 Specially, I was thinking that phenix.refine was now able to refine multiple 
 twin laws, but according to Nat Echols on the phenix mailing list 
 http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html
 phenix.refine only handles 1 twin law at this time. 
 (My typo was that and our second structure was 3nuz with twin fractions 0.38, 
 0.32, 0.16 and 0.14 -- not 2nuz).
 
 A useful search for deposited structures mentioning tetartohedral 
 http://www.ebi.ac.uk/pdbe-srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETARTOHEDRAL
  
 
 Regards,
 Mitch
   
 
 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
 herman.schreu...@sanofi.com
 Sent: Thursday, June 20, 2013 8:04 AM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: [ccp4bb] AW: Twinning problem
 
 Dear Mitch (and Philip and Phil),
 
 It is clear that I should give refmac a go with the non-detwinned F's and 
 just the TWIN command.
 
 Thank you for your suggestions,
 Herman
 
 
 
 -Ursprüngliche Nachricht-
 Von: Miller, Mitchell D. [mailto:mmil...@slac.stanford.edu]
 Gesendet: Donnerstag, 20. Juni 2013 16:18
 An: Schreuder, Herman RD/DE
 Betreff: RE: Twinning problem
 
 Hi Herman,
 Have you considered the possibility of your crystals being tetartohedral 
 twinned.  That is more than one of the twin laws may apply to your crystals.
 E.g. in P32 it is possible to have tetartohedral twinning which would have
 4 twin domains - (h,k,l), (k,h,-l), (-h,-k,l) and (-k,-h,-l). Perfect 
 tetartohedral twinning of P3 would merge in P622 and each twin domain would 
 have a faction of 0.25.
 
  We have had 2 cases like this (the first 2PRX was before there was support 
 for this type of twinning except for in shelxl and we ended up with refined 
 twin fractions of 0.38, 0.28, 0.19, 0.15 for the deposited crystal and a 2nd 
 crystal that we did not deposit had twin fractions of 0.25, 0.27, 0.17, 
 0.31).  The 2nd case we had was after support for twining (including 
 tetartohedral twinning) was added to refmac (and I think phenix.refine can 
 also handle this).  For 2NUZ, it was P32 with refined twin fractions of 0.25, 
 0.27, 0.17, 0.31.
 
  Pietro Roversi wrote a review of tetartohedral twinning for the CCP4 
 proceedings issues of acta D http://dx.doi.org/10.1107/S0907444912006737 
 
  I would try refinement with refmac using the original (non-detwinned F's) 
 with just the TWIN command to see if it ends up keeping twin fractions for 
 all 3 operators (4 domains) -- especially with crystals 1 and 3 which appear 
 to have the largest estimates of the other twin fractions.
 
 Regards,
 Mitch
 
 
 ==
 Mitchell Miller, Ph.D.
 Joint Center for Structural Genomics
 Stanford Synchrotron Radiation Lightsource
 2575 Sand Hill Rd  -- SLAC MS 99
 Menlo Park, CA  94025
 Phone: 1-650-926-5036
 FAX: 1-650-926-3292
 
 
 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
 herman.schreu...@sanofi.com
 Sent: Thursday, June 20, 2013 6:47 AM
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: [ccp4bb] Twinning problem
 
 Dear Bulletin Board,
 
 Prodded by pdb annotators, which are very hesitant to accept coordinate files 
 when their Rfactor does not correspond with our Rfactor, I had a look again 
 into some old data sets, which I suspect are twinned. Below are the results 
 of some twinning tests with the Detwin program (top value: all reflections, 
 lower value: reflections  Nsig*obs (whatever that may mean). The space group 
 is P32, the resolution is 2.3 - 2.6 Å and data are reasonable 

Re: [ccp4bb] Puzzling observation about size exclusion chromatography

2013-06-21 Thread R. M. Garavito
Dear Zhen,

I should also point out that the statement Matt made (Superdex is known to 
have some ion-exchange characteristics, so that it can weakly interact with 
some proteins.) is not completely correct.  Superdex and all chromatographic 
media made from carbohydrates (dextran, agarose, etc.) are also quite 
hydrophobic (which is surprising to some).  The observation Zhen made is the 
classic behavior of hydrophobic interaction with the gel filtration media, 
which led to the development of hydrophobic interaction chromatographic (HIC).  
For HIC, you load the protein in high salt, and elute with low salt or a 
chaotrope (LiCl).

So when you redo you experiment, Zhen, try with AND without salt.  As Matt 
said, if it is an ion-exchange interaction, extra salt will make it elute more 
normally.  But if it gets worse, then the extra salt is increasing the 
hydrophobic interactions, and you should run the column in lower salt (~50 mM 
NaCl) or with a bit of detergent.  Given that you are refolding your protein, a 
partially folded protein may have have more hydrophobic patches. As my lab is 
routinely refolding our target proteins, we are always watching for this 
behavior, including on our analytical Superdex 200 10/300 columns, which is one 
of our favorites.  Excessive hydrophobic interactions can also lead to clogged 
columns, which is not what you want for this expensive column.

Good luck,

Michael


R. Michael Garavito, Ph.D.
Professor of Biochemistry  Molecular Biology
603 Wilson Rd., Rm. 513   
Michigan State University  
East Lansing, MI 48824-1319
Office:  (517) 355-9724 Lab:  (517) 353-9125
FAX:  (517) 353-9334Email:  rmgarav...@gmail.com





On Jun 20, 2013, at 5:38 PM, Zhang, Zhen wrote:

 Hi Matthew,
 
 Thanks a lot. That is a great idea. I will try the high salt and worry about 
 the crystallization later. 
 
 Zhen
 
 -Original Message-
 From: Matthew Franklin [mailto:mfrank...@nysbc.org] 
 Sent: Thursday, June 20, 2013 4:34 PM
 To: Zhang, Zhen
 Cc: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] Puzzling observation about size exclusion chromatography
 
 Hi Zhen -
 
 Superdex is known to have some ion-exchange characteristics, so that it 
 can weakly interact with some proteins.  This is why the manufacturer 
 recommends including a certain amount of salt in the running buffer.  I 
 have had the same experience with a few proteins, including one that 
 came off the column well after the salt peak!  (The protein was very 
 clean after this step; all other proteins had eluted earlier.)
 
 As others have said, you can't rely on molecular weight calibrations in 
 this case, but this behavior alone is no reason to think that the 
 protein is misfolded or otherwise badly behaved. If you don't like the 
 late elution, try increasing the salt concentration of your running 
 buffer to 250 or even 500 mM. You'll probably need to exchange the 
 eluted protein back into a low-salt buffer for your next steps (e.g. 
 crystallization) if you do this.
 
 - Matt
 
 
 On 6/20/13 3:09 PM, Zhang, Zhen wrote:
 Dear all,
 
 I just observed a puzzling phenomenon when purifying a refolded protein with 
 size exclusion chromatography. The protein was solubilized by 8M Urea and 
 refolded by dialysis against 500mM Arginine in PBS. The protein is 40KDal 
 and is expected to be a trimer. The puzzling part is the protein after 
 refolding always eluted at 18ml from the superdex S200 column (10/300), 
 which is calculated to be 5KDal by standard. However, the fractions appear 
 to be at 40KDal with SDS PAGE and the protein is functional in term of in 
 vitro binding to the protein-specific monoclonal antibody. I could not 
 explain the observation and I am wondering if anyone has the similar 
 experience or has an opinion on this. Any comments are welcome.
 
 Thanks.
 
 Zhen
 
 
 The information in this e-mail is intended only for the person to whom it is
 addressed. If you believe this e-mail was sent to you in error and the e-mail
 contains patient information, please contact the Partners Compliance 
 HelpLine at
 http://www.partners.org/complianceline . If the e-mail was sent to you in 
 error
 but does not contain patient information, please contact the sender and 
 properly
 dispose of the e-mail.
 
 
 
 
 -- 
 Matthew Franklin, Ph. D.
 Senior Scientist
 New York Structural Biology Center
 89 Convent Avenue, New York, NY 10027
 (212) 939-0660 ext. 9374



Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Ian Tickle
On 21 June 2013 13:36, Ed Pozharski epozh...@umaryland.edu wrote:

 Replacing Iobs with E(J) is not only unnecessary, it's ill-advised as it
 will distort intensity statistics.  For example, let's say you have
 translational NCS aligned with crystallographic axes, and hence some set of
 reflections is systematically absent.  If all is well, Iobs~0 for the
 subset while E(J) is systematically positive.  This obviously happens
 because the standard Wilson prior is wrong for these reflections, but I
 digress, as usual.


Ed,

If you observe the symptoms of translational NCS in the diffraction pattern
(i.e. systematically weak zones of reflections) you must take it into
account when calculating the averages, i.e. if you do it properly parity
groups should be normalised separately (though I concede there may be a
practical issue in that I'm not aware of any software that currently has
this feature).  In that case E(J) will be ~ 0, as expected.  If you don't
do that then clearly you can't expect to get the right answer!  The
theoretical intensities are based on the assumption that the intensity
distributions are all positive, so it makes no sense to compare them with
an experimental distribution where a significant fraction are negative.
How exactly do you propose to deal properly with the P-Y L test that I
described? - because of course that also inherently assumes that the
intensities are all positive and it's certainly not valid to assume that
E(J) = E(F)^2 !

Another point is that (to paraphrase G. Orwell) not all reflections are
created equal, just some are more equal than others.  What I mean is that
in counting reflections for the cumulative distributions (i.e. you count
the number of reflections in ranges of intensity or in ranges of L), a weak
reflection should be counted as fractional with a contribution to the total
which is less than 1, on a continuous scale from 0 to 1 related to
I/sigma(I).  In fact referring to your original posting reflections with h
 -4 will get such a small weight that it will be effectively zero and they
won't be counted at all (or it won't make the slightest difference whether
you count them or not).  Of course when it comes to outputting reflections
you can't have a fractional reflection, it's either included or it isn't.
So then you may have to have an arbitrary cutoff, though such reflections
would likely end up with zero intensity and large SD (but the programs may
not currently be good at estimating the latter in such a situation, which
is probably why they are currently rejected).

Another point worth mentioning is that the observed distributions of E^n (E
= normalised structure amplitude) tend to be very noisy, particularly for
large n, and I have a suspicion (as yet untested) that this may come from
weak reflections which have made a full contribution to the count when it
should have been fractional (or even zero).

I'm currently working on a revised version of TRUNCATE where some or all of
the above issues will be addressed.

Cheers

-- Ian


Re: [ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Robbie Joosten
Hi Herman, 

Tighter restraints typically close the gap between R and R-free. This does
not mean one should just tighten the restraints to satisfy one's own (or a
referee's) idea of what the gap should be. I don't think there is a clear
target of how large or small the gap should be. If you optimize the
restraints to get the best (free) likelihood, you usually get a reasonable R
gap without explicitly optimizing it. 

Cheers,
Robbie

 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 Eleanor Dodson
 Sent: Friday, June 21, 2013 14:21
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] AW: Twinning problem - almost solved.
 
 At your resolution that seems to me a reasonable gap between R and Rfree?
  Eleanor
 
 On 21 Jun 2013, at 12:28, herman.schreu...@sanofi.com wrote:
 
  Dear Bulletin Board,
 
  After some headbanging (Refmac5 had helpfully created gap records for
all
 insertions and deletions present in the structure), I got refmac5 running
with
 the TWIN option. Refmac5 also found the k,h,-l domain and rejected the
 other possible domains because they were too small. The Rfactor's are now
 extremely good: ~14% and the Rfree's are for me acceptable: ~24%. Since I
 found the difference between R and Rfree somewhat large, I have been
 playing with the weighting. By using a weight of 0.01, I can bring the
Rfactor
 up to 18%, but the Rfree stays about the same or even gets a little worse.
 
  My question: is there a way to bring R and Rfree closer together, or is
it
 related to the twinned data and is it something we have to live with?
 
  Best regards,
  Herman
 
 
  -Ursprüngliche Nachricht-
  Von: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von
 Miller, Mitchell D.
  Gesendet: Donnerstag, 20. Juni 2013 17:43
  An: CCP4BB@JISCMAIL.AC.UK
  Betreff: Re: [ccp4bb] Twinning problem
 
  You are welcome.  Let me also for the benefit of others who may search
 the archives in the future, let me correct two errors below - (typo and a
miss-
 recollection).
 
  Specially, I was thinking that phenix.refine was now able to refine
multiple
 twin laws, but according to Nat Echols on the phenix mailing list
 http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html
  phenix.refine only handles 1 twin law at this time.
  (My typo was that and our second structure was 3nuz with twin fractions
 0.38, 0.32, 0.16 and 0.14 -- not 2nuz).
 
  A useful search for deposited structures mentioning tetartohedral
 http://www.ebi.ac.uk/pdbe-
 srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETAR
 TOHEDRAL
 
  Regards,
  Mitch
 
 
  -Original Message-
  From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 herman.schreu...@sanofi.com
  Sent: Thursday, June 20, 2013 8:04 AM
  To: CCP4BB@JISCMAIL.AC.UK
  Subject: [ccp4bb] AW: Twinning problem
 
  Dear Mitch (and Philip and Phil),
 
  It is clear that I should give refmac a go with the non-detwinned F's
and just
 the TWIN command.
 
  Thank you for your suggestions,
  Herman
 
 
 
  -Ursprüngliche Nachricht-
  Von: Miller, Mitchell D. [mailto:mmil...@slac.stanford.edu]
  Gesendet: Donnerstag, 20. Juni 2013 16:18
  An: Schreuder, Herman RD/DE
  Betreff: RE: Twinning problem
 
  Hi Herman,
  Have you considered the possibility of your crystals being tetartohedral
 twinned.  That is more than one of the twin laws may apply to your
crystals.
  E.g. in P32 it is possible to have tetartohedral twinning which would
have
  4 twin domains - (h,k,l), (k,h,-l), (-h,-k,l) and (-k,-h,-l). Perfect
tetartohedral
 twinning of P3 would merge in P622 and each twin domain would have a
 faction of 0.25.
 
   We have had 2 cases like this (the first 2PRX was before there was
support
 for this type of twinning except for in shelxl and we ended up with
refined
 twin fractions of 0.38, 0.28, 0.19, 0.15 for the deposited crystal and a
2nd
 crystal that we did not deposit had twin fractions of 0.25, 0.27, 0.17,
0.31).
 The 2nd case we had was after support for twining (including tetartohedral
 twinning) was added to refmac (and I think phenix.refine can also handle
 this).  For 2NUZ, it was P32 with refined twin fractions of 0.25, 0.27,
0.17, 0.31.
 
   Pietro Roversi wrote a review of tetartohedral twinning for the CCP4
 proceedings issues of acta D http://dx.doi.org/10.1107/S0907444912006737
 
   I would try refinement with refmac using the original (non-detwinned
F's)
 with just the TWIN command to see if it ends up keeping twin fractions for
all
 3 operators (4 domains) -- especially with crystals 1 and 3 which appear
to
 have the largest estimates of the other twin fractions.
 
  Regards,
  Mitch
 
 
  ==
  Mitchell Miller, Ph.D.
  Joint Center for Structural Genomics
  Stanford Synchrotron Radiation Lightsource
  2575 Sand Hill Rd  -- SLAC MS 99
  Menlo Park, CA  94025
  Phone: 1-650-926-5036
  FAX: 1-650-926-3292
 
 
  -Original 

Re: [ccp4bb] Puzzling observation about size exclusion chromatography

2013-06-21 Thread Zhang, Zhen
Hi Michael,

Thank you very much for your suggestions. I will rerun with both buffer 
conditions and report back later.

Thanks to everyone for your reply.

Zhen

From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of R. M. 
Garavito
Sent: Friday, June 21, 2013 9:28 AM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] Puzzling observation about size exclusion chromatography

Dear Zhen,

I should also point out that the statement Matt made (Superdex is known to 
have some ion-exchange characteristics, so that it can weakly interact with 
some proteins.) is not completely correct.  Superdex and all chromatographic 
media made from carbohydrates (dextran, agarose, etc.) are also quite 
hydrophobic (which is surprising to some).  The observation Zhen made is the 
classic behavior of hydrophobic interaction with the gel filtration media, 
which led to the development of hydrophobic interaction chromatographic (HIC).  
For HIC, you load the protein in high salt, and elute with low salt or a 
chaotrope (LiCl).

So when you redo you experiment, Zhen, try with AND without salt.  As Matt 
said, if it is an ion-exchange interaction, extra salt will make it elute more 
normally.  But if it gets worse, then the extra salt is increasing the 
hydrophobic interactions, and you should run the column in lower salt (~50 mM 
NaCl) or with a bit of detergent.  Given that you are refolding your protein, a 
partially folded protein may have have more hydrophobic patches. As my lab is 
routinely refolding our target proteins, we are always watching for this 
behavior, including on our analytical Superdex 200 10/300 columns, which is one 
of our favorites.  Excessive hydrophobic interactions can also lead to clogged 
columns, which is not what you want for this expensive column.

Good luck,

Michael


R. Michael Garavito, Ph.D.
Professor of Biochemistry  Molecular Biology
603 Wilson Rd., Rm. 513
Michigan State University
East Lansing, MI 48824-1319
Office:  (517) 355-9724 Lab:  (517) 353-9125
FAX:  (517) 353-9334Email:  
rmgarav...@gmail.commailto:garav...@gmail.com





On Jun 20, 2013, at 5:38 PM, Zhang, Zhen wrote:


Hi Matthew,

Thanks a lot. That is a great idea. I will try the high salt and worry about 
the crystallization later.

Zhen

-Original Message-
From: Matthew Franklin [mailto:mfrank...@nysbc.org]
Sent: Thursday, June 20, 2013 4:34 PM
To: Zhang, Zhen
Cc: CCP4BB@JISCMAIL.AC.UKmailto:CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] Puzzling observation about size exclusion chromatography

Hi Zhen -

Superdex is known to have some ion-exchange characteristics, so that it
can weakly interact with some proteins.  This is why the manufacturer
recommends including a certain amount of salt in the running buffer.  I
have had the same experience with a few proteins, including one that
came off the column well after the salt peak!  (The protein was very
clean after this step; all other proteins had eluted earlier.)

As others have said, you can't rely on molecular weight calibrations in
this case, but this behavior alone is no reason to think that the
protein is misfolded or otherwise badly behaved. If you don't like the
late elution, try increasing the salt concentration of your running
buffer to 250 or even 500 mM. You'll probably need to exchange the
eluted protein back into a low-salt buffer for your next steps (e.g.
crystallization) if you do this.

- Matt


On 6/20/13 3:09 PM, Zhang, Zhen wrote:

Dear all,

I just observed a puzzling phenomenon when purifying a refolded protein with 
size exclusion chromatography. The protein was solubilized by 8M Urea and 
refolded by dialysis against 500mM Arginine in PBS. The protein is 40KDal and 
is expected to be a trimer. The puzzling part is the protein after refolding 
always eluted at 18ml from the superdex S200 column (10/300), which is 
calculated to be 5KDal by standard. However, the fractions appear to be at 
40KDal with SDS PAGE and the protein is functional in term of in vitro binding 
to the protein-specific monoclonal antibody. I could not explain the 
observation and I am wondering if anyone has the similar experience or has an 
opinion on this. Any comments are welcome.

Thanks.

Zhen


The information in this e-mail is intended only for the person to whom it is
addressed. If you believe this e-mail was sent to you in error and the e-mail
contains patient information, please contact the Partners Compliance HelpLine at
http://www.partners.org/complianceline . If the e-mail was sent to you in error
but does not contain patient information, please contact the sender and properly
dispose of the e-mail.




--
Matthew Franklin, Ph. D.
Senior Scientist
New York Structural Biology Center
89 Convent Avenue, New York, NY 10027
(212) 939-0660 ext. 9374



[ccp4bb] AW: [ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Herman . Schreuder
Hi Robbie,

That is what I tried. The Rfactor got a lot worse (14%-18%) and the Rfree got 
a little worse (by 0.1-0.2%). My feeling is that that is not the right 
approach. Roger Rowlett suggested to give PDB_REDO a try. Maybe you have some 
instructions available how to get a local version?

Best,
Herman

-Ursprüngliche Nachricht-
Von: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von Robbie 
Joosten
Gesendet: Freitag, 21. Juni 2013 16:21
An: CCP4BB@JISCMAIL.AC.UK
Betreff: Re: [ccp4bb] AW: Twinning problem - almost solved.

Hi Herman, 

Tighter restraints typically close the gap between R and R-free. This does not 
mean one should just tighten the restraints to satisfy one's own (or a
referee's) idea of what the gap should be. I don't think there is a clear 
target of how large or small the gap should be. If you optimize the restraints 
to get the best (free) likelihood, you usually get a reasonable R gap without 
explicitly optimizing it. 

Cheers,
Robbie

 -Original Message-
 From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of 
 Eleanor Dodson
 Sent: Friday, June 21, 2013 14:21
 To: CCP4BB@JISCMAIL.AC.UK
 Subject: Re: [ccp4bb] AW: Twinning problem - almost solved.
 
 At your resolution that seems to me a reasonable gap between R and Rfree?
  Eleanor
 
 On 21 Jun 2013, at 12:28, herman.schreu...@sanofi.com wrote:
 
  Dear Bulletin Board,
 
  After some headbanging (Refmac5 had helpfully created gap records 
  for
all
 insertions and deletions present in the structure), I got refmac5 
 running
with
 the TWIN option. Refmac5 also found the k,h,-l domain and rejected the 
 other possible domains because they were too small. The Rfactor's are 
 now extremely good: ~14% and the Rfree's are for me acceptable: ~24%. 
 Since I found the difference between R and Rfree somewhat large, I 
 have been playing with the weighting. By using a weight of 0.01, I can 
 bring the
Rfactor
 up to 18%, but the Rfree stays about the same or even gets a little worse.
 
  My question: is there a way to bring R and Rfree closer together, or 
  is
it
 related to the twinned data and is it something we have to live with?
 
  Best regards,
  Herman
 
 
  -Ursprüngliche Nachricht-
  Von: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag 
  von
 Miller, Mitchell D.
  Gesendet: Donnerstag, 20. Juni 2013 17:43
  An: CCP4BB@JISCMAIL.AC.UK
  Betreff: Re: [ccp4bb] Twinning problem
 
  You are welcome.  Let me also for the benefit of others who may 
  search
 the archives in the future, let me correct two errors below - (typo 
 and a
miss-
 recollection).
 
  Specially, I was thinking that phenix.refine was now able to refine
multiple
 twin laws, but according to Nat Echols on the phenix mailing list 
 http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html
  phenix.refine only handles 1 twin law at this time.
  (My typo was that and our second structure was 3nuz with twin 
  fractions
 0.38, 0.32, 0.16 and 0.14 -- not 2nuz).
 
  A useful search for deposited structures mentioning tetartohedral
 http://www.ebi.ac.uk/pdbe-
 srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETAR
 TOHEDRAL
 
  Regards,
  Mitch
 
 
  -Original Message-
  From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf 
  Of
 herman.schreu...@sanofi.com
  Sent: Thursday, June 20, 2013 8:04 AM
  To: CCP4BB@JISCMAIL.AC.UK
  Subject: [ccp4bb] AW: Twinning problem
 
  Dear Mitch (and Philip and Phil),
 
  It is clear that I should give refmac a go with the non-detwinned 
  F's
and just
 the TWIN command.
 
  Thank you for your suggestions,
  Herman
 
 
 
  -Ursprüngliche Nachricht-
  Von: Miller, Mitchell D. [mailto:mmil...@slac.stanford.edu]
  Gesendet: Donnerstag, 20. Juni 2013 16:18
  An: Schreuder, Herman RD/DE
  Betreff: RE: Twinning problem
 
  Hi Herman,
  Have you considered the possibility of your crystals being 
  tetartohedral
 twinned.  That is more than one of the twin laws may apply to your
crystals.
  E.g. in P32 it is possible to have tetartohedral twinning which 
  would
have
  4 twin domains - (h,k,l), (k,h,-l), (-h,-k,l) and (-k,-h,-l). 
  Perfect
tetartohedral
 twinning of P3 would merge in P622 and each twin domain would have a 
 faction of 0.25.
 
   We have had 2 cases like this (the first 2PRX was before there was
support
 for this type of twinning except for in shelxl and we ended up with
refined
 twin fractions of 0.38, 0.28, 0.19, 0.15 for the deposited crystal and 
 a
2nd
 crystal that we did not deposit had twin fractions of 0.25, 0.27, 
 0.17,
0.31).
 The 2nd case we had was after support for twining (including 
 tetartohedral
 twinning) was added to refmac (and I think phenix.refine can also 
 handle this).  For 2NUZ, it was P32 with refined twin fractions of 
 0.25, 0.27,
0.17, 0.31.
 
   Pietro Roversi wrote a review of tetartohedral twinning for the 
  CCP4
 proceedings issues of acta D 
 

Re: [ccp4bb] AW: [ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Tim Gruene
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Dear Herman,

a large gap between R and Rfree might indicate a horrible geometry of
your structure, especially if R increased by lowering the matrix
weight in refmac. Or, to put it the other way round: it is easy to
achieve a low R-value by screwing up the geometry of your model. Did
you run the PDB file through molprobity?

My guess is with the R=14% model you get a very red chart...

Best,
Tim

On 06/21/2013 04:45 PM, herman.schreu...@sanofi.com wrote:
 Hi Robbie,
 
 That is what I tried. The Rfactor got a lot worse (14%-18%) and
 the Rfree got a little worse (by 0.1-0.2%). My feeling is that that
 is not the right approach. Roger Rowlett suggested to give PDB_REDO
 a try. Maybe you have some instructions available how to get a
 local version?
 
 Best, Herman
 
 -Ursprüngliche Nachricht- Von: CCP4 bulletin board
 [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von Robbie Joosten 
 Gesendet: Freitag, 21. Juni 2013 16:21 An: CCP4BB@JISCMAIL.AC.UK 
 Betreff: Re: [ccp4bb] AW: Twinning problem - almost solved.
 
 Hi Herman,
 
 Tighter restraints typically close the gap between R and R-free.
 This does not mean one should just tighten the restraints to
 satisfy one's own (or a referee's) idea of what the gap should be.
 I don't think there is a clear target of how large or small the gap
 should be. If you optimize the restraints to get the best (free)
 likelihood, you usually get a reasonable R gap without explicitly
 optimizing it.
 
 Cheers, Robbie
 
 -Original Message- From: CCP4 bulletin board
 [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of Eleanor Dodson Sent:
 Friday, June 21, 2013 14:21 To: CCP4BB@JISCMAIL.AC.UK Subject:
 Re: [ccp4bb] AW: Twinning problem - almost solved.
 
 At your resolution that seems to me a reasonable gap between R
 and Rfree? Eleanor
 
 On 21 Jun 2013, at 12:28, herman.schreu...@sanofi.com wrote:
 
 Dear Bulletin Board,
 
 After some headbanging (Refmac5 had helpfully created gap
 records for
 all
 insertions and deletions present in the structure), I got refmac5
  running
 with
 the TWIN option. Refmac5 also found the k,h,-l domain and
 rejected the other possible domains because they were too small.
 The Rfactor's are now extremely good: ~14% and the Rfree's are
 for me acceptable: ~24%. Since I found the difference between R
 and Rfree somewhat large, I have been playing with the weighting.
 By using a weight of 0.01, I can bring the
 Rfactor
 up to 18%, but the Rfree stays about the same or even gets a
 little worse.
 
 My question: is there a way to bring R and Rfree closer
 together, or is
 it
 related to the twinned data and is it something we have to live
 with?
 
 Best regards, Herman
 
 
 -Ursprüngliche Nachricht- Von: CCP4 bulletin board
 [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von
 Miller, Mitchell D.
 Gesendet: Donnerstag, 20. Juni 2013 17:43 An:
 CCP4BB@JISCMAIL.AC.UK Betreff: Re: [ccp4bb] Twinning problem
 
 You are welcome.  Let me also for the benefit of others who may
  search
 the archives in the future, let me correct two errors below -
 (typo and a
 miss-
 recollection).
 
 Specially, I was thinking that phenix.refine was now able to
 refine
 multiple
 twin laws, but according to Nat Echols on the phenix mailing list
  
 http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html

 
phenix.refine only handles 1 twin law at this time.
 (My typo was that and our second structure was 3nuz with twin 
 fractions
 0.38, 0.32, 0.16 and 0.14 -- not 2nuz).
 
 A useful search for deposited structures mentioning
 tetartohedral
 http://www.ebi.ac.uk/pdbe- 
 srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETAR

 
TOHEDRAL
 
 Regards, Mitch
 
 
 -Original Message- From: CCP4 bulletin board
 [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 herman.schreu...@sanofi.com
 Sent: Thursday, June 20, 2013 8:04 AM To:
 CCP4BB@JISCMAIL.AC.UK Subject: [ccp4bb] AW: Twinning problem
 
 Dear Mitch (and Philip and Phil),
 
 It is clear that I should give refmac a go with the
 non-detwinned F's
 and just
 the TWIN command.
 
 Thank you for your suggestions, Herman
 
 
 
 -Ursprüngliche Nachricht- Von: Miller, Mitchell D.
 [mailto:mmil...@slac.stanford.edu] Gesendet: Donnerstag, 20.
 Juni 2013 16:18 An: Schreuder, Herman RD/DE Betreff: RE:
 Twinning problem
 
 Hi Herman, Have you considered the possibility of your crystals
 being tetartohedral
 twinned.  That is more than one of the twin laws may apply to
 your
 crystals.
 E.g. in P32 it is possible to have tetartohedral twinning which
  would
 have
 4 twin domains - (h,k,l), (k,h,-l), (-h,-k,l) and (-k,-h,-l). 
 Perfect
 tetartohedral
 twinning of P3 would merge in P622 and each twin domain would
 have a faction of 0.25.
 
 We have had 2 cases like this (the first 2PRX was before there
 was
 support
 for this type of twinning except for in shelxl and we ended up
 with
 refined
 twin fractions of 0.38, 0.28, 0.19, 0.15 for the deposited
 crystal 

[ccp4bb] AW: [ccp4bb] AW: [ccp4bb] AW: Twinning problem - almost solved.

2013-06-21 Thread Herman . Schreuder
Dear Tim,

I normally do not use Refmac, so I have no idea what to expect and what would 
be a good weight. I will do the molprobity test, but I do not expect major 
problems. This is a MR structure with a high resolution search model with 100% 
sequence identity. A few amino acids may have problems, but the overall 
structure should be ok. E.g. Refmac lists an RMSD bonds of 0.008 Å and an RMDS 
bond angles of 1.3°. Anyway, if the structure turns out to be really bad, I 
will get punished by the PDB annotators with lots of REMARK 500 records!

Best,
Herman



-Ursprüngliche Nachricht-
Von: Tim Gruene [mailto:t...@shelx.uni-ac.gwdg.de] 
Gesendet: Freitag, 21. Juni 2013 17:30
An: Schreuder, Herman RD/DE
Cc: CCP4BB@JISCMAIL.AC.UK
Betreff: Re: [ccp4bb] AW: [ccp4bb] AW: Twinning problem - almost solved.

-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1

Dear Herman,

a large gap between R and Rfree might indicate a horrible geometry of your 
structure, especially if R increased by lowering the matrix weight in refmac. 
Or, to put it the other way round: it is easy to achieve a low R-value by 
screwing up the geometry of your model. Did you run the PDB file through 
molprobity?

My guess is with the R=14% model you get a very red chart...

Best,
Tim

On 06/21/2013 04:45 PM, herman.schreu...@sanofi.com wrote:
 Hi Robbie,
 
 That is what I tried. The Rfactor got a lot worse (14%-18%) and the 
 Rfree got a little worse (by 0.1-0.2%). My feeling is that that is not 
 the right approach. Roger Rowlett suggested to give PDB_REDO a try. 
 Maybe you have some instructions available how to get a local version?
 
 Best, Herman
 
 -Ursprüngliche Nachricht- Von: CCP4 bulletin board 
 [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von Robbie Joosten
 Gesendet: Freitag, 21. Juni 2013 16:21 An: CCP4BB@JISCMAIL.AC.UK
 Betreff: Re: [ccp4bb] AW: Twinning problem - almost solved.
 
 Hi Herman,
 
 Tighter restraints typically close the gap between R and R-free.
 This does not mean one should just tighten the restraints to satisfy 
 one's own (or a referee's) idea of what the gap should be.
 I don't think there is a clear target of how large or small the gap 
 should be. If you optimize the restraints to get the best (free) 
 likelihood, you usually get a reasonable R gap without explicitly 
 optimizing it.
 
 Cheers, Robbie
 
 -Original Message- From: CCP4 bulletin board 
 [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of Eleanor Dodson Sent:
 Friday, June 21, 2013 14:21 To: CCP4BB@JISCMAIL.AC.UK Subject:
 Re: [ccp4bb] AW: Twinning problem - almost solved.
 
 At your resolution that seems to me a reasonable gap between R and 
 Rfree? Eleanor
 
 On 21 Jun 2013, at 12:28, herman.schreu...@sanofi.com wrote:
 
 Dear Bulletin Board,
 
 After some headbanging (Refmac5 had helpfully created gap records 
 for
 all
 insertions and deletions present in the structure), I got refmac5  
 running
 with
 the TWIN option. Refmac5 also found the k,h,-l domain and rejected 
 the other possible domains because they were too small.
 The Rfactor's are now extremely good: ~14% and the Rfree's are for me 
 acceptable: ~24%. Since I found the difference between R and Rfree 
 somewhat large, I have been playing with the weighting.
 By using a weight of 0.01, I can bring the
 Rfactor
 up to 18%, but the Rfree stays about the same or even gets a little 
 worse.
 
 My question: is there a way to bring R and Rfree closer together, or 
 is
 it
 related to the twinned data and is it something we have to live with?
 
 Best regards, Herman
 
 
 -Ursprüngliche Nachricht- Von: CCP4 bulletin board 
 [mailto:CCP4BB@JISCMAIL.AC.UK] Im Auftrag von
 Miller, Mitchell D.
 Gesendet: Donnerstag, 20. Juni 2013 17:43 An:
 CCP4BB@JISCMAIL.AC.UK Betreff: Re: [ccp4bb] Twinning problem
 
 You are welcome.  Let me also for the benefit of others who may  
 search
 the archives in the future, let me correct two errors below - (typo 
 and a
 miss-
 recollection).
 
 Specially, I was thinking that phenix.refine was now able to refine
 multiple
 twin laws, but according to Nat Echols on the phenix mailing list
  
 http://phenix-online.org/pipermail/phenixbb/2013-March/019538.html

 
phenix.refine only handles 1 twin law at this time.
 (My typo was that and our second structure was 3nuz with twin 
 fractions
 0.38, 0.32, 0.16 and 0.14 -- not 2nuz).
 
 A useful search for deposited structures mentioning tetartohedral
 http://www.ebi.ac.uk/pdbe-
 srv/view/search?search_type=all_texttext=TETARTOHEDRALLY+OR+TETAR

 
TOHEDRAL
 
 Regards, Mitch
 
 
 -Original Message- From: CCP4 bulletin board 
 [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of
 herman.schreu...@sanofi.com
 Sent: Thursday, June 20, 2013 8:04 AM To:
 CCP4BB@JISCMAIL.AC.UK Subject: [ccp4bb] AW: Twinning problem
 
 Dear Mitch (and Philip and Phil),
 
 It is clear that I should give refmac a go with the non-detwinned 
 F's
 and just
 the TWIN command.
 
 Thank you for your suggestions, Herman
 
 
 
 

Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Douglas Theobald
On Jun 21, 2013, at 8:36 AM, Ed Pozharski epozh...@umaryland.edu wrote:

 On 06/20/2013 01:07 PM, Douglas Theobald wrote:
 How can there be nothing wrong with something that is unphysical?  
 Intensities cannot be negative.
 
 I think you are confusing two things - the true intensities and observed 
 intensities.

But I'm not.  Let me try to convince you ...

 True intensities represent the number of photons that diffract off a crystal 
 in a specific direction or, for QED-minded, relative probabilities of a 
 single photon being found in a particular area of the detector when it's 
 probability wave function finally collapses.

I agree. 

 True intensities certainly cannot be negative and in crystallographic method 
 they never are. They are represented by the best theoretical estimates 
 possible, Icalc.  These are always positive.

I also very much agree.  

 Observed intensities are the best estimates that we can come up with in an 
 experiment.  

I also agree with this, and this is the clincher.  You are arguing that 
Ispot-Iback=Iobs is the best estimate we can come up with.  I claim that is 
absurd.  How are you quantifying best?  Usually we have some sort of 
discrepancy measure between true and estimate, like RMSD, mean absolute 
distance, log distance, or somesuch.  Here is the important point --- by any 
measure of discrepancy you care to use, the person who estimates Iobs as 0 when 
IbackIspot will *always*, in *every case*, beat the person who estimates Iobs 
with a negative value.   This is an indisputable fact.  

 These are determined by integrating pixels around the spot where particular 
 reflection is expected to hit the detector.  Unfortunately, science did not 
 yet invent a method that would allow to suspend a crystal in vacuum while 
 also removing all of the outside solvent.  Neither we have included diffuse 
 scatter in our theoretical model.  Because of that, full reflection intensity 
 contains background signal in addition to the Icalc.  This background has to 
 be subtracted and what is perhaps the most useful form of observation is 
 Ispot-Iback=Iobs.

How can that be the most useful form, when 0 is always a better estimate than a 
negative value, by any criterion?

 These observed intensities can be negative because while their true 
 underlying value is positive, random errorsmay result in IbackIspot.  There 
 is absolutely nothing unphysical here.

Yes there is.  The only way you can get a negative estimate is to make 
unphysical assumptions.  Namely, the estimate Ispot-Iback=Iobs assumes that 
both the true value of I and the background noise come from a Gaussian 
distribution that is allowed to have negative values.  Both of those 
assumptions are unphysical.  

 Replacing Iobs with E(J) is not only unnecessary, it's ill-advised as it will 
 distort intensity statistics.  For example, let's say you have translational 
 NCS aligned with crystallographic axes, and hence some set of reflections is 
 systematically absent.  If all is well, Iobs~0 for the subset while E(J) 
 is systematically positive.  This obviously happens because the standard 
 Wilson prior is wrong for these reflections, but I digress, as usual.
 
 In summary, there is indeed nothing wrong, imho, with negative Iobs.  The 
 fact that some of these may become negative is correctly accounted for once 
 sigI is factored into the ML target.
 
 Cheers,
 
 Ed.
 
 -- 
 Oh, suddenly throwing a giraffe into a volcano to make water is crazy?
Julian, King of Lemurs
 


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Ian Tickle
On 21 June 2013 17:10, Douglas Theobald dtheob...@brandeis.edu wrote:

 Yes there is.  The only way you can get a negative estimate is to make
 unphysical assumptions.  Namely, the estimate Ispot-Iback=Iobs assumes that
 both the true value of I and the background noise come from a Gaussian
 distribution that is allowed to have negative values.  Both of those
 assumptions are unphysical.


Actually that's not correct: Ispot and Iback are both assumed to come from
a _Poisson_ distribution which by definition is zero for negative values of
its argument (you can't have a negative number of photons), so are _not_
allowed to have negative values.  For large values of the argument (in fact
the approximation is pretty good even for x ~ 10) a Poisson approximates to
a Gaussian, and then of course the difference Ispot-Iback is also
approximately Gaussian.

But I think that doesn't affect your argument.

Cheers

-- Ian


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Ed Pozharski

On 06/21/2013 10:19 AM, Ian Tickle wrote:
If you observe the symptoms of translational NCS in the diffraction 
pattern (i.e. systematically weak zones of reflections) you must take 
it into account when calculating the averages, i.e. if you do it 
properly parity groups should be normalised separately (though I 
concede there may be a practical issue in that I'm not aware of any 
software that currently has this feature). 


Ian,

I think this is exactly what I was trying to emphasize, that applying 
some conversion to raw intensities may have negative impact when 
conversion is based on incorrect or incomplete assumptions.


Cheers,

Ed.

--
Oh, suddenly throwing a giraffe into a volcano to make water is crazy?
Julian, King of Lemurs


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Douglas Theobald
I kinda think we're saying the same thing, sort of.

You don't like the Gaussian assumption, and neither do I.  If you make the 
reasonable Poisson assumptions, then you don't get the Ispot-Iback=Iobs for the 
best estimate of Itrue.  Except as an approximation for large values, but we 
are talking about the case when IbackIspot, where the Gaussian approximation 
to the Poisson no longer holds.  The sum of two Poisson variates is also 
Poisson, which also can never be negative, unlike the Gaussian.  

So I reiterate: the Ispot-Iback=Iobs equation assumes Gaussians and hence 
negativity.  The Ispot-Iback=Iobs does not follow from a Poisson assumption.  


On Jun 21, 2013, at 1:13 PM, Ian Tickle ianj...@gmail.com wrote:

 On 21 June 2013 17:10, Douglas Theobald dtheob...@brandeis.edu wrote:
 Yes there is.  The only way you can get a negative estimate is to make 
 unphysical assumptions.  Namely, the estimate Ispot-Iback=Iobs assumes that 
 both the true value of I and the background noise come from a Gaussian 
 distribution that is allowed to have negative values.  Both of those 
 assumptions are unphysical.
 
 Actually that's not correct: Ispot and Iback are both assumed to come from a 
 _Poisson_ distribution which by definition is zero for negative values of its 
 argument (you can't have a negative number of photons), so are _not_ allowed 
 to have negative values.  For large values of the argument (in fact the 
 approximation is pretty good even for x ~ 10) a Poisson approximates to a 
 Gaussian, and then of course the difference Ispot-Iback is also approximately 
 Gaussian.
 
 But I think that doesn't affect your argument.
 
 Cheers
 
 -- Ian 


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Douglas Theobald
On Jun 20, 2013, at 2:13 PM, Ian Tickle ianj...@gmail.com wrote:

 Douglas, I think you are missing the point that estimation of the parameters 
 of the proper Bayesian statistical model (i.e. the Wilson prior) in order to 
 perform the integration in the manner you are suggesting requires knowledge 
 of the already integrated intensities!  

Well, that's true, but that's how FW do it.  They allow for negative integrated 
intensities.  I'm arguing that we should not do that, since true intensities 
are positive, and so any estimate of them should also be positive.  

Examples are always better than words, so here goes (and I apologize for the 
length):

The current way of doing things is summarized by Ed's equation: 
Ispot-Iback=Iobs.  Here Ispot is the # of counts in the spot (the area 
encompassing the predicted reflection), and Iback is # of counts in the 
background (usu. some area around the spot).  Our job is to estimate the true 
intensity Itrue.  Ed and others argue that Iobs is a reasonable estimate of 
Itrue, but I say it isn't because Itrue can never be negative, whereas Iobs 
can.  

Now where does the Ispot-Iback=Iobs equation come from?  It implicitly assumes 
that both Iobs and Iback come from a Gaussian distribution, in which Iobs and 
Iback can have negative values.  Here's the implicit data model:

Ispot = Iobs + Iback

There is an Itrue, to which we add some Gaussian noise and randomly generate an 
Iobs.  To that is added some background noise, Iback, which is also randomly 
generated from a Gaussian with a true mean of Ibtrue.  This gives us the 
Ispot, the measured intensity in our spot.  Given this data model, Ispot will 
also have a Gaussian distribution, with mean equal to the sum of Itrue + 
Ibtrue.  From the properties of Gaussians, then, the ML estimate of Itrue will 
be Ispot-Iback, or Iobs.  

Now maybe you disagree with that Gaussian data model.  If so, welcome to my 
POV.  

There are better models, ones that don't give Ispot-Iback as our best estimate 
of Itrue.

Here is a simple example that incorporates our knowledge that Itrue cannot be 
negative (this example is primarily for illustrating the point, it's not 
exactly what I would recommend).  Instead of using Gaussians, we will use Gamma 
distributions, which cannot be negative.  

We assume Iobs is distributed according to a Gamma(Itrue,1).  The mean of this 
distribution is Itrue.  (The Maxwell-Boltzmann energy distribution is also a 
gamma, just for comparison).  

We also assume that the noise is exponential (a special case of the gamma), 
Gamma(1,1).  The mean of this distribution is 1.  (You could imagine that 
you've normalized Ispot relative to its background --- again, just for ease of 
calculation).  

We still assume that Ispot = Iobs + Iback.  Then, Ispot will also have a gamma 
distribution, Gamma(Itrue+1,1).  The mean of the Ispot distribution, as you 
might expect, is Itrue+1.  

Now we measure Ispot.  Given Ispot, the ML estimate of Itrue is:

InvDiGamma[ln(Ispot)]-1   if Ispot0.561
or
0   if Ispot0.561

Note, the ML estimate is no longer Iobs, and the ML estimate cannot be 
negative.  InvDiGamma is the inverse Digamma function --- a bit unusual, but 
easily calculated (actually no weirder than the exponential or logarithm, its a 
relative of factorial and the gamma function).  Not something the Braggs 
would've used, but hey, we've got iPhones now.  We can also estimate the SD of 
of our estimate, but I won't bore you with the equation.

A few examples: 

Ispot   ML Itrue   SD
-      --
0.5 0  0.78
0.6 0.04   0.80
0.8 0.25   0.91
0.9 0.36   0.97
1.0 0.46   1.0
1.5 0.97   1.2
2.0 1.48   1.4
3.0 2.49   1.7
5.0 4.49   2.2
10.09.50   3.2
20.019.5   4.5
100 99.5   10

Note that the first four entries in the table are the case when IspotIback.  
No negative estimates.  You'd get qualitatively similar results if you assume 
Poisson for Iback and Ispot.  

To sum up --- the equation Ispot-Iback=Iobs is unphysical because it is founded 
on unphysical assumptions.  If you make better physical assumptions (i.e., 
Itrue cannot be negative), you end up with different estimates for Itrue.  


 I suppose we could iterate, i.e. assume an approximate prior, integrate, 
 calculate a better prior, re-do the integration with the new prior and so on 
 (hoping of course that the whole process converges), but I think most people 
 would regard that as overkill.  Also dealing with the issue of averaging 
 estimates of intensities that no longer have a Gaussian error distribution, 
 and also crucially outlier rejection, would require some rethinking of the 
 algorithms. The question is would it make any difference in the end compared 
 with the 'post-correction' we're doing now?
 
 Cheers
 
 -- Ian
 
 
 On 20 June 2013 18:14, Douglas Theobald dtheob...@brandeis.edu wrote:
 I still don't see how you get a negative 

Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Douglas Theobald
On Jun 21, 2013, at 2:48 PM, Ed Pozharski epozh...@umaryland.edu wrote:

 Douglas,
 Observed intensities are the best estimates that we can come up with in an 
 experiment.
 I also agree with this, and this is the clincher.  You are arguing that 
 Ispot-Iback=Iobs is the best estimate we can come up with.  I claim that is 
 absurd.  How are you quantifying best?  Usually we have some sort of 
 discrepancy measure between true and estimate, like RMSD, mean absolute 
 distance, log distance, or somesuch.  Here is the important point --- by any 
 measure of discrepancy you care to use, the person who estimates Iobs as 0 
 when IbackIspot will *always*, in *every case*, beat the person who 
 estimates Iobs with a negative value.   This is an indisputable fact.
 
 First off, you may find it useful to avoid such words as absurd and 
 indisputable fact.  I know political correctness may be sometimes overrated, 
 but if you actually plan to have meaningful discussion, let's assume that 
 everyone responding to your posts is just trying to help figure this out.

I apologize for offending and using the strong words --- my intention was not 
to offend.  This is just how I talk when brainstorming with my colleagues 
around a blackboard, but of course then you can see that I smile when I say it. 
 

 To address your point, you are right that J=0 is closer to true intensity 
 then a negative value.  The problem is that we are not after a single 
 intensity, but rather all of them, as they all contribute to electron density 
 reconstruction.  If you replace negative Iobs with E(J), you would 
 systematically inflate the averages, which may turn problematic in some 
 cases.  

So, I get the point.  But even then, using any reasonable criterion, the whole 
estimated dataset will be closer to the true data if you set all negative 
intensity estimates to 0.  

 It is probably better to stick with raw intensities and construct 
 theoretical predictions properly to account for their properties.
 
 What I was trying to tell you is that observed intensities is what we get 
 from experiment.  

But they are not what you get from the detector.  The detector spits out a 
positive value for what's inside the spot.  It is we, as human agents, who 
later manipulate and massage that data value by subtracting the background 
estimate.  A value that has been subjected to a crude background subtraction is 
not the raw experimental value.  It has been modified, and there must be some 
logic to why we massage the data in that particular manner.  I agree, of 
course, that the background should be accounted for somehow.  But why just 
subtract it away?  There are other ways to massage the data --- see my other 
post to Ian.  My argument is that however we massage the experimentally 
observed value should be physically informed, and allowing negative intensity 
estimates violates the basic physics.  

[snip]

 These observed intensities can be negative because while their true 
 underlying value is positive, random errorsmay result in IbackIspot.  
 There is absolutely nothing unphysical here.
 Yes there is.  The only way you can get a negative estimate is to make 
 unphysical assumptions.  Namely, the estimate Ispot-Iback=Iobs assumes that 
 both the true value of I and the background noise come from a Gaussian 
 distribution that is allowed to have negative values.  Both of those 
 assumptions are unphysical.
 
 See, I have a problem with this.  Both common sense and laws of physics 
 dictate that number of photons hitting spot on a detector is a positive 
 number.  There is no law of physics that dictates that under no circumstances 
 there could be IspotIback.  

That's not what I'm saying.  Sure, Ispot can be less than Iback randomly.  That 
does not mean we have to estimate the detected intensity as negative, after 
accounting for background.

 Yes, E(Ispot)=E(Iback).  Yes, E(Ispot-Iback)=0.  But P(Ispot-Iback=0)0, 
 and therefore experimental sampling of Ispot-Iback is bound to occasionally 
 produce negative values.  What law of physics is broken when for a given 
 reflection total number of photons in spot pixels is less that total number 
 of photons in equal number of pixels in the surrounding background mask?
 
 Cheers,
 
 Ed.
 
 -- 
 Oh, suddenly throwing a giraffe into a volcano to make water is crazy?
Julian, King of Lemurs


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Douglas Theobald
On Jun 21, 2013, at 2:52 PM, James Holton jmhol...@lbl.gov wrote:

 Yes, but the DIFFERENCE between two Poisson-distributed values can be 
 negative.  This is, unfortunately, what you get when you subtract the 
 background out from under a spot.  Perhaps this is the source of confusion 
 here?

Maybe, but if you assume Poisson background and intensities, the ML estimate 
when background  measured intensity is not negative, nor is it the difference 
Ispot-Iback.  The ML estimate is 0.  (With a finite non-zero SD, smaller SD the 
smaller the Ispot/Iback ratio).

 On Fri, Jun 21, 2013 at 11:34 AM, Douglas Theobald dtheob...@brandeis.edu 
 wrote:
 I kinda think we're saying the same thing, sort of.
 
 You don't like the Gaussian assumption, and neither do I.  If you make the 
 reasonable Poisson assumptions, then you don't get the Ispot-Iback=Iobs for 
 the best estimate of Itrue.  Except as an approximation for large values, but 
 we are talking about the case when IbackIspot, where the Gaussian 
 approximation to the Poisson no longer holds.  The sum of two Poisson 
 variates is also Poisson, which also can never be negative, unlike the 
 Gaussian.
 
 So I reiterate: the Ispot-Iback=Iobs equation assumes Gaussians and hence 
 negativity.  The Ispot-Iback=Iobs does not follow from a Poisson assumption.
 
 
 On Jun 21, 2013, at 1:13 PM, Ian Tickle ianj...@gmail.com wrote:
 
  On 21 June 2013 17:10, Douglas Theobald dtheob...@brandeis.edu wrote:
  Yes there is.  The only way you can get a negative estimate is to make 
  unphysical assumptions.  Namely, the estimate Ispot-Iback=Iobs assumes 
  that both the true value of I and the background noise come from a 
  Gaussian distribution that is allowed to have negative values.  Both of 
  those assumptions are unphysical.
 
  Actually that's not correct: Ispot and Iback are both assumed to come from 
  a _Poisson_ distribution which by definition is zero for negative values of 
  its argument (you can't have a negative number of photons), so are _not_ 
  allowed to have negative values.  For large values of the argument (in fact 
  the approximation is pretty good even for x ~ 10) a Poisson approximates to 
  a Gaussian, and then of course the difference Ispot-Iback is also 
  approximately Gaussian.
 
  But I think that doesn't affect your argument.
 
  Cheers
 
  -- Ian
 


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Phil
However you decide to argue the point, you must consider _all_ the observations 
of a reflection (replicates and symmetry related) together when you infer Itrue 
or F etc, otherwise you will bias the result even more. Thus you cannot 
(easily) do it during integration

Phil

Sent from my iPad

On 21 Jun 2013, at 20:30, Douglas Theobald dtheob...@brandeis.edu wrote:

 On Jun 21, 2013, at 2:48 PM, Ed Pozharski epozh...@umaryland.edu wrote:
 
 Douglas,
 Observed intensities are the best estimates that we can come up with in an 
 experiment.
 I also agree with this, and this is the clincher.  You are arguing that 
 Ispot-Iback=Iobs is the best estimate we can come up with.  I claim that is 
 absurd.  How are you quantifying best?  Usually we have some sort of 
 discrepancy measure between true and estimate, like RMSD, mean absolute 
 distance, log distance, or somesuch.  Here is the important point --- by 
 any measure of discrepancy you care to use, the person who estimates Iobs 
 as 0 when IbackIspot will *always*, in *every case*, beat the person who 
 estimates Iobs with a negative value.   This is an indisputable fact.
 
 First off, you may find it useful to avoid such words as absurd and 
 indisputable fact.  I know political correctness may be sometimes overrated, 
 but if you actually plan to have meaningful discussion, let's assume that 
 everyone responding to your posts is just trying to help figure this out.
 
 I apologize for offending and using the strong words --- my intention was not 
 to offend.  This is just how I talk when brainstorming with my colleagues 
 around a blackboard, but of course then you can see that I smile when I say 
 it.  
 
 To address your point, you are right that J=0 is closer to true intensity 
 then a negative value.  The problem is that we are not after a single 
 intensity, but rather all of them, as they all contribute to electron 
 density reconstruction.  If you replace negative Iobs with E(J), you would 
 systematically inflate the averages, which may turn problematic in some 
 cases.  
 
 So, I get the point.  But even then, using any reasonable criterion, the 
 whole estimated dataset will be closer to the true data if you set all 
 negative intensity estimates to 0.  
 
 It is probably better to stick with raw intensities and construct 
 theoretical predictions properly to account for their properties.
 
 What I was trying to tell you is that observed intensities is what we get 
 from experiment.  
 
 But they are not what you get from the detector.  The detector spits out a 
 positive value for what's inside the spot.  It is we, as human agents, who 
 later manipulate and massage that data value by subtracting the background 
 estimate.  A value that has been subjected to a crude background subtraction 
 is not the raw experimental value.  It has been modified, and there must be 
 some logic to why we massage the data in that particular manner.  I agree, of 
 course, that the background should be accounted for somehow.  But why just 
 subtract it away?  There are other ways to massage the data --- see my other 
 post to Ian.  My argument is that however we massage the experimentally 
 observed value should be physically informed, and allowing negative intensity 
 estimates violates the basic physics.  
 
 [snip]
 
 These observed intensities can be negative because while their true 
 underlying value is positive, random errorsmay result in IbackIspot.  
 There is absolutely nothing unphysical here.
 Yes there is.  The only way you can get a negative estimate is to make 
 unphysical assumptions.  Namely, the estimate Ispot-Iback=Iobs assumes that 
 both the true value of I and the background noise come from a Gaussian 
 distribution that is allowed to have negative values.  Both of those 
 assumptions are unphysical.
 
 See, I have a problem with this.  Both common sense and laws of physics 
 dictate that number of photons hitting spot on a detector is a positive 
 number.  There is no law of physics that dictates that under no 
 circumstances there could be IspotIback.  
 
 That's not what I'm saying.  Sure, Ispot can be less than Iback randomly.  
 That does not mean we have to estimate the detected intensity as negative, 
 after accounting for background.
 
 Yes, E(Ispot)=E(Iback).  Yes, E(Ispot-Iback)=0.  But P(Ispot-Iback=0)0, 
 and therefore experimental sampling of Ispot-Iback is bound to occasionally 
 produce negative values.  What law of physics is broken when for a given 
 reflection total number of photons in spot pixels is less that total number 
 of photons in equal number of pixels in the surrounding background mask?
 
 Cheers,
 
 Ed.
 
 -- 
 Oh, suddenly throwing a giraffe into a volcano to make water is crazy?
   Julian, King of Lemurs


Re: [ccp4bb] ctruncate bug?

2013-06-21 Thread Terwilliger, Thomas C
I hope I am not duplicating too much of this fascinating discussion with these 
comments:  perhaps the main reason there is confusion about what to do is that 
neither F nor I is really the most suitable thing to use in refinement.  As 
pointed out several times in different ways, we don't measure F or I, we only 
measure counts on a detector.  As a convenience, we process our diffraction 
images to estimate I or F and their uncertainties and model these uncertainties 
as simple functions (e.g., a Gaussian).  There is no need in principle to do 
that, and if we were to refine instead against the raw image data these issues 
about positivity would disappear and our structures might even be a little 
better.

Our standard procedure is to estimate F or I from counts on the detector, then 
to use these estimates of F or I in refinement.  This is not so easy to do 
right because F or I contain many terms coming from many pixels and it is hard 
to model their statistics in detail.  Further, attempts we make to estimate 
either F or I as physically plausible values (e.g., using the fact that they 
are not negative) will generally be biased (the values after correction will 
generally be systematically low or systematically high, as is true for the 
French and Wilson correction and as would be true for the truncation of I at 
zero or above).

Randy's method for intensity refinement is an improvement because the 
statistics are treated more fully than just using an estimate of F or I and 
assuming its uncertainty has a simple distribution.  So why not avoid all the 
problems with modeling the statistics of processed data and instead refine 
against the raw data.  From the structural model you calculate F, from F and a 
detailed model of the experiment (the same model that is currently used in data 
processing) you calculate the counts expected on each pixel. Then you calculate 
the likelihood of the data given your models of the structure and of the 
experiment.  This would have lots of benefits because it would allow improved 
descriptions of the experiment (decay, absorption, detector sensitivity, 
diffuse scattering and other background on the images,on and on) that 
could lead to more accurate structures in the end.  Of course there are some 
minor issues about putting all this in computer memory for refinement

-Tom T

From: CCP4 bulletin board [CCP4BB@JISCMAIL.AC.UK] on behalf of Phil 
[p...@mrc-lmb.cam.ac.uk]
Sent: Friday, June 21, 2013 2:50 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: Re: [ccp4bb] ctruncate bug?

However you decide to argue the point, you must consider _all_ the observations 
of a reflection (replicates and symmetry related) together when you infer Itrue 
or F etc, otherwise you will bias the result even more. Thus you cannot 
(easily) do it during integration

Phil

Sent from my iPad

On 21 Jun 2013, at 20:30, Douglas Theobald dtheob...@brandeis.edu wrote:

 On Jun 21, 2013, at 2:48 PM, Ed Pozharski epozh...@umaryland.edu wrote:

 Douglas,
 Observed intensities are the best estimates that we can come up with in an 
 experiment.
 I also agree with this, and this is the clincher.  You are arguing that 
 Ispot-Iback=Iobs is the best estimate we can come up with.  I claim that is 
 absurd.  How are you quantifying best?  Usually we have some sort of 
 discrepancy measure between true and estimate, like RMSD, mean absolute 
 distance, log distance, or somesuch.  Here is the important point --- by 
 any measure of discrepancy you care to use, the person who estimates Iobs 
 as 0 when IbackIspot will *always*, in *every case*, beat the person who 
 estimates Iobs with a negative value.   This is an indisputable fact.

 First off, you may find it useful to avoid such words as absurd and 
 indisputable fact.  I know political correctness may be sometimes overrated, 
 but if you actually plan to have meaningful discussion, let's assume that 
 everyone responding to your posts is just trying to help figure this out.

 I apologize for offending and using the strong words --- my intention was not 
 to offend.  This is just how I talk when brainstorming with my colleagues 
 around a blackboard, but of course then you can see that I smile when I say 
 it.

 To address your point, you are right that J=0 is closer to true intensity 
 then a negative value.  The problem is that we are not after a single 
 intensity, but rather all of them, as they all contribute to electron 
 density reconstruction.  If you replace negative Iobs with E(J), you would 
 systematically inflate the averages, which may turn problematic in some 
 cases.

 So, I get the point.  But even then, using any reasonable criterion, the 
 whole estimated dataset will be closer to the true data if you set all 
 negative intensity estimates to 0.

 It is probably better to stick with raw intensities and construct 
 theoretical predictions properly to account for their properties.

 What I was trying 

[ccp4bb] peptide planarity restraints for modified residues

2013-06-21 Thread wtempel
Dear CCP4ers/REFMACers,
the C-terminal peptide bond of a modified lysyl, M3L, appears less planar
than other peptide bonds in the same structure. What is the currently
preferred method of restraining such planarity? I have currently not added
any link record in the header of my coordinates file. Would that be
required?
Thank you,
Wolfram Tempel


Re: [ccp4bb] str solving problem

2013-06-21 Thread Francis E. Reyes
Pramod:

[1] Please refrain from posting excessively large (1MB) attachments to the 
ccp4bb. Either use a compression technique or use another means of transmitting 
large files to your recipients without spamming the entire group. 

[2] Your predictions are not overlaying well with the spots .  Well it may be 
overlaying well with a subset of spots (in this case, the strongest spots that 
obey the symmetry *you* chose), but your are leaving out a large number of 
spots. 

As I said before, your diffraction pattern is exhibiting strong reflections 
with weak reflections at a given resolution.  It looks as if there are multiple 
lattices probably due to bad crystal morphology ( a cracked crystal, twinning, 
pseudosymmetry) or multiple crystals in the beam, the list goes on 

I've experienced your case many, many times you'll get a good MR solution, 
but then the refinement becomes 'stuck'. 

I think the idiom... 'garbage in, garbage out'  applies here..

[3]  
 
 I have given several efforts to the XDS but its giving error  data image of 
 particular no. does not exist (initially it was saying 11th image than i 
 change image range then it says 21st and so on) kindly check my data 
 collection profile and XDS.INP file in attachment'
 

Sounds like a file name problem. 

[4]
 Or if the crystal is big enough, you could try shooting it in different areas 
 and 'searching' for a better spot to collect data.
 Or 'grow a better crystal'.
 
 raising the crystals and struggle is on the peak...
 


There are a number of ccp4bb postings about crystal reproducibility or crystal 
diffraction improvement. Search the archives for these. 

F



-
Francis E. Reyes PhD
215 UCB
University of Colorado at Boulder