Re: [ccp4bb] metal ion coordination

2014-04-18 Thread Jrh
Dear Faisal,
When scrutinising such distances do be aware of the possibility of false 
precision in the estimates;  see eg http://dx.doi.org/10.1107/S2052252513031485
Best wishes,
John


Prof John R Helliwell DSc
 
 

On 17 Apr 2014, at 21:13, Faisal Tarique faisaltari...@gmail.com wrote:

 Dear all
 
 Can anybody please explain what is the classical metal ion coordination for 
 Mg2+, Ca+ and Na+ with Oxygen atom and the average distance with these metal 
 ions..does the distance vary with the type of metal ion and its coordination 
 with oxygen atom..what is the best way to identify the correct metal ion in 
 the electron density in the vicinity of negatively charged molecule mostly 
 oxygen containing molecule..In one of my paper the reviewer has asked me to 
 check whether the octahedrally coordinated Mg+ is  Ca+ ion..and similarly 
 raised doubt about the identity of the Na+ ion as well..his argument was 
 based on metal ion to oxygen distance..I am attaching the figure with this 
 mail..i request you to please shed some light on this area and help me in 
 clearing some doubts regarding this. 
 
 -- 
 Regards
 
 Faisal
 School of Life Sciences
 JNU
 
 Fig3.tif


Re: [ccp4bb] Rmerge, Rmeas, I/sigma, Mn(I/sd)

2014-04-18 Thread Kay Diederichs
Hi Roberto,

for my taste the answers so far have not mentioned (or I did not understand 
them) that there is a distinction between indicators of the precision of merged 
data, and those for the precision of unmerged data.

It is not possible to directly compare (or equate) indicators of one category 
with those of the other category. This would be like comparing apples to 
oranges, and is, in my experience, the biggest source of confusion in 
crystallographic statistics, _and_ not clearly explained in writing. The only 
way to do such a comparison numerically is to include a factor of sqrt(n) where 
n is the multiplicity - the ratio of the number of observations and the number 
of unique reflections.

Indicators of precision of unmerged data are: [Rsym=Rmerge (which should be 
deprecated),] Rmeas and the I/sigma(I) of individual observations, as given 
in the first long table in XDS' CORRECT.LP which is fine-grained in resolution. 
Aimless also has this, but it is _not_ the quantity labeled Mn(I)/sd(Mn(I)).

Indicators of precision of merged data are: CC1/2, Rpim and the I/sigma(I) of 
unique reflections after averaging, as given in the repeated (by DATA_RANGE) 
tables in XDS' CORRECT.LP. In aimless, the average signal/noise after averaging 
symmetry-related observations  I/σ(I) is labelled Mn(I)/sd(Mn(I)).

For both categories, there is not much difference in I/sigma(I) and 
I/sigma(I); in particular at high resolution, these are becoming equal.

Thus, 
Rmerge ≈ 0.8/I/σ(I)
can only hold for unmerged data (i.e. observations), not for merged data 
(unique reflections, after averaging over symmetry-related observations).

HTH,

Kay



On Wed, 16 Apr 2014 17:09:28 +0200, Roberto Battistutta 
roberto.battistu...@unipd.it wrote:

Hi,
in the Rupp book the following relation is reported (on pag 415):
Rmerge ≈ 0.8/I/σ(I)
referring to a relation of the linear merging R-value with the 
signal-to-noise ratio.

in a 2006 CCP4bb, Manfred Weiss reported:
Rrim (or Rmeas) = 0.8*sd(i)/I

so, 

First question: is it Rmerge or Rmeas (Rrim) that we are dealing with?

Second question: at the denominator (of the Rupp way to write), it is the 
aimless Mn(I/sd) (usually indicated as the real signal-to-noise ratio) or the 
aimless I/sigma (very different from Mn(I/sd) with high redundancy)?

Thank you very much for the clarification.

Best,
Roberto.

Roberto Battistutta


[ccp4bb] ligand occupancy

2014-04-18 Thread Monica Mittal
Dear all

I have a protein which is dimer having one ligand binding site in each
monomer. I refined the crystal structure with ligand in both sites finally.
I refined will full occupancy of 1 for ligands (same in both). But now i
want to see is there any difference in the occupancy of both ligands in
ligand binding sites of monomer A and B. Is there any way i can get any
information about the occupancy of ligands in two monomers like one is
binding more tightly than another so that i can get an idea about their
differential binding contacts also.

Thank you very much in advance.

Regards
Monica


Re: [ccp4bb] Validation reports for all X-ray structures in the PDB

2014-04-18 Thread Gerard DVD Kleywegt

Bonjour Philippe,

The reports are based on the recommendations of the wwPDB X-ray validation 
task force - see http://www.wwpdb.org/workshop/2011/index.html and the report 
published by this task force at 
http://www.cell.com/structure/abstract/S0969-2126%2811%2900285-1


Of course, if you have specific suggestions for improvements, we'd be 
interested to hear from you - feel free to mail them to 
validat...@mail.wwpdb.org


A stand-alone server already exists - see 
http://www.wwpdb.org/validation-servers.html - hope this turns your Good 
Friday into a Great Friday :-)


Best wishes,

--Gerard





On Wed, 16 Apr 2014, Philippe BENAS wrote:


Dear Gerard CD/DVD/Blue Ray (;-) ),

Yes, these new reports are great although they should/will improve over time.

Another point that would be also really helpful would be to have the 
opportunity to run the associated scripts either locally or on a remote server 
from the PDB, prior to the submission itself. Hence they should provide strong 
guidelines the crystallographers during their rebuilding/refinement stages. I 
know Phenix for instance has already some of these tools, but not all.

Could the PDB provide something in that way for the everyday use of a poor 
X-ray crystallographer ?

Best regards,
Philippe

 


Philippe BENAS, Ph.D.
X-ray diffraction and computing facilities manager

Laboratoire de Cristallographie et RMN Biologiques, UMR 8015 CNRS

E-mails: philippe.be...@parisdescartes.fr, philippe_be...@yahoo.fr
URLs: http://lcrbw.pharmacie.univ-paris5.fr/ , 
http://lcrbw.pharmacie.univ-paris5.fr/spip.php?article18








De : Gerard DVD Kleywegt ger...@xray.bmc.uu.se
? : CCP4BB@JISCMAIL.AC.UK 
Envoy? le : Mercredi 16 avril 2014 19h01

Objet : [ccp4bb] Validation reports for all X-ray structures in the PDB


Hi all,

You may not have noticed, but 19 March 2014 was VR Day - the day that new 
style wwPDB validation reports for all X-ray structures were made publicly 
available - see http://www.wwpdb.org/news/news_2014.html#18-March-2014


The validation-related files for individual X-ray PDB entries can be accessed 
through the web sites and ftp sites of the various wwPDB partners. Speaking 
for PDBe, if you go to the summary page of an X-ray PDB entry, for instance:


          http://pdbe.org/1cbs

you will see the percentile sliders displayed in the PDBportfolio widget 
(http://pdbe.org/portfolio) on the right of the page. (Clicking the big white 
arrow will start a slideshow of images related to this entry.) The legend of 
the percentile-slider plot contains a direct link to the validation report (as 
a PDF file; in this case 
http://www.ebi.ac.uk/pdbe/entry-files/1cbs_validation.pdf).


If you are not yet familiar with these new style validation reports, have a 
look here: http://www.wwpdb.org/validation-reports.html - in particular the 
user guide may be of interest: http://www.wwpdb.org/ValidationPDFNotes.html


If you want to download the full report (which lists all outliers for many of 
the validation criteria, instead of just the worst 5 or the first 5), or a 
graphic image of the percentile-slider plot, or an XML file with all 
validation data in machine-readable form, go to the downloads page of any 
X-ray PDB entry, either through clicking the Downloads link in the menu on 
the left, or directly by going to a URL of the form:


          http://pdbe.org/1cbs/downloads

The section labelled Validation of the table provides the relevant links.

Note that sites that include PDBportfolio in their pages now automatically 
display the percentile-slider plot and download link as well! To see this in 
action, go to the EDS page (if any) of your favourite X-ray PDB entry, e.g.:


          http://eds.bmc.uu.se/cgi-bin/eds/uusfs?pdbCode=1cbs

Please send any comments, questions or suggestions on the new style validation 
reports to validat...@mail.wwpdb.org


Questions about PDBe-specific pages and services can be sent to 
pdbeh...@ebi.ac.uk


--Gerard

---
Gerard J. Kleywegt, PDBe, EMBL-EBI, Hinxton, UK
ger...@ebi.ac.uk . pdbe.org
Secretary: Pauline Haslam  pdbe_ad...@ebi.ac.uk



Best wishes,

--Gerard

**
   Gerard J. Kleywegt

  http://xray.bmc.uu.se/gerard   mailto:ger...@xray.bmc.uu.se
**
   The opinions in this message are fictional.  Any similarity
   to actual opinions, living or dead, is purely coincidental.
**
   Little known gastromathematical curiosity: let z be the
   radius and a the thickness of a pizza. Then the volume
of that pizza is equal to pi*z*z*a !
**

Re: [ccp4bb] Rmerge, Rmeas, I/sigma, Mn(I/sd)

2014-04-18 Thread Bernhard Rupp
[There is] a distinction between indicators of the precision of merged data, 
and those for the precision of unmerged data.

Let's take a step back - definitions matter:

(i) We have multiple observations of the same, already integrated h: the 
'unmerged' data - most important data set which SHOULD BE deposited and rarely 
is.
(ii) Now we weighted average those multiple instances of the same h, sans 
symmetry: 'merged' data - still useful to keep, particularly if one gets the 
metric symmetry/PG wrong 
(iii) Now we merge symmetry related data (generally keeping Friedels apart): 
'unique' data
(iv) both (ii) and (iii) are instances of 'merged' data.

Is that correct? If so, let’s continue the thread (there is more to come...) or 
adjust the definitions.

Best, BR


Re: [ccp4bb] Rmerge, Rmeas, I/sigma, Mn(I/sd)

2014-04-18 Thread Kay Diederichs
On Fri, 18 Apr 2014 12:33:30 +0200, Bernhard Rupp hofkristall...@gmail.com 
wrote:

[There is] a distinction between indicators of the precision of merged data, 
and those for the precision of unmerged data.

Let's take a step back - definitions matter:

(i) We have multiple observations of the same, already integrated h: the 
'unmerged' data - most important data set which SHOULD BE deposited and 
rarely is.

yes, fully agree.

(ii) Now we weighted average those multiple instances of the same h, sans 
symmetry: 'merged' data - still useful to keep, particularly if one gets the 
metric symmetry/PG wrong 
(iii) Now we merge symmetry related data (generally keeping Friedels apart): 
'unique' data
(iv) both (ii) and (iii) are instances of 'merged' data.

I don't quite understand the difference between (ii) and (iii). As soon as you 
take the weighted average, you merge the data, because you create one single 
estimate of the intensity I (and sigma(I)) of a unique reflection from several 
symmetry-related observations of that unique reflection. So, to me, 'taking the 
weighted average' and 'merging' are different words for the same procedure.

best,

Kay


Is that correct? If so, let’s continue the thread (there is more to come...) 
or adjust the definitions.

Best, BR


[ccp4bb] observed criterion sigma

2014-04-18 Thread Faisal Tarique
Dear all

I request you please tell me what is the value of Observed criterion sigma
(F) and Observed criterion sigma (I) for any data processed by imosflm and
scala ?


-- 
Regards

Faisal
School of Life Sciences
JNU


Re: [ccp4bb] Rmerge, Rmeas, I/sigma, Mn(I/sd)

2014-04-18 Thread Bernhard Rupp
(i) We have multiple observations of the same, already integrated h: the 
'unmerged' data - most important data set which SHOULD BE deposited and 
rarely is.

yes, fully agree.

Perfect.

 I don't quite understand the difference between (ii) and (iii). As soon as 
 you take the weighted average, you merge the data, because you create one 
 single estimate of the intensity I (and sigma(I)) of a unique reflection from 
 several symmetry-related observations of that unique reflection. So, to me, 
 'taking the weighted average' and 'merging' are different words for the same 
 procedure.

There is indeed no distinction between (ii) and (iii) form the merging point of 
view, I just wanted to point out the difference between just 'merged' and 
'unique' data.

We return to Kay's original post.

 Indicators of precision of *unmerged* data are: [Rsym=Rmerge (which should 
 be deprecated),] - yes, and I want to iterate:

Here is already where the notational confusion starts - 'unmerged' data (i) 
obviously contain multiple observations of a single reflection h, then how can 
any measure of their quality  logically be called an Rsym (there is no sym in a 
single reflection) ?
A Rsym is per definition of sym a measure producing merged data of type (iii) , 
although it is also AN Rmerge.
Historically this seems to come from the original Arndt definition (c.f. 
Diederichs  Karplus 1997) but it is illogical in the above context. The 
original definition of Rmerge also includes already the summation over a set of 
binned hkls.

Along the same line, that the quality indicator for 'unmerged' data is their 
'merging'  R is also illogical - they have the same quality before they are 
merged.  Not only as a statistic, even as a term Rmerge should be buried (i.e. 
finally BECOME a statistic).

One primary statistic that is valid universally are the i/sigI. None of these 
Rs are robust statistics. Rmeas is an asymptotic target (penalizing you for 
small N) , and Rpim some form of standard error of the mean (rewarding you for 
large N). Choose wisely

Because of its statistical defensibility (clear definition and the association 
with a confidence or significance level) CC1/2 is interesting and perhaps the 
only measure in addition to  primary I/sigi needed - with the juicy bonus of 
having via CC*/work/free a traceable relation to the model quality. That, as 
Kay has pointed out in his papers, is more than you can say about any of these 
Rs. /anti_R_flame

 Thus, Rmerge � 0.8/I/s(I)
can only hold for unmerged data (i.e. observations), not for merged data 
(unique reflections, after averaging over symmetry-related observations).

True. I see that. Which is the reason why it is still close for the low 
redundancy data historically observed, but I think this will change rapidly 
with the PADs  high redundancy becoming standard - another reason to bury 
Rmerge  associates. 

Happy Easter, BR




best,

Kay


Is that correct? If so, let�s continue the thread (there is more to come...) 
or adjust the definitions.

Best, BR


Re: [ccp4bb] ligand occupancy

2014-04-18 Thread George Kontopidis
Hi Monica,

 

Calculate the mean B-factor of all atoms that making interactions with each 
ligand in monomer A and B.

Use those means values as  B-factors for each ligand respectively.

Adjust manually the occupancies, in order  the B-factors for each ligand  to 
stay after refinement close to the above values.

 

In order to calculate occupancies more precisely,  it would help to have the 
un-liganded structure and thus the  location of  the water molecules in each 
binding site.

If you have the above information you could refine water molecules and ligands 
simultaneously in the binding site and get accurate refined occupancies. 

 

George

 

From: CCP4 bulletin board [mailto:CCP4BB@JISCMAIL.AC.UK] On Behalf Of Monica 
Mittal
Sent: Friday, April 18, 2014 1:04 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: [ccp4bb] ligand occupancy

 

Dear all

 

I have a protein which is dimer having one ligand binding site in each monomer. 
I refined the crystal structure with ligand in both sites finally. I refined 
will full occupancy of 1 for ligands (same in both). But now i want to see is 
there any difference in the occupancy of both ligands in ligand binding sites 
of monomer A and B. Is there any way i can get any information about the 
occupancy of ligands in two monomers like one is binding more tightly than 
another so that i can get an idea about their differential binding contacts 
also. 

 

Thank you very much in advance.

 

Regards

Monica 



Re: [ccp4bb] ligand occupancy

2014-04-18 Thread Boaz Shaanan



Hi Monica,

You can refine the ligand occupancy in refmac as explained here:
http://www2.mrc-lmb.cam.ac.uk/groups/murshudov/content/refmac/refmac_keywords.html

or in phenix, whichever program you're using.

 Cheers,

 Boaz




Boaz Shaanan, Ph.D.

Dept. of Life Sciences 
Ben-Gurion University of the Negev 
Beer-Sheva 84105 
Israel 
 
E-mail: bshaa...@bgu.ac.il
Phone: 972-8-647-2220Skype: boaz.shaanan 
Fax: 972-8-647-2992 or 972-8-646-1710










From: CCP4 bulletin board [CCP4BB@JISCMAIL.AC.UK] on behalf of Monica Mittal [monica.mitta...@gmail.com]
Sent: Friday, April 18, 2014 1:03 PM
To: CCP4BB@JISCMAIL.AC.UK
Subject: [ccp4bb] ligand occupancy




Dear all


I have a protein which is dimer having one ligand binding site in each monomer. I refined the crystal structure with ligand in both sites finally. I refined will full occupancy of 1 for ligands (same in both). But now i want to see is there any difference
 in the occupancy of both ligands in ligand binding sites of monomer A and B. Is there any way i can get any information about the occupancy of ligands in two monomers like one is binding more tightly than another so that i can get an idea about their differential
 binding contacts also.


Thank you very much in advance.


Regards
Monica








Re: [ccp4bb] Rmerge, Rmeas, I/sigma, Mn(I/sd)

2014-04-18 Thread Edward A. Berry

Roberto Battistutta wrote:

Hi,
in the Rupp book the following relation is reported (on pag 415):
Rmerge ≈ 0.8/I/σ(I)
referring to a relation of the linear merging R-value with the signal-to-noise 
ratio.

in a 2006 CCP4bb, Manfred Weiss reported:
Rrim (or Rmeas) = 0.8*sd(i)/I



Bernhard Rupp wrote:

 0.8*sd(i)/I = 0.8/(I/sd(i))

---

Yes, but in this context it is worth pointing out that
I/σ(I) != 1/σ(I)/I

especially if there is a wide range in values of I/σ(I),
which would be narrower but still significant in the outer shells.

ave(100, 10, 1) = 37
ave(0.01, 0.1, 1) = .37 = 1/(2.7) != 1/37

So while of course Rupp's equation is correct, if we try to apply it to average 
values,
which we have to do to compare with Rmerge, it is no longer correct, so even at high 
multiplicity the two equations quoted in original post seem incompatible.

(Unless I'm badly confused again)
eab


Re: [ccp4bb] Rmerge, Rmeas, I/sigma, Mn(I/sd)

2014-04-18 Thread Kay Diederichs
Hi Ed,

your example seems to be designed to show that the average of reciprocal values 
is not the same as the reciprocal of an average value? If that is what you are 
alluding to, then please not that the (relatively narrow) Wilson distribution 
of intensities has the effect of making the relation I/σ(I) ~ 1/σ(I)/I work 
fairly well in practice.

The relation Rmeas ≈ 0.8/I/σ(I) (where I refers to the intensity of unmerged, 
individual observations) is obviously not an exact one ... rather it depends on 
how the σ(I) are calculated (the error model) and some other things. But it 
should _not_ depend on the multiplicity, and it should hold fairly well at high 
resolution.

Kay

On Fri, 18 Apr 2014 10:12:10 -0400, Edward A. Berry ber...@upstate.edu wrote:

Roberto Battistutta wrote:
 Hi,
 in the Rupp book the following relation is reported (on pag 415):
 Rmerge ≈ 0.8/I/σ(I)
 referring to a relation of the linear merging R-value with the 
 signal-to-noise ratio.

 in a 2006 CCP4bb, Manfred Weiss reported:
 Rrim (or Rmeas) = 0.8*sd(i)/I

 
Bernhard Rupp wrote:
 
  0.8*sd(i)/I = 0.8/(I/sd(i))
 
---

Yes, but in this context it is worth pointing out that
   I/σ(I) != 1/σ(I)/I

especially if there is a wide range in values of I/σ(I),
which would be narrower but still significant in the outer shells.

ave(100, 10, 1) = 37
ave(0.01, 0.1, 1) = .37 = 1/(2.7) != 1/37

So while of course Rupp's equation is correct, if we try to apply it to 
average values,
which we have to do to compare with Rmerge, it is no longer correct, so even 
at high
multiplicity the two equations quoted in original post seem incompatible.
(Unless I'm badly confused again)
eab


Re: [ccp4bb] crystallographic confusion

2014-04-18 Thread Lavie, Arnon
Dear Kay.

Arguably, the resolution of a structure is the most important number to
look at; it is definitely the first to be examined, and often the only one
examined by non-structural biologists.

Since this number conveys so much concerning the quality/reliability of
the the structure, it is not surprising that we need to get this one
parameter right.

Let us examine a hypothetical situation, in which a data set at the
2.2-2.0 resolution shell has 20% completeness. Is this a 2.0 A resolution
structure?  While you make a sound argument that including that data may
result in a better refined model (more observations, more restraints), I
would not consider that model the same quality as one refined against a
data set that has 90% completeness at that resolution shell.

As I see it, there are two issues here: one, is whether to include such
data in refinement?  I am not sure if low completeness (especially if not
random) can be detrimental to a correct model, but I will let other weigh
in on that.

The second question is where to declare the resolution limit of a
particular data set?  To my mind, here high completeness (the term high
needs a precise definition) better describes the true resolution limit of
the diffraction, and with this what I can conclude about the quality of
the refined model.

My two cents.

Arnon Lavie

On Fri, April 18, 2014 6:51 pm, Kay Diederichs wrote:
 Hi everybody,

 since we seem to have a little Easter discussion about crystallographic
 statistics anyway, I would like to bring up one more topic.

 A recent email sent to me said: Another referee complained that the
 completeness in that bin was too low at 85% - my answer was that I
 consider the referee's assertion as indicating a (unfortunately not
 untypical case of) severe statistical confusion. Actually, there is no
 reason at all to discard a resolution shell just because it is not
 complete, and what would be a cutoff, if there were one? What
 constitutes too low?

 The benefit of including also incomplete resolution shells is that every
 reflection constitutes a restraint in refinement (and thus reduces
 overfitting), and contributes its little bit of detail to the electron
 density map. Some people may be mis-lead by a wrong understanding of the
 cats and ducks examples by Kevin Cowtan: omitting further data from
 maps makes Fourier ripples/artifacts worse, not better.

 The unfortunate consequence of the referee's opinion (and its
 enforcement and implementation in papers) is that the structures that
 result from the enforced re-refinement against truncated data are
 _worse_ than the original data that included the incomplete resolution
 shells.

 So could we as a community please abandon this inappropriate and
 un-justified practice - of course after proper discussion here?

 Kay




Re: [ccp4bb] crystallographic confusion

2014-04-18 Thread Dale Tronrud
-BEGIN PGP SIGNED MESSAGE-
Hash: SHA1


   I see no problem with saying that the model was refined against every
spot on the detector that the data reduction program said was observed
(and I realize there is argument about this) but declare that the
resolution of the model is a number based on the traditional criteria.
This solution allows for the best possible model to be constructed and
the buyer is still allowed to make quality judgements the same way
as always.

Dale Tronrud

On 4/18/2014 5:22 PM, Lavie, Arnon wrote:
 Dear Kay.
 
 Arguably, the resolution of a structure is the most important
 number to look at; it is definitely the first to be examined, and
 often the only one examined by non-structural biologists.
 
 Since this number conveys so much concerning the
 quality/reliability of the the structure, it is not surprising that
 we need to get this one parameter right.
 
 Let us examine a hypothetical situation, in which a data set at
 the 2.2-2.0 resolution shell has 20% completeness. Is this a 2.0 A
 resolution structure?  While you make a sound argument that
 including that data may result in a better refined model (more
 observations, more restraints), I would not consider that model the
 same quality as one refined against a data set that has 90%
 completeness at that resolution shell.
 
 As I see it, there are two issues here: one, is whether to include
 such data in refinement?  I am not sure if low completeness
 (especially if not random) can be detrimental to a correct model,
 but I will let other weigh in on that.
 
 The second question is where to declare the resolution limit of a 
 particular data set?  To my mind, here high completeness (the term
 high needs a precise definition) better describes the true
 resolution limit of the diffraction, and with this what I can
 conclude about the quality of the refined model.
 
 My two cents.
 
 Arnon Lavie
 
 On Fri, April 18, 2014 6:51 pm, Kay Diederichs wrote:
 Hi everybody,
 
 since we seem to have a little Easter discussion about
 crystallographic statistics anyway, I would like to bring up one
 more topic.
 
 A recent email sent to me said: Another referee complained that
 the completeness in that bin was too low at 85% - my answer was
 that I consider the referee's assertion as indicating a
 (unfortunately not untypical case of) severe statistical
 confusion. Actually, there is no reason at all to discard a
 resolution shell just because it is not complete, and what would
 be a cutoff, if there were one? What constitutes too low?
 
 The benefit of including also incomplete resolution shells is
 that every reflection constitutes a restraint in refinement (and
 thus reduces overfitting), and contributes its little bit of
 detail to the electron density map. Some people may be mis-lead
 by a wrong understanding of the cats and ducks examples by
 Kevin Cowtan: omitting further data from maps makes Fourier
 ripples/artifacts worse, not better.
 
 The unfortunate consequence of the referee's opinion (and its 
 enforcement and implementation in papers) is that the structures
 that result from the enforced re-refinement against truncated
 data are _worse_ than the original data that included the
 incomplete resolution shells.
 
 So could we as a community please abandon this inappropriate and 
 un-justified practice - of course after proper discussion here?
 
 Kay
 
 
-BEGIN PGP SIGNATURE-
Version: GnuPG v2.0.22 (MingW32)
Comment: Using GnuPG with Thunderbird - http://www.enigmail.net/

iEYEARECAAYFAlNRz14ACgkQU5C0gGfAG138HwCfYbUXb5MgQvC/8iCftiuuP1pn
H0AAn24ej2FSBxbNbndjnHoJ/xAKCitK
=Xh7C
-END PGP SIGNATURE-


Re: [ccp4bb] crystallographic confusion

2014-04-18 Thread William G. Scott
Dear Arnon et al:

My understanding of the Shannon/Nyquist sampling theorem is admittedly 
extremely rudimentary, but I think aliasing can result if an arbitrary 
brick-wall resolution cut-off to the data is applied.

So let’s say there are real data are to 2.0 Å resolution. Applying the 2.2 Å 
cutoff will result in aliasing artifacts in the electron density map 
corresponding to an outer shell reciprocal space volume equal but opposite to 
the cut out data.

The alternative, which is to process and keep all the measured reflections, 
should help to minimize this.  An effective resolution can be calculated and 
quoted.  This becomes a significant problem with nucleic acids and their 
complexes, which often diffract with significant anisotropy.

The idea that 85% completeness in the outer shell should dictate its rejection 
seems rather surprising and arbitrary. The aliasing artifacts in that case 
would probably be significant.  The map image quality, after all, is what we 
are after, not beautiful Table 1 statistics.

Bill


William G. Scott
Professor
Department of Chemistry and Biochemistry
and The Center for the Molecular Biology of RNA
University of California at Santa Cruz
Santa Cruz, California 95064
USA
http://scottlab.ucsc.edu/scottlab/



On Apr 18, 2014, at 5:22 PM, Lavie, Arnon la...@uic.edu wrote:

 Dear Kay.
 
 Arguably, the resolution of a structure is the most important number to
 look at; it is definitely the first to be examined, and often the only one
 examined by non-structural biologists.
 
 Since this number conveys so much concerning the quality/reliability of
 the the structure, it is not surprising that we need to get this one
 parameter right.
 
 Let us examine a hypothetical situation, in which a data set at the
 2.2-2.0 resolution shell has 20% completeness. Is this a 2.0 A resolution
 structure?  While you make a sound argument that including that data may
 result in a better refined model (more observations, more restraints), I
 would not consider that model the same quality as one refined against a
 data set that has 90% completeness at that resolution shell.
 
 As I see it, there are two issues here: one, is whether to include such
 data in refinement?  I am not sure if low completeness (especially if not
 random) can be detrimental to a correct model, but I will let other weigh
 in on that.
 
 The second question is where to declare the resolution limit of a
 particular data set?  To my mind, here high completeness (the term high
 needs a precise definition) better describes the true resolution limit of
 the diffraction, and with this what I can conclude about the quality of
 the refined model.
 
 My two cents.
 
 Arnon Lavie
 
 On Fri, April 18, 2014 6:51 pm, Kay Diederichs wrote:
 Hi everybody,
 
 since we seem to have a little Easter discussion about crystallographic
 statistics anyway, I would like to bring up one more topic.
 
 A recent email sent to me said: Another referee complained that the
 completeness in that bin was too low at 85% - my answer was that I
 consider the referee's assertion as indicating a (unfortunately not
 untypical case of) severe statistical confusion. Actually, there is no
 reason at all to discard a resolution shell just because it is not
 complete, and what would be a cutoff, if there were one? What
 constitutes too low?
 
 The benefit of including also incomplete resolution shells is that every
 reflection constitutes a restraint in refinement (and thus reduces
 overfitting), and contributes its little bit of detail to the electron
 density map. Some people may be mis-lead by a wrong understanding of the
 cats and ducks examples by Kevin Cowtan: omitting further data from
 maps makes Fourier ripples/artifacts worse, not better.
 
 The unfortunate consequence of the referee's opinion (and its
 enforcement and implementation in papers) is that the structures that
 result from the enforced re-refinement against truncated data are
 _worse_ than the original data that included the incomplete resolution
 shells.
 
 So could we as a community please abandon this inappropriate and
 un-justified practice - of course after proper discussion here?
 
 Kay