Re: [CF-metadata] More background on the burned area (Jonathan Gregory)

2011-07-27 Thread Schultz, Martin
Jonathan, Kevin,

I don't think that it is necessary to further qualify burned_area. If you 
do an internet search for this term you always come up with hits related to 
wildfire which would suggest that there is little ambiguity in this term. I 
propose to add the vegetation fire relationship in the definition of the term 
but keep the term short and concise. Should there ever be a conflict we can 
still create an alias or re-define. (Just imagine you start arguing about 
air_temperature: is this the temperature of air inside a box or a house? ...)

Cheers,

Martin



Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDirig Dr. Karl Eugen Huthmacher
Geschaeftsfuehrung: Prof. Dr. Achim Bachem (Vorsitzender),
Dr. Ulrich Krafft (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt,
Prof. Dr. Sebastian M. Schmidt


___
CF-metadata mailing list
CF-metadata@cgd.ucar.edu
http://mailman.cgd.ucar.edu/mailman/listinfo/cf-metadata


Re: [CF-metadata] new TEOS-10 standard names

2011-07-27 Thread Paul.Durack
To follow on from Trevor's comments, most CMIP3-grade ocean models were 
initialised, at least in their pre-coupled ocean state from a variant of the 
World Ocean Atlas (WOA). Listed on the CMIP3 Climate Model Documentation, 
References, and Links webpages 
(http://www-pcmdi.llnl.gov/ipcc/model_documentation/ipcc_model_documentation.php)
 these suggest that initialisation fields were sourced mostly from:
LEV82: Levitus (1982);
LEV94: Levitus et al., (1994), Levitus and Boyer (1994);
LEV98: Conkright et al.,(1999);
LEV: unspecified Levitus database

Just selecting one model from the ~24 or so which comprise the CMIP3 suite, 
CNRM-CM3, under section 5 - Simulation Details indicates:
Picntrl/Run_1:
This preindustrial control simulation was initialized from a  coupled 
simulation of  a previous version of CNRM coupled model initialized an ocean at 
rest with temperature and salinity profiles specified from Levitus (1982) 
climatology, integrated for 30 years with a relaxation of surface temperature 
to the monthly mean Reynolds climatology for 1950.

Subsequent runs are often then initialised from a previously completed spinup..

So I think it is fair to say that most model fields are then a modelled version 
of practical salinity..

Cheers,

P

From: McDougall, Trevor (CMAR, Hobart)
Sent: Wednesday, 27 July 2011 6:27 AM
To: Jonathan Gregory
Cc: ngalbra...@whoi.edu; CF-metadata@cgd.ucar.edu; Durack, Paul (CMAR, Hobart); 
Barker, Paul (CMAR, Hobart); rainer.feis...@io-warnemuende.de; r...@eos.ubc.ca; 
b...@noc.soton.ac.uk; stephen.griff...@noaa.gov
Subject: RE: [CF-metadata] new TEOS-10 standard names

A couple of quick comments following on from Jonathan's post.

(1) I know of at least 6 pre-TEOS-10 expressions for density used in models, 
with authors like
 Fofonoff  Millard,
 Cox,
 Wright,
 Jackett  McDougall,
 McDougall et al.
 Jackett et al.
and they are all written in terms of Practical Salinity.  I know of none used 
in ocean models
that use any other type of salinity (until TEOS-10 has come along).  So we can 
safely say that
ocean and climate models have had their sea water equations of state written in 
terms of Practical Salinity.

(2)  The fact that a model variable drifts should not be a reason to use a 
different name for that variable.  For example, we do not change the name 
potential temperature to something else just because model temperatures are 
not perfect and they drift.

Trevor


-Original Message-
From: Jonathan Gregory [mailto:j.m.greg...@reading.ac.uk]
Sent: Wednesday, 27 July 2011 1:45 AM
To: McDougall, Trevor (CMAR, Hobart)
Cc: ngalbra...@whoi.edu; CF-metadata@cgd.ucar.edu; Durack, Paul (CMAR, Hobart); 
Barker, Paul (CMAR, Hobart); rainer.feis...@io-warnemuende.de; r...@eos.ubc.ca; 
b...@noc.soton.ac.uk; stephen.griff...@noaa.gov
Subject: Re: [CF-metadata] new TEOS-10 standard names

Dear all

I understand the need to be clear, with new standard names, which observational
quantity is being collected in future. I do not agree, however, that we should
make the plain salinity name an alias for something more precise. This is
partly because that might change the meaning of existing data, possibly
incorrectly as Trevor points out. Partly it is also because I think it is
quite possible that models, perhaps idealised, may be used in which it would
not be meaningful to be more precise than just salinity.

Trevor argues that existing ocean models use practical salinity because (a)
they are initialised with observations of that and (b) they assume so in their
equation of state. I don't think (a) is necessarily so. In some cases, they
might not be initialised with observations, for instance in idealised
investigations of spin-up. Even when initialised from obs,
they will almost certainly drift to a less realistic state. I don't know enough
about it to be sure about (b). Unless we could be certain this is always the
case, I think plain salinity should be retained for possible use in models.
However, we could certainly recommend that models should use one of the new
more precise terms if definitely appropriate. This recommendation could be
included in the standard_name definition of plain salinity.

I understand the existing standard name sea_water_temperature to mean in-situ
temperature, as it does for air temperature. This could be stated in the
definition.

The purpose of standard names themselves is not to prescribe or recommend what
quantities people should store in netCDF files. It is to allow them to describe
with sufficient precision the quantities they have chosen to store, in order
to make it possible to decide which quantities from different datasets should
be regarded as comparable.

Standard names are all in lower case, regardless of what case is used in
ordinary writing. This is for simplicity in matching strings. Case-sensitive
matching would inevitably trip people up and cause a 

Re: [CF-metadata] the need to store lat/lon coordinates in a CF metadata compliant netCDF file

2011-07-27 Thread David Blodgett
 Without the grid_mapping, the lat and lon still make sense in the common case
 (and original CF case) of GCM data, and in many other cases, the intended
 usage of the data does not require precision about the figure of the Earth.
 Although this metadata could be valuable if it can be defined, I think it 
 would
 be too onerous to require it.

I hope to present on this very issue at AGU. The problem we see with ambiguous 
definition of datums is a cascade of non-recognition of datums through 
processing algorithms and in the output of some processes that generate very 
detailed data. 

The prime example is downscaled climate data. Because the climate modelers 
involved generally consider lat/lon to be a lowest common denominator, the 
datum used to geolocate historical data (like rain gages) is neglected. What 
results is, in our case, a 1/8deg (12km) grid with no datum. This is 
unacceptable. As at this resolution, the errors in a wrong assumption of datum 
for the grid can cause very substantial (a full grid cell or more) geolocation 
errors.

If the CF community intends to consume any ground based data, then datums must 
be preserved from ingest of ground based forcing throughout data storage and 
processing. This is fundamental information that is required for ALL data 
comparison operations.

I would argue that CF compliance should require this information. This puts the 
requirement to make metadata assumptions on data publishers/producers rather 
than data consumers. It is unacceptable to have different data consumers making 
different assumptions of geolocation on the same data. 

Off soapbox.

Dave Blodgett
Center for Integrated Data Analytics (CIDA)
USGS WI Water Science Center
8505 Research Way Middleton WI 53562
608-821-3899 | 608-628-5855 (cell)
http://cida.usgs.gov


On Jul 26, 2011, at 5:24 AM, Jonathan Gregory wrote:

 Dear all
 
 For datasets which are intended for analysis by end-users I think it would be
 undesirable to remove the requirement of providing explicit lat and lon
 coords even if a grid_mapping is provided. I think it is unrealistic to expect
 all software which someone might use to analyse netCDF files to be able to
 recognise and act upon all possible values of the CF grid_mapping attribute,
 and without the lat and lon information the user would have a problem. If the
 issue is storage space in the file I think the much better choice is to store
 the explicit coordinates in another file, by extending the CF convention to
 allow datasets to be distributed over several linked files, as gridspec does
 for example.
 
 Steve appears to suggest that grid_mapping is required in some circumstances,
 but I don't think it is at present. However, the text Steve quotes may not be
 quite right:
 
   /When the coordinate variables for a horizontal grid are not
   longitude and latitude,*_it is required that the true latitude and
   longitude coordinates be supplied_* via the coordinates attribute/.
 
 The text should make it clear that this requirement applies when the data has 
 a
 geolocated horizontal grid. It doesn't necessarily apply to idealised cases.
 We could clarify this with a defect ticket.
 
 Cheers
 
 Jonathan
 ___
 CF-metadata mailing list
 CF-metadata@cgd.ucar.edu
 http://mailman.cgd.ucar.edu/mailman/listinfo/cf-metadata

___
CF-metadata mailing list
CF-metadata@cgd.ucar.edu
http://mailman.cgd.ucar.edu/mailman/listinfo/cf-metadata


Re: [CF-metadata] the need to store lat/lon coordinates in a CF metadata compliant netCDF file

2011-07-27 Thread David Blodgett
Jonathan and Karl,

I don't disagree with the GCM examples or lack of model resolution being talked 
about here, I just don't think that is the only argument.

There are cases, like downscaled climate projections, reanalysis products, or 
radar indicated rainfall, where a datum is critical to interpreting the data 
accurately. My concern is that the justified disregard for datums at course 
resolution (and lack of requirement for it in the spec) fosters a lack of 
awareness that datums become critically important at finer scales. 

Jonathan, you have a point that an analyst should be free to make their own 
assumptions. However, I would prefer that they know they are making 
assumptions. It would be helpful if the CF specification paid some tribute to 
this issue rather than treating lat/lon coordinates as base-equivalent 
coordinates.

Are there any arguments against CF recommending a standard datum assumption 
when intersecting data without a datum specified with data that does have a 
datum specified?

Cheers,

Dave 

On Jul 27, 2011, at 12:14 PM, Karl Taylor wrote:

 Hi all,
 
 (I think the horse still shows a few signs of life)
 
 what I'm arguing is that if the results the scientists are using come from a 
 GCM, then although the two scientists got differences, those differences (no 
 matter how large) should not be considered significant in terms of their 
 reliability (as opposed to in some statistical sense).  Just as one wouldn't 
 rely on a climate model to predict global mean temperature to within 0.001 K 
 (and surely it would be silly and perhaps misleading to report temperatures 
 to this precision) one shouldn't expect to pin down the *location* of the 
 grid cell temperature reported by a GCM to a point given with higher 
 precision than the spacing of the grid cells (at least when comparing with 
 observations).
 
 I will grant you that at some time far in the future, it is possible our 
 models' resolution and accuracy will have improved to the point that we might 
 have to alter the precision with which we report the locations of their 
 output values, but we're not there yet.
 
 Best regards,
 Karl
 
 
 
 On 7/27/11 9:23 AM, David Blodgett wrote:
 
 Not to beat a dead horse, but this issue has been a huge stumbling block in 
 our work to integrate data and software across the climate and geographic 
 communities.
 
 The argument here is: Since CF data is usually so coarse and low precision 
 complete geolocation metadata should not be required. 
 
 An example of why this matters: Two scientists take the same downscaled 
 climate data that doesn't have a datum specification and import it into 
 their application. One application assumes one datum, the other assumes 
 another datum. Scientist 1's results differ from scientist 2's results. In 
 situations where their are steep gradients in downscaled data, these 
 differences may be substantial. 
 
 One solution would be to adopt a default datum for data lacking datum 
 definition. So, given a file that uses lat/lon and claims to follow CF spec, 
 a scientist could follow the specs guidance on what datum to assume. Without 
 this type guidance or a requirement to include the information, lat/lon 
 without a datum amounts to providing any other value without units.
 
 Dave
 
 On Jul 27, 2011, at 10:59 AM, Karl Taylor wrote:
 
 Dear all,
 
 another view:
 
 Can't remember *all* the issues here, but certainly reporting the latitude 
 and longitude points for GCM grids without further precision (e.g., 
 information on the figure of the Earth) is sufficient for any comparison 
 with observations.  Only certain (usually prescribed) conditions at the 
 earth's surface (e.g., surface height) coming from a GCM should be trusted 
 at the individual grid point scale, and no sub-grid scale information is 
 directly available from the GCM (normally).  So, even if a station data is 
 near the boundary of a GCM's grid-cell, it should hardly matter which of 
 the grid cells it straddles you compare it to.  The GCM sort of gives you a 
 grid cell average value that applies to some region in the vicinity of the  
 cell.  So, it doesn't matter where you think it is precisely located.
 
 Down-scaled output from the GCM will be at higher resolution, but again 
 since the original data doesn't apply at a point but for a general region 
 (usually quite a bit larger than 12 km, and even if it weren't we wouldn't 
 believe stuff going on at that scale), so where the cell is exactly located 
 again doesn't matter.
 
 best regards,
 Karl
 
 
 On 7/27/11 4:38 AM, David Blodgett wrote:
 
 Without the grid_mapping, the lat and lon still make sense in the common 
 case
 (and original CF case) of GCM data, and in many other cases, the intended
 usage of the data does not require precision about the figure of the 
 Earth.
 Although this metadata could be valuable if it can be defined, I think it 
 would
 be too onerous to require it.
 
 I hope to present on this very