I think this is an excellent reference to work from, when dealing
with the issue of URIs in RDF generation & processing.
As I have always seen it (this is admittedly a the view of an RDF
naif), DOIs and LSIDs both seek to fulfill the role one would expect
to be played by URIs in the STM literature and biomedical object
domains, respectively.
For those who had the chance to read the paper, I would specifically
point to the discussion of the CrossRef & OpenURL projects. Both
relate to how you resolve a DOI tied to very practical Use Cases.
One is very much focussed on the commercial issue of dis-ambiguating
which journals a given library system has a subscription to. The
goal for this (OpenURL - http://www.exlibrisgroup.com/
sfx_openurl.htm) was to create an infrastructure for publishers (and
aggregators) to resolve this issue in a way that is transparent for
the user as they click on a link to an article (HTML or PDF). The
SFX system many may be familiar with seeing in the search engines
hosted by their library systems.
CrossRef (http://www.crossref.org/) is more designed to address the
core issue on the article of how you both maintain stable pointers to
inherently unstable online resources, and also providing a URI-like
generic resource pointer which can be resolved to the actual resource
the moment a reader clicks on the reference in a bibliography.
CrossRef is much more focussed on dealing with the many different
scenarios related to the latter task and coming up with a way that -
again - from the user's point transparently gets them to the correct
resource. CrossRef the organization seems to pitch themselves as the
service designed to de-reference DOIs - which obviously makes the
work they've done very relevant to this conversation.
Clearly, both of these issues are ones the folks from BMC & PLoS can
give us some very practical insight into.
The one major project related to the topic in the article that the
author seemed to neglect is the Internet Archive (http://
archive.org). This is a long standing project (in Internet time,
anyway - going back to 1996). They trawl the entire public net and
backup it up as often as possible. They have massive, robotics-based
tape drive systems working round the clock. The original archive
took almost a year to crawl the entire "public" net (it still takes
about 2 months to cover everything, though there is a lot of effort
they've put into to categorizing the rate at which contents changes -
with content having a more rapid turnover getting more frequent
observation). After the end of the 1996 presidential campaign -
within weeks, the only source for historians to analyze use of the
web in the election was the Internet Archive. This has continued to
be the case for many research projects focussed on the use or and
evolution of web content. The IA has set up to donate periodic dumps
to the Library of Congress. They're technology has greatly improved
over the years (they now have PetaByte storage racks and much a much
more mature software layer). Though IA doesn't solve the issue of
the "hidden"/dynamic web all that much better than the other search
engines (which is the space in which most if not all scientific
literature lives), they clearly provide a great utility to difficult
to manage mess the HTML web often devolves into. IA is also
intimately involved in the discussions in the library science
community on this issue of digital reference resolution and
archiving, as well as the critical issue of FIXING IP law - very much
aligned with the efforts of the Creative Commons. Some CC folks are
also directly involved with IA. IA has set up a specific group to
help researchers make better use of IA content (http://
www.archive.org/web/researcher/researcher.php).
Cheers,
Bill
On Jun 18, 2006, at 12:52 PM, John Madden wrote:
Alan et al,
Wow, great topic. I'll need to get my thoughts together on this.
Meanwhile, operationally what a uri "means" is clearly related to
the question of its (non)persistence. I recently found a wonderful
historical review of this topic from the point of view of a library
scientist. The group might enjoy it:
http://www.aallnet.org/products/pub_llj_v97n04/2005-42.pdf
John
On Jun 18, 2006, at 1234, Alan Ruttenberg wrote:
[It was on this list: http://lists.w3.org/Archives/Public/public-
semweb-lifesci/2006Jun/0149]
-Alan'
On Jun 18, 2006, at 12:20 PM, John Madden wrote:
I can't locate the beginning of this thread. Did the discussion
start on another list?
Thanks.
John
On Jun 17, 2006, at 1708, Eric Neumann wrote:
This is a very useful and important discussion thread, and I
would like to see others on the list to contribute their
thoughts/concerns as well.
May I ask all the contributors to include HTML links to any
acronyms they reference (e.g., NAPTR)? This will make it easier
for the rest of us to catch up quickly, and to eventually
collect the approaches out there into a comprehensive list of
viable implementations.
thanks,
Eric
--- Sean Martin <[EMAIL PROTECTED]> wrote:
> MW>
> MW> I believe this SRV-redirection behaviour is part
> of the LSID spec, and
> MW> we use it for all of the BioMOBY LSIDs...
> MW>
>
> It also uses NAPTR's as described in IETF RFC's
> 3401->3405 to traverse the
> URN namespace, allowing the dereferencing process to
> bridge the gap that
> separates authority name strings from service
> locations. From what I
> recall, the URN specs specifically do not permit
> names and locations to be
> confounded.
>
> Kindest regards, Sean
>
> --
> Sean Martin
> IBM Corp.
>
> [EMAIL PROTECTED] wrote on
> 06/16/2006 12:59:12 PM:
>
> >
> > On Fri, 2006-06-16 at 10:41 -0400, Alan Ruttenberg
> wrote:
> >
> > > something, but as far as I can see, the only
> authority related to
> > > namespaces in URLs is the DNS, and while there
> is the SRV field which
> > > might be used to direct someone to information
> about the namespace, I
> > > don't know whether anyone does.
> >
> >
> > I believe this SRV-redirection behaviour is part
> of the LSID spec, and
> > we use it for all of the BioMOBY LSIDs...
> >
> > M
> >
> >
> >
> > --
> >
> > --
> > Mark Wilkinson
> > Asst. Professor, Dept. of Medical Genetics
> > University of British Columbia
> > PI in Bioinformatics, iCAPTURE Centre
> > St. Paul's Hospital, Rm. 166, 1081 Burrard St.
> > Vancouver, BC, V6Z 1Y6
> > tel: 604 682 2344 x62129
> > fax: 604 806 9274
> >
> > "For most of this century we have viewed
> communications as a conduit,
> > a pipe between physical locations on the
> planet.
> > What's happened now is that the conduit has become
> so big and
> interesting
> > that communication has become more than a
> conduit,
> > it has become a destination in its own
> right..."
> >
> > Paul Saffo - Director, Institute
> for the Future
> >
> >
>
Eric Neumann, PhD
co-chair, W3C Healthcare and Life Sciences,
and Senior Director Product Strategy
Teranode Corporation
83 South King Street, Suite 800
Seattle, WA 98104
+1 (781)856-9132
www.teranode.com
Bill Bug
Senior Analyst/Ontological Engineer
Laboratory for Bioimaging & Anatomical Informatics
www.neuroterrain.org
Department of Neurobiology & Anatomy
Drexel University College of Medicine
2900 Queen Lane
Philadelphia, PA 19129
215 991 8430 (ph)
610 457 0443 (mobile)
215 843 9367 (fax)
Please Note: I now have a new email - [EMAIL PROTECTED]
This email and any accompanying attachments are confidential.
This information is intended solely for the use of the individual
to whom it is addressed. Any review, disclosure, copying,
distribution, or use of this email communication by others is strictly
prohibited. If you are not the intended recipient please notify us
immediately by returning this message to the sender and delete
all copies. Thank you for your cooperation.