Re: [CODE4LIB] Math or the other math?

2013-02-27 Thread Michael Hopwood
>From a physics point of view, "computer science" looks about 50% discrete 
>math, and 50% engineering (since computers, fancy as they may be, are simply 
>machines, and have specific physical constraints that it may be helpful to 
>understand).

Actual coding nowadays, I assume, may sometimes actually have a lot in common 
with language arts (UK readers: we don't study "language arts". Sorry...) but 
it's worth noting that logic is the common factor between language arts and 
math.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Al 
Matthews
Sent: 27 February 2013 14:28
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Math or the other math?

+1 mostly to the thread

Programming seems to me -- just me here -- stratified like any other 
profession, in particular by access or lack of access to computer science 
within software dev.

There are other factors. But computer science seems now heavily invested in 
math.

--
Al Matthews

Software Developer, Digital Services Unit Atlanta University Center, Robert W. 
Woodruff Library
email: amatth...@auctr.edu; office: 1 404 978 2057





On 2/27/13 9:17 AM, "Michael Hopwood"  wrote:

>You mean discrete mathematics?
>
>http://en.wikipedia.org/wiki/Discrete_mathematics
>
>I always kicked myself for not taking that course at high school (UK 
>readers, I mean secondary school) but at least I picked up the basics 
>during my physics MSci (a lot of physics these days is coding).
>
>Cheers,
>
>m
>
>-Original Message-
>From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
>Ken Irwin
>Sent: 27 February 2013 13:53
>To: CODE4LIB@LISTSERV.ND.EDU
>Subject: Re: [CODE4LIB] back to minorities question, seeking guidance
>
>What both Kelly and David say is true here:
>David: programming needs math, not arithmetic.
>Kelly: computers are good at arithmetic on their own.
>
>To which I'll add: the related skill that I see as necessary here is 
>quantitative reasoning - not the crunching of numbers but the correct 
>assembly of the formulae, articulating the systematization of the problem.
>
>What I'm less certain of is what sort of training tend to lead to that 
>sort of conceptual skill.
>
>Ken
>
>
>
>On Feb 27, 2013, at 8:44 AM, "David Faler"  wrote:
>
>> I think math is essential, but what they teach in schools these days 
>> isn't math.  It's arithmetic.  Some intro philosophy courses teach 
>> math.  I'll stop before I start ranting.
>>
>> On Wed, Feb 27, 2013 at 12:04 AM, Kelly Lucas 
>>wrote:
>>
>>>
>>>
>>>
>>> On Sat, Feb 23, 2013 at 2:57 AM, Thomas Krichel 
>>> 
>>> wrote:
>>>
>>>>  Wilhelmina Randtke writes
>>>>
>>>>> Pretty much the whole entire entry level programming class for the
>>>> average
>>>>> class covers using code to do things that you can do much more 
>>>>> easily without code.
>>>>
>>>>  Probably it was the wrong course. I think coding should start with 
>>>> building web pages. A calculator can't do that.
>>>>
>>>>  Cheers,
>>>>
>>>>  Thomas Krichelhttp://openlib.org/home/krichel
>>>>  http://authorprofile.org/pkr1
>>>>   skype: thomaskrichel
>>>>
>>>
>>>
>>>
>>> --
>>> Kelly R. Lucas
>>> Senior Developer
>>> Isovera, Inc.
>>> klu...@isovera.com
>>> http://www.isovera.com
>>> http://drupal.org/user/271780
>>> twitter: @bp1101
>>>


-
**
The contents of this email and any attachments are confidential.
They are intended for the named recipient(s) only.
If you have received this email in error please notify the system manager or  
the sender immediately and do not disclose the contents to anyone or make 
copies.

** IronMail scanned this email for viruses, vandals and malicious content. **
**


[CODE4LIB] Math or the other math?

2013-02-27 Thread Michael Hopwood
You mean discrete mathematics?

http://en.wikipedia.org/wiki/Discrete_mathematics

I always kicked myself for not taking that course at high school (UK readers, I 
mean secondary school) but at least I picked up the basics during my physics 
MSci (a lot of physics these days is coding).

Cheers,

m

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Ken 
Irwin
Sent: 27 February 2013 13:53
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] back to minorities question, seeking guidance

What both Kelly and David say is true here:
David: programming needs math, not arithmetic. 
Kelly: computers are good at arithmetic on their own. 

To which I'll add: the related skill that I see as necessary here is 
quantitative reasoning - not the crunching of numbers but the correct assembly 
of the formulae, articulating the systematization of the problem.

What I'm less certain of is what sort of training tend to lead to that sort of 
conceptual skill. 

Ken



On Feb 27, 2013, at 8:44 AM, "David Faler"  wrote:

> I think math is essential, but what they teach in schools these days 
> isn't math.  It's arithmetic.  Some intro philosophy courses teach 
> math.  I'll stop before I start ranting.
> 
> On Wed, Feb 27, 2013 at 12:04 AM, Kelly Lucas  wrote:
> 
>> 
>> 
>> 
>> On Sat, Feb 23, 2013 at 2:57 AM, Thomas Krichel 
>> wrote:
>> 
>>>  Wilhelmina Randtke writes
>>> 
 Pretty much the whole entire entry level programming class for the
>>> average
 class covers using code to do things that you can do much more 
 easily without code.
>>> 
>>>  Probably it was the wrong course. I think coding should start with  
>>> building web pages. A calculator can't do that.
>>> 
>>>  Cheers,
>>> 
>>>  Thomas Krichelhttp://openlib.org/home/krichel
>>>  http://authorprofile.org/pkr1
>>>   skype: thomaskrichel
>>> 
>> 
>> 
>> 
>> --
>> Kelly R. Lucas
>> Senior Developer
>> Isovera, Inc.
>> klu...@isovera.com
>> http://www.isovera.com
>> http://drupal.org/user/271780
>> twitter: @bp1101
>> 


Re: [CODE4LIB] XMP Metadata to tab-delemited file

2013-01-12 Thread Michael Hopwood
Hello Misty, Andrea,

Thanks for your discussion on extracting useful XMP.

I have been trying to do this in the context of the 
www.linkedheritage.eu<http://www.linkedheritage.eu> project.

I got as far as producing XMP RDF/XML files but the problem then remains; how 
to usefully manage these via XSLT transforms?

The problem is that XMP uses an RDF syntax that comes in many flavours and 
doesn't result in a predictable set of xpaths to apply the XSLT to.

I wonder if you got any solutions beyond this point? A de-customising XSLT to 
extract a "flat" XMP flavour before applying the XSLT of choice?

Best,

Michael Hopwood
Linked Heritage Project Lead
EDItEUR
United House, North  Road
London N7 9DP
UK

Tel: +44 20 7503 6418
Mob: +44 7811 591036
Skype: michael.hopwood.editeur
http://www.linkedheritage.org/
http://editeur.org/

The information contained in this e-mail is confidential and may be privileged. 
It is intended for the addressee only. If you are not the intended recipient, 
please inform the sender and delete this e-mail immediately. The contents of 
this e-mail must not be disclosed or copied without the sender's consent. We 
cannot accept any responsibility for viruses, so please scan all attachments. 
The statements and opinions expressed in this message are those of the author 
and do not necessarily reflect those of the company.

EDItEUR Limited is a company limited by guarantee, registered in England no 
2994705. Registered Office:
United House, North Road, London N7 9DP, United Kingdom


Re: [CODE4LIB] What is a "coder"?

2012-11-29 Thread Michael Hopwood
Thought process of an information professional:

1- I need to open a file in my program
2 - I'll ask someone who can do it for me...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
Friscia, Michael
Sent: 29 November 2012 13:46
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] What is a "coder"?

Thought process of a coder:
1- I need to open a file in my program
2- ok, I'll import IO into my application and read the definition
3- i create methods and functions around the definition and open my file Total 
time to deliver code: 5 mins

Thought process of a non-coder
1- I need to open a file in my program
2- I open up a web browser and go to google
3- search "open file in java"
4- copy/paste the code I find
5- can't figure out why it doesn't work, go back to step 3 and try a different 
person's code
6- really stuck, contemplates changing the programming language
7- runs some searches on easier programming languages
8- goes back to Google and tries new search terms and gets different results
9- finally get it working
10- remove all comments from the copy/paste code so it looks like I wrote it.
Total time to deliver code: 5 hours


___
Michael Friscia
Manager, Digital Library & Programming Services 

Yale University Library
(203) 432-1856


-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Mark A. 
Matienzo
Sent: Wednesday, November 28, 2012 10:03 PM
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] What is a "coder"?

Some discussion (both on-list and otherwise) has referred to "coders,"
and some discussion as such has raised the question whether
"non-coders" are welcome at code4lib.

What's a coder? I'm not trying to be difficult - I want to make
code4lib as inclusive as possible.

Mark A. Matienzo 
Digital Archivist, Manuscripts and Archives, Yale University Library
Technical Architect, ArchivesSpace


Re: [CODE4LIB] OpenURL linking but from the content provider's point of view

2012-11-21 Thread Michael Hopwood
> The same principle should apply for any bibliographic resource that has a 
> Linked Data identifier.

...the difference being that with a DOI you have at least in principle some 
guarantee of what sort of stuff is on the end of the URI (and of it still being 
there).

See http://www.doi.org/doi_handbook/5_Applications.html#5.4 

> PS it occurs to me that the other part of the question is 'what 
> metadata should be included in the OpenURL to give it the best chance 
> of working with a link resolver'?

For DOIs there is a simple answer to this: 
http://www.doi.org/doi_handbook/4_Data_Model.html#4.3.1 - for specific 
implementations, there's a domain-specific set.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
Young,Jeff (OR)
Sent: 21 November 2012 15:09
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] OpenURL linking but from the content provider's point 
of view

If the referent has a DOI, then I would argue that
rft_id=http://dx.doi.org/10.1145/2132176.2132212 is all you need. The 
descriptive information that typically goes in the ContextObject can be 
obtained (if necessary) by content-negotiating for application/rdf+xml.
OTOH, if someone pokes this same URI from a browser instead, you will generally 
get redirected to the publisher's web site with the full-text close at hand.

The same principle should apply for any bibliographic resource that has a 
Linked Data identifier.

Jeff

> -Original Message-
> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf
Of
> Owen Stephens
> Sent: Wednesday, November 21, 2012 9:55 AM
> To: CODE4LIB@LISTSERV.ND.EDU
> Subject: Re: [CODE4LIB] OpenURL linking but from the content
provider's
> point of view
> 
> The only difference between COinS and a full OpenURL is the addition
of
> a link resolver address. Most databases that provide OpenURL links 
> directly (rather than simply COinS) use some profile information - 
> usually set by the subscribing library, although some based on 
> information supplied by an individual user. If set by the library this 
> is then linked to specific users by IP or by login.
> 
> There are a couple(?) of generic base URLs you can use which will try 
> to redirect to an appropriate link resolver based on IP range of the 
> requester, with fallback options if it can't find an appropriate 
> resolver (I think this is how the WorldCat resolver works? The
'OpenURL
> Router' in the UK definitely works like this)
> 
> The LibX toolbar allows users to set their link resolver address, and 
> then translates COinS into OpenURLs when you view a page - all user 
> driven, no need for the data publisher to do anything beyond COinS
> 
> There is also the 'cookie pusher' solution which ArXiv uses - where
the
> user can set a cookie containing the base URL, and this is picked up 
> and used by ArXiV (http://arxiv.org/help/openurl)
> 
> Owen
> 
> 
> Owen Stephens
> Owen Stephens Consulting
> Web: http://www.ostephens.com
> Email: o...@ostephens.com
> Telephone: 0121 288 6936
> 
> On 20 Nov 2012, at 19:39, David Lawrence 
> wrote:
> 
> > I have some experience with the library side of link resolver code.
> > However, we want to implement OpenURL hooks on our open access 
> > literature database and I can not find where to begin.
> >
> > SafetyLit is a free service of San Diego State University in 
> > cooperation with the World Health Organization. We already provide 
> > embedded metadata in both COinS and unAPI formats to allow its
> capture
> > by Mendeley, Papers, Zotero, etc. Over the past few months, I have 
> > emailed or talked with many people and read everything I can get my 
> > hands on about this but I'm clearly not finding the right people or
> information sources.
> >
> > Please help me to find references to examples of the code that is 
> > required on the literature database server that will enable library 
> > link resolvers to recognize the SafetyLit.org metadata and allow 
> > appropriate linking to full text.
> >
> > SafetyLit.org receives more than 65,000 unique (non-robot) visitors 
> > and the database responds to almost 500,000 search queries every
> week.
> > The most frequently requested improvement is to add link resolver
> capacity.
> >
> > I hope that code4lib users will be able to help.
> >
> > Best regards,
> >
> > David
> >
> > David W. Lawrence, PhD, MPH, Director Center for Injury Prevention 
> > Policy and Practice San Diego State University, School of Public 
> > Health
> > 6475 Alvarado Road, Suite 105
> > San Diego, CA  92120  usadavid.lawre...@sdsu.edu
> > V 619 594 1994   F 619 594 1995  Skype: DWL-SDCAwww.CIPPP.org  --
> > www.SafetyLit.org


Re: [CODE4LIB] COinS

2012-11-21 Thread Michael Hopwood
Owen,

Good points well made - I don't think that any one commercial service has yet 
solved all the issues involved in these questions of appropriate copy, but I 
would caution that libraries and publishers need to work together on extending 
existing broad-based standards into these areas, if anything  to avoid 
proprietary solutions becoming de facto standards.

Cheers,

M

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Owen 
Stephens
Sent: 21 November 2012 10:37
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] COinS

Agreed.

The SchemaBibex group is having some of this discussion, and I think the 
'appropriate copy' problem is one the library community can potentially bring 
to the table. There are no guarantees, and it could be we end up with yet 
another set of standards/guidelines/practices that the wider world/web doesn't 
care about - but I think there is an opportunity to position this so that other 
services can see the benefits of pushing relevant data out, and search engines 
can see how it can be used to enhance their services. I suspect that discussing 
this and coming up with proposals in the context of Schema.org is the best bet 
(for the moment at least) at moving this kind of work from the current niche to 
a more mainstream position.

I'd argue that matching resources (via descriptions) to availability to is now 
a more general problem than when OpenURL was conceived as the growth of 
subscription based services like Netflix/Kindle lending/Spotify etc. lead to 
the same issues. This is expressed on the SchemaBibex wiki 
http://www.w3.org/community/schemabibex/wiki/Why_Extend. Also several of the 
use cases described are in this area - 
http://www.w3.org/community/schemabibex/wiki/Use_Cases#Use_case:_Describe_library_holding.2Favailability_information,
 alongside use cases that look at how to describe scholarly articles 
http://www.w3.org/community/schemabibex/wiki/Use_Cases#Use_case:_journal_articles_and_other_periodical_publications

If we are going to see adoption, I strongly believe the outcomes we are 
describing have to be compelling to search engines, and their users, as well as 
publishers and other service providers. It would be great to get more 
discussion of what a compelling proposal might look like on the SchemaBibex 
list http://lists.w3.org/Archives/Public/public-schemabibex/ or wiki 
http://www.w3.org/community/schemabibex/wiki/Main_Page

Owen


Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com
Telephone: 0121 288 6936

On 21 Nov 2012, at 07:37, Dave Caroline  wrote:

>> 
>> 
>> In terms of vocabulary, Schema.org is "extensible" via several mechanisms 
>> including mashups with other vocabularies or, ideally, direct integration 
>> into the Schema.org namespace such as we've seen with RNews 
>>  , 
>> JobPostings 
>>  , 
>> and GoodRelations 
>>  . This is 
>> a win/win scenario, but it requires communities to prove they can articulate 
>> a sensible set of extensions and deliver the information in that model. 
>> Within the "bibliographic" community, this is the mandate set for the 
>> http://www.w3.org/community/schemabibex/ group. If you are disappointed with 
>> OpenURL metadata formats, poor support for COinS, and disappointing 
>> probabilities for content resolution, here's your chance for leveraging SEO 
>> for those purposes.
> 
> But... it is no good choosing a random extension if the Search engine 
> is or will be blind to that particular method.
> As someone who likes to leverage SEO the "right" way so one does not 
> get penalised, some standardisation  is needed.
> 
> Dave Caroline, waiting


Re: [CODE4LIB] Schema.org and its extension for bib data (was Coins)

2012-11-21 Thread Michael Hopwood
Speaking as a metadata (cataloguing in the old money) librarian working in the 
heart of the publishing industry (as I should do, because that's what I am) I 
would add:

1. On standardisation, I hope Bib Extend will largely draw on the existing 
widely used standard ONIX for Books. Two reasons:

a) apart from publishers and their data service partners sitting on mountains 
of this data that could potentially be opened up somewhat, it's already mapped 
to MaRC21 so there's automatic cross-sector appeal (see 
http://www.oclc.org/research/news/2012/05-21.html and 
http://www.oclc.org/research/news/2010/04-09.html)

b) it's "linked-data ready" because it latches into the DOI model - see e.g. 
http://www.doi.org/doi_handbook/5_Applications.html#5.4 and 
http://bitwacker.com/2010/01/19/the-doi-datacite-and-linked-data-made-for-each-other/
 - etc. etc.

2. More broadly, the other DOI implementations (CrossRef, DataCite) already 
have nice ways to output bib citations - is there any way to get Schema.org 
interested and talking with them?

Schema.org and DOI really ought to play well together - the DOI kernels and 
extensions are very parsimonious and "lightweight" but have the extra "data 
smarts" that will make Linked Data less onerous in future...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Richard 
Wallis
Sent: 21 November 2012 08:27
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Schema.org and its extension for bib data (was Coins)

On 21 November 2012 07:37, Dave Caroline wrote:

>
> But... it is no good choosing a random extension if the Search engine 
> is or will be blind to that particular method.
> As someone who likes to leverage SEO the "right" way so one does not 
> get penalised, some standardisation  is needed.
>
>
This is exactly what is behind the Schema.org initiative - 'the search engines' 
(Google, Bing, Yahoo and Yandex) have agreed to recognise structured markup 
using the Schema.org vocabulary and will expand that recognition over time as 
the vocabulary evolves.

There is a process, under the wing of the W3C, to propose extensions to the 
vocabulary to improve it's descriptive capabilities for particular domains.
 As Jeff mentioned, this has already occurred in the areas of news,
commerce, jobs.   Coming mostly from a groups organisations in those
domains, these proposals were successful as they came with some authority to 
the [by definition] broadly focused group behind schema.org.

It is for that reason, I formed the W3C Group Schema Bib Extend < 
http://www.w3.org/community/schemabibex/> to create such a consensus in the 
community concerned with publishing bibliographic data on the web.  All are 
welcome to join this group, membership of the W3C is not a requirement.
 Elements of this COinS conversation are obviously relevant to such proposals.

*Bit of Background for those new to this:*

* Schema.org  introduced in mid 2011 by Google, Yahoo, 
Bing, and Yandex.
* A generic vocabulary for describing most things in structured data on the web 
that the search engines will recognise
* By June 2012, Google & Bing report that 7%-10% of crawled pages contain 
schema.org markup < 
http://dataliberate.com/2012/06/schema-org-consensus-at-semtechbiz/>
* Schema Bib Extend W3C Group formed Sept 2012 as a short lived group to 
propose bibliographic (in the widest sense) extensions to Schema.
* SchemaBibEx not just focusing on library needs, includes publishers etc.
- anyone wanting to publish bibliographic structured data on the web.
* Schema.org, due to its broad generic nature will only complement, not 
replace, other detailed library standards.
* By publishing bib metadata in this way we have at last a way to tell the 
world, not only the search engines, about our resources using markup that will 
be broadly understood.
* Using markup 'the world will understand, and can use' underpins the release 
of WorldCat linked data earlier this year < 
http://dataliberate.com/2012/06/oclc-worldcat-linked-data-release-significant-in-many-ways/
>

~Richard.
--
Richard Wallis
Founder, Data Liberate
http://dataliberate.com

Technology Evangelist, OCLC

Tel: +44 (0)7767 886 005

Linkedin: http://www.linkedin.com/in/richardwallis
Skype: richard.wallis1
Twitter: @rjw
IM: rjw3...@hotmail.com


Re: [CODE4LIB] Easiest way to tag thousands of images

2012-11-21 Thread Michael Hopwood
Hello Kyle,

In addition to the excellent suggestions of Kari and Shaun, may I also point 
you to a special interest group of photometadata/cultural heritage experts who 
are working on very similar issues:

http://groups.yahoo.com/group/Cultural-Heritage-Metadata/

One of the problems they are tabling is an extended schema for embedding rich 
descriptions in file headers.

I think they would love to hear your use case. To join the group you need to 
"apply" to the moderators; that would be Sarah Saunders 
(sa...@electriclane.co.uk) and David Riecks (da...@riecks.com).

I'm also open to discussing this on or off list as I put together a 
comprehensive mapping of photo data (IPTC/XMP) to the heritage schema LIDO 
(www.lido-schema.org) recently and made some recommendations around VRA: 
http://www.linkedheritage.org/getFile.php?id=394

Cheers,

Michael

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Kyle 
Banerjee
Sent: 20 November 2012 19:55
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Easiest way to tag thousands of images

I am in the process of examining how photo collections maintained by campus 
units can be incorporated into the library's repository. In all cases that I've 
had to deal with so far, they're just using the file system -- i.e.
traversing folders that arrange images thematically to file names that indicate 
the content.

Each of these collections contains many thousands of images. This means that 
it's a hassle for them to find images, but also that there's no way library 
staff alone will be able to handle all the metadata creation.

I'd like to use something slick like picasa to help them out (facial 
recognition is an especially big deal for us). But I'm finding the metadata to 
be both minimalist and clunky to work with so I wanted a reality check to see 
if I'm not doing this the dumb way. Things I've noticed:


   1. Picasa appears to store info in xmp rather than exif which is great
   given the limitations of exif. However, I haven't yet found a way to use
   more than a couple fields. The caption shows up in a description field, and
   they tag show up in subjects. But aside from that, I'm at a loss of how to
   populate other DC fields through the interface.

   2. Facial recognition metadata doesn't show up in xmp at all. However, I
   can get that by parsing .picasa.ini and contacts.xml (clunky, but doable).
   I'm kind of tempted to tell people to go into albums and batch tag the
   people albums since it's going to be fun explaining how to locate these
   hidden files.

My real question is whether anyone has come up with a really good way to assign 
metadata to thousands of photos, preferably in batch fashion? Thanks,

kyle


Re: [CODE4LIB] PBCore 2.0 to MARC XML?

2012-11-07 Thread Michael Hopwood
John,

To put a final piece on this jigsaw, I would also note that if you want to 
generate schema-to-schema stylesheets, especially for complex formats like 
MARCXML, in a user-friendly way, you could do worse than MINT:

http://mint.image.ece.ntua.gr/redmine/projects/mint/wiki/Mint

I am currently working with the developers of MINT and I'm happy to put you in 
touch.

You might need some developer resource to implement a MARC21 instance, but 
there may be one out there and/or the newer open source release may allow you 
to specify source schema directly.

I can't imagine that the need to map from X to MARC21 is that uncommon... 
probably why there's also http://www.extensiblecatalog.org/

Cheers,

Michael

> On Mon, Nov 5, 2012 at 2:15 PM, john passmore 
> wrote:
>
> > Hi,
> > Before I start reinventing the wheel, does anyone know of any 
> > stylesheets out there that convert PBCore 2.0 XML 
> >  to MARC XML?
> >
> > Thanks!
> > John
> > WNYC Archives
> >
>


Re: [CODE4LIB] PBCore 2.0 to MARC XML?

2012-11-06 Thread Michael Hopwood
How about mapping PBCore to the CIDOC-CRM as a hub format then going into MARC?

http://www.cidoc-crm.org/crm_mappings.html - I don't see a PBCore mapping yet, 
but it should be doable.

You may have to use the FRBRoo extension 
(http://www.cidoc-crm.org/frbr_inro.html) for some fields.

The beauty of doing it that way would be that your intermediate data would also 
have a longer term usability if you convert to another schema later on.

Or if your institution is really generous, get them to fund a mapping of PBCore 
and (the relevant bits of) MARC21 to the VMF: 
http://www.doi.org/VMF/registering.html and put it out there for everyone 
(including any DOI registrations) to link to?

Cheers,

Michael

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of john 
passmore
Sent: 06 November 2012 15:28
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] PBCore 2.0 to MARC XML?

Ah, that makes sense. I could try  PBCore>Dublin Core>MARC 21.


On Tue, Nov 6, 2012 at 9:01 AM, Tom Keays  wrote:

> PBCore lists a bunch of crosswalk mapping schemes that they've identified.
>
> http://www.pbcore.org/PBCore/PBCore_Mappings.html
>
> Some of them are listed but aren't done yet, including a direct 
> mapping to MARC 21. However, a mapping of PBCore to Dublin Core 
> exists, so you could probably get there by using DC as an intermediate.
>
> http://www.pbcore.org/PBCore/mappings/PBCore-DublinCore_Mapping.html
>
> They also identify the Metadata Advisory Group of the MIT Libraries as 
> a source of mapping information, so you could possibly consult them 
> for help in this project.
>
> http://libraries.mit.edu/guides/subjects/metadata/mappings.html
>
>
> On Mon, Nov 5, 2012 at 2:15 PM, john passmore 
> wrote:
>
> > Hi,
> > Before I start reinventing the wheel, does anyone know of any 
> > stylesheets out there that convert PBCore 2.0 XML 
> >  to MARC XML?
> >
> > Thanks!
> > John
> > WNYC Archives
> >
>


Re: [CODE4LIB] Book metadata source

2012-10-26 Thread Michael Hopwood
With reference to...

>>> I realize that different manifestations of the same work will have 
>>> different ISBNs, so we'd be seeking any work in print format, ideally the 
>>> most commonly held.

...have you had a look at ISTC?

http://www.istc-international.org/html/

If you register the "works" (there's a format for that: 
http://www.editeur.org/106/ONIX-ISTC-Registration-Format/) you can collate the 
titles by content rather than by manifestation/product ID while linking to the 
various ISBNs of available versions...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Cab 
Vinton
Sent: 25 October 2012 20:06
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Book metadata source

I have a list of several hundred book titles & corresponding authors, 
comprising our State Library's book group titles, & am looking for ways of 
putting these titles online in a way that would be useful to librarians & 
patrons. Something along the lines of a LibraryThing collection or Amazon 
wishlist.

Without ISBNs, however, the process could be very labor-intensive.

Any suggestions for how we could handle this as part of a batch process?

I realize that different manifestations of the same work will have different 
ISBNs, so we'd be seeking any work in print format, ideally the most commonly 
held.

The only thought I've had is to do a Z39.50 search using the author & title 
Bib-1 attributes EG @and @attr 1=4 mansfield @attr 1=1003 austen.

Thanks for your thoughts,

Cab Vinton, Director
Sanbornton Public Library
Sanbornton, NH


Re: [CODE4LIB] Oral History Metadata Best Practices

2012-10-25 Thread Michael Hopwood
Hello Jane, Priscilla,

I would recommend looking at www.lido-schema.org as more interoperable, 
extensible and generally longer-term value-adding schema for collection of a 
lot of historical / heritage data.

It has the same capabilities and easy entry level (only three mandatory 
sections; object/work type - title/name - record details) as Dublin Core to 
collate a lot of data, potentially from different source, but it also has the 
optional depth and breadth required for enriching data with links, and the 
specific semantics used precisely by archives and historians, rather than 
libraries.

Data created / collected in LIDO will have greater reuse potential than less 
contextual schemas.

It's based on the ISO standard CIDOC-CRM (see 
http://www.cidoc-crm.org/uses_applications.html) which is itself the result of 
painstaking work by historical and archives data people.

The CIDOC-CRM itself is probably worth looking at too, maybe in terms of "CRM 
CORE" (a lightweight model using just key parts) and the many real life 
applications in archive contexts...

Disclaimer: I work (as a librarian!) on one part of www.linkedheritage.eu which 
does have some of the world LIDO experts as partners.

Best,

Michael Hopwood

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
Priscilla Caplan
Sent: 25 October 2012 14:17
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Oral History Metadata Best Practices

You might want to look at the section on Cataloging in the best practices guide 
on Florida Voices: 
http://www.fcla.edu/FloridaVoices/index.htm

Priscilla

On 10/25/2012 8:57 AM, Jacobs, Jane W wrote:
> Hi Library-Coders,
>
> My colleagues and I are researching best practices in recording metadata for 
> Oral Histories for an article tentatively accepted for publication.  We're 
> looking for input from practicing librarians, archivists, and historians.  In 
> particular we'd like to know what encodings (e.g. MARC, EAD, METS, etc.) 
> people are using and how happy (or unhappy) they are with them.  Also what 
> fields are people using to enter their data? Any data-dictionaries or 
> templates showing required, repeatable, non-repeatable fields would be 
> welcome.
>
> So far we've discovered that with new digital technologies allowing much 
> easier collection and retransmission of oral histories, creation is booming; 
> standards not so much.
>
> We would appreciate input from anyone who is willing to share their 
> procedures.  As mentioned above, we are planning to publish an article, but 
> we will, of course, ask permission, before quoting anyone directly.  Off-list 
> responses are welcome.
>
> Please excuse duplication (cross-posting) and forward to interested 
> colleagues.
>
> Thanks in advance for your help.
>
> JJ
>
>
>
> **Views expressed by the author do not necessarily represent those of 
> the Queens Library.**
>
> Jane Jacobs
> Asst. Coord., Catalog Division
> Queens Borough Public Library
> 89-11 Merrick Blvd.
> Jamaica, NY 11432
> tel.: (718) 990-0804
> e-mail: 
> jane.w.jac...@queenslibrary.org<mailto:jane.w.jac...@queenslibrary.org
> >
> FAX. (718) 990-8566
>
>
>
> Connect with Queens Library:
>   
> *  QueensLibrary.org
>  http://www.queenslibrary.org/
>
>   *  Facebook
>   http://www.facebook.com/queenslibrarynyc
>
>   *  Twitter
>   http://www.twitter.com/queenslibrary
>
>   *  LinkedIn
>   http://www.linkedin.com/company/queens-library
>
>   *  Google+
>   https://plus.google.com/u/0/116278397527253207785
>
>   *  Foursquare
>   https://foursquare.com/queenslibrary
>
>   *  YouTube
>   http://www.youtube.com/queenslibrary
>
>   *  Flickr
>   http://www.flickr.com/photos/qbpllid/
>
>   *  Goodreads
>   http://www.goodreads.com/group/show/58240.Queens_Library
>
>
> The information contained in this message may be privileged and 
> confidential and protected from disclosure. If the reader of this 
> message is not the intended recipient, or an employee or agent 
> responsible for delivering this message to the intended recipient, you 
> are hereby notified that any dissemination, distribution or copying of 
> this communication is strictly prohibited. If you have received this 
> communication in error, please notify us immediately by replying to 
> the message and deleting it from your computer.
>

--
Priscilla Caplan
Assistant Director for Digital Library Services Florida Virtual Campus
5830 NW 39th Avenue
Gainesville, FL 32606
(352) 392-9020 x324
(352) 392-9185 (fax)


Re: [CODE4LIB] Oral History Metadata Best Practices

2012-10-25 Thread Michael Hopwood
PS - there is also a paper on oral traditions in FRBR...

archivesic.ccsd.cnrs.fr/docs/00/06/26/91/PDF/sic_1629.pdf

-Original Message-
From: Michael Hopwood 
Sent: 25 October 2012 14:44
To: 'Jacobs, Jane W'; 'Priscilla Caplan'
Cc: CODE4LIB@LISTSERV.ND.EDU
Subject: RE: [CODE4LIB] Oral History Metadata Best Practices

Hello Jane, Priscilla,

I would recommend looking at www.lido-schema.org as more interoperable, 
extensible and generally longer-term value-adding schema for collection of a 
lot of historical / heritage data.

It has the same capabilities and easy entry level (only three mandatory 
sections; object/work type - title/name - record details) as Dublin Core to 
collate a lot of data, potentially from different source, but it also has the 
optional depth and breadth required for enriching data with links, and the 
specific semantics used precisely by archives and historians, rather than 
libraries.

Data created / collected in LIDO will have greater reuse potential than less 
contextual schemas.

It's based on the ISO standard CIDOC-CRM (see 
http://www.cidoc-crm.org/uses_applications.html) which is itself the result of 
painstaking work by historical and archives data people.

The CIDOC-CRM itself is probably worth looking at too, maybe in terms of "CRM 
CORE" (a lightweight model using just key parts) and the many real life 
applications in archive contexts...

Disclaimer: I work (as a librarian!) on one part of www.linkedheritage.eu which 
does have some of the world LIDO experts as partners.

Best,

Michael Hopwood

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
Priscilla Caplan
Sent: 25 October 2012 14:17
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Oral History Metadata Best Practices

You might want to look at the section on Cataloging in the best practices guide 
on Florida Voices: 
http://www.fcla.edu/FloridaVoices/index.htm

Priscilla

On 10/25/2012 8:57 AM, Jacobs, Jane W wrote:
> Hi Library-Coders,
>
> My colleagues and I are researching best practices in recording metadata for 
> Oral Histories for an article tentatively accepted for publication.  We're 
> looking for input from practicing librarians, archivists, and historians.  In 
> particular we'd like to know what encodings (e.g. MARC, EAD, METS, etc.) 
> people are using and how happy (or unhappy) they are with them.  Also what 
> fields are people using to enter their data? Any data-dictionaries or 
> templates showing required, repeatable, non-repeatable fields would be 
> welcome.
>
> So far we've discovered that with new digital technologies allowing much 
> easier collection and retransmission of oral histories, creation is booming; 
> standards not so much.
>
> We would appreciate input from anyone who is willing to share their 
> procedures.  As mentioned above, we are planning to publish an article, but 
> we will, of course, ask permission, before quoting anyone directly.  Off-list 
> responses are welcome.
>
> Please excuse duplication (cross-posting) and forward to interested 
> colleagues.
>
> Thanks in advance for your help.
>
> JJ
>
>
>
> **Views expressed by the author do not necessarily represent those of 
> the Queens Library.**
>
> Jane Jacobs
> Asst. Coord., Catalog Division
> Queens Borough Public Library
> 89-11 Merrick Blvd.
> Jamaica, NY 11432
> tel.: (718) 990-0804
> e-mail: 
> jane.w.jac...@queenslibrary.org<mailto:jane.w.jac...@queenslibrary.org
> >
> FAX. (718) 990-8566
>
>
>
> Connect with Queens Library:
>   
> *  QueensLibrary.org
>  http://www.queenslibrary.org/
>
>   *  Facebook
>   http://www.facebook.com/queenslibrarynyc
>
>   *  Twitter
>   http://www.twitter.com/queenslibrary
>
>   *  LinkedIn
>   http://www.linkedin.com/company/queens-library
>
>   *  Google+
>   https://plus.google.com/u/0/116278397527253207785
>
>   *  Foursquare
>   https://foursquare.com/queenslibrary
>
>   *  YouTube
>   http://www.youtube.com/queenslibrary
>
>   *  Flickr
>   http://www.flickr.com/photos/qbpllid/
>
>   *  Goodreads
>   http://www.goodreads.com/group/show/58240.Queens_Library
>
>
> The information contained in this message may be privileged and 
> confidential and protected from disclosure. If the reader of this 
> message is not the intended recipient, or an employee or agent 
> responsible for delivering this message to the intended recipient, you 
> are hereby notified that any dissemination, distribution or copying of 
> this communication is strictly prohibited. If you have received this 
> communication in error, please notify us immediately by replying to 
> the message and deleting it from your computer.
>

--
Priscilla Caplan
Assistant Director for Digital Library Services Florida Virtual Campus
5830 NW 39th Avenue
Gainesville, FL 32606
(352) 392-9020 x324
(352) 392-9185 (fax)


[CODE4LIB] FW: [CODE4LIB] Q.: software for vendor title list processing

2012-10-17 Thread Michael Hopwood
Hi Godmar,

There is also ONIX for Serials Online Holdings 
(http://www.editeur.org/120/ONIX-SOH/). I'm copying in Tim Devenport who might 
say more.

Best wishes,

Michael

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Owen 
Stephens
Sent: 16 October 2012 23:09
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Q.: software for vendor title list processing

I'm working on the JISC KB+ project that Tom mentioned.

As part of the project we've been collating journal title lists from various 
sources. We've been working with members of the KBART steering group and have 
used KBART where possible, although we've been collecting data not covered by 
KBART.

All the data we have at this level is published under a CC0 licence at 
http://www.kbplus.ac.uk/kbplus/publicExport - including a csv that uses the 
KBART data elements. The focus so far has been on packages negotiated by JISC 
in the UK - although in many cases the title lists may be the same as are made 
available in other markets. We also include what we call 'Master lists' which 
are an attempt to capture the complete list of titles and coverage offered by a 
content provider. We'd very much welcome any feedback on these exports, and of 
course be interested to know if anyone makes use of them.

So far a lot of the work on collating/coverting/standardising the data has been 
done by hand - which is clearly not ideal. In the next phase of the project the 
KB+ project is going to work with the GoKB project http://gokb.org - as part of 
this collaboration we are currently working on ways of streamlining the data 
processing from publisher files or other sources, to standardised data. While 
we are still working on how this is going to be implemented, we are currently 
investigating the possibility of using Google/Open Refine to capture and re-run 
sets of rules across data sets from specific sources. We should be making 
progress on this in the next couple of months.

Hope that's helpful

Owen

Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com
Telephone: 0121 288 6936

On 16 Oct 2012, at 20:23, Tom Pasley  wrote:

> You might also be interested in the work at http://www.kbplus.ac.uk . 
> The site is up at the moment, but I can't reach it for some reason... 
> they have a public export page which you might want to know about 
> http://www.kbplus.ac.uk/kbplus/publicExport
> 
> Tom
> 
> On Wed, Oct 17, 2012 at 8:12 AM, Jonathan Rochkind  wrote:
> 
>> I think KBART is such an effort.  As with most library standards 
>> groups, there may not be online documentation of their most recent 
>> efforts or successes, but: http://www.uksg.org/kbart
>> 
>> http://www.uksg.org/kbart/s5/**guidelines/data_format> .org/kbart/s5/guidelines/data_format>
>> 
>> 
>> 
>> On 10/16/2012 2:16 PM, Godmar Back wrote:
>> 
>>> Hi,
>>> 
>>> at our library, there's an emerging need to process title lists from 
>>> vendors for various purposes, such as checking that the titles 
>>> purchased can be discovered via discovery system and/or OPAC. It 
>>> appears that the formats in which those lists are provided are 
>>> non-uniform, as is the process of obtaining them.
>>> 
>>> For example, one vendor - let's call them "Expedition Scrolls" - 
>>> provides title lists for download to Excel, but which upon closer 
>>> inspection turn out to be HTML tables. They are encoded using an odd 
>>> mixture of CP1250 and HTML entities. Other vendors use entirely different 
>>> formats.
>>> 
>>> My question is whether there are efforts, software, or anything 
>>> related to streamlining the acquisition and processing of vendor 
>>> title lists in software systems that aid in the collection 
>>> development and maintenance process. Any pointers would be appreciated.
>>> 
>>>  - Godmar
>>> 
>>> 
>>> 


Re: [CODE4LIB] CODE4LIB equivalent in UK?

2012-10-08 Thread Michael Hopwood
I think this answer came close to what I was "thinking" when I wrote the 
question last week, and also reminds me of another similar vague suggestion I 
saw recently:

http://blog.ldodds.com/2012/07/11/uk-eu-linked-data-consultant-network/ 

-Original Message-
From: Ian Ibbotson [mailto:ian.ibbot...@k-int.com] 
Sent: 08 October 2012 10:09
To: Michael Hopwood
Cc: Code for Libraries
Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?

hehe ah you caught me out :)

It's kinda hard as my existing "channels" are more a diffuse collection of 
lists and twitter accounts if I'm really honest.. Deffo might be worth starting 
a wiki page with UK specific resources, just not sure I can handle another 
channel..

FWIW tho, I'd start by looking in places like

http://helibtech.com/Open+Source

which points pretty directly to the JISCMail list LIS-OSS 
(https://www.jiscmail.ac.uk/cgi-bin/webadmin?A0=LIS-OSS). Whilst you might 
think of this as being a little too HE specific, I think it's a pretty diverse 
community.

But really, I'd take my primary information channel as twitter these days (for 
localised event news and information), checking out the #UKLibChat discussions 
might be an idea. @ostephens, @daveyp and @benosteen usually have their finger 
on the pulse of all the good stuff (If we could persuade them to keep their 
lanyrd pages up to date that would be a pretty amazing resource for events).

I think for generic subject oriented questions, it's still good to have a wider 
community and this list serves that purpose reasonably well.

Guess thats where I'd start really, not sure it's much help tho.

e


On Mon, Oct 8, 2012 at 9:54 AM, Michael Hopwood  wrote:
> What's your favoured UK channel? Or say, the top 3 to watch?
>
> -Original Message-
> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf 
> Of Ian Ibbotson
> Sent: 08 October 2012 09:48
> To: CODE4LIB@LISTSERV.ND.EDU
> Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?
>
> +1 another UK lurker here..
>
> FWIW tho, I'd be pro reusing one of the existing (many) uk channels 
> rather than starting up another list to subscribe to :)
>
> e
>
>
> On Mon, Oct 8, 2012 at 9:14 AM, Richard Wallis 
>  wrote:
>> The Mashed Library folks might be fertile ground for gaining interest 
>> in a code4libuk.
>>  http://www.mashedlibrary.com
>>
>> ~Richard.
>>
>> On 7 October 2012 16:28, Tim Hill  wrote:
>>
>>> Here's another lurking UK code4libber! I work for a UK/US company, 
>>> but I spend the bulk of my time in the UK (and never enough in the 
>>> US to coincide with a code4lib meetup). I'd certainly be interested 
>>> in getting the/a community more active in the UK.
>>>
>>> Tim Hill
>>>
>>> On Tue, Oct 2, 2012 at 9:12 AM, Simeon Warner 
>>> 
>>> wrote:
>>> > Have a look at http://devcsi.ukoln.ac.uk/ . This is mainly focused 
>>> > on repositories but seems somewhat similar from an outside view.
>>> >
>>> > Cheers,
>>> > Simeon (lurking expat Brit)
>>> >
>>> >
>>> > On 10/2/12 4:11 AM, Michael Hopwood wrote:
>>> >>
>>> >> Yes - my question was implicitly aimed at lurking UKavians.
>>> >>
>>> >> -Original Message-
>>> >> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On 
>>> >> Behalf Of Dave Caroline
>>> >> Sent: 02 October 2012 09:08
>>> >> To: CODE4LIB@LISTSERV.ND.EDU
>>> >> Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?
>>> >>
>>> >> On Tue, Oct 2, 2012 at 8:55 AM, Michael Hopwood 
>>> >> 
>>> >> wrote:
>>> >>>
>>> >>> I know that CODE4LIB isn't per se "in" the USA but it seems like 
>>> >>> a
>>> large
>>> >>> number of its active users are.
>>> >>>
>>> >>> Is there an equivalent list that you folks know of?
>>> >>
>>> >>
>>> >> I dont know of an equivalent British list but there are a few of 
>>> >> us
>>> brits
>>> >> about lurking in #cod4lib too (archivist)
>>> >>
>>> >> Dave Caroline
>>> >>
>>> >
>>>
>>
>>
>>
>> --
>> Richard Wallis
>> Founder, Data Liberate
>> http://dataliberate.com
>> Tel: +44 (0)7767 886 005
>>
>> Linkedin: http://www.linkedin.com/in/richardwallis
>> Skype: richard.wallis1
>> Twitter: @rjw
>> IM: rjw3...@hotmail.com


Re: [CODE4LIB] CODE4LIB equivalent in UK?

2012-10-08 Thread Michael Hopwood
What's your favoured UK channel? Or say, the top 3 to watch?

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Ian 
Ibbotson
Sent: 08 October 2012 09:48
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?

+1 another UK lurker here..

FWIW tho, I'd be pro reusing one of the existing (many) uk channels rather than 
starting up another list to subscribe to :)

e


On Mon, Oct 8, 2012 at 9:14 AM, Richard Wallis 
 wrote:
> The Mashed Library folks might be fertile ground for gaining interest 
> in a code4libuk.
>  http://www.mashedlibrary.com
>
> ~Richard.
>
> On 7 October 2012 16:28, Tim Hill  wrote:
>
>> Here's another lurking UK code4libber! I work for a UK/US company, 
>> but I spend the bulk of my time in the UK (and never enough in the US 
>> to coincide with a code4lib meetup). I'd certainly be interested in 
>> getting the/a community more active in the UK.
>>
>> Tim Hill
>>
>> On Tue, Oct 2, 2012 at 9:12 AM, Simeon Warner 
>> 
>> wrote:
>> > Have a look at http://devcsi.ukoln.ac.uk/ . This is mainly focused 
>> > on repositories but seems somewhat similar from an outside view.
>> >
>> > Cheers,
>> > Simeon (lurking expat Brit)
>> >
>> >
>> > On 10/2/12 4:11 AM, Michael Hopwood wrote:
>> >>
>> >> Yes - my question was implicitly aimed at lurking UKavians.
>> >>
>> >> -Original Message-
>> >> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On 
>> >> Behalf Of Dave Caroline
>> >> Sent: 02 October 2012 09:08
>> >> To: CODE4LIB@LISTSERV.ND.EDU
>> >> Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?
>> >>
>> >> On Tue, Oct 2, 2012 at 8:55 AM, Michael Hopwood 
>> >> 
>> >> wrote:
>> >>>
>> >>> I know that CODE4LIB isn't per se "in" the USA but it seems like 
>> >>> a
>> large
>> >>> number of its active users are.
>> >>>
>> >>> Is there an equivalent list that you folks know of?
>> >>
>> >>
>> >> I dont know of an equivalent British list but there are a few of 
>> >> us
>> brits
>> >> about lurking in #cod4lib too (archivist)
>> >>
>> >> Dave Caroline
>> >>
>> >
>>
>
>
>
> --
> Richard Wallis
> Founder, Data Liberate
> http://dataliberate.com
> Tel: +44 (0)7767 886 005
>
> Linkedin: http://www.linkedin.com/in/richardwallis
> Skype: richard.wallis1
> Twitter: @rjw
> IM: rjw3...@hotmail.com


Re: [CODE4LIB] CODE4LIB equivalent in UK?

2012-10-08 Thread Michael Hopwood
That's a good point Richard. I wonder how much overlap there is between 
mashed/JISC Dev?

Perhaps between them they cover a good number of people with such interests...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Richard 
Wallis
Sent: 08 October 2012 09:14
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?

The Mashed Library folks might be fertile ground for gaining interest in a 
code4libuk.
 http://www.mashedlibrary.com

~Richard.

On 7 October 2012 16:28, Tim Hill  wrote:

> Here's another lurking UK code4libber! I work for a UK/US company, but 
> I spend the bulk of my time in the UK (and never enough in the US to 
> coincide with a code4lib meetup). I'd certainly be interested in 
> getting the/a community more active in the UK.
>
> Tim Hill
>
> On Tue, Oct 2, 2012 at 9:12 AM, Simeon Warner 
> 
> wrote:
> > Have a look at http://devcsi.ukoln.ac.uk/ . This is mainly focused 
> > on repositories but seems somewhat similar from an outside view.
> >
> > Cheers,
> > Simeon (lurking expat Brit)
> >
> >
> > On 10/2/12 4:11 AM, Michael Hopwood wrote:
> >>
> >> Yes - my question was implicitly aimed at lurking UKavians.
> >>
> >> -Original Message-
> >> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On 
> >> Behalf Of Dave Caroline
> >> Sent: 02 October 2012 09:08
> >> To: CODE4LIB@LISTSERV.ND.EDU
> >> Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?
> >>
> >> On Tue, Oct 2, 2012 at 8:55 AM, Michael Hopwood 
> >> 
> >> wrote:
> >>>
> >>> I know that CODE4LIB isn't per se "in" the USA but it seems like a
> large
> >>> number of its active users are.
> >>>
> >>> Is there an equivalent list that you folks know of?
> >>
> >>
> >> I dont know of an equivalent British list but there are a few of us
> brits
> >> about lurking in #cod4lib too (archivist)
> >>
> >> Dave Caroline
> >>
> >
>



--
Richard Wallis
Founder, Data Liberate
http://dataliberate.com
Tel: +44 (0)7767 886 005

Linkedin: http://www.linkedin.com/in/richardwallis
Skype: richard.wallis1
Twitter: @rjw
IM: rjw3...@hotmail.com


Re: [CODE4LIB] CODE4LIB equivalent in UK?

2012-10-02 Thread Michael Hopwood
Yes - my question was implicitly aimed at lurking UKavians.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Dave 
Caroline
Sent: 02 October 2012 09:08
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] CODE4LIB equivalent in UK?

On Tue, Oct 2, 2012 at 8:55 AM, Michael Hopwood  wrote:
> I know that CODE4LIB isn't per se "in" the USA but it seems like a large 
> number of its active users are.
>
> Is there an equivalent list that you folks know of?

I dont know of an equivalent British list but there are a few of us brits about 
lurking in #cod4lib too (archivist)

Dave Caroline


[CODE4LIB] CODE4LIB equivalent in UK?

2012-10-02 Thread Michael Hopwood
I know that CODE4LIB isn't per se "in" the USA but it seems like a large number 
of its active users are.

Is there an equivalent list that you folks know of?

Best,

Michael Hopwood
Linked Heritage Project Lead
EDItEUR
United House, North  Road
London N7 9DP
UK

Tel: +44 20 7503 6418
Mob: +44 7811 591036
Skype: michael.hopwood.editeur
http://www.linkedheritage.org/
http://editeur.org/

The information contained in this e-mail is confidential and may be privileged. 
It is intended for the addressee only. If you are not the intended recipient, 
please inform the sender and delete this e-mail immediately. The contents of 
this e-mail must not be disclosed or copied without the sender's consent. We 
cannot accept any responsibility for viruses, so please scan all attachments. 
The statements and opinions expressed in this message are those of the author 
and do not necessarily reflect those of the company.

EDItEUR Limited is a company limited by guarantee, registered in England no 
2994705. Registered Office:
United House, North Road, London N7 9DP, United Kingdom


Re: [CODE4LIB] Corrections to Worldcat/Hathi/Google

2012-08-29 Thread Michael Hopwood
Thanks for this pointer Owen.

It's a nice illustration of the fact that what users actually want (well, I 
know I did back when I actually worked in large information services 
departments!) is something more like an intranet where the content I find is 
weighted towards me, the "audience" e.g. the intranet knows I'm a 2nd year 
medical student and one of my registered preferred languages is Mandarin 
already, or it "knows" that I'm a rare books cataloguer and I want to see what 
"nine out of ten" other cataloguers recorded for this obscure and confusing 
title.

However, this stuff is quite intense for linked data, isn't it? I understand 
that it would involve lots of quads, named graphs or whatever...

In a parallel world, I'm currently writing up recommendations for aggregating 
ONIX for Books records. ONIX data can come from multiple sources who 
potentially assert different things about a given "book" (i.e. something with 
an ISBN to keep it simple).

This is why *every single ONIX data element* can have option attributes of

@datestamp
@sourcename
@sourcetype [e.g. publisher, retailer, data aggregator... library?]

...and the ONIX message as a whole is set up with "header" and "product record" 
 segments that each include some info about the sender/recipient/data record in 
question.

How people in the book supply chain are implementing these is a distinct issue, 
but could these capabilities have some relevance to what you're discussing?

Do you have any other pointers to "intranet-like" catalogues?

In the museum space, there is of course this: http://www.researchspace.org/

Cheers,

Michael 

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Owen 
Stephens
Sent: 28 August 2012 21:37
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Corrections to Worldcat/Hathi/Google

The JISC funded CLOCK project did some thinking around cataloguing processes 
and tracking changes to statements and/or records - e.g. 
http://clock.blogs.lincoln.ac.uk/2012/05/23/its-a-model-and-its-looking-good/

Not solutions of course, but hopefully of interest

Owen

Owen Stephens
Owen Stephens Consulting
Web: http://www.ostephens.com
Email: o...@ostephens.com
Telephone: 0121 288 6936

On 28 Aug 2012, at 19:43, Simon Spero  wrote:

> On Aug 28, 2012, at 2:17 PM, Joe Hourcle wrote:
> 
>> I seem to recall seeing a presentation a couple of years ago from someone in 
>> the intelligence community, where they'd keep all of their intelligence, but 
>> they stored RDF quads so they could track the source.
>> 
>> They'd then assign a confidence level to each source, so they could get an 
>> overall level of confidence on their inferences.
>> [...]
>> It's possible that it was in the context of provenance, but I'm getting 
>> bogged down in too many articles about people storing provenance information 
>> using RDF-triples (without actually tracking the provenance of the triple 
>> itself)
> 
> Provenance is of great importance in the IC and related sectors.   
> 
> An good overview of the nature of evidential reasoning is David A Schum 
> (1994;2001). Evidential Foundations of Probabilistic Reasoning. Wiley & Sons, 
> 1994; Northwestern University Press, 2001 [Paperback edition].
> 
> There are usually papers on provenance and associated semantics at the GMU 
> Semantic Technology for Intelligence, Defense, and Security (STIDS).  This 
> years conference is 23 - 26 October 2012; see http://stids.c4i.gmu.edu/ for 
> more details. 
> 
> Simon


Re: [CODE4LIB] How do you get plain language, plain English out of the .sgstn stenograph stenonote record of the public meeting?... [see other message]

2012-05-23 Thread Michael Hopwood
Watch out... MEATadata jokes up ahead...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Doran, 
Michael D
Sent: 23 May 2012 22:41
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] How do you get plain language, plain English out of the 
.sgstn stenograph stenonote record of the public meeting?... [see other message]

Hi Roy,

> Since we already control the Bacon Stamp of Approval, baloney seems 
> like the next logical step.

We should be thinking ahead to future use cases.  I say go for a broader "Cured 
Meats" Stamp of Approval.  Or perhaps "Charcuterie" to lend it some class.  To 
do otherwise could lead to a proliferation of stamps.


Re: [CODE4LIB] free source for issn->periodical-type data?

2012-04-17 Thread Michael Hopwood
Just a quick note:

The correct URL for ONIX for Serials is 
http://www.editeur.org/17/ONIX-for-Serials/ - note that this is a family of 
standards, so it covers a very wide range of data types and content. The code 
lists Tom mentioned are available there in human-readable form.

Also: it sounded to me that Ken was after an actual database of the journal 
product type information - something like a "serials in print" database?

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Tom 
Pasley
Sent: 16 April 2012 22:15
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] free source for issn->periodical-type data?

Hi Ken,

Actually, I'm not sure this will answer all of your needs - although it does 
cover peer-review:

Metadata fields for an ISSN

A number of metadata fields can be associated with an ISSN number:

   - form: Each ISSN has a production form, indicated by an ONIX production
   form code . Current supported
   values include: JB ( Printed serial ), JC ( Serial distributed
   electronically by carrier ) ,JD ( Electronic serial distributed online ),
   MA ( Microform )
   - oclcnum: Oclcnum
   - peerreview: Peerreview, 'Y' if the ISSN is peer-reviewed, 'N' if the
   ISSN is not peer-reviewed.
   - publisher: Publisher
   - rawcoverage: Human-readable Coverage
   - title: Title
   - issnl: Linking ISSN, as defined
here
   - rssurl: Journal feed URL, data obtained from 
ticTOCS

T.

On Tue, Apr 17, 2012 at 1:33 AM, Ken Irwin  wrote:

> Hi folks,
>
> Does anyone know of a free data source that correlates ISSNs with data 
> that includes "what kind of publication is this?" e.g.
>
> *Academic journal (+/- peer review?)
>
> *Popular magazine
>
> *Newspaper
>
> *Trade journal
>
> *Etc
>
> Obviously, there's some wiggle room in these designations, and I don't 
> need a super-solid answer.
>
> I've been asked to supply information about our academic journal 
> collection, and I don't have a particularly good way of 
> differentiating between our e-journals and e-magazines, for instance. 
> Individual suppliers might make these distinctions, but I'm really 
> hoping that a query-able (or,
> better: downloadable) file exists.
>
> Any ideas?
>
> Thanks
> Ken
>


Re: [CODE4LIB] Representing geographic hiearchy in linked data

2012-04-10 Thread Michael Hopwood
Exactly - and how do we know about these ancient geographic entities (AGEs)?

They are attested in a variety of sources on various types of evidence, and it 
maybe useful to be able to document which "version" of an AGE you are talking 
about because maybe different people had different views on the matter (or 
different finds point to different cultures in a strictly geographic 
region?!)...

This is the kind of stuff that CIDOC-CRM-SIG had in mind when creating CRM 
because, well, that's what museum data is about...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Ethan 
Gruber
Sent: 10 April 2012 00:13
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Representing geographic hiearchy in linked data

Ancient geographic entities.  Athens is in Attica.  Sardis is in Lydia (in 
Anatolia, for example).  If these were modern geopolitical entities, I would 
use geonames.  We're linking cities to Pleiades, but Pleiades does not maintain 
parent::child geographic relationships.

Ethan
On Apr 9, 2012 5:53 PM, "Simon Spero"  wrote:

> Are you talking about geographical entities, or geopolitical ones? For 
> example,  is there an answer to the question "what country is 
> constantinople located in?"
>
> Simon
> On Apr 8, 2012 8:02 PM, "Ethan Gruber"  wrote:
>
> > CIDOC-CRM may be the answer here. I will look over the documentation 
> > in greater detail tomorrow.
> >
> > Thanks,
> > Ethan
> > On Apr 8, 2012 7:56 PM, "Ethan Gruber"  wrote:
> >
> > > The data is modeled, but I want to use an ontology for geographic
> > concepts
> > > that already exists, if possible.  If anything, my issue 
> > > highlights the point that linked data can be *too* flexible.
> > > On Apr 8, 2012 3:54 PM, "Michael Hopwood"  wrote:
> > >
> > >> I think this highlights the point that, at some point, you have 
> > >> to
> model
> > >> the data.
> > >>
> > >> -Original Message-
> > >> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On 
> > >> Behalf
> Of
> > >> Ethan Gruber
> > >> Sent: 08 April 2012 15:44
> > >> To: CODE4LIB@LISTSERV.ND.EDU
> > >> Subject: Re: [CODE4LIB] Representing geographic hiearchy in 
> > >> linked
> data
> > >>
> > >> Hi,
> > >>
> > >> Thanks for the info, but it's not quite what I'm looking for.  
> > >> We've established authority control for ancient places, but I'm 
> > >> looking for
> an
> > >> ontology I can use to describe the child:parent relationship 
> > >> between
> > city
> > >> and region or region and larger region (in any way that isn't 
> > >> dcterms:partOf).  Geonames has defined their own vocabulary that 
> > >> can't really be reused in other geographic contexts, e.g. with
> gn:countryCode,
> > >> gn:parentCountry.
> > >>
> > >> Thanks,
> > >> Ethan
> > >>
> > >> On Fri, Apr 6, 2012 at 11:40 AM, Karen Coyle 
> wrote:
> > >>
> > >> > Also, there is Geonames (http://www.geonames.org), which is the 
> > >> > primary geographic data set on the Semantic Web. Here is the 
> > >> > link to
> > >> Athens:
> > >> >
> > >> > http://www.geonames.org/**search.html?q=athens&country=**GR<
> > http://www
> > >> > .geonames.org/search.html?q=athens&country=GR>
> > >> >
> > >> > kc
> > >> >
> > >> >
> > >> > On 4/6/12 4:54 PM, Karen Miller wrote:
> > >> >
> > >> >> Ethan, have you considered Getty's Thesaurus of Geographic Names?
>  It
> > >> >> does provide a geographic hierarchy, although the data for 
> > >> >> Athens they provide isn't quite the one you've described:
> > >> >>
> > >> >> http://www.getty.edu/vow/**TGNHierarchy?find=athens&**
> > >> >> place=&nation=&prev_page=1&**english=Y&subjectid=7001393<
> > http://www.g
> > >> >>
> > etty.edu/vow/TGNHierarchy?find=athens&place=&nation=&prev_page=1&eng
> > l
> > >> >> ish=Y&subjectid=7001393>
> > >> >>
> > >> >> This vocabulary is available in XML here:
> > >> >>
> > >> >>
&

Re: [CODE4LIB] Representing geographic hiearchy in linked data

2012-04-08 Thread Michael Hopwood
Sorry, a more useful response would have been - have you tried asking the 
CIDOC-CRM people? www.cidoc-crm.org/ - it seems to me unbelievable that they 
haven't hit up against this problem in their extensive work on museum / 
heritage / history ontologies.

The LIDO schema I'm working with has explicit relations between parts of places 
in various shapes and forms, and it is based directly on CIDOC-CRM, so there is 
some precedent there. I haven't looked up where the relationships come from but 
you could potentially derive some should they not yet be in use in someone's 
code list (probably in a museum context?)...

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Ethan 
Gruber
Sent: 08 April 2012 15:44
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Representing geographic hiearchy in linked data

Hi,

Thanks for the info, but it's not quite what I'm looking for.  We've 
established authority control for ancient places, but I'm looking for an 
ontology I can use to describe the child:parent relationship between city and 
region or region and larger region (in any way that isn't dcterms:partOf).  
Geonames has defined their own vocabulary that can't really be reused in other 
geographic contexts, e.g. with gn:countryCode, gn:parentCountry.

Thanks,
Ethan

On Fri, Apr 6, 2012 at 11:40 AM, Karen Coyle  wrote:

> Also, there is Geonames (http://www.geonames.org), which is the 
> primary geographic data set on the Semantic Web. Here is the link to Athens:
>
> http://www.geonames.org/**search.html?q=athens&country=**GR .geonames.org/search.html?q=athens&country=GR>
>
> kc
>
>
> On 4/6/12 4:54 PM, Karen Miller wrote:
>
>> Ethan, have you considered Getty's Thesaurus of Geographic Names?  It 
>> does provide a geographic hierarchy, although the data for Athens 
>> they provide isn't quite the one you've described:
>>
>> http://www.getty.edu/vow/**TGNHierarchy?find=athens&**
>> place=&nation=&prev_page=1&**english=Y&subjectid=7001393> etty.edu/vow/TGNHierarchy?find=athens&place=&nation=&prev_page=1&engl
>> ish=Y&subjectid=7001393>
>>
>> This vocabulary is available in XML here:
>>
>> http://www.getty.edu/research/**tools/vocabularies/obtain/**index.htm
>> l
>>
>> I have looked at it but not used it; it's a big tangled mess of XML.
>>
>> MODS mimics a hierarchy (the subject/hierarchicalGeographic element 
>> has these children: continent, country, province, region, state, 
>> territory, county, city, island, area, extraterrestrialArea, 
>> citySection). The VRA Core location element provides a similar mapping.
>>
>> I try to stay away from Dublin Core, but I did venture onto the DC 
>> Terms page just now and saw TGN listed in the vocabulary encoding 
>> schemes there, so probably someone has implemented it.
>>
>> Karen
>>
>>
>> Karen D. Miller
>> Monographic/Digital Projects Cataloger Bibliographic Services Dept.
>> Northwestern University Library
>> Evanston, IL
>> k-mill...@northwestern.edu
>> 847-467-3462
>>
>>
>>
>>
>> -Original Message-
>> From: Code for Libraries 
>> [mailto:code4...@listserv.nd.**EDU]
>> On Behalf Of Ethan Gruber
>> Sent: Thursday, April 05, 2012 12:49 PM
>> To: CODE4LIB@LISTSERV.ND.EDU
>> Subject: [CODE4LIB] Representing geographic hiearchy in linked data
>>
>> Hi all,
>>
>> I have a dilemma that needs to be sorted out.  I'm looking for an
>> ontology that can describe geographic hierarchy, and hopefully someone on
>> the list has experience with this.  For example, if I have an RDF record
>> that describes Athens, I want to point Athens to Attica, and Attica to
>> Greece, and so on.  The current proposal is to use dcterms:partOf, but the
>> problem with this is that our records will also use dcterms:partOf to
>> describe a completely different type of relational concept, and it will be
>> almost impossible for scripts to recognize the difference between these two
>> uses of the same DC term.
>>
>> Thanks,
>> Ethan
>>
>
> --
> Karen Coyle
> kco...@kcoyle.net http://kcoyle.net
> ph: 1-510-540-7596
> m: 1-510-435-8234
> skype: kcoylenet
>


Re: [CODE4LIB] Representing geographic hiearchy in linked data

2012-04-08 Thread Michael Hopwood
I think this highlights the point that, at some point, you have to model the 
data.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Ethan 
Gruber
Sent: 08 April 2012 15:44
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Representing geographic hiearchy in linked data

Hi,

Thanks for the info, but it's not quite what I'm looking for.  We've 
established authority control for ancient places, but I'm looking for an 
ontology I can use to describe the child:parent relationship between city and 
region or region and larger region (in any way that isn't dcterms:partOf).  
Geonames has defined their own vocabulary that can't really be reused in other 
geographic contexts, e.g. with gn:countryCode, gn:parentCountry.

Thanks,
Ethan

On Fri, Apr 6, 2012 at 11:40 AM, Karen Coyle  wrote:

> Also, there is Geonames (http://www.geonames.org), which is the 
> primary geographic data set on the Semantic Web. Here is the link to Athens:
>
> http://www.geonames.org/**search.html?q=athens&country=**GR .geonames.org/search.html?q=athens&country=GR>
>
> kc
>
>
> On 4/6/12 4:54 PM, Karen Miller wrote:
>
>> Ethan, have you considered Getty's Thesaurus of Geographic Names?  It 
>> does provide a geographic hierarchy, although the data for Athens 
>> they provide isn't quite the one you've described:
>>
>> http://www.getty.edu/vow/**TGNHierarchy?find=athens&**
>> place=&nation=&prev_page=1&**english=Y&subjectid=7001393> etty.edu/vow/TGNHierarchy?find=athens&place=&nation=&prev_page=1&engl
>> ish=Y&subjectid=7001393>
>>
>> This vocabulary is available in XML here:
>>
>> http://www.getty.edu/research/**tools/vocabularies/obtain/**index.htm
>> l
>>
>> I have looked at it but not used it; it's a big tangled mess of XML.
>>
>> MODS mimics a hierarchy (the subject/hierarchicalGeographic element 
>> has these children: continent, country, province, region, state, 
>> territory, county, city, island, area, extraterrestrialArea, 
>> citySection). The VRA Core location element provides a similar mapping.
>>
>> I try to stay away from Dublin Core, but I did venture onto the DC 
>> Terms page just now and saw TGN listed in the vocabulary encoding 
>> schemes there, so probably someone has implemented it.
>>
>> Karen
>>
>>
>> Karen D. Miller
>> Monographic/Digital Projects Cataloger Bibliographic Services Dept.
>> Northwestern University Library
>> Evanston, IL
>> k-mill...@northwestern.edu
>> 847-467-3462
>>
>>
>>
>>
>> -Original Message-
>> From: Code for Libraries 
>> [mailto:code4...@listserv.nd.**EDU]
>> On Behalf Of Ethan Gruber
>> Sent: Thursday, April 05, 2012 12:49 PM
>> To: CODE4LIB@LISTSERV.ND.EDU
>> Subject: [CODE4LIB] Representing geographic hiearchy in linked data
>>
>> Hi all,
>>
>> I have a dilemma that needs to be sorted out.  I'm looking for an
>> ontology that can describe geographic hierarchy, and hopefully someone on
>> the list has experience with this.  For example, if I have an RDF record
>> that describes Athens, I want to point Athens to Attica, and Attica to
>> Greece, and so on.  The current proposal is to use dcterms:partOf, but the
>> problem with this is that our records will also use dcterms:partOf to
>> describe a completely different type of relational concept, and it will be
>> almost impossible for scripts to recognize the difference between these two
>> uses of the same DC term.
>>
>> Thanks,
>> Ethan
>>
>
> --
> Karen Coyle
> kco...@kcoyle.net http://kcoyle.net
> ph: 1-510-540-7596
> m: 1-510-435-8234
> skype: kcoylenet
>


Re: [CODE4LIB] presenting merged records?

2012-03-30 Thread Michael Hopwood
Hi Peter, Graham,

I'm interested! Myself and a few friends and colleagues in the UK here are 
looking at this and similar questions as part of an ongoing interest in linked 
library data, and the schema/data model issues that underlie how useful it 
could be.

What we have here is certainly related to the data model (mostly implicit) 
behind MaRC records, which is an interesting blend of expression, manifestation 
and maybe item-level data.

Options for merging, linking and otherwise combining various records is of 
interest to us both "theoretically" and practically i.e. technically, so we 
would be pleased to contribute to and benefit from the conversation on this.

Best wishes,

Michael Hopwood

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Peter 
Noerr
Sent: 30 March 2012 01:09
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] presenting merged records?

Hi Graham,

What we do in our federated search system, and have been doing for some few 
years, is basically give the "designer" a choice of what options the user gets 
for "de-duped" records.

Firstly de-duping can be of a number of levels of sophistication, and a many of 
them lead to the situation you have - records which are "similar" rather than 
identical. On the web search side of things there are a surprising number of 
real duplicates (well maybe not surprising if you study more than one page of 
web search engine results), and on Twitter the duplicates well outnumber the 
original posts (many thanks 're-tweet').

Where we get duplicate records the usual options are: 1) keep the first and 
just drop all the rest. 2) keep the largest (assumed to have the most 
information) and drop the rest. These work well for WSE results where they are 
all almost identical (the differences often are just in the advertising 
attached to the pages and the results), but not for bibliographic records.

Less draconian is 3) Mark all the duplicates and keep them in the list (so you 
get 1, 2, 3, 4, 5, 5.1, 5.2, 5.3, 6, ...). This groups all the similar records 
together under the sort key of the first one, and does enable the user to 
easily skip them.

More user friendly is 4) Mark all duplicates and hide them in a sub-list 
attached to the "head" record. This gets them out of the main display, but 
allows the user who is interested in that "record" to expand the list and see 
the variants. This could be of use to you.

After that we planned to do what you are proposing and actually merge record 
content into a single virtual record, and worked on algorithms to do it. But 
nobody was interested. All our partners (who provide systems to lots of 
libraries, both public, academic, and special) decided that it would confuse 
their users more than it would help. I have my doubts, but they spoke and we 
put the development on ice.

I'm not sure this will help, but it has stood the test of time, and is well 
used in its various guises. Since no-one else seems interested in this topic, 
you could email me off list and we could discuss what we worked through in the 
way of algorithms, etc.

Peter


> -Original Message-
> From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf 
> Of graham
> Sent: Wednesday, March 28, 2012 8:05 AM
> To: CODE4LIB@LISTSERV.ND.EDU
> Subject: Re: [CODE4LIB] presenting merged records?
> 
> Hi Michael
> 
> On 03/27/12 11:50, Michael Hopwood wrote:
> > Hi Graham, do I know you from RHUL?
> >
> Yes indeed :-)
> 
> > My thoughts on "merged records" would be:
> >
> > 1. don't do it - use separate IDs and just present links between 
> > related manifestations; thus
> avoiding potential confusions.
> 
> In my case, I can't avoid it as it's a specific requirement: I'm doing 
> a federated search across a large number of libraries, and if closely 
> similar items aren't merged, the results become excessively large and 
> repetitive. I'm merging all the similar items, displaying a summary of 
> the merged bibliographic data, and providing links to each of the 
> libraries with a copy.  So it's not really FRBRization in the normal 
> sense, I just thought that FRBRization would lead to similar problems, so 
> that there might be some well-known discussion of the issues around... The 
> merger of the records does have advantages, especially if some libraries have 
> very underpopulated records (eg subject fields).
> 
> Cheers
> Graham
> 
> >
> > http://www.bic.org.uk/files/pdfs/identification-digibook.pdf
> >
> > possible relationships - see 
> > http://www.editeur.org/ONIX/book/codelists/current.html - lists 51
> (manifestation)and 164 (work).
> >
> > 2. c.f. the way Amazon displa

[CODE4LIB] FW: open-humanities Digest, Vol 20, Issue 11

2012-03-27 Thread Michael Hopwood
Just thought it may interest some :)

Date: Mon, 26 Mar 2012 23:19:20 -0700
From: Jonathan Gray 
Subject: [open-humanities] On Machine Readable Reading Lists
To: open-philosophy 
Cc: open-humanities 
Message-ID:

Content-Type: text/plain; charset=ISO-8859-1

Some thoughts on creating machine readable reading lists - in the first 
instance for philosophy departments in the UK:

http://jonathangray.org/2012/03/26/on-machine-readable-reading-lists/

J.

--
Jonathan Gray
http://jonathangray.org



--

___
open-humanities mailing list
open-humanit...@lists.okfn.org
http://lists.okfn.org/mailman/listinfo/open-humanities


End of open-humanities Digest, Vol 20, Issue 11
***


Re: [CODE4LIB] presenting merged records?

2012-03-27 Thread Michael Hopwood
Hi Graham, do I know you from RHUL?

My thoughts on "merged records" would be:

1. don't do it - use separate IDs and just present links between related 
manifestations; thus avoiding potential confusions.

http://www.bic.org.uk/files/pdfs/identification-digibook.pdf

possible relationships - see 
http://www.editeur.org/ONIX/book/codelists/current.html - lists 51 
(manifestation)and 164 (work).

2. c.f. the way Amazon displays rough and ready categories (paperback, 
hardback, audiobooks, *ahem* ebooks of some sort...)

On dissection and reconstitution of records - there is a lot of talk going on 
about RDFizing MaRC records and re-using in various ways, e.g.:

http://www.slideshare.net/JenniferBowen/moving-library-metadata-toward-linked-data-opportunities-provided-by-the-extensible-catalog

Cheers,

Michael

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of graham
Sent: 27 March 2012 11:06
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] presenting merged records?

Hi

There seems to be a general trend to presenting merged records to users, as 
part of the move towards FRBRization. If records need merging this generally 
means they weren't totally identical to start with, so you can end up with 
conflicting bibliographic data to display.

Two examples I've come across with this: Summon can merge print/electronic 
versions of texts, so uses a new 'merged' material type of 'book/ebook' (it 
doesn't yet seem to have all the other possible permutations, eg 
book/audiobook). Pazpar2 (which I'm working with at the
moment) has a merge option for publication dates which presents dates as a 
period eg 1997-2002.

The problem is not with the underlying data (the original unmerged values can 
still be there in the background) but how to present them to the user in an 
intuitive way. With the date example, presenting dates in this format sometimes 
throws people as it looks too much like the author birth/death dates you might 
see with a record.

I guess people must generally be starting to run into this kind of display 
problem, so it has maybe been discussed to death on ... wherever it is people 
talk about FRBRIzation. Any suggestions? Any mailing lists, blogs etc any can 
recommend for me to look at?

Thanks for any ideas
Graham


Re: [CODE4LIB] Learning Microsoft SQL

2012-03-19 Thread Michael Hopwood
Hi Bill,

A google search limited to US universities' sites like "sql basics site:.edu" 
turned up some nice results, including:

http://www.suu.edu/it/admin/summit2007/cdsite/color/BTechnical/1414/1414.pdf

http://myweb.brooklyn.liu.edu/gnarra/database/downloads/SQLTutorial.pdf

Michael H

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Wilfred 
Drew
Sent: 16 March 2012 18:31
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] Learning Microsoft SQL

I am setting up my laptop to teach myself Microsoft SQL. I am installing SQL 
Express. I purchased "Microsoft SQL Server 2008 All-in-one desk reference for 
Dummies." Any suggestions on other tools to add to my laptop to learn SQL?  
Preferably free.

Bill Drew
Web: BillTheLibrarian.com
Voice/SMS/: 607-745-4461
Email: bill.d...@gmail.com
G+: gplus.to/BillDrew
Twitter/Skype: BillDrew4

Web Design, Social Media,
New Tech, Assessment,
Change Management, Innovation, Mobile Tech, and more.


[cid:image001.png@01CD0381.754C6DA0]


Re: [CODE4LIB] NON-MARC ILS?

2012-03-14 Thread Michael Hopwood
That was my impression...

I should have mentioned the efforts just starting within my current project to 
map MaRC (or flavour[s] thereof!) to the open LIDO standard:

http://network.icom.museum/cidoc/working-groups/data-harvesting-and-interchange.html

It's interesting in that it will hopefully express a library standard (or 
standardS; the situation is complicated in Europe by the many MaRC variants in 
use) in terms of a schema developed for museum objects.

In theory this is definitely possible (LIDO is built to cope with any type of 
objects, including "Manifestation Product Types" [cf. 
http://www.cidoc-crm.org/docs/frbr_oo/frbr_docs/FRBRoo_V1.0_2009_june_.pdf]) 
but it will be a challenge.

M

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Owen, 
Will
Sent: 14 March 2012 13:38
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] NON-MARC ILS?

The Endeca implementation at the Triangle Research Libraries Network (and 
indeed, in general) is an *index* of information about items in our libraries' 
collections.  The format of the data that's fed into the index can be (and is) 
variable: we're about to start loading items from our digital collections that 
are MARC-based into our Endeca index, where they will co-exist with traditional 
bibliographic information that comes out of our ILS systems.

Endeca provides a public interface to library holdings: it is not and could not 
be an ILS, performing functions like circulation, accounting control, etc.  In 
this respect it's more akin to Blacklight that to an ILS.

Will


On 3/14/12 9:28 AM, "Michael Hopwood"  wrote:

>Matt,
>
>Looks to me at a cursory glance that at least one Endeca implementation 
>is still drawing on MaRC data:
>
>http://www.lib.ncsu.edu/endeca/technology.html
>
>...even if not directly using MaRC for search:
>
>http://www.lib.ncsu.edu/endeca/faqs.html
>
>MaRC is simply the widest-used library standard. If you could get hold 
>of ONIX (2.1 or 3.0) feeds, you could in theory build an LMS around 
>that instead, or any other data format you prefer.
>
>There is at least a fair amount of interest on interoperating at least 
>these 2 major formats:
>
>http://www.oclc.org/research/news/2010-04-09.htm (ONIX 3.0 version 
>pending).
>
>-Original Message-
>From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of 
>Matt Amory
>Sent: 14 March 2012 13:00
>To: CODE4LIB@LISTSERV.ND.EDU
>Subject: [CODE4LIB] NON-MARC ILS?
>
>Is there a full-featured ILS that is not based on MARC records?
>I know we love complexity, but it seems to me that my public library 
>and its library network and maybe even every public library could 
>probably do without 95% of MARC Fields and encoding, streamline 
>workflows and save $ if there were a simpler standard.
>Is this what an Endeca-based system is about, or do those rare birds 
>also use MARC in the background?
>Forgive me if the question has been hashed and rehashed over the years...
>
>--
>Matt Amory
>(917) 771-4157
>matt.am...@gmail.com
>http://www.linkedin.com/pub/matt-amory/8/515/239


Re: [CODE4LIB] NON-MARC ILS?

2012-03-14 Thread Michael Hopwood
Matt,

Looks to me at a cursory glance that at least one Endeca implementation is 
still drawing on MaRC data:

http://www.lib.ncsu.edu/endeca/technology.html

...even if not directly using MaRC for search:

http://www.lib.ncsu.edu/endeca/faqs.html

MaRC is simply the widest-used library standard. If you could get hold of ONIX 
(2.1 or 3.0) feeds, you could in theory build an LMS around that instead, or 
any other data format you prefer.

There is at least a fair amount of interest on interoperating at least these 2 
major formats:

http://www.oclc.org/research/news/2010-04-09.htm (ONIX 3.0 version pending).

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Matt 
Amory
Sent: 14 March 2012 13:00
To: CODE4LIB@LISTSERV.ND.EDU
Subject: [CODE4LIB] NON-MARC ILS?

Is there a full-featured ILS that is not based on MARC records?
I know we love complexity, but it seems to me that my public library and its 
library network and maybe even every public library could probably do without 
95% of MARC Fields and encoding, streamline workflows and save $ if there were 
a simpler standard.
Is this what an Endeca-based system is about, or do those rare birds also use 
MARC in the background?
Forgive me if the question has been hashed and rehashed over the years...

--
Matt Amory
(917) 771-4157
matt.am...@gmail.com
http://www.linkedin.com/pub/matt-amory/8/515/239


Re: [CODE4LIB] Metadata

2012-02-14 Thread Michael Hopwood
Having done research, and now working in a very varied metadata role, I don't 
quite understand this discussion about data that is or isn't metadata. 
Scientific data is a great example of structured data, but it's not impossible 
to distinguish it from metadata purely describing a dataset.

However, if you have scientific research data created during the experiments, 
even if it's "operational", it's clearly part of "the" data. This doesn't mean 
there can't be metadata describing *that data*. Just because it's not glamorous 
data doesn't mean it's not essential to the scientific process. Similarly, just 
being about mundane or procedural things doesn't make data into metadata...!

You're absolutely right, the contextual information is certainly part of the 
experimental outcome in this example; otherwise it would be abstract data such 
as one might use in a textbook example.

Metadata would describe the dataset itself, not the scientific research. 
There's always a certain ambiguity involved in identifying "the data" as 
distinct from the metadata, and it's a false dichotomy to suggest metadata is 
not useful at all for the domain expert. It's contextual, and the definition is 
always at least partly based on your use case for the data and its description.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Nate 
Vack
Sent: 14 February 2012 14:45
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Metadata

On Tue, Feb 14, 2012 at 1:22 AM, Graham Triggs  wrote:

> That's an interesting distinction though. Do you need all that data in 
> order to make sense of the results? You don't [necessarily] need to 
> know who conducted some research, or when they conducted it in order 
> to analyse and make sense of the data. In the context of having the 
> data, this other information becomes irrelevant in terms of 
> understanding what that data says.

It is *essential* to understanding what the data says. Perhaps you find out 
your sensor was on the fritz during a time period -- you need to be able to 
know what datasets are suspect. Maybe the blood pressure effect you're looking 
at is mediated by circadian rhythms, and hence, times of day.

Not all of your data is necessary in every analysis, but a bunch of blood 
pressure measurements in the absence of contextual information is universally 
useless.

The metadata is part of the data.

-n


Re: [CODE4LIB] Metadata

2012-02-14 Thread Michael Hopwood
>> The actual data (the novel) is not in the catalog (which is composed only of 
>> metadata).

>>> That's a technical limitation.

It's also a legal/commercial limitation, as well as a question of provenance.

To summarise a lot of good points made already:

" An item of metadata is a relationship that someone claims to exist between 
two entities."
 - source: http://www.doi.org/topics/indecs/indecs_framework_2000.pdf

Cheers,

M

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Nate 
Vack
Sent: 13 February 2012 22:39
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Metadata

On Mon, Feb 13, 2012 at 4:25 PM, Genny Engel  wrote:

> You simply can't use the average library catalog to look up Author X's novel 
> that starts with the sentence "So a string walks into a bar."  The actual 
> data (the novel) is not in the catalog (which is composed only of metadata).

That's a technical limitation. If you're Google Books (or any other fulltext 
index), the actual data *is* in the catalog, and data and metadata are again 
functionally identical.

The best working definition of metadata I've come up with is "something I have 
a field for in my data cataloging program."

I think it's kind of a circular issue: We know metadata and data are separate 
because our software and workflow require it. Software and workflows are 
designed to separate metadata and data because we know they're separate.

-n


Re: [CODE4LIB] Job: Head, Digital Projects & Metadata, Beinecke Rare Book and Manuscript Library at Yale University

2012-02-08 Thread Michael Hopwood
If anyone's interested in a view from across the pond, I'm a qualified 
librarian with an MLIS-equivalent from the UK... I feel there's a significant 
grey area between "library" and "IT", which partly originates in the failures 
of both professional areas to address some of the areas of overlap, or 
basically to keep up with the times.

Having studied physics at university and learned basically how to build 
computers from the ground up, as well as program them and use them in 
"real-world" contexts, and then made the jump into the "softer" world of 
information/library management, with its concepts of "information literacy" and 
more or less organically-developed classification structures, and now working 
in the commercial world on (meta)data and identifier standards 
interoperability, I've found that there are plenty of these dichotomies in the 
working world too (although in commercial data you can find a surprising level 
of coherence and universality that was a bit of a utopian dream back on Library 
World).

The sooner we build bridges of understanding, standards and systems across 
these divides, the better.

-Original Message-
From: Code for Libraries [mailto:CODE4LIB@LISTSERV.ND.EDU] On Behalf Of Carol 
Bean
Sent: 08 February 2012 13:52
To: CODE4LIB@LISTSERV.ND.EDU
Subject: Re: [CODE4LIB] Job: Head, Digital Projects & Metadata, Beinecke Rare 
Book and Manuscript Library at Yale University

Sometimes it is politically complex.  Here, in the Federal Courts, there are 
good reasons for distinguishing between library and IT, and the library degree 
keeps the job in the realm of the library (when it comes to turf wars), which 
is a good thing.

The position I am about to leave will (hopefully) be posted soon.  I wrote up 
the job requirements, requiring a library degree, specifically distinguishing 
it from the type of work typically done by IT, although part of the job will be 
doing some IT help-desk type work. Having worked intimately with IT the last 
six months, I am convinced they just don't get it the way library people do, 
and the only way to ensure the position gets filled by a library-type person, 
in this situation, is to require the degree.

Carol

On Tue, Feb 7, 2012 at 4:46 PM, Hugh Cayless  wrote:

> I can't speak for Yale, of course, but when I worked for UNC, there 
> were requirements in place set by General Administration that 
> "Librarians" had to have library degrees, and they were very picky 
> about it. It's unnecessarily exclusionary for most tech-in-libraries 
> positions in my opinion. Institutional cultures are slow to recognize 
> the need for change-and the Library itself may not be responsible for the 
> requirement.
>
> H
>
> On Feb 7, 2012, at 4:27PM, Ethan Gruber wrote:
>
> > Why are MLS degrees always required for these sorts of jobs?
> >
> > Ethan
> >
> > On Tue, Feb 7, 2012 at 4:21 PM,  wrote:
> >
> >> Yale University offers exciting opportunities for achievement and
> growth in
> >> New Haven, Connecticut. Conveniently located between Boston and New
> York,
> >> New
> >> Haven is the creative capital of Connecticut with cultural 
> >> resources
> that
> >> include two major art museums, a critically-acclaimed repertory 
> >> theater, state-of-the-art concert hall, and world-renowned schools 
> >> of
> Architecture,
> >> Art, Drama, and Music.
> >>
> >> **The University and the Library**
> >> The Yale University Library, as one of the world's leading research 
> >> libraries, collects, organizes, preserves, and provides access to 
> >> and services for
> a
> >> rich
> >> and unique record of human thought and creativity. It fosters
> intellectual
> >> growth and is a highly valued partner in the teaching and research 
> >> missions of Yale University and scholarly communities worldwide. A 
> >> distinctive strength is its rich spectrum of resources, including 
> >> more than 12.5 million volumes and information in all media, 
> >> ranging from ancient papyri to early printed books to electronic 
> >> databases. The Library is engaged in numerous digital initiatives 
> >> designed to provide access to a full array of scholarly 
> >> information. Housed in the Sterling Memorial Library and twenty 
> >> school
> and
> >> departmental libraries, it employs a dynamic, diverse, and 
> >> innovative staff of over 500who have the opportunity to work with 
> >> the highest caliber of faculty and students, participate on 
> >> committees, and are involved in other
> areas of
> >> staff development. For additional information on the Yale 
> >> University Library, please visit the Library's web site 
> >> at[http://www.library.y ale.edu/](http://www.library.yale.edu/).
> >>
> >> **Beinecke Rare Book and Manuscript Library** The Beinecke Library 
> >> is Yale's principal repository for literary papers
> and
> >> early manuscripts and rare books. In addition to distinguished 
> >> general collections, the library houses the Osborn Collection, 
> >> noted for its Britis