Thanks for this pointer Owen.
It's a nice illustration of the fact that what users actually want (well, I know I did back when I actually worked in large information services departments!) is something more like an intranet where the content I find is weighted towards me, the "audience" e.g. the intranet knows I'm a 2nd year medical student and one of my registered preferred languages is Mandarin already, or it "knows" that I'm a rare books cataloguer and I want to see what "nine out of ten" other cataloguers recorded for this obscure and confusing title.
However, this stuff is quite intense for linked data, isn't it? I understand that it would involve lots of quads, named graphs or whatever...
In a parallel world, I'm currently writing up recommendations for aggregating ONIX for Books records. ONIX data can come from multiple sources who potentially assert different things about a given "book" (i.e. something with an ISBN to keep it simple).
This is why *every single ONIX data element* can have option attributes of
@sourcetype [e.g. publisher, retailer, data aggregator... library?]
...and the ONIX message as a whole is set up with "header" and "product record" segments that each include some info about the sender/recipient/data record in question.
How people in the book supply chain are implementing these is a distinct issue, but could these capabilities have some relevance to what you're discussing?
Do you have any other pointers to "intranet-like" catalogues?
In the museum space, there is of course this: http://www.researchspace.org/
From: Code for Libraries [mailto:[log in to unmask]] On Behalf Of Owen Stephens
Sent: 28 August 2012 21:37
To: [log in to unmask]
Subject: Re: [CODE4LIB] Corrections to Worldcat/Hathi/Google
The JISC funded CLOCK project did some thinking around cataloguing processes and tracking changes to statements and/or records - e.g.
Not solutions of course, but hopefully of interest
Owen Stephens Consulting
Email: [log in to unmask]
Telephone: 0121 288 6936
On 28 Aug 2012, at 19:43, Simon Spero <[log in to unmask]> wrote:
> On Aug 28, 2012, at 2:17 PM, Joe Hourcle wrote:
>> I seem to recall seeing a presentation a couple of years ago from someone in the intelligence community, where they'd keep all of their intelligence, but they stored RDF quads so they could track the source.
>> They'd then assign a confidence level to each source, so they could get an overall level of confidence on their inferences.
>> It's possible that it was in the context of provenance, but I'm getting bogged down in too many articles about people storing provenance information using RDF-triples (without actually tracking the provenance of the triple itself)
> Provenance is of great importance in the IC and related sectors.
> An good overview of the nature of evidential reasoning is David A Schum (1994;2001). Evidential Foundations of Probabilistic Reasoning. Wiley & Sons, 1994; Northwestern University Press, 2001 [Paperback edition].
> There are usually papers on provenance and associated semantics at the GMU Semantic Technology for Intelligence, Defense, and Security (STIDS). This years conference is 23 - 26 October 2012; see http://stids.c4i.gmu.edu/ for more details.