Print

Print


On 2/23/2012 3:53 PM, Karen Coyle wrote:
> Jonathan, while having these thoughts your Umlaut service did come to 
> mind.  If you ever have time to expand on how it could work in a wide 
> open web environment, I'd love to hear it. (I know you explain below, 
> but I don't know enough about link resolvers to understand what it 
> really means from a short 
> explanation. Diagrams are always welcome!)

I'm not entirely sure what is meant by 'wide open web environment.'

I mean, part of the current environment is that there's lots of stuff on 
the web that is NOT free/open access, it's only available to certain 
licensed people. AND that libraries license a lot of this stuff on 
behalf of their user group. (Not just content, but sometimes services 
too).  It's really that environment Umlaut is focused on, if that 
changed, what would be required would have little to do with Umlaut as 
it is now, I think.

But I don't think anyone anticipates that changing anytime soon, I don't 
think that's what Karen means by 'wide open web environment.'

So if that continues to be the case.... I think Umlaut has a role 
working pretty much as it does now, it would work how it works.  (Maybe 
I'm not sufficiently forward-thinking).

I will admit that, while I come accross lots of barriers in implementing 
Umlaut, I have yet to come accross anything that makes me think "this 
would be a lot easier if only there was more RDF."  Maybe it's a failure 
of imagination on my part.  More downloadable data, sure. More http 
APIs, even more so.  And Umlaut already takes advantage of such things, 
especially the API's more than the downloadable data (it turns out it's 
a lot more 'expensive' to try to download data and do something with it 
yourself, compared to using an API someone else provides to do the heavy 
lifting for you).  But has it ever been much of a problem that the data 
is in some format other than RDF, such that it would be easier in RDF? 
Not from my perspective, not really. (In some ways, RDF is harder to 
deal with than other formats, from where I'm coming from. If a service 
does offer data in RDF triples as well as something else, I'm likely to 
choose the something else).

This may be ironic because Umlaut is very concerned with 'linking data', 
in the sense of figuring out whether this record from the local catalog 
represents 'the same thing' as this record from Amazon, as this record 
from Google Books, or HathiTrust. If this citation that came in as an 
OpenURL represents the 'same thing' as a record in a vendor database, or 
mendeley, or whatever.

There are real barriers in making this determination; they wouldn't be 
solved if everything was just in RDF, but they _would_ be solved if 
there were more consistent use of identifiers, for sure. I DO think 
"this would be easier if only there were more consistent use of 
identifiers" all the time.

That experience with Umlaut is also what leads me to believe that the 
WEMI ontology is not only not contradictory to "linked data 
applications", but _crucial_ for it. Realizing that without it, it's 
very hard to tell when something is "the same thing". There are lots of 
times Umlaut ends up saying "Okay, I found something that I _think_ is 
at least an edition of the same thing you care about, but I really can't 
tell you if it's the _same_ edition you are interested in or not."

So, yeah, Umlaut would work _better_ with more widespread use 
identifiers, and even better with consistent use of common identifiers. 
I guess that's maybe where RDF could come in, in expressing 
determinations people have made of "this identifier in system X 
represents the same 'thing' as this other identifier in system Y" 
(someone would still have to MAKE those determinations, RDF would just 
be one way to then convey that determination, and I wouldn't 
particularly care if it was conveyed in RDF or something else). So 
anyway, it would work better with some of that stuff, but would it work 
substantially _differently_? Not so much.

Ah, if web pages started having more embedded machine readable data with 
citations and identifiers of "what is being looked at" (microdata, RDFa, 
whatever), that would make it easier to get a user from some random web 
page _to_ an institution's Umlaut, that's one thing that would be nice.

You may (or may not) find the "What is Umlaut, Anyway?" article on the 
Umlaut wiki helpful.

https://github.com/team-umlaut/umlaut/wiki/What-is-Umlaut-anyway


And there's really not much to understand about 'link resolvers' for 
these purposes, except that there's this thing called OpenURL (really 
bad name), which is really just a way for one website to hyperlink to 
another website and pass a machine-readable citation to it. This 
application receiving the machine readable citation then tries to get 
the user to appropriate access or services for it, with regard to 
institutional entitlements. That's about it, if you understand that, you 
understand enough.  Except that most commercially available 'link 
resolvers' do a so-so job with scholarly article citations and full 
text, and don't even try much with anything else. (In part, because it's 
_hard_, especially to provide an out of the box solution, because of 
libraries diverse messed up infrastructures).