On 4/29/10 12:32 PM, MJ Suhonos wrote:
>> What I hope for is that OpenURL 1.0 eventually takes a place alongside SGML as a too-complex standard that directly paves the way for a universally adopted foundational technology like XML. What I fear is that it takes a place alongside MARC as an anachronistic standard that paralyzes an entire industry.
>>
> Hear hear.
>
> I'm actually encouraged by Benjamin's linking (har har) to the httpRange-14 issue as being relevant to the concept of "link resolution", or at least redirection (indirection?) using URL surrogates for resources. Many are critical of the TAG's "resolution" (har har har) of the issue, and think it places too much on the 303 redirect.
>
> I'm afraid I still don't understand the issue fully enough to comment — though I'd love to hear from any who can. I agree with Eric's hope that the library world can look to W3C's thinking to inform a "better way" forward for link resolving, though.
>
One key thing to remember with the W3C work is that URL's have to be
dereference-able. I can't lookup (without an OpenLink resolver or Google
or the like) a url:isbn:{whatever}, but I can dereference
http://dbpedia.org/resource/The_Lord_of_the_Rings -- which 303's to
http://dbpedia.org/page/The_Lord_of_the_Rings -- which is full of more
/resource/ dereferencable (via 303 See Other) URL's.
The main thing that the W3C was trying to avoid was RDF that
inadvertently talks about online documents when what it really wants to
talk about is the "real thing." Real things (like books) need a URI, but
ideally a URI that can be dereferenced (via HTTP in this case) to give
them some information about that real thing--which isn't possible with
the urn:isbn style schemes.
That's my primitive understanding of it anyway. Apologies if any
overlaps with library tech are off. :)
|