A suggestion on how to get a prof to enter a url. I use this bookmarklet to add a URL to Hacker News: javascript:window.location=%22http://news.ycombinator.com/submitlink?u=%22+encodeURIComponent(document.location)+%22&t=%22+encodeURIComponent(document.title) I'm tempted to suggest an api based on OpenURL, but I fear the 10 emails it would provoke. On Sep 15, 2009, at 10:56 AM, Ross Singer wrote: > Owen, I might have missed it in this message -- my eyes are starting > glaze over at this point in the thread, but can you describe how the > input of these resources would work? > > What I'm basically asking is -- what would the professor need to do to > add a new: citation for a 70 year old book; journal on PubMed; URL to > CiteSeer? > > How does their input make it into your database? > > -Ross. > > On Tue, Sep 15, 2009 at 5:04 AM, O.Stephens <[log in to unmask]> > wrote: >>> True. How, from the OpenURL, are you going to know that the rft is >>> meant >>> to represent a website? >> I guess that was part of my question. But no one has suggested >> defining a new metadata profile for websites (which I probably >> would avoid tbh). DC doesn't seem to offer a nice way of doing this >> (that is saying 'this is a website'), although there are perhaps >> some bits and pieces (format, type) that could be used to give some >> indication (but I suspect not unambiguously) >> >>> But I still think what you want is simply a purl server. What >>> makes you >>> think you want OpenURL in the first place? But I still don't really >>> understand what you're trying to do: "deliver consistency of >>> approach >>> across all our references" -- so are you using OpenURL for it's more >>> "conventional" use too, but you want to tack on a purl-like >>> functionality to the same software that's doing something more >>> like a >>> conventional link resolver? I don't completely understand your >>> use case. >> >> I wouldn't use OpenURL just to get a persistent URL - I'd almost >> certainly look at PURL for this. But, I want something slightly >> different. I want our course authors to be able to use whatever URL >> they know for a resource, but still try to ensure that the link >> works persistently over time. I don't think it is reasonable for a >> user to have to know a 'special' URL for a resource - and this >> approach means establishing a PURL for all resources used in our >> teaching material whether or not it moves in the future - which is >> an overhead it would be nice to avoid. >> >> You can hit delete now if you aren't interested, but ... >> >> ... perhaps if I just say a little more about the project I'm >> working on it may clarify... >> >> The project I'm working on is concerned with referencing and >> citation. We are looking at how references appear in teaching >> material (esp. online) and how they can be reused by students in >> their personal environment (in essays, later study, or something >> else). The references that appear can be to anything - books, >> chapters, journals, articles, etc. Increasingly of course there are >> references to web-based materials. >> >> For print material, references generally describe the resource and >> nothing more, but for digital material references are expected not >> only to describe the resource, but also state a route of access to >> the resource. This tends to be a bad idea when (for example) >> referencing e-journals, as we know the problems that surround this >> - many different routes of access to the same item. OpenURLs work >> well in this situation and seem to me like a sensible (and perhaps >> the only viable) solution. So we can say that for journals/articles >> it is sensible to ignore any URL supplied as part of the reference, >> and to form an OpenURL instead. If there is a DOI in the reference >> (which is increasingly common) then that can be used to form a URL >> using DOI resolution, but it makes more sense to me to hand this >> off to another application rather than bake this into the reference >> - and OpenURL resolvers are reasonably set to do this. >> >> If we look at a website it is pretty difficult to reference it >> without including the URL - it seems to be the only good way of >> describing what you are actually talking about (how many people >> think of websites by 'title', 'author' and 'publisher'?). For me, >> this leads to an immediate confusion between the description of the >> resource and the route of access to it. So, to differentiate I'm >> starting to think of the http URI in a reference like this as a >> URI, but not necessarily a URL. We then need some mechanism to >> check, given a URI, what is the URL. >> >> Now I could do this with a script - just pass the URI to a script >> that checks what URL to use against a list and redirects the user >> if necessary. On this point Jonathan said "if the usefulness of >> your technique does NOT count on being inter-operable with existing >> link resolver infrastructure... PERSONALLY I would be using >> OpenURL, I don't think it's worth it" - but it struck me that if we >> were passing a URI to a script, why not pass it in an OpenURL? I >> could see a number of advantages to this in the local context: >> >> Consistency - references to websites get treated the same as >> references to journal articles - this means a single approach on >> the course side, with flexibility >> Usage stats - we could collect these whatever, but if we do it via >> OpenURL we get this in the same place as the stats about usage of >> other scholarly material and could consider driving personalisation >> services off the data (like the bX product from Ex Libris) >> Appropriate copy problem - for resources we subscribe to with >> authentication mechanisms there is (I think) an equivalent to the >> 'appropriate copy' issue as with journal articles - we can push a >> URI to 'Web of Science' to the correct version of Web of Science >> via a local authentication method (using ezproxy for us) >> >> The problem with the approach (as Nate and Eric mention) is that >> any approach that relies on the URI as a identifier (whether using >> OpenURL or a script) is going to have problems as the same URI >> could be used to identify different resources over time. I think >> Eric's suggestion of using additional information to help >> differentiate is worth looking at, but I suspect that this is going >> to cause us problems - although I'd say that it is likely to cause >> us much less work than the alternative, which is allocating every >> single reference to a web resource used in our course material it's >> own persistent URL. >> >> The use case we are currently looking at is only with our own >> (authenticated) learning environment - so these OpenURLs are not >> going to appear in the wild, so to some extent perhaps it doesn't >> matter what we do - but it still seems sensible to me to look at >> what 'good practice' might look like. >> >> I hope this is clear - I'm still struggling with some of this, and >> sometimes it doesn't make complete sense to me, but that's my best >> stab at explaining my thinking at the moment. Again, I appreciate >> the comments. Jonathan said "But you seem to understand what's up". >> I wish I did! I guess that I'm reasonably confident that the >> approach I'm describing has some chance of doing the job - whether >> it is the best approach I'm not so sure about. >> >> Owen >> >> >> The Open University is incorporated by Royal Charter (RC 000391), >> an exempt charity in England & Wales and a charity registered in >> Scotland (SC 038302). >> Eric Hellman President, Gluejar, Inc. 41 Watchung Plaza, #132 Montclair, NJ 07042 USA [log in to unmask] http://go-to-hellman.blogspot.com/