Interesting -- thanks, Birkin -- and tell us what you think when you get it implemented! :) -Jodi On Mon, Sep 30, 2013 at 5:19 PM, Birkin Diana <[log in to unmask]>wrote: > > ...you'd want to create a caching service... > > > One solution for a relevant particular problem (not full-blown linked-data > caching): > > http://en.wikipedia.org/wiki/XML_Catalog > > excerpt: "However, if they are absolute URLs, they only work when your > network can reach them. Relying on remote resources makes XML processing > susceptible to both planned and unplanned network downtime." > > We'd heard about this a while ago, but, Jodi, you and David Riordan and > Congress have caused a temporary retreat from normal sprint-work here at > Brown today to investigate implementing this! :/ > > The particular problem that would affect us: if your processing tool > checks, say, an loc.gov mods namespace url, that processing will fail if > the loc.gov url isn't available, unless you've implemented xml catalog, > which is a formal way to locally resolve such external references. > > -b > --- > Birkin James Diana > Programmer, Digital Technologies > Brown University Library > [log in to unmask] > > > On Sep 30, 2013, at 7:15 AM, Uldis Bojars <[log in to unmask]> wrote: > > > What are best practices for preventing problems in cases like this when > an > > important Linked Data service may go offline? > > > > --- originally this was a reply to Jodi which she suggested to post on > the > > list too --- > > > > A safe [pessimistic?] approach would be to say "we don't trust > [reliability > > of] linked data on the Web as services can and will go down" and to cache > > everything. > > > > In that case you'd want to create a caching service that would keep > updated > > copies of all important Linked Data sources and a fall-back strategy for > > switching to this caching service when needed. Like archive.org for > Linked > > Data. > > > > Some semantic web search engines might already have subsets of Linked > Data > > web cached, but not sure how much they cover (e.g., if they have all of > LoC > > data, up-to-date). > > > > If one were to create such a service how to best update it, considering > > you'd be requesting *all* Linked Data URIs from each source? An efficient > > approach would be to regularly load RDF dumps for every major source if > > available (e.g., LoC says - here's a full dump of all our RDF data ... > and > > a .torrent too). > > > > What do you think? > > > > Uldis > > > > > > On 29 September 2013 12:33, Jodi Schneider <[log in to unmask]> wrote: > > > >> Any best practices for caching authorities/vocabs to suggest for this > >> thread on the Code4Lib list? > >> > >> Linked Data authorities & vocabularies at Library of Congress ( > id.loc.gov) > >> are going to be affected by the website shutdown -- because of lack of > >> government funds. > >> > >> -Jodi >