FYI - this also means that there's a very good chance that the MARC standards site [1] and the Source Codes site [2] will be down as well. I don't know if there are any mirror sites out there for these pages. [1] http://www.loc.gov/marc/ [2] http://www.loc.gov/standards/sourcelist/index.html Thanks, Becky, about to be (forcefully) departed with her standards documentation On Mon, Sep 30, 2013 at 11:39 AM, Jodi Schneider <[log in to unmask]>wrote: > Interesting -- thanks, Birkin -- and tell us what you think when you get it > implemented! > > :) -Jodi > > > On Mon, Sep 30, 2013 at 5:19 PM, Birkin Diana <[log in to unmask] > >wrote: > > > > ...you'd want to create a caching service... > > > > > > One solution for a relevant particular problem (not full-blown > linked-data > > caching): > > > > http://en.wikipedia.org/wiki/XML_Catalog > > > > excerpt: "However, if they are absolute URLs, they only work when your > > network can reach them. Relying on remote resources makes XML processing > > susceptible to both planned and unplanned network downtime." > > > > We'd heard about this a while ago, but, Jodi, you and David Riordan and > > Congress have caused a temporary retreat from normal sprint-work here at > > Brown today to investigate implementing this! :/ > > > > The particular problem that would affect us: if your processing tool > > checks, say, an loc.gov mods namespace url, that processing will fail if > > the loc.gov url isn't available, unless you've implemented xml catalog, > > which is a formal way to locally resolve such external references. > > > > -b > > --- > > Birkin James Diana > > Programmer, Digital Technologies > > Brown University Library > > [log in to unmask] > > > > > > On Sep 30, 2013, at 7:15 AM, Uldis Bojars <[log in to unmask]> wrote: > > > > > What are best practices for preventing problems in cases like this when > > an > > > important Linked Data service may go offline? > > > > > > --- originally this was a reply to Jodi which she suggested to post on > > the > > > list too --- > > > > > > A safe [pessimistic?] approach would be to say "we don't trust > > [reliability > > > of] linked data on the Web as services can and will go down" and to > cache > > > everything. > > > > > > In that case you'd want to create a caching service that would keep > > updated > > > copies of all important Linked Data sources and a fall-back strategy > for > > > switching to this caching service when needed. Like archive.org for > > Linked > > > Data. > > > > > > Some semantic web search engines might already have subsets of Linked > > Data > > > web cached, but not sure how much they cover (e.g., if they have all of > > LoC > > > data, up-to-date). > > > > > > If one were to create such a service how to best update it, considering > > > you'd be requesting *all* Linked Data URIs from each source? An > efficient > > > approach would be to regularly load RDF dumps for every major source if > > > available (e.g., LoC says - here's a full dump of all our RDF data ... > > and > > > a .torrent too). > > > > > > What do you think? > > > > > > Uldis > > > > > > > > > On 29 September 2013 12:33, Jodi Schneider <[log in to unmask]> > wrote: > > > > > >> Any best practices for caching authorities/vocabs to suggest for this > > >> thread on the Code4Lib list? > > >> > > >> Linked Data authorities & vocabularies at Library of Congress ( > > id.loc.gov) > > >> are going to be affected by the website shutdown -- because of lack of > > >> government funds. > > >> > > >> -Jodi > > >