FYI - this also means that there's a very good chance that the MARC
standards site [1] and the Source Codes site [2] will be down as well. I
don't know if there are any mirror sites out there for these pages.
[1] http://www.loc.gov/marc/
[2] http://www.loc.gov/standards/sourcelist/index.html
Thanks,
Becky, about to be (forcefully) departed with her standards documentation
On Mon, Sep 30, 2013 at 11:39 AM, Jodi Schneider <[log in to unmask]>wrote:
> Interesting -- thanks, Birkin -- and tell us what you think when you get it
> implemented!
>
> :) -Jodi
>
>
> On Mon, Sep 30, 2013 at 5:19 PM, Birkin Diana <[log in to unmask]
> >wrote:
>
> > > ...you'd want to create a caching service...
> >
> >
> > One solution for a relevant particular problem (not full-blown
> linked-data
> > caching):
> >
> > http://en.wikipedia.org/wiki/XML_Catalog
> >
> > excerpt: "However, if they are absolute URLs, they only work when your
> > network can reach them. Relying on remote resources makes XML processing
> > susceptible to both planned and unplanned network downtime."
> >
> > We'd heard about this a while ago, but, Jodi, you and David Riordan and
> > Congress have caused a temporary retreat from normal sprint-work here at
> > Brown today to investigate implementing this! :/
> >
> > The particular problem that would affect us: if your processing tool
> > checks, say, an loc.gov mods namespace url, that processing will fail if
> > the loc.gov url isn't available, unless you've implemented xml catalog,
> > which is a formal way to locally resolve such external references.
> >
> > -b
> > ---
> > Birkin James Diana
> > Programmer, Digital Technologies
> > Brown University Library
> > [log in to unmask]
> >
> >
> > On Sep 30, 2013, at 7:15 AM, Uldis Bojars <[log in to unmask]> wrote:
> >
> > > What are best practices for preventing problems in cases like this when
> > an
> > > important Linked Data service may go offline?
> > >
> > > --- originally this was a reply to Jodi which she suggested to post on
> > the
> > > list too ---
> > >
> > > A safe [pessimistic?] approach would be to say "we don't trust
> > [reliability
> > > of] linked data on the Web as services can and will go down" and to
> cache
> > > everything.
> > >
> > > In that case you'd want to create a caching service that would keep
> > updated
> > > copies of all important Linked Data sources and a fall-back strategy
> for
> > > switching to this caching service when needed. Like archive.org for
> > Linked
> > > Data.
> > >
> > > Some semantic web search engines might already have subsets of Linked
> > Data
> > > web cached, but not sure how much they cover (e.g., if they have all of
> > LoC
> > > data, up-to-date).
> > >
> > > If one were to create such a service how to best update it, considering
> > > you'd be requesting *all* Linked Data URIs from each source? An
> efficient
> > > approach would be to regularly load RDF dumps for every major source if
> > > available (e.g., LoC says - here's a full dump of all our RDF data ...
> > and
> > > a .torrent too).
> > >
> > > What do you think?
> > >
> > > Uldis
> > >
> > >
> > > On 29 September 2013 12:33, Jodi Schneider <[log in to unmask]>
> wrote:
> > >
> > >> Any best practices for caching authorities/vocabs to suggest for this
> > >> thread on the Code4Lib list?
> > >>
> > >> Linked Data authorities & vocabularies at Library of Congress (
> > id.loc.gov)
> > >> are going to be affected by the website shutdown -- because of lack of
> > >> government funds.
> > >>
> > >> -Jodi
> >
>
|