I scrape data from a large book retailer we sometimes order from to make
our acquisitions workflow a bit easier. In particular, I ask people to make
wishlists and scrape those (I can do individual items too, but the large
retailer doesn't like that, even though we're trying to purchase from
them), check those titles against our holdings, then line them up with some
info in a spreadsheet for people. It is NOT the way I would suggest going
about things if you can help it, as pages change frequently, large
retailers block IP addresses, large retailers time out a lot, and so on,
but when up against a rock and a hard place.
Not for libraries, but I also recently had to put together a scraper when
looking for daycares in Massachusetts as the page listing daycares was
helpful but really, really clunky. Saved me quite a bit of sanity.
On Wed, Nov 29, 2017 at 8:50 AM, Ross Singer <[log in to unmask]> wrote:
> Due the absence of APIs, we have to scrape III WebBridge and EBSCO
> LinkSource link resolver results to determine electronic holdings for
> Neither of them make it particularly easy, since they don't provide many
> semantic clues in the markup as to what you're looking at and there are all
> kinds of other conditions you have to account for (e.g. direct linking to
> certain sources, etc.).
> It's generally one of those things I avoid at all costs since the pages you
> want/need to scrape are the most likely to be the most frustrating to work
> On Tue, Nov 28, 2017 at 1:26 PM Brad Coffield <[log in to unmask]
> > I think there's likely a lot of possibilities out there and was hoping to
> > hear examples of web scraping for libraries. Your example might just
> > inspire me or another reader to do something similar. At the very least,
> > the ideas will be interesting!
> > Brad
> > --
> > Brad Coffield, MLIS
> > Assistant Information and Web Services Librarian
> > Saint Francis University
> > 814-472-3315 <(814)%20472-3315>
> > [log in to unmask]