I leveraged the IMPORTXML() and xpath features in Google Sheets to pull
information from a large university website to help create a set of weeding
lists for a branch campus. They needed extra details about what was in
off-site storage and what was held at the central campus library.
This was very much like Jason's FIFO API, the central reporting group had
sent me a spreadsheet with horrible data that I would have had to sort out
almost completely manually, but the call numbers were pristine. I used the
call numbers as a key to query the catalog with limits for each campus I
needed to check, and then it dumped all of the necessary content (holdings,
dates, etc) into the spreadsheet.
I've also used Feed43 as a way to modify certain RSS feeds and scrape
websites to only display the content I want.
Brett Williams
On Tue, Nov 28, 2017 at 1:24 PM, Brad Coffield <[log in to unmask]>
wrote:
> I think there's likely a lot of possibilities out there and was hoping to
> hear examples of web scraping for libraries. Your example might just
> inspire me or another reader to do something similar. At the very least,
> the ideas will be interesting!
>
> Brad
>
>
> --
> Brad Coffield, MLIS
> Assistant Information and Web Services Librarian
> Saint Francis University
> 814-472-3315
> [log in to unmask]
>
|