You might want to investigate the E-Resource Access Checker script [1]
Kristina Spurgin wrote and documented in the 2014 code4lib article "Getting
What We Paid for: a Script to Verify Full Access to E-Resources" [2] -
sounds like it would achieve your goals (and beyond).
1. https://github.com/UNC-Libraries/Ebook-Access-Checker
2. http://journal.code4lib.org/articles/9684
On Fri, Jul 7, 2017 at 12:19 PM, Ken Irwin <[log in to unmask]> wrote:
> Hi folks,
>
> I'm looking for a tool to do automated unit-style testing on urls. I want
> to know about HTTP errors, but also be able to test for things like "does
> this page include particular text", "does this page fall within the
> expected length parameters" and things of that sort. I have a few dozen
> URLs that I want to be able to run tests periodically in our dev process
> for a project that's been having some unexpected affects on a variety of
> the library's online tools.
>
> By preference, a web-based or PHP-based tool would be ideal, but I'm open
> to less familiar-to-me options as well. I could probably write something
> with cURL, but maybe somebody's already got an elegant solution.
>
> Any ideas?
>
> Thanks
> Ken
>
|