You might want to investigate the E-Resource Access Checker script [1] Kristina Spurgin wrote and documented in the 2014 code4lib article "Getting What We Paid for: a Script to Verify Full Access to E-Resources" [2] - sounds like it would achieve your goals (and beyond). 1. https://github.com/UNC-Libraries/Ebook-Access-Checker 2. http://journal.code4lib.org/articles/9684 On Fri, Jul 7, 2017 at 12:19 PM, Ken Irwin <[log in to unmask]> wrote: > Hi folks, > > I'm looking for a tool to do automated unit-style testing on urls. I want > to know about HTTP errors, but also be able to test for things like "does > this page include particular text", "does this page fall within the > expected length parameters" and things of that sort. I have a few dozen > URLs that I want to be able to run tests periodically in our dev process > for a project that's been having some unexpected affects on a variety of > the library's online tools. > > By preference, a web-based or PHP-based tool would be ideal, but I'm open > to less familiar-to-me options as well. I could probably write something > with cURL, but maybe somebody's already got an elegant solution. > > Any ideas? > > Thanks > Ken >