Hi Eric!

I'm a long time lurker on this listserv, but seldom reply.  For JSTOR in particular, you might be able to get some of what you want using our Data For Research tool: .  While this particular tool won't give you pdfs of the content, it does support the retrieval of OCR and object metadata for some of the research purposes you enumerate as I understand them.  You can download datasets of up to 25k with only registering an account and larger datasets by request.  I wish I could offer you a tool for all the needs you describe, but we have some way to go yet.

My personal focus is supporting the development and curation of internal datasets, but I will forward your message to my colleagues focusing on product development for good measure.  


Devin O'Hara

´╗┐On 2/5/20, 6:03 PM, "Code for Libraries on behalf of Eric Lease Morgan" <[log in to unmask] on behalf of [log in to unmask]> wrote:

    Do you find it difficult to get content? I do, and I sometimes feel as if I've been sold a bill of goods.
    With the advent of the Internet (and Google), it is relatively easy to find content, but it is still very difficult to actually get content, especially at scale; content is very often hidden behind obscure links, splash pages, etc.
    Take for example a mature open access publication with all the right intentions, Theological Librarianship. There you will find a pointer to the current issue and links to the archives. Cool & wonderful. But what are the actual links (URLs) to the articles? What is a link that will actually download an article to my desktop? I want to "save the time of the reader", and share a link with my colleague. Go ahead. Try to figure it out. I'll wait...
    "So what?", you might say. Yes, but what if I want to download the whole of Theological Librarianship for the purposes of distant reading? What if I want to study trends in the journal? What if I want to compare & contrast Theological Librarianship with other open access publications? Downloading all of those articles one by one would deter me from ever getting started. In the past I could go to the shelf, see all the bound issues, and begin to read.
    Got tired of looking for the links? Well, the links look like this, and there are about 350 of them:
    Given such a list and saved in a file, it is trivial to download all the PDF documents in less than 60 seconds, all 350 of them. [2]
    Suppose you maintain an institutional repository. Suppose it suports search. Do the search results point to the actual identified items, or do the search results point to some sort of "splash" page or "about" page? Again, for single items splash pages are not bad things, but what if I want to download all those preprints from a specific author, department, or school? What if I want to use & understand the whole of the College of Arts & Letters dissertation output? What if you wanted to download all those images, drop them into a machine learning process, and output metadata tags? Your research is stymied because, while you can find the content, you can not actually get it.
    The HaitiTrust is FULL of content. Do cool search. List results. Show me the links to download even plain text (OCR) versions of the open access content. They don't exist. Instead, one must identify a given book's key, and then programmatically download each page of the document one by one. [3]
    Our licensed databases are just as bad, if not worse. For example, do cool search against JSTOR. Get a list of results. Go the extra step and use some sort of browser extension to list all the URLs on a given page. [4] Copy the list. Paste it into a text editor. Sort the list. Remove the duplicates. Remove all the navigation links, and eventually discover that links to documents look like this:
    Fine, but when you go there you are presented with a splash page and another link:
    So you get smart, and you perform a find/replace operation against your links to point to the PDF files, but when you go to these links you are presented with a challenge which is (by design) very difficult to circumvent. By this time you are tired and give up. But still you have done the perfect search, identified the perfect set of twenty five articles, and despite all of this cool Internet hipe, you can not get the content.
    Other examples are innumerable. 
    With the advent of the Internet I feel as if we have gone one step forward and half a step back. "Look at all the glorious content that is here, but you can't have it... unless you pay." We pay in terms of time, energy, or real money, and even then it is not enough. Intellectually I understand, especially from a for-profit publisher's point of view, but I don't think this makes very much sense when it comes to content from libraries, other cultural heritage institutions, publishers who simply what to get their content out there, or content which was licensed and paid for.
    As people who do Internet stuff in libraries, I think we can do better. 
    [1] Theological Librarianship -
    [2] cat urls.txt | parallel wget
    [3] A hack to do just this is located at
    [4] For example, Link Grabber at  which is simply a really bad URL in and of itself
    Eric Lease Morgan, Librarian