On Jul 24, 2009, at 2:20 PM, [Chris Stockwell] wrote: > Over the next few years, I am tasked to download 30,000 archival > masters > from Internet Archive into an archive for long-term staff access > that we may > preserve with LOCKSS. These are masters of Montana state > publications. I > have a hierarchy in mind to receive these files. The hierarchy is > state > agency\year\title\pub_date\*.pdf. > > I am intending to download the files in batches of 200 - 500 pdfs, > but am > thinking that if I slot them automatically into the archive > hierarchy, misplaced > or missing files could be very hard to find as the total grows. I > will be logging > the downloads, which should give me some control. Are there other > strategies > for ensuring that I can readily correct download errors? I am > looking for > recommendations for the simplest way to maintain reasonable control > over the > download process. A couple things: If you already have archive.org identifiers picked out, you can use something like this python script to download them all from IA: http://blog.openlibrary.org/2008/11/24/bulk-access-to-ocr-for-1-million-books/ You can use the archive.org advanced search engine to produce xml, json, or csv file with all identifiers for a particular contributing institution: http://www.archive.org/advancedsearch.php eg. all identifier for Montana State Library (http://www.archive.org/details/MontanaStateLibrary ) as an xml file (change rows=10 to rows=10000 to get them all): http://www.archive.org/advancedsearch.php?q=collection%3A%22montanastatelib%22&fl%5B%5D=creator&fl%5B%5D=identifier&fl%5B%5D=title&sort%5B%5D=&sort%5B%5D=&sort%5B%5D=&rows=10&fmt=xml&xmlsearch=Search Also, if you have an archive.org identifier, then you can get the files.xml that contains md5 and sha1 hashes, so you can verify your download: To pull the files.xml, use a /download/id/id_files.xml url. e.g.: http://www.archive.org/download/librariesoffutur00lickuoft/librariesoffutur00lickuoft_files.xml -raj