Wget will pull the pages, but it won't help you backup the database that runs the wordpress website. Wordpress has a page dedicated to helping you with backups, have you looked at it? For easy reference it is http://codex.wordpress.org/WordPress_Backups -----Original Message----- From: Code for Libraries [mailto:[log in to unmask]] On Behalf Of Alex Armstrong Sent: Monday, October 06, 2014 3:48 AM To: [log in to unmask] Subject: Re: [CODE4LIB] wget archiving for dummies I wanted a quick-and-dirty solution to archiving our old LibGuides site a few months ago. wget was my first port of call also. I don't have good notes as to what went wrong, but I ended up using httrack: http://www.httrack.com/ It basically worked out of the box. HTH, Alex On 10/06/2014 09:44 AM, Eric Phetteplace wrote: > Hey C4L, > > If I wanted to archive a Wordpress site, how would I do so? > > More elaborate: our library recently got a "donation" of a remote > Wordpress site, sitting one directory below the root of a domain. I > can tell from a cursory look it's a Wordpress site. We've never > archived a website before and I don't need to do anything fancy, just > download a workable copy as it presently exists. I've heard this can be as simple as: > > wget -m $PATH_TO_SITE_ROOT > > but that's not working as planned. Wget's convert links feature > doesn't seem to be quite so simple; if I download the site, disable my > network connection, then host locally, some 20 resources aren't > available. Mostly images which are under the same directory. Possibly loaded via AJAX. Advice? > > (Anticipated) pertinent advice: I shouldn't be doing this at all, we > should outsource to Archive-It or similar, who actually know what they're doing. > Yes/no? > > Best, > Eric