I wanted a quick-and-dirty solution to archiving our old LibGuides site a few months ago. wget was my first port of call also. I don't have good notes as to what went wrong, but I ended up using httrack: http://www.httrack.com/ It basically worked out of the box. HTH, Alex On 10/06/2014 09:44 AM, Eric Phetteplace wrote: > Hey C4L, > > If I wanted to archive a Wordpress site, how would I do so? > > More elaborate: our library recently got a "donation" of a remote Wordpress > site, sitting one directory below the root of a domain. I can tell from a > cursory look it's a Wordpress site. We've never archived a website before > and I don't need to do anything fancy, just download a workable copy as it > presently exists. I've heard this can be as simple as: > > wget -m $PATH_TO_SITE_ROOT > > but that's not working as planned. Wget's convert links feature doesn't > seem to be quite so simple; if I download the site, disable my network > connection, then host locally, some 20 resources aren't available. Mostly > images which are under the same directory. Possibly loaded via AJAX. Advice? > > (Anticipated) pertinent advice: I shouldn't be doing this at all, we should > outsource to Archive-It or similar, who actually know what they're doing. > Yes/no? > > Best, > Eric