Wget will pull the pages, but it won't help you backup the database that runs the wordpress website.
Wordpress has a page dedicated to helping you with backups, have you looked at it?
For easy reference it is http://codex.wordpress.org/WordPress_Backups
-----Original Message-----
From: Code for Libraries [mailto:[log in to unmask]] On Behalf Of Alex Armstrong
Sent: Monday, October 06, 2014 3:48 AM
To: [log in to unmask]
Subject: Re: [CODE4LIB] wget archiving for dummies
I wanted a quick-and-dirty solution to archiving our old LibGuides site a few months ago.
wget was my first port of call also. I don't have good notes as to what went wrong, but I ended up using httrack:
http://www.httrack.com/
It basically worked out of the box.
HTH,
Alex
On 10/06/2014 09:44 AM, Eric Phetteplace wrote:
> Hey C4L,
>
> If I wanted to archive a Wordpress site, how would I do so?
>
> More elaborate: our library recently got a "donation" of a remote
> Wordpress site, sitting one directory below the root of a domain. I
> can tell from a cursory look it's a Wordpress site. We've never
> archived a website before and I don't need to do anything fancy, just
> download a workable copy as it presently exists. I've heard this can be as simple as:
>
> wget -m $PATH_TO_SITE_ROOT
>
> but that's not working as planned. Wget's convert links feature
> doesn't seem to be quite so simple; if I download the site, disable my
> network connection, then host locally, some 20 resources aren't
> available. Mostly images which are under the same directory. Possibly loaded via AJAX. Advice?
>
> (Anticipated) pertinent advice: I shouldn't be doing this at all, we
> should outsource to Archive-It or similar, who actually know what they're doing.
> Yes/no?
>
> Best,
> Eric
|