And while this is veering off-topic, it's also worth noting that the
development version of wget has support for WARC, the website archiving
format that the wayback machine is based around.
On 12-05-23 8:27 AM, "Tom Keays" <[log in to unmask]> wrote:
>I haven't tried it on a wiki, but the command-line Unix utility wget can
>be
>used to mirror a website.
>
>http://www.gnu.org/software/wget/manual/html_node/Advanced-Usage.html
>
>I usually call it like this:
>
>wget -m -p http://www.site.com/
>
>common flags:
> -m = mirroring on/off
> -p = page_requisites on/off
> -c = continue - when download is interrupted
> -l5 = reclevel - Recursion level (depth) default = 5
>
>On Tue, May 22, 2012 at 5:04 PM, Carol Hassler
><[log in to unmask]>wrote:
>
>> My organization would like to archive/export our internal wiki in some
>> kind of end-user friendly format. The concept is to copy the wiki
>> contents annually to a format that can be used on any standard computer
>> in case of an emergency (i.e. saved as an HTML web-style archive, saved
>> as PDF files, saved as Word files).
>>
>> Another way to put it is that we are looking for a way to export the
>> contents of the wiki into a printer-friendly format - to a document that
>> maintains some organization and formatting and can be used on any
>> standard computer.
>>
>> Is anybody aware of a tool out there that would allow for this sort of
>> automated, multi-page export? Our wiki is large and we would prefer not
>> to do this type of backup one page at a time. We are using JSPwiki, but
>> I'm open to any option you think might work. Could any of the web
>> harvesting products be adapted to do the job? Has anyone else backed up
>> a wiki to an alternate format?
>>
>> Thanks!
>>
>>
>> Carol Hassler
>> Webmaster / Cataloger
>> Wisconsin State Law Library
>> (608) 261-7558
>> http://wilawlibrary.gov/
>>
>>
|