Is this for a one-shot project, or will it be ongoing? For a medium- to long-term initiative, I would suggest a subscription service like Archive-It; for a one-time effort, it would make more sense to use open-source tools like wget to generate a local copy + WARC. If it's the latter, I'll be happy to take a look at the page and walk you through the process.
I'm not really aware of a set of best practices, beyond the usual tenets of digital preservation (show your work, maintain authenticity, do minimal harm, document, document, document, etc). The model I've used in the past is generating a WARC alongside the access copy (using wget's WARC output), using that as the preservation master+technical metadata, and hosting the access copy on a front-facing machine. --Alex On Thu, Mar 20, 2014 at 11:05 AM, Kari R Smith <smit...@mit.edu> wrote: > Also, contact the SAA (Society of American Archivists) Web Archiving > round table. Lots of experience and help from that list of folks. > I'm forwarding your question to that list. > > Kari > > -----Original Message----- > From: Kim, Bohyun <b...@hshsl.umaryland.edu> > Sent: Thursday, March 20, 2014 8:26 AM > To: CODE4LIB@listserv.nd.edu > Subject: [CODE4LIB] Archiving a website - best practices > > I am not up to date with archiving practices. So I may be asking about a > well-known problem. > > But anyone archiving an old website and if so, what method do you use? We > are discussing taking screenshots and/or creating a zip file of the whole > site and uploading to a repository at MPOW. Both seem to have some > shortcomings. > > Thank you! > Bohyun >