Thursday 31 July 2008

DPC's preservation planning workshop

Earlier in the week I attended a DPC workshop on preservation planning, which was largely constructed of material coming out of the European project called Planets, which is now half-way through its four-year programme. There were also interesting contributions from Natalie Walters of the Wellcome Library and Matthew Woollard of the UK Data Archive.

A preservation system for the Wellcome Library?
Much of what Natalie had to say about the curation of born-digital archives chimed with our experiences here. Unlike us though, Wellcome are in the process of evaluating 'off the shelf' systems to manage digital preservation. They put out a tender earlier this year and received five responses that seem, in the main, to demonstrate a misunderstanding of archival requirements and the immaturity of the digital curation/preservation marketplace. One criticism was that the responses offered systems for 'access' or 'institutional repositories' (of the kind associated with open access HE content - academic papers and e-theses). This is something we also felt when we evaluated the Fedora and DSpace repositories on the Paradigm project (admittedly, this evaluation becomes a bit more obsolete day by day). Balancing access and preservation requirements has long been an issue for archivists, since we often have to preserve material that is embargoed for a period of time. I still believe that systems providing preservation services and systems providing researcher access are doing different things, but we do of course need some form of access to embargoed material for management and processing purposes. I also find the adoption of new meanings for words, like 'repository' and 'archive', tricky to negotiate at times. These issues aside, one of the systems offered seems to have held Wellcome's interest and I'll be keen to find out which one when this information can be revealed.

Preservation policy at UKDA
Matthew spoke about the evolution of preservation policy at the UKDA, which had no preservation policy until 2003 despite celebrating its 40th anniversary last year. The first two editions of the policy were more or less exclusively concerned with the technical aspects of preserving digital material, specifying such things as acceptable storage conditions and the frequency with which tape should be re-tensioned. The latest (third) edition embraces wider requirements including organisational/business need, user requirements (designated community and others), standards, legislation, technology and security. The new policy increases emphasis on data integrity and archival standards, it defines archival packages more closely to provide for their verification, and it pays attention to the curation of metadata describing the resources to be preserved.

If I understood correctly, the UKDA preserves datasets in their original form (SIP), migrates them to a neutral format (AIP1) and creates usable versions from the neutral format (AIP2). All these versions are preserved and dissemination versions of the dataset are created from AIP2. The degree of processing applied to a dataset is determined by applying a matrix which assigns a value on the basis of likely use and value. These processes feel similar to those evolving here, though we need to do more work to formalise them.

Matthew also showed us a nice little diagram from 1976, which was created to document UKDA workflow from initial acquisition of a dataset to its presentation to the final user. The fundamentals of professional archival, or OAIS-like, practice are evident. The UKDA's analysis of its own conformance with the OAIS model undertaken under the JISC 04/04 Programme is worth a look for those who haven't seen it.

Towards the end of the talk Matthew reminded us that having written a policy, one must implement it. It's not normally possible to implement every new thing in a policy at once, but the policy is valueless without mechanisms in place to audit it. Steps must be taken to progress those aspects of the policy that are new and to audit compliance more generally. The policy must also be available to relevant audiences who can evaluate the degree to which the archive complies with its own policy for themselves. I found this a very useful overview of the key issues involved in developing a preservation policy and the resulting policy itself is very clear and concise.

Planets tools for preservation planning
It's great to see the promise of Planets starting to be realised, especially since we plan to build on the project's work in relation to characterising material, planning and executing preservation strategies. Andreas Rauber kicked things off with an overview of the Planets project, which helped to demonstrated how the various components fit together.What is uncertain at the moment is how the software and services being developed by Planets will be sustained beyond the project's life. Neither is it clear what licensing model/s will be adopted for different components in the project, since there are the needs of commercial partners to consider as well as those of national archives, libraries and universities.

Plato
Christoph Becker gave us an overview of Plato, a tool which allows the user to develop preservation strategies for specific kind of objects. In Plato, users can design experiments to determine the best available preservation strategy for a particular type of material. This involves a formal definition of constraints and objectives, which includes an assessment of the relative importance of each of these factors. Factors might include:

* object migration time - max. 1 second
* object migration cost - max £0.05 per object
* preserve footnotes - 5
* preserve images- 5
* preserve headings - 4
* open format required - 5
* preserve font - 3
* and so on...

These are expressed in an 'objective tree', which can be created directly in Plato or in the Freemind mind mapping tool and uploaded to Plato. Objective tress can be very simple, but the process of creating a good and detailed objective tree is quite demanding (we had a go at doing this ourselves in the afternoon). In future we should be able to build on previous objective trees as these are developed and that will ease the process. For the moment the templates provided are minimal because the Plato team don't want to preempt user requirements!

The user must also supply a sample of material which can be used to assess the effectiveness of different strategies. This should be the bare minimum of objects required to represent the range of factors expressed in the objective tree. The user then selects different strategies to apply to the sample material, sets the experiment in motion, and compares the results against the objective tree. The process of evaluating results is manual at present, but there are plans to begin automating aspects of this too. Once the evaluation is complete, Plato can produce a report of the experiment which should demonstrate why one preservation strategy was chosen over another in respect of a particular class of material.

Plato is available for offline use, which will be necessary for us when processing embargoed material, but it is also offered as an online service where users can perform experiments in one place and benefit from working with the results of experiments performed by others.

Characterisation
The Planets work on characterisation was introduced by Manfred Thaller. This work develops two formal characterisation languages - the extensible characterisation extraction language (XCEL) and the extensible characterisation description language (XCDL). The work should make it possible to perform more automatically determine whether a preservation action, such as migration, has preserved an object's essential characteristics (or significant properties). It is expected that the Microsoft family of formats, PDF formats and common image formats will treated before the end of the project.

One of the interesting aspects of the characterisation work is developing an understanding of what is preserved or not in a particular process and how a file format impacts on this. Thaller demonstrated this (using a little tool for *shooting* files) by deliberately causing a small amount of damage to a png file and a tif file. A small amount of damage to the png file had severe consequences for its rendering, while the tif file could be damaged much more extensively and still retain some of its informational value. Thaller also used the example of migrating a MS Word 2003 document to the Open Document Text format. The migration to ODT seemed to lose a footnote in the document. Thaller then showed the same MS Word 2003 document migrated to PDF, where the footnote appears to be retained. In actual fact the footnote isn't lost in the migration to ODT, it's just not rendered. On the other hand, the footnote is structurally lost in the PDF file, but visually present. Thaller is proposing a solution which allows structure and appearance to be preserved.

Testbed
The final element of planets on show was the testbed developed at HATII, demonstrated by Matthew Barr. The testbed looks very useful and, like Plato, will be available for use online and offline. There did seem to be some overlap in aims and functionality with Plato, but there are differences too. It's essential objectives seem similar - users should be able to perform experiments with select data and tools, evaluate those experiments and draw conclusions to inform their preservation strategy; the testbed will also tools and services to be benchmarked. It struck me is that the process of conducting an experiment was simpler than with Plato, since a granular expression of objectives is not necessary. It's more quick and dirty, which may suit some scenarios better, but will the result be as good? Aspects I found particularly interesting were the development of a corpora and the ability to add new services (tools are deployed and accessed using web services) for testing.

Monday 21 July 2008

Seeking a Software Engineer

We are looking for a Software Engineer to work on the futureArch project. You can read the job advertisement at the vacancies section of the University of Oxford's website; there is also a link to the further particulars from here. Closing date is 29 August 2008.

Wednesday 16 July 2008

Annotating sound and video

At the JISC innovation forum earlier this week, I was fortunate enough to run into an improptu demo of Synote by Mike Wald of ECS, who had hijacked the British Library's sound archive project stand. Well, perhaps 'hijack' is a little strong - the BL demo was pretty much done and Peter Findlay was happy to tune in to what Mike was showing.

Synote is a rather nifty tool which lets users add annotations to specific points in a digital sound or video recording. These annotations might be notes, tags, or images; they act like bookmarks - they can be returned to easily as and when the need arises. Synote uses a transcript of the audio, which can be generated by speech recognition software if the audio is clean enough, or compiled by hand if not. The transcript plays alongside the content, and the users' annotations are highlighted in it; clicking on a word in the transcript allows the user to skip ahead to the bookmark and, of course, the transcript is searchable. It's been designed as a teaching and learning tool, but I think it has a lot of possibilities as a means of interacting with audio and video content present in archival collections. The project has a sourceforge page, so hopefully we'll be able to have a go ourselves in due course.

Wednesday 2 July 2008

Greetings!

Hello world, as they say. Welcome to this new blog, which is a place for those of us working with born-digital archives at the Bodleian Library to share our thoughts, frustrations and successes. We'll also be making a note of interesting or useful things we stumble upon.

We've been working on issues relating to the long-term preservation of digital archives for a few years now. If you take a look at our Paradigm and Cairo projects, that should give you an idea of the kinds of issues we're dealing with.

This blog is being born as we launch an important phase of development at the Library. We're about to begin the futureArch project, which will see us move the curation of born-digital archives and manuscripts from a series of small projects to a sustainable activity integrated with other aspects of the Library's operations. When futureArch concludes, in just over three years time, we aim to have embedded the curation of born-digital archives and manuscripts into the way we do things.

That's probably more than enough for a first post...