Webarc:Main: Difference between revisions
From Adapt
No edit summary |
No edit summary |
||
Line 21: | Line 21: | ||
==Our Approaches== | ==Our Approaches== | ||
* [[Webarc:PISA|PISA:Persistent Indexing Structure for Archives] | * [[Webarc:PISA|PISA:Persistent Indexing Structure for Archives]] | ||
* [[Webarc:Packaging|Web Container Packaging] | * [[Webarc:Packaging|Web Container Packaging]] |
Revision as of 21:05, 12 September 2008
Introduction
In the era of digital information, the efforts to preserve valuable human activities have broadened to also include documents, images, audio and video in their digital form. An unprecedented amount of information encompassing almost every facet of human activity across the world exists in the form of zeros and ones, and is also growing at an extremely fast pace. Moreover, the digital representation is often the only form in which such information is recorded.
However, the long-term preservation of digital information involves many complex issues. Digital information is generally very fragile over long term and susceptible to various threats. For example, various media technologies degrade over time, potentially causing random bit errors. Also hardware/software failure can lead to loss of data. Technology changes make current hardware, software, and digital formats invalid. There also can be malicious computer/network security attacks. Accidental operational errors cannot be overlooked either.
Moreover, many digital objects face another set of challenges. They are dynamically updated at an unknown frequency, and often are interrelated to each other with temporal dependence. For these objects, the data collecting process in an archive needs to be able to determine whether or not an update occurred when it encounters a previously archived object. Otherwise, the archive can significantly waste its storage space by storing duplicate copies over and over again. Also, an archive needs to be aware of the interlinking relationships among the archived objects to better organize and manage the holdings, possibly accelerating the access performance. For example, in a system where small objects are packaged together in a container, and accesses are made on a container basis, placing heavily interlinked objects together in the same container will greatly improve the overall access speed.
Another important issue in the long-term preservation pertains to discovery and delivery of the preserved contents. In essence, the major purpose of preservation is to provide the preserved knowledge to the users who need it in the future. It is, thus, vital for any preservation system to provide an easy way to find and access the relevant contents. However, it is not a trivial matter to provide an effective, yet cost-effective, method to find the requested information mainly due to the large and ever-growing size of the preserved data. Preservation systems that solely rely on a relational database with well defined schemas may allow their users to find information more easily using well-structured queries. However, fitting every type of digital objects into a fixed set of schemas is often impossible. Clearly, we need a more general framework to enable effective information discovery and access to the archived contents.
In this research, we address two important problems in long-term preservation. First, we devise a methodology to efficiently store and index inter-related objects. Two, we devise a methodology to discover requested information from the preserved contents. While having played a major role in the past, traditional methods cannot be directly applied for long-term preservation of digital objects. To our best knowledge, no significant research effort has been made to deal with a methodology that can efficiently and temporally handle inter-related objects for storage and indexing in a long-term archive. Also, most existing access schemes for a preservation system lack in convenience and/or effectiveness.
Objectives
In this research, we address some of the fundamental issues in web archiving storage and access. Specifically, we focus on providing technical methodologies to deal with the following two questions.
- How do we compactly store and index contents in a way to preserve complex structures and linked information, and retrieve contents quickly?
- How do we enable effective information discovery and access to the archived contents?
First, we aim to devise a methodology to efficiently store and index interrelated digital objects that may change over time. Ingested (or crawled) digital objects are sometimes interrelated to one another. A typical example of such interrelated objects is the web objects where the objects are hyperlinked to one another. Thus, this research considers the problem of storing a linked set of web objects into web containers in such a way as to minimize the number of containers accessed during a typical browsing session. We also consider a temporal indexing scheme that can locate web objects based on their URLs and the time of acquisition. One of the major benefits that come from an efficient temporal indexing scheme is to detect any duplicates before ingesting newly crawled web objects into an archive. This duplicate detection capability can potentially save a considerable amount of the storage space by not storing the identical objects more than once. For storing hyperlinked web objects, we have developed an approach based on a link-analysis technique and a graph portioning technique. This approach contrasts to the conventional approach where web objects are simply packaged into web containers on a first-come-first-served basis. The experimental results showed significant improvement over the conventional approach, and, at the same time, the overhead incurred for the link analysis and graph partitioning was very low. For the temporal indexing scheme, we have also developed and implemented a persistent indexing scheme called PISA (Persistent Indexing Structure for Archives). PISA has shown to be very fast and space-efficient; it demonstrated a considerable storage space saving with a negligible overhead.
Second, we aim to devise an efficient information discovery scheme. The ever growing size of a digital archive makes it harder over the long-term to pin-point the relevant information that needs to be accessed. However, many existing archives, such as the Internet Archive, simply rely on the user’s prior knowledge about what archive documents need to be retrieved. For example, when a user enters a URL, the Internet Archive gives a list of dates on which the contents at the URL were archived. From the list, the user then selects a date to view the archived page on the date. For archives based on a relational database with well-defined schemas, structured database queries can work in some cases. However, not every archive uses a relational database, and more importantly, even with such a database, it is often practically impossible to know the exact attributes of a certain dataset with which queries can be made. We, therefore, do not assume any particular infrastructure or data storage – such as the existence of database and schemas. We only have the following two assumptions: 1) each archived object has a unique handle (such as a combination of a URL and a time, or a unique OID) with which we can retrieve the object from the system, 2) and we can inspect every object in the system. Based on the two assumptions above, we plan to build an efficient information discovery system that will allow users to find and access more relevant contents in a long-term archive. In particular, we plan to develop a search scheme that adopts some of traditional information retrieval methodologies and more recent document ranking techniques based on link analysis. We also plan to incorporate a temporal dimension that the archived objects possess through version updates when computing the relevancy ranking.