Personal tools

Webarc:Main: Difference between revisions

From Adapt

Jump to: navigation, search
No edit summary
No edit summary
 
(11 intermediate revisions by the same user not shown)
Line 1: Line 1:
==Overview==
==Overview==
In the era of digital information, the efforts to preserve valuable human activities have broadened to also include documents, images, audio and video in their digital form. An unprecedented amount of information encompassing almost every facet of human activity across the world exists in the form of zeros and ones, and is also growing at an extremely fast pace. Moreover, the digital representation is often the only form in which such information is recorded.
<center>[[Image:web.jpg]]</center>


However, many digital objects face another set of challenges. They are dynamically updated at an unknown frequency, and often are interrelated to each other with temporal dependence. For these objects, the data collecting process in an archive needs to be able to determine whether or not an update occurred when it encounters a previously archived object. Otherwise, the archive can significantly waste its storage space by storing duplicate copies over and over again. Also, an archive needs to be aware of the interlinking relationships among the archived objects to better organize and manage the holdings, possibly accelerating the access performance. For example, in a system where small objects are packaged together in a container, and accesses are made on a container basis, placing heavily interlinked objects together in the same container will greatly improve the overall access speed.
An unprecedented amount of information encompassing almost every facet of human activity across the world is currently available on the web and is growing at an extremely fast pace. In many cases, the web is the only medium where such information is recorded. However, the web is an ephemeral medium whose contents are constantly changing and new information is rapidly replacing old information, resulting in the disappearance of a large number of web pages every day and in a permanent loss of part of our cultural and scientific heritage on a regular basis. A number of efforts, currently underway, are trying to develop methodologies and tools for capturing and archiving some of the web’s contents that are deemed critical. However there are major technical, social, and political challenges that are confronting these efforts. Major technical challenges include automatic tools to identify, find, and collect web contents to be archived, automatic extraction of metadata and context for such contents including linking structures that are inherent to the web, the organization and indexing of the data and the metadata, and the development of preservation and access mechanisms for current and future users, all at unprecedented scale and complexity.


Another important issue in the long-term preservation pertains to discovery and delivery of the preserved contents. In essence, the major purpose of preservation is to provide the preserved knowledge to the users who need it in the future. It is, thus, vital for any preservation system to provide an easy way to find and access the relevant contents. However, it is not a trivial matter to provide an effective, yet cost-effective, method to find the requested information mainly due to the large and ever-growing size of the preserved data. Preservation systems that solely rely on a relational database with well defined schemas may allow their users to find information more easily using well-structured queries. However, fitting every type of digital objects into a fixed set of schemas is often impossible. Clearly, we need a more general framework to enable effective information discovery and access to the archived contents.
Leaving aside dynamic and deep contents, web contents involve a wide variety of objects such as html pages, documents, multimedia files, scripts, etc., as well as, linking structures involving these objects. While the size of most web pages is small, the total number of web pages on a single web site can range from one to several millions. For example, as of Oct 30, 2006, Wikipedia.org alone claims to have about 1.4 million articles, each making up a distinct web page. A critical piece of web archiving is to capture the linking structures and organize the archived pages in such a way that future generations of users will be able to access and navigate through the archived web information in the same way as in the original linked structure. Note that by that time, the archived web contents may have migrated through several generations of hardware and software upgrades, including migration through different types of media, different file systems, and different formats.


In this research, we address three important problems in long-term preservation. First, we devise a methodology to efficiently store and index inter-related objects [[Webarc:Packaging|[1]]]. Second, we devise a methodology to temporally locate multi-version archived contents, as well as to detect duplicates [[Webarc:PISA|[2]].] Third, we devise a methodology to discover requested information from the preserved contents [[Webarc:Temporal Text-Search Index|[3]]].
While many challenges for archiving web contents exist, we focus in our work on ''scalable'' solutions supporting ''compact storage'' and ''fast access'' to large scale web archives. Our efforts to achieve the goal have been made around answering the following two questions:
 
# How do we compactly store contents in a way to retrieve contents quickly?
# How do we enable effective information discovery and access to the archived contents?
 
To address the first question, we devise methodologies to put together tightly tightly related objects [[Webarc:Packaging|[1]]], and index their locations temporarily [[Webarc:PISA|[2]]]. Using the fast location index [[Webarc:PISA|[2]]], a quick duplicate detection scheme is also designed to maintain compact archive storage. As the first step toward answering the second question,  we devise a temporal text-search index [[Webarc:Temporal Text-Search Index|[3]]] using a similar idea that we use for the fast location index.


==Our Approaches==
==Our Approaches==
* [[Webarc:Packaging|Web Container Packaging]]
* [[Webarc:Packaging|[1] Web Container Packaging]]
* [[Webarc:PISA|PISA:Persistent Indexing Structure for Archives]]
* [[Webarc:PISA|[2] PISA:Persistent Indexing Structure for Archives]]
* [[Webarc:Temporal Text-Search Index|Temporal Text-Search Index]]
* [[Webarc:Search and Access Strategies |[3] Search and Access Strategies]]
<!-- * [[Webarc:Temporal Scoring |[4] Temporal Scoring]] -->

Latest revision as of 20:23, 24 November 2009

Overview

Web.jpg

An unprecedented amount of information encompassing almost every facet of human activity across the world is currently available on the web and is growing at an extremely fast pace. In many cases, the web is the only medium where such information is recorded. However, the web is an ephemeral medium whose contents are constantly changing and new information is rapidly replacing old information, resulting in the disappearance of a large number of web pages every day and in a permanent loss of part of our cultural and scientific heritage on a regular basis. A number of efforts, currently underway, are trying to develop methodologies and tools for capturing and archiving some of the web’s contents that are deemed critical. However there are major technical, social, and political challenges that are confronting these efforts. Major technical challenges include automatic tools to identify, find, and collect web contents to be archived, automatic extraction of metadata and context for such contents including linking structures that are inherent to the web, the organization and indexing of the data and the metadata, and the development of preservation and access mechanisms for current and future users, all at unprecedented scale and complexity.

Leaving aside dynamic and deep contents, web contents involve a wide variety of objects such as html pages, documents, multimedia files, scripts, etc., as well as, linking structures involving these objects. While the size of most web pages is small, the total number of web pages on a single web site can range from one to several millions. For example, as of Oct 30, 2006, Wikipedia.org alone claims to have about 1.4 million articles, each making up a distinct web page. A critical piece of web archiving is to capture the linking structures and organize the archived pages in such a way that future generations of users will be able to access and navigate through the archived web information in the same way as in the original linked structure. Note that by that time, the archived web contents may have migrated through several generations of hardware and software upgrades, including migration through different types of media, different file systems, and different formats.

While many challenges for archiving web contents exist, we focus in our work on scalable solutions supporting compact storage and fast access to large scale web archives. Our efforts to achieve the goal have been made around answering the following two questions:

  1. How do we compactly store contents in a way to retrieve contents quickly?
  2. How do we enable effective information discovery and access to the archived contents?

To address the first question, we devise methodologies to put together tightly tightly related objects [1], and index their locations temporarily [2]. Using the fast location index [2], a quick duplicate detection scheme is also designed to maintain compact archive storage. As the first step toward answering the second question, we devise a temporal text-search index [3] using a similar idea that we use for the fast location index.

Our Approaches