An ongoing multiple volume set with updated index.
Publisher
West Group
Publication Location
St. Paul, MN
Critical Arguements
CA "Contains over 400 separate titles on a broad range of legal topics which, taken together, systematically describe the entire field of American legal doctrine. Documents available for each topic may include a summary, topic contents, each (TRUNCATED)
Type
Book Whole
Title
Words and Phrases: All Judicial Constructions and Definitions of Words and Phrases
by the States and Federal Courts
Artiste is a European project developing a cross-collection search system for art galleries and museums. It combines image content retrieval with text based retrieval and uses RDF mappings in order to integrate diverse databases. The test sites of the Louvre, Victoria and Albert Museum, Uffizi Gallery and National Gallery London provide their own database schema for existing metadata, avoiding the need for migration to a common schema. The system will accept a query based on one museumÔÇÖs fields and convert them, through an RDF mapping into a form suitable for querying the other collections. The nature of some of the image processing algorithms means that the system can be slow for some computations, so the system is session-based to allow the user to return to the results later. The system has been built within a J2EE/EJB framework, using the Jboss Enterprise Application Server.
Secondary Title
WWW2002: The Eleventh International World Wide Web Conference
Publisher
International World Wide Web Conference Committee
ISBN
1-880672-20-0
Critical Arguements
CA "A key aim is to make a unified retrieval system which is targeted to usersÔÇÖ real requirements and which is usable with integrated cross-collection searching. Museums and Galleries often have several digital collections ranging from public access images to specialised scientific images used for conservation purposes. Access from one gallery to another was not common in terms of textual data and not done at all in terms of image-based queries. However the value of cross-collection access is recognised as important for example in comparing treatments and conditions of paintings. While ARTISTE is primarily designed for inter-museum searching it could equally be applied to museum intranets. Within a MuseumÔÇÖs intranet there may be systems which are not interlinked due to local management issues."
Conclusions
RQ "The query language for this type of system is not yet standardised but we hope that an emerging standard will provide the session-based connectivity this application seems to require due to the possibility of long query times." ... "In the near future, the project will be introducing controlled vocabulary support for some of the metadata fields. This will not only make retrieval more robust but will also facilitate query expansion. The LouvreÔÇÖs multilingual thesaurus will be used in order to ensure greater interoperability. The system is easily extensible to other multimedia types such as audio and video (eg by adding additional query items such as "dialog" and "video sequence" with appropriate analysers). A follow-up project is scheduled to explore this further. There is some scope for relating our RDF query format to the emerging query standards such as XQuery and we also plan to feed our experience into standards such as the ZNG initiative.
SOW
DC "The Artiste project is a European Commission funded collaboration, investigating the use of integrated content and metadata-based image retrieval across disparate databases in several major art galleries across Europe. Collaborating galleries include the Louvre in Paris, the Victoria and Albert Museum in London, the Uffizi Gallery in Florence and the National Gallery in London." ... "Artiste is funded by the European CommunityÔÇÖs Framework 5 programme. The partners are: NCR, The University of Southampton, IT Innovation, Giunti Multimedia, The Victoria and Albert Museum, The National Gallery, The research laboratory of the museums of France (C2RMF) and the Uffizi Gallery. We would particularly like to thank our collaborators Christian Lahanier, James Stevenson, Marco Cappellini, John Cupitt, Raphaela Rimabosci, Gert Presutti, Warren Stirling, Fabrizio Giorgini and Roberto Vacaro."
Type
Conference Proceedings
Title
Integrating Metadata Schema Registries with Digital Preservation Systems to Support Interoperability: A Proposal
There are a large number of metadata standards and initiatives that have relevance to digital preservation, e.g. those designed to support the work of national and research libraries, archives and digitization initiatives. This paper introduces some of these, noting that the developers of some have acknowledged the importance of maintaining or re-using existing metadata. It is argued here that the implementation of metadata registries as part of a digital preservation system may assist repositories in enabling the management and re-use of this metadata and may also help interoperability, namely the exchange of metadata and information packages between repositories.
Publisher
2003 Dublin Core Conference: Supporting Communities of Discourse and Practice-Metadata Research & Applications
Publication Location
Seatle, WA
Critical Arguements
CA "This paper will introduce a range of preservation metadata initiatives including the influential Open Archival Information System (OAIS) reference model and a number of other initiatives originating from national and research libraries, digitization projects and the archives community. It will then comment on the need for interoperability between these specifications and propose that the implementation of metadata registries as part of a digital preservation system may help repositories manage diverse metadata and facilitate the exchange of metadata or information packages between repositories."
Conclusions
RQ "The plethora of metadata standards and formats that have been developed to support the management and preservation of digital objects leaves us with several questions about interoperability. For example, will repositories be able to cope with the wide range of standards and formats that exist? Will they be able to transfer metadata or information packages containing metadata to other repositories? Will they be able to make use of the 'recombinant potential' of existing metadata?" ... "A great deal of work needs to be done before this registry-based approach can be proved to be useful. While it would undoubtedly be useful to have registries of the main metadata standards developed to support preservation, it is less clear how mapping-based conversions between them would work in practice. Metadata specifications are based on a range of different models and conversions often lead to data loss. Also, much more consideration needs to be given to the practical issues of implementation." 
SOW
DC Michael Day is a research officer at UKOLN, which is based at the University of Bath. He belongs to UKOLN's research and development team, and works primarily on projects concerning metadata, interoperability and digital preservation. 
Type
Conference Proceedings
Title
Preserving the Fabric of Our Lives: A Survey of Web Preservation Initiatives
This paper argues that the growing importance of the World Wide Web means that Web sites are key candidates for digital preservation. After an [sic] brief outline of some of the main reasons why the preservation of Web sites can be problematic, a review of selected Web archiving initiatives shows that most current initiatives are based on combinations of three main approaches: automatic harvesting, selection and deposit. The paper ends with a discussion of issues relating to collection and access policies, software, costs and preservation.
Secondary Title
Research and Advanced Technology for Digital Libraries, 7th European Conference, ECDL 2003, Trondheim, Norway, August 2003 Proceedings
Publisher
Springer
Publication Location
Berlin
Critical Arguements
CA "UKOLN undertook a survey of existing Web archiving initiatives as part of a feasibility study carried out for the Joint Information Systems Committee (JISC) of the UK further and higher education funding councils and the Library of the Wellcome Trust. After a brief description of some of the main problems with collecting and preserving the Web, this paper outlines the key findings of this survey." (p. 462) Addresses technical, legal and organizational challenges to archiving the World Wide Web. Surveys major attempts that have been undertaken to archive the Web, highlights the advantages and disadvantages of each, and discusses problems that remain to be addressed.
Conclusions
RQ "It is hoped that this short review of existing Web archiving initiatives has demonstrated that collecting and preserving Web sites is an interesting area of research and development that has now begun to move into a more practical implementation phase. To date, there have been three main approaches to collection, characterised in this report as 'automatic harvesting,' 'selection' and 'deposit.' Which one of these has been implemented has normally depended upon the exact purpose of the archive and the resources available. Naturally, there are some overlaps between these approaches but the current consensus is that a combination of them will enable their relative strengths to be utilised. The longer-term preservation issues of Web archiving have been explored in less detail." (p. 470)
SOW
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.
Type
Conference Proceedings
Title
Place, Interface and Cyberspace: Archives at the Edge, Proceedings of the 1998 Conference of the Australian Society of Archivists
CA Online, humans act both as universal everymen and as community members with their own cultural assumptions. When people transact business online, the legal and social relationships engendered place on each participant "a range of rights and responsibilities that underpin the regulation of the net as a community." (p.104) So while the interrelations may seem more complex in cyberspace, in the end establishing the relationships between key parties is still crucial to ascertaining their legal obligations, whether they are online or offline. (p.120)
Conclusions
RQ In order to ensure that evidential requirements are extended to net transactions, we must address the following questions: Are we revisiting the problems of electronic information systems without recordkeeping functionality in the cyberspace environment? Can intranet systems linked to the Net retrieve transactions with all their context intact?
CA "Ironically, electronic records systems make it both possible to more fully capture provenance than paper recrods systems did and at the same time make it more likely that provenance will be lost and that archives, even if they are preserved, will therefore lack evidential value. This paper explores the relationship between provenance and evidence and its implications for management of paper or electronic information systems." (p. 177)
Conclusions
"Electronic information systems, therefore, present at least two challenges to archivists. The first is that the designers of these systems may have chosen to document less contextual information than may be of interest to archivists when they designed the system. The second is that the data recorded in any given information system will, someday, need to be transferred to another system. ... [A]rchivists will need to return to fundamental archival principles to determine just what they really wanted to save anyway. ... It may be that archivists will be satisfied with the degree of evidential historicity they were able to achieve in paper based record systems, in which case there are very few barriers to implementing successful electronic based archival environments. Or archivists may decide that the fuller capability of tracking the actual participation of electronic data objects in organizational activities needs to be documented by archivally satisfactory information systems, in which case they will need to define those levels of evidential historicity that must be attained, and specify the systems requirements for such environments. ... At a meeting on electronic records management research issues sponsored by the National Historical Publications and Records Commission in January 1991, participants identified the concept of technological and economic plateaux in electronic data capture and archiving as an important arena for research ... Hopefully this research will produce information to help archivists make decisions regarding the amount of contextual information they can afford to capture and the requirements of systems designed to document context along with managing data content. ... I will not be surprised as we refine our concepts of evidential historicity to discover that the concept of provenance takes on even greater granularity." (p. 192-193)
CA Discusses the ways traditional archival science can inform IT, and the ways IT can help the goals of archival science be achieved more easily and efficiently.
Conclusions
<RQ> "When archivists work with information technologies or electronic archiving specialists, they have a lot to offer. They are the ones who have the conceptual key to the analysis and design of the new archiving systems." (p. 174)
Type
Journal
Title
Strategies for managing electronic records: A new archival paradigm? An affirmation of our archival traditions?
CA It is still too early to tell which models (like Pitt or UBC) actually work until we have had time to evaluate them. Archivists need to learn new skills in order to be effective in the electronic environment, and cannot wait for an out-of-the-box solution. Most likely, any solution will require a combination of strategies. Most of all, one must remain flexible and open to new ways of doing things.
Phrases
<P1> Unlike paper documents where context and physical form are united in a medium that provides the record of the transaction, and where relationships among documents can be observed, electronic records are not physical but are logically constructed and often "virtual" entities. Therefore, it is argued, efforts to document business transactions based on the examination of "views" or of automated forms will fail to reveal the nature of the business transactions. Consequently, methods other than direct observation and review must be employed to properly document automated systems. (p.25) <P2> System metadata typically do not contain all the information archivists need to describe electronic records, in particular, all the necessary contextual data required to understand the context of the transaction are not present. Therefore it is suggested that archivists will need to know which metadata elements are required to fully describe these records and must be in a position to add these descriptive elements to the system, preferably at the design stage. (p.26)
Conclusions
RQ What is involved in effecting a major shift from creating descriptive data to capturing, managing and adding value to system metadata?
Type
Journal
Title
Reality and Chimeras in the Preservation of Electronic Records
CA An emulation approach is not viable for e-records preservation because it preserves the "wrong thing": systems functionality rather than records. Consequently, an emulation solution would not preserve e-records as evidence "even if it could be made to work."
Phrases
<P1>Electronic records that are not moved out of obsolete hardware and software environments are very likely to die with them. <P2> Failure to examine in detail what makes an electronic record evidence over time has led Rothenberg, and many others, to assume they want to preserve system functionality. (p.2) <P3> The state of a database at any given moment is not a record. (p.2) <P4> If we want to preserve electronic records, what we really want are records of the actual inputs and outputs from the system to be maintained as evidence over time. This does not require the information system to function as it once did. All (!) it requires is that we can capture all transactions entering and leaving the system when they are created, ensuring that the original context of their creation and content is documented, and that the requirements of evidence are preserved over time. (p.2)
Conclusions
RQ Metadata encapsualtion strategies need to identify how metadata will be captured at the time of a record's creation, how it will be stored over time while supporting the use of the record by authorized users and more generally how the recordkeeping infrastructure will be constructed and maintained.
Type
Journal
Title
Capturing records' metadata: Unresolved questions and proposals for research
The author reviews a range of the research questions still unanswered by research on the capture of metadata required for recordness. These include how to maintain inviolable linkages between records and their metadata in a variety of architectures, what structure metadata content should take, the semantics of records metadata and that of other electronic sources, how new metadata can be acquired by records over time, maintaining the meaning of contextual metadata over time, the use of metadata in records management and the design of environments in which Business Acceptable Communications ÔÇô BAC ÔÇô (those with appropriate evidential metadata) can persist.
Critical Arguements
CA "My research consists of model building which enables the construction of theories and parallel implementations based on shared assumptions. Some of these models are now being tested in applications, so this report reflects both what we do not yet know from abstract constructs and questions being generated by field testing. " ... Bearman overviews research questions such as semantics, syntax, structure and persistence of metadata that still need to be addressed.
Phrases
<P1> Records are evidence when they are bound to appropriate metadata about their content, structure and context. <P2> The metadata required for evidence is described in the Reference Model for Business Acceptable Communications (BAC). <P3> Metadata which is required for evidence must continue to be associated with the record to which it relates over time and neither it nor the record content can be alterable. <P4> To date we have only identified three implementations which, logically, could allow metadata to retain this inviolable connection. Metadata can be: kept in a common envelope WITH a record (encapsulated), bound TO a record (by integrity controls within an environment), or LINKED with a record through a technical and/or social process (registration, key deposit, etc.). <P5> Metadata content was defined in order to satisfy a range of functional requirements of records, hence it ought to have a structure which enables it to serve these functions effectively and in concrete network implementations. <warrant> <P6> Clusters of metadata are must operate together. Clusters of metadata are required by different processes which take place at different times, for different software clients, and within a variety of processes. Distinct functions will need access to specified metadata substructures and must be able to act on these appropriately. Structures have been proposed in the Reference Model for Business Acceptable Communications. <P7> Metadata required for recordness must, logically, be standard; that required for administration of recordkeeping systems is extensible and locally variable. <P8> Records metadata must be semantically homogenous but it is probably desirable for it to be syntactically heterogeneous and for a range of protocols to operate against it. Records metadata management system requirements have both an internal and external aspect; internally they satisfy management requirements while externally they satisfy on-going recordness requirements. <P9> The metadata has to come either from a specific user/session or from rules defined to extract data either from a layer in the application or a layer between the application and the recording event. <P10> A representation of the business context must exist from which the record-creating event can obtain metadata values. <P11> Structural metadata must both define the dependent structures and identify them to a records management environment which is ÔÇ£patrollingÔÇØ for dependencies which are becoming risky in the evolving environment in order to identify needs for migration. <P12> BAC conformant environments could reduce overheads and, if standards supported the uniform management of records from the point of issue to the point of receipt. Could redundancy now imposed by both paper and electronic processes be dramatically reduced if records referenced other records? <P13>
Conclusions
RQ "All the proposed methods have some degree of external dependency. What are the implications software dependencies? Encapsulation, integrity controls and technico-social process are all software dependent. Is this avoidable? Can abstract reference models of the metadata captured by these methods serve to make them effectively software independent? " ... "What are the relative overhead costs of maintaining the systems which give adequate societal assurances of records retention following any of these approaches? Are there some strategies that are currently more efficient or effective? What are the organizational requirements for implementing metadata capture systems? In particular, what would the costs of building such systems within a single institution be versus the costs of implementing records metadata adhering communications servers on a universal scale?" ... "Can we model mechanisms to enable an integrated environment of recordkeeping throughout society for all electronically communicated transactions?" ... "Are the BAC structures workable? Complete? Extensible in ways that are known to be required? For example, metadata required for ÔÇ£recordnessÔÇØ is created at the time of the creation of the records but other metadata, as premised by the Warwick Framework, 2 may be created subsequently. Are these packets of metadata orthogonal with respect to recordness? If not, how are conflicts dealt with? " ... "Not all metadata references fixed facts. Thus, for example, we have premised that proper reference to a retention schedule is a citation to an external source rather than a date given within the metadata values of a record. Similar external references are required for administration of shifting access permissions. What role can registries (especially rights clearinghouses) play in a world of electronic records? How well do existing languages for permission management map to the requirements of records administration, privacy and confidentiality protection, security management, records retention and destruction, etc." ... "Not all records will be created with equally perfect metadata. Indeed risk-based decisions taken by organizations in structuring their recordsÔÇÖ capture are likely to result in conscious decisions to exclude certain evidential metadata. What are the implications of incomplete metadata on an individual organization level and on a societal level? Does the absence of data as a result of policy need to be noted? And if so, how?" ... "Since metadata has owners, howdo owners administer recordsÔÇÖ metadata over time? In particular, since records contain records, how are the layers of metadata exposed for management and administrative needs (if internal metadata documenting dependencies can slip through the migration process, we will end up with records that cannot serve as evidence. If protected records within unprotected records are not protected, we will end up with insecure records environments, etc. etc.)." ... "In principle, the BAC could be expressed as Dublin metadata 3 and insofar as it cannot be, the Dublin metadata will be inadequate for evidence. What other syntax could be used? How could these be comparatively tested?" .. "Could Dublin Core metadata, if extended by qualifying schema, serve the requirements of recordness? Records are, after all, documents in the Dublin sense of fixed information objects. What would the knowledge representation look like?" ... "Strategies for metadata capture currently locate the source of metadata either in the API layer, or the communications system, using data provided by the application (an analysis supports defining which data and where they can be obtained), from the user interface layer, or from the business rules defined for specified types of communication pathways. Can all the required metadata be obtained by some combination of these sources? In other words, can all the metadata be acquired from sources other than content created by the record-creator for the explicit and sole purpose of documentation (since such data is both suspect in itself and the demand for it is annoying to the end user)? " ... "Does the capture of metadata from the surrounding software layers require the implementation of a business-application specific engine, or can we design generic tools that provide the means by which even legacy computing systems can create evidential records if the communication process captures the interchange arising from a record-event and binds it with appropriate metadata?" ... "What kinds of representations of business processes and structures can best carry contextualizing metadata at this level of granularity and simultaneously serve end user requirements? Are the discovery and documentation representations of provenance going to have to be different? " ... "Can a generic level of representation of context be shared? Do standards such a STEP 4 provide adequate semantic rules to enable some meaningful exchange of business context information? " ... "Using past experiences of expired standards as an indicator, can the defined structural metadata support necessary migrations? Are the formal standards of the source and target environments adequate for actual record migration to occur?" ... "What metadata is required to document a migration itself?" ... "Reduction of redundancy requires record uses to impose post-creation metadata locks on records created with different retention and access controls. To what extent is the Warwick Framework relevant to these packets and can architectures be created to manage these without their costs exceeding the savings?" ... "A number of issues about proper implementation depend on the evolution (currently very rapid) of metadata strategies in the broader Internet community. Issues such as unique identification of records, external references for metadata values, models for metadata syntax, etc. cannot be resolved for records without reference to the ways in which the wider community is addressing them. Studies that are supported for metadata capture methods need to be aware of, and flexible in reference to, such developments."
The Getty Art History Information Program: Research Agenda for Cultural Heritage on Information Networks
Publication Year
1995
Critical Arguements
CA The inability to effectively preserve and authenticate electronic records presents a significant problem for the humanities research, which depends on correct attribution and the ability to view resources long after they were created.
Phrases
<P1> Current research on software dependence and interoperability is not largely driven by archival concerns and takes a relatively short view on the requirement to preserve functionality. Little research has been done on modeling the information loss that accompanies multiple migrations or the risks inherent in the use of commercial systems before standards are developed, yet these are the critical questions being posed by archives. (p.2) <P2> The metadata required for recordness and the means to capture this data and ensure that it is bonded to electronic communications is the most significant area for research in the near future. (p.3) <P3> Within organizations, archivists must find automatic means of identifying the business process for which a record is generated. Such data modeling will become increasingly critical in an era of ongoing business re-engineering. If records are retained for their evidential significance and for a period associated with risk, then certain knowledge of their functional source is essential to their rational control. If they are retained for long-term informational value, knowledge of context is necessary to understand their significance. (p.3) <warrant>
Conclusions
RQ We need to research what value e-records have other than as a means of assessing accountability. How are they used, and what value do users derive from them? What do we need to know about a record's content to support the discovery of billions of records? How can our preservation solutions be made scaleable?
CA Makes a distinction between archival description of the record at hand and documentation of the context of its creation. Argues the importance of the latter in establishing the evidentiary value of records, and criticizes ISAD(G) for its failure to account for context. "(1) The subject of documentation is, first and foremost, the activity that generated the records, the organizations and individuals who used the records, and the purposes to which the records were put. (2). The content of the documentation must support requirements for the archival management of records, and the representations of data should support life cycle management of records. (3) The requirements of users of archives, especially their personal methods of inquiry, should determine the data values in documentation systems and guide archivists in presenting abstract models of their systems to users." (p. 45-46)
Phrases
<P1> [T]he ICA Principles rationalize existing practice -- which the author believes as a practical matter we cannot afford; which fail to provide direct access for most archives users; and which do not support the day-to-day information requirements of archivists themselves. These alternatives are also advanced because of three, more theoretical, differences with the ICA Principles: (1) In focusing on description rather than documentation, they overlook the most salient characteristic of archival records: their status as evidence. (2) In proposing specific content, they are informed by the bibliographic tradition rather than by concrete analysis of the way in which information is used in archives. (3) In promoting data value standardization without identifying criteria or principles by which to identify appropriate language or structural links between the objects represented by such terms, they fail adequately to recognize that the data representation rules they propose reflect only one particular, and a limiting, implementation. (p. 33-34) <P2> Archives are themselves documentation; hence I speak here of "documenting documentation" as a process the objective of which is to construct a value-added representation of archives, by means of strategic information capture and recording into carefully structured data and information access systems, as a mechanism to satisfy the information needs of users including archivists. Documentation principles lead to methods and practices which involve archivists at the point, and often at the time, of records creation. In contrast, archival description, as described in the ICA Principles[,] is "concerned with the formal process of description after the archival material has been arranged and the units or entities to be described have been determined." (1.7) I believe documentation principles will be more effective, more efficient and provide archivists with a higher stature in their organizations than the post accessioning description principles proposed by the ICA. <warrant> (p. 34) <P3> In the United States, in any case, there is still no truly theoretical formulation of archival description principles that enjoys a widespread adherence, in spite of the acceptance of rules for description in certain concrete application contexts. (p. 37) <P4> [T]he MARC-AMC format and library bibliographic practices did not adequately reflect the importance of information concerning the people, corporate bodies and functions that generated records, and the MARC Authority format did not support appropriate recording of such contexts and relations. <warrant> (p. 37) <P5> The United States National Archives, even though it had contributed to the data dictionary which led to the MARC content designation, all the data which it believed in 1983 that it would want to interchange, rejected the use of MARC two years later because it did not contain elements of information required by NARA for interchange within its own information systems. <warrant> (p. 37) <P6> [A]rchivists failed to understand then, just as the ISAD(G) standard fails to do now, that rules for content and data representation make sense in the context of the purposes of actual exchanges or implementation, not in the abstract, and that different rules or standards for end-products may derive from the same principles. (p. 38) <P7> After the Committee on Archival Information Exchange of the Society of American Archivists was confronted with proposals to adopt many different vocabularies for a variety of different data elements, a group of archivists who were deeply involved in standards and description efforts within the SAA formed an Ad Hoc Working Group on Standards for Archival Description (WGSAD) to identify what types of standards were needed in order to promote better description practices.  WSAD concluded that existing standards were especially inadequate to guide practice in documenting contexts of creation.  Since then, considerable progress has been made in developing frameworks for documentation, archival information systems architecture and user requirements analysis, which have been identified as the three legs on which the documenting documentation platform rests. <warrant> (p. 38) <P8> Documentation of organizational activity ought to begin long before records are transferred to archives, and may take place even before any records are created -- at the time records are created -- at the time when new functions are assigned to an organization. (p. 39) <P9> It is possible to identify records which will be created and their retention requirements before they are created, because their evidential value and informational content are essentially predetermined. (p. 39) <P10> Archivists can actively intervene through regulation and guidance to ensure that the data content and values depicting activities and functions are represented in such a way that will make them useful for subsequent management and retrieval of the records resulting from these activities. This information, together with systems documentation, defines the immediate information system context out of which the records were generated, in which they are stored, and from which they were retrieved during their active life. (p. 39) <P11> Documentation of the link between data content and the context of creation and use of the records is essential if records (archives or manuscripts) are to have value as evidence. (p. 39) <P12> [C]ontextual documentation capabilities can be dramatically improved by having records managers actively intervene in systems design and implementation.  The benefits of proactive documentation of the context of records creation, however, are not limited to electronic records; the National Archives of Canada has recently revised its methods of scheduling to ensure that such information about important records systems and contexts of records creation will be documented earlier. <warrant> (p. 39) <P13> Documentation of functions and of information systems can be conducted using information created by the organization in the course of its own activity, and can be used to ensure the transfer of records to archives and/or their destruction at appropriate times. It ensures that data about records which were destroyed as well as those which were preserved will be kept, and it takes advantage of the greater knowledge of records and the purposes and methods of day-to-day activity that exist closer to the events. (p. 40) <P14> The facts of processing, exhibiting, citing, publishing and otherwise managing records becomes significant for their meaning as records, which is not true of library materials. (p. 41) <P15> [C]ontent and data representation requirements ought to be derived from analysis of the uses to which such systems must be put, and should satisfy the day to day information requirements of archivists who are the primary users of archives, and of researchers using archives for primary evidential purposes. (p. 41) <P16> The ICA Commission proposes a principle by which archivists would select data content for archival descriptions, which is that "the structure and content of representations of archival material should facilitate information retrieval." (5.1) Unfortunately, it does not help us to understand how the Commission selected the twenty-five elements of information identified as its standard, or how we could apply the principle to the selection of additional data content. It does, however, serve as a prelude to the question of which principles should guide archivists in choosing data values in their representations. (p. 42) <P17> Libraries have found that subject access based on titles, tables of contents, abstracts, indexes and similar formal subject analysis by-products of publishing can support most bibliographic research, but the perspectives brought to materials by archival researchers are both more varied and likely to differ from those of the records creators. (p. 43) <P18> The user should not only be able to employ a terminology and a perspective which are natural, but also should be able to enter the system with a knowledge of the world being documented, without knowing about the world of documentation. (p. 44) <P19> Users need to be able to enter the system through the historical context of activity, construct relations in that context, and then seek avenues down into the documentation. This frees them from trying to imagine what records might have survived -- documentation assists the user to establish the non-existence of records as well as their existence -- or to fathom how archivists might have described records which did survive. (p. 44) <P20> When they departed from the practices of Brooks and Schellenberg in order to develop means for the construction of union catalogues of archival holdings, American archivists were not defining new principles, but inventing a simple experiment. After several years of experience with the new system, serious criticisms of it were being leveled by the very people who had first devised it. (p. 45)
Conclusions
RQ "In short, documentation of the three aspects of records creation contexts (activities, organizations and their functions, and information systems), together with representation of their relations, is essential to the concept of archives as evidence and is therefore a fundamental theoretical principle for documenting documentation. Documentation is a process that captures information about an activity which is relevant to locating evidence of that activity, and captures information about records that are useful to their ongoing management by the archival repository. The primary source of information is the functions and information systems giving rise to the records, and the principal activity of the archivist is the manipulation of data for reference files that create richly-linked structures among attributes of the records-generating context, and which point to the underlying evidence or record." (p. 46)
Type
Journal
Title
Archival Issues in Network Electronic Publications
"Archives are retained information systems that are developed according to professional principles to meet anticipated demands of user clienteles in the context of the changing conditions created by legal environments and electronic or digital technologies. This article addresses issues in electronic publishing, including authentication, mutability, reformatting, preservation, and standards from an archival perspective. To ensure continuing access to electronically published texts, a special emphasis is placed on policy planning in the development and implementation of electronic systems" (p.701).
Critical Arguements
<P1> Archives are established, administered, and evaluated by institutions, organizations, and individuals to ensure the retention, preservation, and utilization of archival holdings (p.701) <P2> The three principal categories of archival materials are official files of institutions and organizations, publications issued by such bodies, and personal papers of individuals. . . . Electronic information technologies have had profound effects on aspects of all these categories (p.702) <P3> The primary archival concern with regard to electronic publishing is that the published material should be transferred to archival custody. When the transfer occurs, the archivist must address the issues of authentication, appraisal, arrangement, description, and preservation or physical protection (p.702) <P4> The most effective way to satisfy archival requirements for handling electronic information is the establishment of procedures and standards to ensure that valuable material is promptly transferred to archival custody in a format which will permit access on equipment that will be readily available in the future (p.702) <P5> Long-term costs and access requirements are the crucial factors in determining how much information should be retained in electronic formats (p.703) <P6> Authentication involves a determination of the validity or integrity of information. Integrity requires the unbroked custody of a body of information by a responsible authority or individual <warrant> (p.703) <P7> From an archival perspective, the value of information is dependent on its content and the custodial responsibility of the agency that maintains it -- e.g., the source determines authenticity. The authentication of archival information requires that it be verified as to source, date, and content <warrant> (p.704) <P8> Information that is mutable, modifiable, or changeable loses its validity if the persons adding, altering, or deleting information cannot be identified and the time, place and nature of the changes is unknown (p.704) <P9> [P]reservation is more a matter of access to information than it is a question of survival of any physical information storage media (p.704) <P10> [T]o approach the preservation of electronic texts by focusing on physical threats will miss the far more pressing matter of ensuring continued accessibility to the information on such storage media (p.706) <P11> If the information is to remain accessible as long as paper, preservation must be a front-end, rather than an ex post facto, action (p.708) <P12> [T]he preservation of electronic texts is first and foremost a matter of editorial and administrative policy rather than of techniques and materials (p.708) <P13> Ultimately, the preservation of electronic publications cannot be solely an archival issue but an administrative one that can be addressed only if the creators and publishers take an active role in providing resources necessary to ensure that ongoing accesibility is part of initial system and product design (p.709) <P14> An encouraging development is that SGML has been considered to be a critical element for electronic publishing because of its transportability and because it supports multiple representations of a single text . . . (p.711) <P15> Underlying all questions of access is the fundamental consideration of cost (p.711)
Type
Journal
Title
The Management of Digital Data: A metadata approach
CA "Super-metadata may well play a crucial role both in facilitating access to DDOs and in providing a means of selecting and managing the maintenance of these DDOs over time."
Phrases
<P1> The preservation of the intellectual content of DDOs brings into focus a major issue: "the integrity and authenticity of the information as originally recorded" (Graham, 1997). (p.365). <P2> The emergence of dynamic and living DDOs is presenting challenges to the conventional understanding of the preservation of digital resources and is forcing many organizations to reevaluate their strategies in the light of these rapid advances in information sources. The use of appropriate metadata is recognized to be essential in ensuring continued access to dynamic and living DDOs, but the standards for such metadata are not yet fully understood or developed. (p.369)
Conclusions
RQ How can we decide what to preserve ? How can we assure long-term access? What will be the cost of electronic archiving? Which metadata schema will be in use 10 years from now, and how will migration be achieved?
CA Unless organizations find a means to preserve e-records, the long-term memory of modern institutions will be at great risk. As society now moves from written records to virtual documents, archivists are offering their traditional understanding of the structure and context of recorded evidence as protection against the widespread amnesia now threatening our electronic world.
Phrases
<P1> For electronic media, the content, structure and context of the record change significantly from that of the paper world. The only approximate match with paper is the content element, where the letters and numbers look much the same on the computer screen as on paper. But the structure and especially the context of electronic records are not apparent when retrieved from the text only. (p.4)
Type
Journal
Title
Warrant and the Defintion of Electronic Records: Questions Arising from the Pittsburgh Project
The University of Pittsburgh Electronic Recordkeeping Research Project established a model for developing functional requirements and metadata specifications based on warrant, defined as the laws, regulations, best practices, and customs that regulate recordkeeping. Research has shown that warrant can also increase the acceptance by records creators and others of functional requirements for recordkeeping. This article identifies areas related to warrant that require future study. The authors conclude by suggesting that requirements for recordkeeping may vary from country to country and industry to industry because of differing warrant.
Publisher
Kluwer Academic Publishers
Publication Location
The Netherlands
Critical Arguements
CA Poses a long series of questions and issues concerning warrant and its ability to increase the acceptance of recordkeeping requirements. Proposes that research be done to answer these questions. Discusses two different views about whether warrant can be universal and/or international.
Phrases
<P1> As we proceeded with the project [the University of Pittsburgh Electronic Recordkeeping Research Project] we ultimately turned our attention to the idea of the literary warrant -- defined as the mandate from law, professional best practices, and other social sources requiring the creation and continued maintenance of records. Wendy Duff's doctoral research found that warrant can increase the acceptance of some recordkeeping functional requirements, and therefore it has the potential to build bridges between archival professionals and others concerned with or responsible for recordkeeping. We did not anticipate the value of the literary warrant and, in the hindsight now available to us, the concept of the warrant may turn out to be the most important outcome of the project. <P2> In Wendy Duff's dissertation, legal, auditing and information science experts evluated the authority of the sources of warrant for recordkeeping. This part of the study provided evidence that information technology standards may lack authority, but this finding requires further study. Moreover, the number of individuals who evaluated the sources of warrant was extremely small. A much larger number of standards should be included in a subsequent study and a greater number of subjects are needed to evaluate these standards. <P3> We found a strong relationship between warrant and the functional requirements for electronic recordkeeping systems. Research that studies this relationship and determines the different facets that may affect it might provide more insights into the relationship between the warrant and the functional requirements. <P4> [W]e need to develop a better understanding of the degree to which the warrant for recordkeeping operates in various industries, disciplines, and other venues. Some institutions operate in a much more regulated environment than others, suggesting that the imporance of records and the understanding of records may vary considerably between institutional types, across disciplines and from country to country. <P5> We need to consider whether the recordkeeping functional requirements for evidence hold up or need to be revised for recordkeeping requirements for corporate memory, accountability, and cultural value -- the three broad realms now being used to discuss records and recordkeeping. <P6> The warrant gathered to date has primarily focused on technical, legal or the administrative value of records. A study that tested the effectiveness of warrant that supported the cultural or historical mandate of archives might help archivists gain support for their archival programs. <P7> This concern leads us to a need for more research about the understanding of records and recordkeeping in particular institutions, disciplines, and societies. <P8> A broader, and perhaps equally important question, is whether individual professionals and workers are even aware of their regulatory environment. <P9> How do the notion of the warrant and the recordkeeping functional requirements relate to the ways in which organizations work and the management tools they use, such as business process reengineering and data warehousing? <P10> What are the economic implications for organizations to comply with the functional requirements for recordkeeping in evidence? <P11> Is there a warrant and separate recordkeeping functional requirements for individual or personal recordkeeping? <P12> As more individuals, especially writers, financial leaders, and corporate and societal innovators, adopt electronic information technologies for the creation of their records, an understanding of the degree of warrant for such activity and our ability to use this warrant to manage these recordkeeping systems must be developed. <P13> We believe that archivists and records managers can imporve their image if they become experts in all aspects of recordkeeping. This will require a thorough knowledge of the legal, auditing, information technology, and management warrant for recordkeeping. <P14> The medical profession emphasizes that [sic] need to practice evidence-based medicine. We need to find out what would happen if records managers followed suit, and emphasized and practiced warrant-based recordkeeping. Would this require a major change in what we do, or would it simply be a new way to describe what we have always done? <P15> More work also has to be done on the implications of warrant and the functional requirements for the development of viable archives and records management programs. <P16> The warrant concept, along with the recordkeeping functional requirements, seem to possess immense pedagogical implications for what future archivists or practicing archivists, seeking to update their skills, should or would be taught. <P17> We need to determine the effectiveness of using the warrant and recordkeeping functional requirements as a basis for graduate archival and records management education and for developing needed topics for research by masters and doctoral students. <P18> The next generation of educational programs might be those located in other professional schools, focusing on the particular requirements for records in such institutions as corporations, hospitals, and the courts. <P19> We also need to determine the effectiveness of using the warrant and recordkeeping functional requirements in continuing education, public outreach, and advocacy for helping policy makers, resource allocators, administrators, and others to understand the importance of archives and records. Can the warrant and recordkeeping functional requirements support or foster stronger partnerships with other professions, citizen action groups, and other bodies interested in accountability in public organizations and government? <P20> Focusing on the mandate to keep and manage records, instead of the records as artifacts or intersting stuff, seems much more relevant in late twentieth century society. <P21> We need to investigate the degree to which records managers and archivists can develop a universal method for recordkeeping. ... Our laws, regulations, and best practices are usually different from country to country. Therefore, must any initiative to develop warrant also be bounded by our borders? <P22> A fundamental difference between the Pittsburgh Project and the UBC project is that UBC wishes to develop a method for managing and preserving electronic records that is applicable across all juridical systems and cultures, while the Pittsburgh Project is proposing a model that enables recordkeeping to be both universal and local at the same time. <P23> We now have a records management standard from Australia which is relevant for most North American records programs. It has been proposed as an international standard, although it is facing opposition from some European countries. Can there be an international standard for recordkeeping and can we develop one set of procedures which will be accepted across nations? Or must methods of recordkeeping be adapted to suit specific cultures, juridical systems, or industries?
Conclusions
RQ See above.
Type
Journal
Title
Digital preservation: Where we are, where we're going, where we need to be
CA Digital preservation will begin to come into its own. The past five years were about building access; now standards are coalescing and more focus is being paid to actual preservation strategies. Major legal obstacles include the DMCA, which restricts what institutions can do to preserve digital information. There are economic challenges, and we do not really know how much digital preservation will cost.
Phrases
<P1> There will be change, there is no guarantee that you can pick a technology and stay with it for ten years. We have to have an awareness of technological change and what's coming -- we listen to peers and the larger institutions that are taking leading and bleeding edge roles, and we make wise decisions. So in this case it is OK to be trailing edge and choose something that is well-established." (p.3)
SOW
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.
Over the last decade a number of writers have encouraged archivists to develop strategies and tactics to redefine their role and to insert themselves into the process of designing recordkeeping systems. This paper urges archivists to exploit the authority inherent in the laws, regulations, standards, and professional best practices that dictate recordkeeping specifications to gain great acceptance for the requirements for electronic evidence. Furthermore, it postulates that this proactive approach could assist in gaining greater respect for the archival profession.
Critical Arguements
CA The use of authoritative sources of warrant would improve acceptance of electronic records as evidence and create greater respect for the archival profession.
Phrases
<P1> The legal, administrative, fiscal, or information value of records is dependent upon the degree of trust society places in records as reliable testimony or evidence of the acts they purport to document. In turn, this trust is dependent on society's faith in the procedures that control the creation and maintenance of the record. <P2> [S]ociety bestows some methods of recordkeeping and record creating with an authority or 'warrant' for generating reliable records. <P3> David Bearman first proposed the idea of "literary warrant." <P4> [S]tatements of warrant provide clear instructions on how records should be kept and delineate elements needed for the records to be complete. <P5> The information technology field promulgates standards, but in North America adherence to them is voluntary rather than obligatory. <P6> The University of Pittsburgh Electronic Recordkeeping Project suggested that requirements for electronic recordkeeping should derive from authoritative sources, such as the law, customs, standards, and professional best practices accepted by society and codified in the literature of different professions concerned with records and recordkeeping rather than developed in isolation. <P7> On their own, archival requirements for recordkeeping have very little authority as no authoritative agencies such as standards boards or professional associations have yet to endorse them [sic] and few archivists have the authority to insist that their organizations follow them. <P8> An NHPRC study suggested that archivists have not been involved in the process of meeting the challenges of electronic records because they are undervalued by their colleagues, or, in other words, are not viewed as a credible source.
Conclusions
RQ "By highlighting the similarity between recordkeeping requirements and the requirements delineated in authoritative statements in the law, auditing standards, and professional best practices, archivists will increase the power of their message. ... If archivists are to take their rightful place as regulators of an organization's documentary requirements, they will have to reach beyond their own professional literature and understand the requirements for recordkeeping imposed by other professions and society in general. Furthermore, they will have to study methods of increasing the accpetance of their message and the impact and power of warrant."
Type
Journal
Title
Will Metadata Replace Archival Description: A Commentary
CA Before archival description can be replaced by metadata, "archivists must first study their user needs, identify processes that protect the integrity and impartiality of records, and ensure the capture of important contextual information." (p.38)
Phrases
<P1> Unfortunately, information systems often do not create records, concentrating instead on the preservation of information to the detriment of recordkeeping. Concern over this issue has lead Wallace to promote a new role for archivists, one that places them at the conception of the life cycle, establishing standards for record preservation and management as well as dictating record creation. Demarcation between archivists and records managers disappears in this new paradigm, and a new role as auditor, system designer, and regulator begins to emerge. (p.34) <P2> "Metadata are essential if archivists are to maintain the integrity and authenticity of evidence of actions. McNeil likens metadata systems to protocol registers and sees metadata itself as evidence, as well as a means of preserving evidence." (p.35)
Conclusions
RQ Will metadata replace archival description? Will metadata requirements fulfill the needs of secondary users? Will metadata require secondary descriptions?
Type
Journal
Title
Ensuring the Preservation of Reliable Evidence: A Research Project Funded by the NHPRC
CA Archivists need to propogate research projects that delineate means to engender trust and accountability for our e-records.
Phrases
<P1> "The task of preserving evidence in a hardware and software dependent environment challenges archivists to develop new techniques and new ways of of thinking about what to capture and how to preserve it. The development of the functional requirements, including the production rules, the literary warrant, and the metadata reference model, is a first step toward solving some of the most pressing problems that archivists face in the new electronic world. (p.39) <P2> As records migrate from a stable paper reality to an intangible electronic existence, their physical attributes, vital for establishing the authenticity and reliability of the evidence they contain, are threatened. (p. 29) <P3> Unfortunately, systems that create and maintain electronic records often fail to preserve the structure or the context essential for the evidentiary nature of records. (p.30)
Conclusions
RQ Can warrant increase the credibility of the functional requirements for recordkeeping? Can one type of warrant be more influential than others? Is the warrant from a person's specific profession seen by him or her as more important than others?
Type
Journal
Title
Law, evidence and electronic records: A strategic perspective from the global periphery
CA A recordkeeping paradigm set up around the records continuum will take us into the future, because it sees opportunities, not problems, in e-environments. It fosters accountability through evidence-generating recordkeeping practices.
Phrases
<P1> This challenge is being addressed by what Chris Hurley has called second-generation archival law, which stretches the reach of archival jurisdictions into the domain of the record-creator. A good example of such archival law is South Africa's National Archives Act of 1996, which gives the National Archives regulatory authority over all public records from the moment of their creation. The Act provides a separate definition of "electronic records systems" and accords the National Archives specific powers in relation to their management. Also significant is that the Act brings within the National Archives' jurisdiction those categories of record-creators commonly allowed exclusion -- the security establishment, public services outside formal structures of government, and "privatized" public service agencies. (p.34) <P2> A characteristic (if an absence can be a characteristic) of most archival laws, first and second generation, is a failure to define either the conditions/processes requiring "recording" or the generic attributes of a "record." (p.34) <P3> Archival law, narrowly defined, is not at the cutting edge and is an increasingly small component of broader recordkeeping regimes. This is one of the many signs of an emerging post-custodial era, which Chris Hurley speculates will be informed by a third generation of archival law. Here, the boundaries between recordkeeping domains dissolve, with all of them being controlled by universal rules. (p.34)
Conclusions
RQ What is the relationship between the event and the record? Is the idea of evidence pivotal to the concept of "recordness"? Should evidence be privileged above all else in defining a record? What about remembering, forgetting, imagining?
Type
Journal
Title
Building record-keeping systems: Archivists are not alone on the wild frontier
CA The digital environment offers archivists a host of new tools that can be adapted and used for recordkeeping. However, archivists must choose their tools judisciously while considering the long-term implications of their use as well as research and development. Ultimately, they must pick tools and strategies that dovetail with their institutions' specific needs while working to produce reliable and authentic records.
Phrases
<P1> Evidence from this review of emerging methods for secure and authentic electronic communications shows that the division of responsibility, accountability, and jurisdiction over recordkeeping is becoming more complex than a clear line between the records creator and the records preserver. (p.66) <P2> Storage of records in encrypted form is another area of concern because encryption adds additional levels of systems dependency on access to keys, proprietary encryption algorithims, hardware, and software. (p.62) <P3> It is important for archivists and records managers to understand parallel developments, because some new strategies and methods may support recordkeeping, while others may impede the achievement of archival objectives. (p.45) <P4> The concept of warrant and subsequent research on it by Wendy Duff is a significant contribution, because it situates the mandates for creating and maintaining records in a legal, administrative, and professional context, and it presents a methodology for locating, compiling, and presenting the rules governing proper and adequate documentation in modern organizations. (p. 48)
Conclusions
RQ Are electronic recordkeeping systems truly inherently inferior to paper-based systems in their capacity to maintain authentic records over time? How tightly can recordkeeping be integrated into normal business processes, and where does one draw the line between how a business does its work and how it does its recordkeeping?
Type
Journal
Title
How Do Archivists Make Electronic Archives Usable and Accessible?
CA In order to make electronic archives useable, archivists will need to enhance and link access systems to facilitate resource discovery while making the whole process as seamless and low-cost (or no-cost) as possible for the user.
Phrases
<P1> Rather than assuming that the archival community will succeed in transferring all valuable electronic records to archival institutions for preservation and future access, archivists must develop strategies and methods for accessibility and usability that can span a variety of custodial arrangements. (p.9) <P2> Maintaining linkages between different formats of materials will become increasingly burdensome if archvists do not find ways to develop integrated access systems. (p.10) <P3> Archivists must also think about ways to teach users the principles of a new digital diplomatics so that they can apply these principles themselves to make educated judgements about the accuracy, reliability, and authenticity of the documents they retrieve from electronic archives. (p.15)
Type
Journal
Title
Documenting digital images: Textual meta-data at the Blake Archive
The Electronic Library: The International Journal for Minicomputer, Microcomputer, and Software Applications in Libraries
Publication Year
1998
Volume
16
Issue
4
Pages
239
Critical Arguements
CA One of the critical issues in the future development of digital libraries is the provision of documentary metadata for non-textual electronic files.
Phrases
<P1> When libraries create digital image collections, however, documentation becomes more problematic. Even if an image is surrounded by a robust network of supporting materials, the functionality of client-server networks such as the World Wide Web permits the image to become detached from the documentary process." (p. 239)
Type
Journal
Title
When Documents Deceive: Trust and Provenance as New Factors for Information Retrieval in a Tangled Web
Journal of the American Society for Information Science and Technology
Periodical Abbreviation
JASIST
Publication Year
2001
Volume
52
Issue
1
Pages
12
Publisher
John Wiley & Sons
Critical Arguements
"This brief and somewhat informal article outlines a personal view of the changing framework for information retrieval suggested by the Web environment, and then goes on to speculate about how some of these changes may manifest in upcoming generations of information retrieval systems. It also sketches some ideas about the broader context of trust management infrastructure that will be needed to support these developments, and it points towards a number of new research agendas that will be critical during this decade. The pursuit of these agendas is going to call for new collaborations between information scientists and a wide range of other disciplines." (p. 12) Discusses public key infrastructure (PKI) and Pretty Good Practice (PGP) systems as steps toward ensuring the trustworthiness of metadata online, but explains their limitations. Makes a distinction between the identify of providers of metadata and their behavior, arguing that it is the latter we need to be concerned with.
Phrases
<P1> Surrogates are assumed to be accurate because they are produced by trusted parties, who are the only parties allowed to contribute records to these databases. Documents (full documents or surrogate records) are viewed as passive; they do not actively deceive the IR system.... Compare this to the realities of the Web environment. Anyone can create any metadata they want about any object on the net, with any motivation. (p. 13) <P2> Sites interested in manipulating the results of the indexing process rapidly began to exploit the difference between the document as viewed by the user and the document as analyzed by the indexing crawler through a set of techniques broadly called "index spamming." <P3> Pagejacking might be defined generally as providing arbitrary documents with independent arbitrary index entries. Clearly, building information retrieval systems to cope with this environment is a huge problem. (p. 14) <P4> [T]he tools are coming into place that let one determine the source of a metadata assertion (or, more precisely and more generally) the identity of the person or organization that stands behind the assertion, and to establish a level of trust in this identity. (p. 16) <P5> It is essential to recognize that in the information retrieval context one is not concerned so much with identity as with behavior. ... This distinction is often overlooked or misunderstood in discussions about what problems PKI is likely to solve: identity alone does not necessarily solve the problem of whether to trust information provided by, or warranted by, that identity. ... And all of the technology for propagating trust, either in hierarchical (PKI) or web-of-trust identity management, is purely about trust in identity. (p. 16) <P6> The question of formalizing and recording expectations about behavior, or trust in behavior, are extraordinarily complex, and as far as I know, very poorly explored. (p. 16) <P7> [A]n appeal to certification or rating services simply shifts the problem: how are these services going to track, evaluate, and rate behavior, or certify skills and behavior? (p. 16) <P8> An individual should be able to decide how he or she is willing to have identity established, and when to believe information created by or associated with such an identity. Further, each individual should be able to have this personal database evolve over time based on experience and changing beliefs. (p. 16) <P9> [T]he ability to scale and to respond to a dynamic environment in which new information sources are constantly emerging is also vital.<P10> In determining what data a user (or an indexing system, which may make global policy decisions) is going to consider in matching a set of search criteria, a way of defining the acceptable level of trust in the identity of the source of the data will be needed. (p. 16) <P10> Only if the data is supported by both sufficient trust in the identity of the source and the behavior of that identity will it be considered eligible for comparison to the search criteria. Alternatively, just as ranking of result sets provided a more flexible model of retrieval than just deciding whether documents or surrogates did or did not match a group of search criteria, one can imagine developing systems that integrate confidence in the data source (both identity and behavior, or perhaps only behavior, with trust in identity having some absolute minimum value) into ranking algorithms. (p. 17) <P11> As we integrate trust and provenance into the next generations of information retrieval systems we must recognize that system designers face a heavy burden of responsibility. ... New design goals will need to include making users aware of defaults; encouraging personalization; and helping users to understand the behavior of retrieval systems <warrant> (p. 18) <P12> Powerful paternalistic systems that simply set up trust-related parameters as part of the indexing process and thus automatically apply a fixed set of such parameters to each search submitted to the retrieval system will be a real danger. (p. 17)
Conclusions
RQ "These developments suggest a research agenda that addresses indexing countermeasures and counter-countermeasures; ways of anonymously or pseudononymously spot-checking the results of Web-crawling software, and of identifying, filtering out, and punishing attempts to manipulate the indexing process such as query-source-sensitive responses or deceptively structured pages that exploit the gap between presentation and content." (p. 14) "Obviously, there are numerous open research problems in designing such systems: how can the user express these confidence or trust constraints; how should the system integrate them into ranking techniques; how can efficient index structures and query evaluation algorithms be designed that integrate these factors. ... The integration of trust and provenance into information retrieval systems is clearly going to be necessary and, I believe, inevitable. If done properly, this will inform and empower users; if done incorrectly, it threatens to be a tremendously powerful engine of censorship and control over information access. (p. 17)
Type
Journal
Title
Metadata Strategies and Archival Description: Comparing Apples to Oranges
Advocates of a "metadata systems approach" to the description of electronic records argue that metadata's capacity to provide descriptive information about the context of electronic records creation will obviate, or reduce significantly, the need for traditional archival description. This article examines the assumptions about the nature of archival description and of metadata on which metadata strategies are grounded, for the purposes of ascertaining the following: whether the skepticism concerning the capacity of traditional description to meet the challenges posed by the so-called "second generation" of electronic records is justified; whether the use of metadata as archival description is consistent with their nature and purpose; and whether metadata are capable of servinng archival descriptive purposes.
Critical Arguements
CA "Before the archival profession assigns to traditional archival description the diminished role of "added value" (i.e. accessory) or abandons it altogether, the assumptions about the nature of archival description and of metadata on which metadata strategies are grounded ought to be carefully examined. Such an examination is necessary to ascertain the following: whether the skepticism concerning the capacity of traditional description to meet the challenges posed by the so-called "second generation" of electronic records is justified, whether the use of metadata as archival description is consistent with their nature and purpose, and whether metadata are acapable of serving archival purposes."
Phrases
<P1> In an article published in Archivaria, David Wallace summarized recent writing on the subject of metadata and concluded that "[d]ata dictionaries and the types of metadata that they house and can be built to house should be seriously evaluated by archivists" because of their potential to signficantly improve and ultimately transform traditional archival practice in the areas of appraisal, arrangement, description, reference, and access. <warrant> <P2> In the area of description, specifically, advocates of "metadata management" or a "metadata systems approach" believe that metadata's capacity to provide descriptive information about the context of electronic records creation will obviate, or reduce significantly, the need for traditional description. <P3> Charles Dollar maintains that archival participation in the IRDS standard is essential to ensure that archival requirements, including descriptive requirements, are understood and adopted within it. <warrant> <P4> According to David Wallace, "archivists will need to concentrate their efforts on metadata systems creation rather than informational content descriptions, since in the electronic realm, archivists' concern for informational value will be eclipsed by concern for the evidential value of the system." <warrant> <P5> Charles Dollar, for his part, predicts that, rather than emphasize "the products of an information system," a metadata systems approach to description will focus on "an understanding of the information system context that supports organization-wide information sharing." <P6> Because their scope and context are comparitively narrow, metadata circumscribe and atomize these various contexts of records creation. Archival description, on the other hand, enlarges and integrates them. In so doing it reveals continuities and discontinuities in the matrix of function, structure, and record-keeping over time. <P7> Metadata are part of this broader context, since they constitute a series within the creator's fonds. The partial context provided by metadata should not, however, be mistaken for the whole context. <P8> Metadata, for example, may be capable of explaining contextual attributes of the data within an electronic records system, but they are incapable of describing themselves -- i.e., their own context of creation and use -- because they cannot be detached from themselves. For this reason, it is necessary to describe the context in which the metadata are created so that their meaning also will be preserved over time. <P9> A metadata system is like a diary that, in telegraphic style, records the daily events that take place in the life of an individual as they occur and from the individual's perspective. <P10> Archival description, it could be said, is the view from the plane; metadata, the view from the field as it is plowed. <P11> While a close-up shot-- such as the capture of a database view -- may be necessary for the purposes of preserving record context and system functionality, it does not follow that such a snapshot is necessary or even desirable for the purposes of description. <P12> Because the context revealed by metadata systems is so detailed, and the volume of transactions they capture is so enormous, metadata may in fact obscure, rather than illuminate, the broader administrative context and thereby bias the users' understanding of the records' meaning. In fact, parts of actions and transactions may develop entirely outside of the electronic system and never be included in the metadata. <P13> If the metadata are kept in their entirety, users searching for documents will have to wade through a great deal of irrelevant data to find what they need. If the metadata are chopped up into bits corresponding to what has been kept, how comprehensible will they be to the uesr? <P14> The tendency to describe metadata in metaphorical terms, e.g., in relation to archival inventories, has distracted attention from consideration of what metadata are in substantial, concrete terms. They are, in fact, records created and used in the conduct of affairs of which they form a part. <P15> The transactions captured by metadata systems may be at a more microscopic level than those captured in registers and the context may be more detailed, given the technological complexity of electronic record-keeping environments. Nevertheless, their function remains the same. <P16> And, like protocol registers, whose permanent retention is legislated, metadata need to be preserved in perpetuity because they are concrete evidence of what documents were made and received, who handled them, with what results, and the transactions to which they relate. <warrant> <P17> While it is true that metadata systems show or reveal the context in which transactions occur in an electronic system and therefore constitute a kind of description of it -- Jenkinson made the same observation about registers -- their real object is to record the fact of these transactions; they should be, like registers, "preserved as a [record] of the proceedings in that connection." <P18> Viewing metadata systems as tools for achieving archival purposes, rather than as tools for achieving the creators' purposes is dangerous because it encourages us to, in effect, privilege potential secondary uses of metadata over their actual primary use; in so doing, we could reshape such use for purposes other than the conduct of affairs of which they are a part. <P19> Metadata strategies risk compromising, specifically, the impartiality of the records' creation. <P20> For archivists to introduce in the formation of metadata records requirements directed toward the future needs of archivists and researchers rather than toward the current needs of the creator would contribute an element of self-consciousness into the records creation process that is inconsistent with the preservation of the records' impartiality. <P21> If the impartiality of the metadata is compromised, their value as evidence will be compromised, which means, ultimately, that the underlying objective of metadata strategies -- the preservation of evidence -- will be defeated. <P22> None of these objections should be taken to suggest that archivists do not have a role to play in the design and maintenance of metadata systems. It is, rather, to suggest that that role must be driven by our primary obligation to protect and preserve, to the extent possible, the essential characterisitcis of the archives. <P23> The proper role of an archivist in the design of a metadata system, then, is to assist the organization in identifying its own descriptive needs as well as to ensure that the identification process is driven, not by narrowly defined system requirements, but by the organization's overarching need and obligation to create and maintain complete, reliable, and authentic records. <P24> That is why it is essential that information holdings are identified and described in a meaningful way, organized in a logical manner that fascilitates their access, and preserved in a manner that permits their continuing use. <P25> Record-keeing requirements for electronic records must address the need to render documentary relationships wisible and to build in procedures for authentication and preservation; such measures will ensure that record-keeping systems meet the criteris of "intergrity, currency an relevancy" necessary to the records creator. <P26> In other words, effective description is a consequence of effective records management and intelligent appraisal, not their purpose. If the primary objectives of metadata are met, description will be fascilitated and the need for description at lower levels (e.g., below the series level) may even be obviated. <P27> Metadata systems cannot and should not replace archival description. To meet the challenges posed by electronic records, it is more important than ever that we follow the dictates of archival science, which begin from a consideration of the nature of archives. <P28> Archival participation in the design and maintenance of metadata systems must be driven by the need to preserve them as archival documents, that is, as evidence of actions and transactions, not as descriptive tools. Our role is not to promote our own intersts, but to deepen the creator's understanding of its interests in preserving the evidence of its own actions and transactions. We can contribute to that understanding because we have a broader view of the creator's needs over time. In supporting these interests, we indirectly promote our own. <P29> To ensure that our descriptive infrastructure is sound -- that is to say, comprehensible, flexible, efficient, and effective -- we need equally to analyze our own information management methods and, out of that analysis, to develop complementary systems of administrative and intellectual control that will build upon each other. By these means we will be able to accomodate the diversity and complexity of the record-keeping environments with which we must deal.
Conclusions
RQ "Since 'current metadata systems do not account for the provenancial and contextual information needed to manage archival records,' archivists are exhorted [by Margaret Hedstrom] to direct their research efforts (and research dollars) toward the identification of the types of metadata that ought to be captured and created to meet archival descriptive requirements. "
SOW
DC Dr. Heather MacNeil is an Assistant Professor at the School of Library, Archival, and Information Studies at the University of British Columbia. Dr. MacNeilÔÇÖs major areas of interests include: trends and themes in archival research & scholarship; arrangement and description of archival documents; management of current records; trustworthiness of records as evidence; protection of personal privacy; interdisciplinary perspectives on record trustworthiness; and archival preservation of authentic electronic records
Type
Journal
Title
Challenges for service providers when importing metadata in digital libraries
CA Problems in implementing metadata for online resource discovery, in this case for digital libraries, will not be solved simply by adopting a common schema. Intellectual property rights remain another major obstacle to be dealt with.
Phrases
RQ Under what circumstances can metadata be altered? How should the copyright information of a resource be distinguished from the copyright information of its metadata? Will an audit trail be used as metadata shared with other repositories?
This article provides an overview of evolving Australian records continuum theory and the records continuum model, which is interpreted as both a metaphor and a new worldview, representing a paradigm shift in Kuhn's sense. It is based on a distillation of research findings drawn from discourse, literary warrant and historical analysis, as well as case studies, participant observation and reflection. The article traces the emergence in Australia in the 1990s of a community of practice which has taken continuum rather than life cycle based perspectives, and adopted postcustodial approaches to recordkeeping and archiving. It "places" the evolution of records continuum theory and practice in Australia in the context of a larger international discourse that was reconceptualizing traditional theory, and "reinventing" records and archives practice.
Publisher
Kluwer Academic Publishers
Publication Location
The Netherlands
Critical Arguements
CA Looks at the development of the Australian community of practice that led to records continuum theory: an approach that, in contrast to the North American life cycle approach, sees recordkeeping and archival practices as part of the same continuum of activities. Since the 1990s, there has been a lively debate between proponents of these two different ways of thinking. The second part of the article is highly theoretical, situating records continuum theory in the larger intellectual trend toward postmodernism and postpositivism.
Phrases
<P1> The model was built on a unifying concept of records inclusive of archives, which are defined as records of continuing value. It also drew on ideas about the "fixed" and "mutable" nature of records, the notion that records are ÔÇ£always in a process of becoming." (p. 334). <P2> Continuum ideas about the nature of records and archives challenge traditional understandings which differentiate "archives" from "records" on the basis of selection for permanent preservation in archival custody, and which focus on their fixed nature. Adopting a pluralist view of recorded information, continuum thinking characterises records as a special genre of documents in terms of their intent and functionality. It emphasises their evidentiary, transactional and contextual nature, rejecting approaches to the definition of records which focus on their subject content and informational value. (p. 335) <P3> [R]ecordkeeping and archiving processes ... help to assure the accessibility of meaningful records for as long as they are of value to people, organisations, and societies ÔÇô whether that be for a nanosecond or millennia. (p. 336) <P4> [I]f North American understandings of the term record keeping, based on life cycle concepts of records management, are used to interpret the writings of members of the Australian recordkeeping community, there is considerable potential for misunderstanding. <P5> Members of the recordkeeping and archiving community have worked together, often in partnership with members of other records and archives communities, on a range of national policy and standards initiatives, particularly in response to the challenge of electronic recordkeeping. These collaborative efforts resulted in AS 4390, the Australian Standard: Records Management (1996), the Australian Council of Archives' Common Framework for Electronic Recordkeeping (1996), and the Australian Records and Archives Competency Standards (1997). In a parallel and interconnected development, individual archival organisations have been developing electronic recordkeeping policies, standards, system design methodologies, and implementation strategies for their jurisdictions, including the National Archives of Australia's suite of standards, policies, and guidelines under the e-permanence initiative launched in early 2000. These developments have been deliberately set within the broader context of national standards and policy development frameworks. Two of the lead institutions in these initiatives are the National Archives of Australia and the State Records Authority of New South Wales, which have based their work in this area on exploration of fundamental questions about the nature of records and archives, and the role of recordkeeping and archiving in society. <warrant> (p. 339) <P6> In adopting a continuum-based worldview and defining its "place" in the world, the Australian recordkeeping and archiving community consciously rejected the life cycle worldview that had dominated records management and archives practice in the latter half of the 20th century in North America. ... They were also strong advocates of the nexus between accountable recordkeeping and accountability in a democratic society, and supporters of the dual role of an archival authority as both a regulator of current recordkeeping, and preserver of the collective memory of the state/nation. (p. 343-344) <P7> [P]ost-modern ideas about records view them as dynamic objects that are fixed in terms of content and meaningful elements of their structure, but linked to ever-broadening layers of contextual metadata that manages their meanings, and enables their accessibility and useability as they move through "spacetime." (p. 349) <P8> In exploring the role of recordkeeping and archiving professionals within a postcustodial frame of reference, archival theorists such as Brothman, Brown, Cook, Harris, Hedstrom, Hurley, Nesmith, and Upward have concluded that they are an integral part of the record and archive making and keeping process, involved in society's remembering and forgetting. (p. 355) <P9> Writings on the societal context of functional appraisal have gone some way to translate into appraisal policies and strategies the implications of the shifts in perception away from seeing records managers as passive keepers of documentary detritus ... and archivists as Jenkinson's neutral, impartial custodians of inherited records. (p. 355-356)
Conclusions
RQ "By attempting to define, to categorise, pin down, and represent records and their contexts of creation, management, and use, descriptive standards and metadata schema can only ever represent a partial view of the dynamic, complex, and multi-dimensional nature of records, and their rich webs of contextual and documentary relationships. Within these limitations, what recordkeeping metadata research is reaching towards are ways to represent records and their contexts as richly and extensively as possible, to develop frameworks that recognise their mutable and contingent nature, as well as the role of recordkeeping and archiving professionals (records managers and archivists) in their creation and evolution, and to attempt to address issues relating to time and space." (p. 354)
Type
Journal
Title
Accessing essential evidence on the web: Towards an Australian recordkeeping metadata standard
CA Standardized recordkeeping metadata allows for access to essential evidence of business activities and promotes reliability and authenticity. The Australian records and metadata community have been working hard to define standards and identify requirements as well as support interoperability.
Phrases
<P1> But records, as accountability traces and evidence of business activity, have additional metadata requirements. Authoritative, well-structured metadata which specifies their content, structure, context, and essential management needs must be embedded in, wrapped around and otherwise persistently linked to them from the moment they are created if they are to continue to function as evidence. (p.2) <P2> People do business in social and organizational contexts that are governed by external mandates (e.g. social mores, laws) and internal mandates (e.g. policies, business rules). Mandates establish who is responsible for what, and govern social and organizational activity, including the creation of full and accurate records. <warrant> (p.3)
Type
Journal
Title
Describing Records in Context in the Continuum: The Australian Recordkeeping Metadata Schema
CA RKMS is based on traditional recordkeeping thinking. However, it also looks to the future by viewing records as active agents of change, as intelligent information objects, which are supported by the metadata that RKMS' framework provides. Through RKMS, the dynamic world of business can be linked to the more passive world of cyberspace resource management.
Phrases
<P1> As long as records remain in the local domains in which they are created, a lot of broader contextual metadata is "in the air," carried in the minds of the corporate users of the records. When records move beyond the boundaries of the local domain in which they are created or, as is increasingly the case in networked environments, they are created in the first place in a global rather than a local domain, then this kind of metadata needs to be made explicit -- that is, captured and persistently linked to the record. This is essential so that users in the broader domain can uniquely identify, retrieve and understand the meanings of records. (p.7) <P2> The broader social context of the project is the need for individuals, society, government, and commerce to continually access the information they need to conduct their business, protect their rights and entitlements, and securely trace the trail of responsibility and action in distributed enterprises. ... Maintaining reliable, authentic and useable evidence of transactions through time and space has significant business, social, and cultural implications, as records provide essential evidence for purposes of governance, accountability, memory and identity. (p.6)
Conclusions
RQ There is a need to develop typologies of recordkeeping relationships such as agent to record and better ways to express them through metadata.
Type
Journal
Title
Towards Frameworks for Standardising Recordkeeping Metadata
CA There are many challenges to devising metadata schema to manage records over time. Continuum thinking provides a conceptual framework to identify these problems.
Phrases
<P1> It is clear from the SPIRT Project definition that recordkeeping and archival control systems have always been about capturing and managing recordkeeping metadata. (p.30) <P2> One of the keys to understanding the Project's approach to what metadata needs to be captured, persistently linked to documentation of social and business activity, and managed through space and time, lies in the continuum view of records. In continuum thinking, [records] are seen not as 'passive objects to described retrospectively,' but as agents of action, 'active participants in business processes and technologies.'" (p.37)
CA Information and communications technology (ICTs) can positively or negatively affect the availability of records for accountability. ICTs challenges basic organizational values in records management, such as finding a balance between central control and autonomy. A well-designed electronic records management plan needs to take into account the various values of an organization's "accountability situation."
Phrases
<P1> A systematic records management policy is required to ensure that the appropriate records will be available for accountability processes. This involves developing and maintaining policies, procedures, and methodologies, appointing experts, and creating and managing specialized departments." (p .260) The fact that many records are not converted and are handled exclusively in an electronic form may have an important impact on the content and quality of records kept for purposes of accountability." (p. 260)
Conclusions
RQ We must look at the tools of information and communications technology closely. Often researchers assume that tools only need to be shaped or adapted in a certain way in order to address a variety of problems in electronic records management, while ignoring the fact that the shaping of a tool is in itself highly controversial.
CA The want for hard unassailable recordkeeping rules ignores the fact that recordkeeping is contingent upon unique needs of each organization as far as acceptable risks and context. Reed argues that aiming to achieve basic agreement on a minimal set of metadata attributes is an important start.
Phrases
<P1> Recordkeeping must be tailored to the requirements of specific business functions and activities linked to related social and legal requirements, incorporated into particular business processes, and maintained through each change to those processes. (p. 222) <P2> A record core or metadata set which lacks such specificity, detailing only requirements for a unique identifier, will not support interpretation of the record outside the creating domain. To enable that, we need more detailed specification of the domain itself, data which is redundant when you know where you are, but essential to understanding and interpreting records where the domain is not explicit. (p. 229)
Conclusions
RQ To establish requirements for viable core elements, the big challenge is the issue of time and that data will change over time ÔÇöespecially as far as individual competence, business function and language.
One such expedient could be more structured and more integrated use of formal and institutional data on records and archives. I cannot offer any completed model of this enhanced perspective, and as far as I know, one does not exist. However, it is a new way of thinking and looking at the problems we encounter. What I would like to do is draw attention to some of the approaches now being developed in The Netherlands. In a way, this presentation will therefore be a report on the Dutch arvhival situation.
Critical Arguements
CA "In a world defined by the enormous size of archives, where the multiplicity of records is in turn driven by the growing complexity of society and its administration, and by the proliferation of types of 'information carriers', it is becoming increasingly difficult fpr archivists to fulfill their primary tasks. It is therefore necessary to study carefully the development of maintenance and control mechanisms for archives. We cannot afford waste or overlook any possibility. It is also necessary to look around us, to discover what other archivists in other countries are doing, and what others in related fields, such as libraries and museums, have accomplished. Essentially, we all deal with the same problems and must try to find new solutions to master these problems."
Phrases
<P1> Document forms can be regarded as forms of objects. We probably need to gain more experience in recognizing different forms of documents and interpreting them, but once we have this knowledge, we can use it in the same way as we now use 'form' in its archival sense: to distinguish one object from another. <P2> In fact, by extension, one can even construct and defend the thesis that all decisions in an administration are reached using standard procedures and forms. Once this is realized, one can ask: what use to archivists make of this knowledge in their daily work? What are the possibilities? <P3> Often the forms of materials created prove to be of a more consistent nature than the offices that use them. If an office ceases its activity, another will take over its tasks and for the most part will use the same or almost the same forms of material. <P4> Understanding the functions of the organization will provide archivists not only with information about the material involved, but also with knowledge of the procedures, which in turn provides information about the records and their different forms. This kind of sympathetic understanding enables archivists to make all kinds of decisions, and it is important to note that at least part of this knowledge should be provided to the users, so that they can decide which records might be of interest to them. <warrant> <P5> We are increasingly aware that we must distinguish between processing an archive (i.e. organizing records according to archival principles after appraisal) and making the contents available for users through finding aids, indexes and other means. <P6> With respect to the latter, it is clear that archivists should make use of both context- and provenance-based indexing. They should take advantage of the possiblities offered by the structures and forms of material -- something which the librarian cannot do. Furthermore, they should also use content indexing in a selective way, only when they think it necessary [to] better serve researchers. <warrant> <P7> The National Archives in The Hague has responded to these new perspectives by developing a computer programme called MAIS (Micro Archives Inventory System), which is a formal way of processing archives based on provenance. <P8> The object of this presentation has been to show that use of structure, forms of material and functions, can aid the archivist in his/her work.
Conclusions
RQ "While these initial Dutch efforts have been produced in a rather unorganized way, it should nevertheless be possible to approach the work more systematically in [the] future, building up a body of knowledge of forms for users of archives. David Bearman has offered some preliminary suggestions in this direction, in the article cited above; it is now a matter of more research required to realize something positive in this field."
SOW
DC J. Peter Sigmond is Director of Collections at the Rijksmuseum in Amsterdam, the Netherlands
Type
Journal
Title
Structuring the Records Continuum Part Two: Structuration Theory and Recordkeeping
In the previous issue of Archives and Manuscripts I presented the first part of this two part exploration. It dealt with some possible meanings for 'post' in the term postcustodial. For archivists, considerations of custody are becoming more complex because of changing social, technical and legal considerations. These changes include those occurring in relation to access and the need to document electronic business communications reliably. Our actions, as archivists, in turn become more complex as we attempt to establish continuity of custody in electronic recordkeeping environments. In this part, I continue the case for emphasising the processes of archiving in both our theory and practice. The archives as a functional structure has dominated twentieth century archival discourse and institutional ordering, but we are going through a period of transformation. The structuration theory of Anthony Giddens is used to show that there are very different ways of theorising about our professional activities than have so far been attempted within the archival profession. Giddens' theory, at the very least, provides a useful device for gaining insights into the nature of theory and its relationship with practice. The most effective use of theory is as a way of seeing issues. When seen through the prism of structuration theory, the forming processes of the virtual archives are made apparent.
Critical Arguements
CA "This part of my exploration of the continuum will continue the case for understanding 'postcustodial' as a bookmark term for a major transition in archival practice. That transition involves leaving a long tradition in which continuity was a matter of sequential control. Electronic recordkeeping processes need to incorporate continuity into the essence of recordkeeping systems and into the lifespan of documents within those systems. In addressing this issue I will present a structurationist reading of the model set out in Part 1, using the sophisticated theory contained in the work of Anthony Giddens. Structuration theory deals with process, and illustrates why we must constantly re-assess and adjust the patterns for ordering our activities. It gives some leads on how to go about re-institutionalising these new patterns. When used in conjunction with continuum thinking, Giddens' meta-theory and its many pieces can help us to understand the complexities of the virtual archives, and to work our way towards the establishment of suitable routines for the control of document management, records capture, corporate memory, and collective memory."
Phrases
<P1> Broadly the debate has started to form itself as one between those who represent the structures and functions of an archival institution in an idealised form, and those who increasingly concentrate on the actions and processes which give rise to the record and its carriage through time and space. In one case the record needs to be stored, recalled and disseminated within our institutional frameworks; in the other case it is the processes for storing, recalling, and disseminating the record which need to be placed into a suitable framework. <P2> Structure, for Giddens, is not something separate from human action. It exists as memory, including the memory contained within the way we represent, recall, and disseminate resources including recorded information. <P3> Currently in electronic systems there is an absence of recordkeeping structures and disconnected dimensions. The action part of the duality has raced ahead of the structural one; the structuration process has only just begun. <P4> The continuum model's breadth and richness as a conceptual tool is expanded when it is seen that it can encompass action-structure issues in at least three specialisations within recordkeeping: contemporary recordkeeping - current recordkeeping actions and the structures in which they take place; regulatory recordkeeping - the processes of regulation and the enabling and controlling structures for action such as policies, standards, codes, legislation, and promulgation of best practices; historical recordkeeping - explorations of provenance in which action and structure are examined forensically as part of the data sought about records for their storage, recall and dissemination. <P5> The capacity to imbibe information about recordkeeping practices in agencies will be crucial to the effectiveness of the way archival 'organisations' set up their postcustodial programs. They will have to monitor the distribution and exercise of custodial responsibilities for electronic records from before the time of their creation. <warrant> <P6> As John McDonald has pointed out, recordkeeping activities need to occur at desktop level within systems that are not dependent upon the person at the desktop understanding all of the details of the operation of that system. <P7> Giddens' more recent work on reflexivity has many parallels with metadata approaches to recordkeeping. What if the records, as David Bearman predicts, can be self-managing? Will they be able to monitor themselves? <P8> He rejects the life cycle model in sociology, based on ritualised passages through life, and writes of 'open experience thresholds'. Once societies, for example, had rites for coming of age. Coming of age in a high modern society is now a complex process involving a host of experiences and risks which are very different to that of any previous generation. Open experience threshholds replace the life cycle thresholds, and as the term infers, are much less controlled or predictable. <P9> There is a clear parallel with recordkeeping in a high modern environment. The custodial thresholds can no longer be understood in terms of the spatial limits between a creating agency and an archives. The externalities of the archives as place will decline in significance as a means of directly asserting the authenticity and reliability of records. The complexities of modern recordkeeping involve many more contextual relationships and an ever increasing network of relationships between records and the actions that take place in relation to them. We have no need for a life cycle concept based on the premise of generational repetition of stages through which a record can be expected to pass. We have entered an age of more recordkeeping choices and of open experience thresholds. <P10> It is the increase in transactionality, and the technologies being used for those transactions, which are different. The solution, easier to write about than implement, is for records to parallel Giddens' high modern individual and make reflexive use of the broader social environment in which they exist. They can reflexively monitor their own action and, with encoding help from archivists and records managers, resolve their own crises as they arise. <warrant> <P11> David Bearman's argument that records can be self-managing goes well beyond the easy stage. It is supported by the Pittsburgh project's preliminary set of metadata specifications. The seeds of self-management can be found in object oriented programming, java, applets, and the growing understanding of the importance and nature of metadata. <P12> Continuum models further assist us to conceive of how records, as metadata encapsulated objects, can resolve many of their own life crises as they thread their way through time and across space. <P13> To be effective monitors of action, archival institutions will need to be recognised by others as the institutions most capable of providing guidance and control in relation to the integration of the archiving processes involved in document management, records capture, the organisation of corporate memory and the networking of archival systems. <warrant> <P14> Signification, in the theoretical domain, refers to our interpretative schemes and the way we encode and communicate our activities. At a macro level this includes language itself; at a micro level it can include our schemes for classification and ordering. <P15> The Pittsburgh project addressed the three major strands of Giddens' theoretical domain. It explored and set out functional requirements for evidence - signification. It sought literary warrants for archival tasks - legitimation. It reviewed the acceptability of the requirements for evidence within organisational cultures - domination. <P16> In Giddens' dimensional approach, the theoretical domain is re-defined to be about coding, organising our resources, and developing norms and standards. In this area the thinking has already begun to produce results, which leads this article in to a discussion of structural properties. <P17> Archivists deal with structural properties when, for example, they analyse the characteristics of recorded information such as the document, the record, the archive and the archives. The archives as a fortress is an observable structural property, as is the archives as a physical accumulation of records. Within Giddens' structuration theory, when archivists write about their favourite features, be they records or the archives as a place, they are discussing structural properties. <P18> Postcustodial practice in Australia is already beginning to put together a substantial array of structural properties. These developments are canvassed in the article by O'Shea and Roberts in the previous issue of Archives and Manuscripts. They include policies and strategies, standards, recordkeeping regimes, and what has come to be termed distributed custody. <P19> As [Terry] Eastwood comments in the same article, we do not have adequate electronic recordkeeping systems. Without them there can be no record in time-space to serve any form of accountability. <warrant> <P20> In the Pittsburgh project, for example, the transformation of recordkeeping processes is directed towards the creation and management of evidence, and possible elements of a valid rule-resource set have emerged. Elements can include the control of recordkeeping actions, accountability, the management of risk, the development of recordkeeping regimes, the establishment of recordkeeping requirements, and the specification of metadata. <P21> In a postcustodial approach it is the role of archival institutions to foster better recordkeeping practices within all the dimensions of recordkeeping. <warrant>
Conclusions
RQ "Best practice in the defence of the authoritative qualities of records can no longer be viewed as a linear chain, and the challenge is to establish new ways of legitimating responsibilities for records storage and custody which recognise the shifts which have occurred." ... "The recordkeeping profession should seek to establish itself as ground cover, working across terrains rather than existing tree-like in one spot. Beneath the ground cover there are shafts of specialisation running both laterally and vertically. Perhaps we can, as archivists, rediscover something that a sociologist like Giddens has never forgotten. Societies, including their composite parts, are the ultimate containers of recorded information. As a place in society, as Terry Cook argues, the archives is a multiple reality. We can set in train policies and strategies that can help generate multiplicity without losing respect for particular mine shafts. Archivists have an opportunity to pursue policies which encourage the responsible exercising of a custodial role throughout society, including the professions involved in current, regulatory and historical recordkeeping. If we take up that opportunity, our many goals can be better met and our concerns will be addressed more effectively."
SOW
DC "Frank Upward is a senior lecturer in the Department of Librarianship, Archives and Records at Monash University. He is an historian of the ideas contained in the Australian records continuum approach, and an ex- practitioner within that approach." ... "These two articles, and an earlier one on Ian Maclean and the origins of Australian continuum thinking, have not, so far, contained appropriate acknowledgements. David Bearman provided the necessary detonation of certain archival practices, and much more. Richard Brown and Terry Cook drew my attention to Anthony Giddens' work and their own work has helped shape my views. I have many colleagues at Monash who encourage my eccentricities. Sue McKemmish has helped shape my ideas and my final drafts and Barbara Reed has commented wisely on my outrageous earlier drafts. Livia Iacovino has made me stop and think more about the juridical tradition in recordkeeping. Chris Hurley produced many perspectives on the continuum during the 1996 seminars which have helped me see the model more fully. Don Schauder raised a number of key questions about Giddens as a theorist. Bruce Wearne of the Sociology Department at Monash helped me lift the clarity of my sociological explanations and made me realise how obsessed Giddens is with gerunds. The structural-functionalism of Luciana Duranti and Terry Eastwood provided me with a counterpoint to many of my arguments, but I also owe them debts for their respective explorations of recordkeeping processes and the intellectual milieu of archival ideas, and for their work on the administrative-juridical tradition of recordkeeping. Glenda Acland has provided perceptive comments on my articles - and supportive ones, for which I am most grateful given how different the articles are from conventional archival theorising. Australian Archives, and its many past and present staff members, has been important to me."
Type
Journal
Title
Structuring the Records Continuum Part One: Post-custodial principles and properties
The records continuum is becoming a much used term, but has seldom been defined in ways which show it is a time/space model not a life of the records model. Dictionary definitions of a continuum describe such features as its continuity, the indescernibility of its parts, and the way its elements pass into each other. Precise definitions, accordingly, have to discern the indiscernible, identify points that are not distinct, and do so in ways which accomodate the continuity of change. This article, and a second part to be published in the next volume, will explore the continuum in time/space terms supported by a theoretical mix of archival science, postmodernity and the 'structuration theory' of Anthony Giddens. In this part the main objectives are to give greater conceptual firmness to the continuum; to clear the way for broader considerations of the nature of the continuum by freeing archivists from the need to debate custody; to show how the structural principles for archival practice are capable of different expression without losing contact with something deeper that can outlive the manner of expression.
Critical Arguements
CA "This is the first instalment of a two part article exploring the records continuum. Together the articles will build into a theory about the constitution of the virtual archives. In this part I will examine what it can mean to be 'postcustodial', outline some possible structural principles for the virtual archives, and present a logical model for the records continuum." ... "In what follows in the remainder of this article (and all of the next) , I will explore the relevance of [Anthony] Giddens' theory to the structuring of the records continuum."
Phrases
<P1> If the archival profession is to avoid a fracture along the lines of paper and electronic media, it has to be able to develop ways of expressing its ideas in models of relevance to all ages of recordkeeping, but do so in ways which are contemporaneous with our own society. <warrant> <P2> We need more of the type of construct provided by the Pittsburgh Project's functional requirements for evidence which are 'high modern' but can apply to recordkeeping over time. <P3> What is essential is for electronic records to be identified, controlled and accessible for as long as they have value to Government and the Community. <warrant> <P4> We have to face up to the complexification of ownership, possession, guardianship and control within our legal system. Even possession can be broken down into into physical possession and constructed possession. We also have to face the potential within our technology for ownership, possession, custody or control to be exercised jointly by the archives, the organisation creating the records, and auditing agencies. The complexity requires a new look at our way of allocating authorities and responsibilities. <P5> In what has come to be known as the continuum approach Maclean argued that archivists should base their profession upon studies of the characteristics of recorded information, recordkeeping systems, and classification (the way the records were ordered within recordkeeping systems and the way these were ordered through time). <P6> A significant role for today's archival institution is to help to identify and establish functional requirements for recordkeeping that enable a more systematic approach to authentication than that provided by physical custody. <warrant> <P7> In an electronic work environment it means, in part, that the objectivity, understandability, availability, and usability of records need to be inherent in the way that the record is captured. In turn the documents need to be captured in the context of the actions of which they are part, and are recursively involved. <warrant> <P8>A dimensional analysis can be constructed from the model and explained in a number of ways including a recordkeeping system reading. When the co-ordinates of the continuum model are connected, the different dimensions of a recordkeeping system are revealed. The dimensions are not boundaries, the co-ordinates are not invariably present, and things may happen simultaneously across dimensions, but no matter how a recordkeeping system is set up it can be analysed in terms such as: first dimensional analysis: a pre- communication system for document creation within electronic systems [creating the trace]; second dimensional analysis: a post- communication system, for example traditional registry functionality which includes registration, the value adding of data for linking documents and disseminating them, and the maintenance of the record including disposition data [capturing trace as record]; third dimensional analysis: a system involving building, recalling and disseminating corporate memory [organising the record as memory]; fourth dimensional analysis: a system for building, recalling and disseminating collective memory (social, cultural or historical) including information of the type required for an archival information system [pluralizing the memory]. <P9> In the high modern recordkeeping environment of the 1990's a continuum has to take into account a different array of recordkeeping tools. These tools, plucking a few out at random but ordering the list dimensionally, include: document management software, Australian records system software, the intranet and the internet. <P10> In terms of a records continuum which supports an evidence based recordkeeping approach, the second dimension is crucial. This is where the document is disembedded from the immediate contexts of the first dimension. It is this disembedding process that gives the record its value as a 'symbolic token'. A document is embedded in an act, but the document as a record needs to be validatable using external reference points. These points include the operation of the recordkeeping system into which it was received, and information pertaining to the technical, social (including business) and communication processes of which the document was part.
Conclusions
RQ "Postcustodial approaches to archives and records cannot be understood if they are treated as a dualism. They are not the opposite of custody. They are a response to opportunities for asserting the role of an archives - and not just its authentication role - in many re-invigorating ways, a theme which I will explore further in the next edition of Archives and Manuscripts."
SOW
DC "Frank Upward is a senior lecturer in the Department of Librarianship, Archives and Records at Monash University. He is an historian of the ideas contained in the Australian records continuum approach, and an ex-practitioner within that approach."
CA Strength of a metadata structure lies in its ability to layer and exchange information from a wide variety of creators in a "loosely coupled system of organization." This review covers metadata literature from mid 1996 through early 1998. It focuses on the development and application of metadata standards used by the LIS community for resource description, discovery and retrieval within a digital environment.
Type
Journal
Title
Archives and the information superhighway: Current status and future challenges
CA One struggle facing us is to convince the rest of society that the ÔÇ£information superhighwayÔÇØ is very much about records, evidence and ÔÇ£recordnessÔÇØ.
Phrases
<P1> It has been argued that existing computer software applications harm recordkeeping because they are remiss in capturing the full breadth of contextual information required to document transactions and create records -- records which can serve as reliable evidence of the transactions which created them. In place of records, these systems are producing data which fails to relate the who, what, when, where, and why of human communications -- attributes which are required for record evidence. This argument has found both saliency and support in other work conducted by the Netherlands and the World Bank, which have both noted that existing software applications fail to provide for the capture of the required complement of descriptive attributes required for proper recordkeeping. These examples point to the vast opportunity presented to archivists to position themselves as substantive contributors to information infrastructure discussions. Archivists are capable of pointing out what will be necessary to create records in the electronic environment which, in the words of David Bearman, meet the requirements of ÔÇ£business acceptable commincation. (p.87) <warrant>
Conclusions
RQ Can archivists provide access to information in the unstable electronic records environment we find ourselves in today?
Type
Journal
Title
Managing the Present: Metadata as Archival Description
Traditional archival description undertaken at the terminal stages of the life cycle has had two deleterious effects on the archival profession. First, it has resulted in enormous, and in some cases, insurmountable processing backlogs. Second, it has limited our ability to capture crucial contextual and structural information throughout the life cycle of record-keeping systems that are essential for fully understanding the fonds in our institutions. This shortcoming has resulted in an inadequate knowledge base for appraisal and access provision. Such complications will only become more magnified as distributed computering and complex software applications continue to expand throughout organizations. A metadata strategy for archival description will help mitigate these problems and enhance the organizational profile of archivists who will come to be seen as valuable organizational knowledge and accountability managers.
Critical Arguements
CA "This essay affirms this call for evaluation and asserts that the archival profession must embrace a metadata systems approach to archival description and management." ... "It is held here that the requirements for records capture and description are the requirements for metadata."
Phrases
<P1> New archival organizational structures must be created to ensure that records can be maintained in a usable form. <warrant> <P2> The recent report of Society of American Archivists (SAA) Committee on Automated Records and Techniques (CART) on curriculum development has argued that archivists need to "understand the nature and utility of metadata and how to interpret and use metadata for archival purposes." <warrant> <P3> The report advises archivists to acquire knowledge on the meanings of metadata, its structures, standards, and uses for the management of electronic records. Interestingly, the requirements for archival description immediately follow this section and note that archivists need to isolate the descriptive requirements, standards, documentiation, and practices needed for managing electronic records. <warrant> <P4> Clearly, archivists need to identify what types of metadata will best suit their descriptive needs, underscoring the need for the profession to develop strategies aand tactics to satisfy these requirements within active software environments. <warrant> <P5> Underlying the metadata systems strategy for describing and managing electronic information technologies is the seemingly universal agreement amongst electronic records archivists on the requirement to intervene earlier in the life cycle of electronic information systems. <warrant> <P6> Metadata has loomed over the archival management of electronic records for over five years now and is increasingly being promised as a basic control strategy for managing these records. <warrant> <P7> However, she [Margaret Hedstrom] also warns that as descriptive practices shift from creating descriptive information to capturing description along with the records, archivists may discover that managing the metadata is a much greater challenge than managing the records themselves. <P8> Archivists must seek to influence the creation of record-keeping systems within organizations by connecting the transaction that created the data to the data itself. Such a connection will link informational content, structure, and the context of transactions. Only when these conditions are met will we have records and an appropriate infrastructure for archival description. <warrant> <P9> Charles Dollar has argued that archivists increasingly will have to rely upon and shape the metadata associated with electronic records in order to fully capture provenance information about them. <warrant> <P10> Bearman proposes a metadata systems strategy, which would focus more explicitly on the context out of which records arise, as opposed to concentrating on their content. This axiom is premised on the assumption that "lifecycle records systems control should drive provenance-based description and link to top-down definitions of holdings." <warrant> <P11> Bearman and Margaret Hedstrom have built upon this model and contend that properly specified metadata capture could fully describe sytems while they are still active and eliminate the need for post-hoc description. The fundamental change wrought in this approach is the shift from doing things to records (surveying, scheduling, appraising, disposing/accessioning, describing, preserving, and accessing) to providing policy direction for adequate documentation through management of organizational behavior (analyzing organizational functions, defining business transactions, defining record metadata, indentifying control tactics, and establishing the record-keeping regime). Within this model archivists focus on steering how records will be captured (and that they will be captured) and how they will be managed and described within record-keeping systems while they are still actively serving their parent organization. <P12> Through the provision of policy guidance and oversight, organizational record-keeping is managed in order to ensure that the "documentation of organizational missions, functions, and responsibilities ... and reporting relationships within the organization, will be undertaken by the organizations themselves in their administrative control systems." <warrant> <P13> Through a metadata systems approach, archivists can realign themselves strategically as managers of authoritative information about organizational record-keeping systems, providing for the capture of information about each system, its contextual attributes, its users, its hardware configurations, its software configurations, and its data configurations. <warrant> <P14> The University of Pittsburgh's functional requirements for record-keeping provides a framework for such information management structure. These functional requirements are appropriately viewed as an absolute ideal, requiring testing within live systems and organizations. If properly implemented, however, they can provide a concrete model for metadata capture that can automatically supply many of the types of descriptive information both desired by archivists and required for elucidating the context out of which records arise. <P15> It is possible that satisfying these requirements will contribute to the development of a robust archival description process integrating "preservation of meaning, exercise of control, and provision of access'" within "one prinicipal, multipurpose descriptive instrument" hinted at by Luciana Duranti as a possible outcome of the electronic era. <P16> However, since electronic records are logical and not physical entities, there is no physical effort required to access and process them, just mental modelling. <P17> Depending on the type of metadata that is built into and linked to electronic information systems, it is possible that users can identify individual records at the lowest level of granularity and still see the top-level process it is related to. Furthermore, records can be reaggregated based upon user-defined criteria though metadata links that track every instance of their use, their relations to other records, and the actions that led to their creation. <P18> A metadata strategy for archival description will help to mitigate these problems and enhance the organizational profile of archivists, who will come to be seen as valuable organizational knowledge and accountability managers. <warrant>
Conclusions
RQ "First and foremost, the promise of metadata for archival description is contingent upon the creation of electronic record-keeping systems as opposed to a continuation of the data management orientation that seems to dominate most computer applications within organizations." ... "As with so many other aspects of the archival endeavour, these requirements and the larger metadata model for description that they are premised upon necessitate further exploration through basic research."
SOW
DC "In addition to New York State, recognition of the failure of existing software applications to capture a full compliment of metadata required for record-keeping and the need for such records management control has also been acknowledged in Canada, the Netherlands, and the World Bank." ... "In conjunction with experts in electronic records managment, an ongoing research project at the University of Pittsburgh has developed a set of thirteen functional requirements for record-keeping. These requirements provide a concrete metadata tool sought by archivists for managing and describing electronic records and electronic record-keeping systems." ... David A. Wallace is an Assistant Professor at the School of Information, University of Michigan, where he teaches in the areas of archives and records management. He holds a B.A. from Binghamton University, a Masters of Library Science from the University at Albany, and a doctorate from the University of Pittsburgh. Between 1988 and 1992, he served as Records/Systems/Database Manager at the National Security Archive in Washington, D.C., a non-profit research library of declassified U.S. government records. While at the NSA he also served as Technical Editor to their "The Making of U.S. Foreign Policy" series. From 1993-1994, he served as a research assistant to the University of Pittsburgh's project on Functional Requirements for Evidence in Recordkeeping, and as a Contributing Editor to Archives and Museum Informatics: Cultural Heritage Informatics Quarterly. From 1994 to 1996, he served as a staff member to the U.S. Advisory Council on the National Information Infrastructure. In 1997, he completed a dissertation analyzing the White House email "PROFS" case. Since arriving at the School of Information in late 1997, he has served as Co-PI on an NHPRC funded grant assessing strategies for preserving electronic records of collaborative processes, as PI on an NSF Digital Government Program funded planning grant investigating the incorporation of born digital records into a FOIA processing system, co-edited Archives and the Public Good: Accountability and Records in Modern Society (Quorum, 2002), and was awarded ARMA International's Britt Literary Award for an article on email policy. He also serves as a consultant to the South African History Archives Freedom of Information Program and is exploring the development of a massive digital library of declassified imaged/digitized U.S. government documents charting U.S. foreign policy.
Annual Review of Information Science and technology
Periodical Abbreviation
ARIST
Publication Year
2001
Volume
35
Pages
337
Critical Arguements
CA Yakel gives an overview of the literature on digital preservation from the early 1980s through 2000.
Phrases
<P1> The immediate question is whether the industry will adopt these standards, strategies, and functional elements to create evidence-based recordkeeping systems that ensure their authenticity and reliability. If the model of the DOD guidelines is any indication, some sectors of the vendor population will respond to recordkeeping specifications if there is sufficient customer leverage. Recordkeeping requirements rely on various metadata schemes and the viability of standards. (p.366) <warrant>
Conclusions
RQ More research into the hybrid approach (emulation and migration) is needed to determine criteria that will ensure the preservation of authenticity and reliability.
Type
Electronic Journal
Title
Keeping Memory Alive: Practices for Preserving Digital Content at the National Digital Library Program of the Library of Congress
CA An overview of the major issues and initiatives in digital preservation at the Library of Congress. "In the medium term, the National Digital Library Program is focusing on two operational approaches. First, steps are taken during conversion that are likely to make migration or emulation less costly when they are needed. Second, the bit streams generated by the conversion process are kept alive through replication and routine refreshing supported by integrity checks. The practices described here provide examples of how those steps are implemented to keep the content of American Memory alive."
Phrases
<P1> The practices described here should not be seen as policies of the Library of Congress; nor are they suggested as best practices in any absolute sense. NDLP regards them as appropriate practices based on real experience, the nature and content of the originals, the primary purposes of the digitization, the state of technology, the availability of resources, the scale of the American Memory digital collection, and the goals of the program. They cover not just the storage of content and associated metadata, but also aspects of initial capture and quality review that support the long-term retention of content digitized from analog sources. <P2> The Library recognizes that digital information resources, whether born digital or converted from analog forms, should be acquired, used, and served alongside traditional resources in the same format or subject area. Such responsibility will include ensuring that effective access is maintained to the digital content through American Memory and via the Library's main catalog and, in coordination with the units responsible for the technical infrastructure, planning migration to new technology when needed. <P3> Refreshing can be carried out in a largely automated fashion on an ongoing basis. Migration, however, will require substantial resources, in a combination of processing time, out-sourced contracts, and staff time. Choice of appropriate formats for digital masters will defer the need for large-scale migration. Integrity checks and appropriate capture of metadata during the initial capture and production process will reduce the resource requirements for future migration steps. <warrant> We can be certain that migration of content to new data formats will be necessary at some point. The future will see industrywide adoption of new data formats with functional advantages over current standards. However, it will be difficult to predict exactly which metadata will be useful to support migration, when migration of master formats will be needed, and the nature and extent of resource needs. Human experts will need to decide when to undertake migration and develop tools for each migration step. <P4> Effective preservation of resources in digital form requires (a) attention early in the life-cycle, at the moment of creation, publication, or acquisition and (b) ongoing management (with attendant costs) to ensure continuing usability. <P5> The National Digital Library Program has identified several categories of metadata needed to support access and management for digital content. Descriptive metadata supports discovery through search and browse functions. Structural metadata supports presentation of complex objects by representing relationships between components, such as sequences of images. In addition, administrative metadata is needed to support management tasks, such as access control, archiving, and migration. Individual metadata elements may support more than one function, but the categorization of elements by function has proved useful. <P6> It has been recognized that metadata representations appropriate for manipulation and long-term retention may not always be appropriate for real-time delivery. <P7> It has also been realized that some basic descriptive metadata (at the very least a title or brief description) should be associated with the structural and administrative metadata. <P8> During 1999, an internal working group reviewed past experience and prototype exercises and compiled a core set of metadata elements that will serve the different functions identified. This set will be tested and refined as part of pilot activities during 2000. <P9> Master formats are well documented and widely deployed, preferably formal standards and preferably non-proprietary. Such choices should minimize the need for future migration or ensure that appropriate and affordable tools for migration will be developed by the industry. <warrant>
Conclusions
RQ "Developing long-term strategies for preserving digital resources presents challenges associated with the uncertainties of technological change. There is currently little experience on which to base predictions of how often migration to new formats will be necessary or desirable or whether emulation will prove cost-effective for certain categories of resources. ... Technological advances, while sure to present new challenges, will also provide new solutions for preserving digital content."
Type
Electronic Journal
Title
A Spectrum of Interoperability: The Site for Science Prototype for the NSDL
"Currently, NSF is funding 64 projects, each making its own contribution to the library, with a total annual budget of about $24 million. Many projects are building collections; others are developing services; a few are carrying out targeted research.The NSDL is a broad program to build a digital library for education in science, mathematics, engineering and technology. It is funded by the National Science Foundation (NSF) Division of Undergraduate Education. . . . The Core Integration task is to ensure that the NSDL is a single coherent library, not simply a set of unrelated activities. In summer 2000, the NSF funded six Core Integration demonstration projects, each lasting a year. One of these grants was to Cornell University and our demonstration is known as Site for Science. It is at http://www.siteforscience.org/ [Site for Science]. In late 2001, the NSF consolidated the Core Integration funding into a single grant for the production release of the NSDL. This grant was made to a collaboration of the University Corporation for Atmospheric Research (UCAR), Columbia University and Cornell University. The technical approach being followed is based heavily on our experience with Site for Science. Therefore this article is both a description of the strategy for interoperability that was developed for Site for Science and an introduction to the architecture being used by the NSDL production team."
ISBN
1082-9873
Critical Arguements
CA "[T]his article is both a description of the strategy for interoperability that was developed for the [Cornell University's NSF-funded] Site for Science and an introduction to the architecture being used by the NSDL production team."
Phrases
<P1> The grand vision is that the NSDL become a comprehensive library of every digital resource that could conceivably be of value to any aspect of education in any branch of science and engineering, both defined very broadly. <P2> Interoperability among heterogeneous collections is a central theme of the Core Integration. The potential collections have a wide variety of data types, metadata standards, protocols, authentication schemes, and business models. <P3> The goal of interoperability is to build coherent services for users, from components that are technically different and managed by different organizations. This requires agreements to cooperate at three levels: technical, content and organizational. <P4> Much of the research of the authors of this paper aims at . . . looking for approaches to interoperability that have low cost of adoption, yet provide substantial functionality. One of these approaches is the metadata harvesting protocol of the Open Archives Initiative (OAI) . . . <P5> For Site for Science, we identified three levels of digital library interoperability: Federation; Harvesting; Gathering. In this list, the top level provides the strongest form of interoperability, but places the greatest burden on participants. The bottom level requires essentially no effort by the participants, but provides a poorer level of interoperability. The Site for Science demonstration concentrated on the harvesting and gathering, because other projects were exploring federation. <P6> In an ideal world all the collections and services that the NSDL wishes to encompass would support an agreed set of standard metadata. The real world is less simple. . . . However, the NSDL does have influence. We can attempt to persuade collections to move along the interoperability curve. <warrant> <P7> The Site for Science metadata strategy is based on two principles. The first is that metadata is too expensive for the Core Integration team to create much of it. Hence, the NSDL has to rely on existing metadata or metadata that can be generated automatically. The second is to make use of as much of the metadata available from collections as possible, knowing that it varies greatly from none to extensive. Based on these principles, Site for Science, and subsequently the entire NSDL, developed the following metadata strategy: Support eight standard formats; Collect all existing metadata in these formats; Provide crosswalks to Dublin Core; Assemble all metadata in a central metadata repository; Expose all metadata records in the repository for service providers to harvest; Concentrate limited human effort on collection-level metadata; Use automatic generation to augment item-level metadata. <P8> The strategy developed by Site for Science and now adopted by the NSDL is to accumulate metadata in the native formats provided by the collections . . . If a collection supports the protocols of the Open Archives Initiative, it must be able to supply unqualified Dublin Core (which is required by the OAI) as well as the native metadata format. <P9> From a computing viewpoint, the metadata repository is the key component of the Site for Science system. The repository can be thought of as a modern variant of the traditional library union catalog, a catalog that holds comprehensive catalog records from a group of libraries. . . . Metadata from all the collections is stored in the repository and made available to providers of NSDL service.
Conclusions
RQ 1 "Can a small team of librarians manage the collection development and metadata strategies for a very large library?" RQ 2 "Can the NSDL actually build services that are significantly more useful than the general web search services?"
CA Through OAI, access to resources is effected in a low-cost, interoperable manner.
Phrases
<P1> The need for a metadata format that would support both metadata creation by authors and interoperability across heterogeneous repositories led to the choice of unqualified Dublin Core. (p.16) <P2> OAI develops and promotes a low-barrier interoperability framework and associated standards, originally to enhance access to e-print archives, but now taking into account access to other digital materials. (p.16)
Conclusions
RQ The many players involved in cultural heritage need to work together to define standards and best practices.
Type
Electronic Journal
Title
Metadata Corner: Working Meeting on Electronic Records Research
CA Just as the digital library has forced librarians to rethink their profession, so has the electronic record done the same for archivists and recordkeepers. Much of the debate centers around what e-records are and how to deal with them.
Phrases
<P1> Their presentations at the Working Meeting elaborated on the concept of "literary warrant," which can be defined as the mandate from outside the archives profession -- from law, professional best practices and other special sources -- which requires the creation and maintenance of records. (p.1) <P2> Systems that records professionals devise to maintain these concepts over time, may, therefore, be of interest to information professionals interested in digital preservation. Authenticity, indeed, is a key component of Peter Graham's description of "intellectual preservation." (p.3)
Conclusions
RQ How can metadata be linked to its record over time? How can we ensure the "least-loss" migration of metadata over time? Collaboration with warrant creators from other fields such as lawyers and auditors is desirable.
CA Metadata is a key part of the information infrastructure necessary to organize and classify the massive amount of information on the Web. Metadata, just like the resources they describe, will range in quality and be organized around different principles. Modularity is critical to allow metadata schema designers to base their new creations on established schemas, thereby benefiting from best practices rather than reinventing elements each time. Extensibility and cost-effectiveness are also important factors. Controlled vocabularies provide greater precision and access. Multilingualism (translating specification documents into many languages) is an important step in fostering global metadata architecture(s).
Phrases
<P1> The use of controlled vocabularies is another important approach to refinement that improves the precision for descriptions and leverages the substantial intellectual investment made by many domains to improve subject access. (p.4) <P2> Standards typically deal with these issues through the complementary processes of internalization and localization: the former process relates to the creation of "neutral" standards, whereas the latter refers to the adaptation of such a neutral standard to a local context. (p.4)
Conclusions
RQ In order for the full potential of resource discovery that the Web could offer to be realized, a"convergence" of standards and semantics must occur.
Type
Electronic Journal
Title
The National Digital Information Infrastructure Preservation Program: Expectations, Realities, Choices and Progress to Date
CA The goals of this plan include the continued collecting of materials regardless of evolving digital formats, the long-term preservation of said materials and ensuring access to them for the American people.
Phrases
<P1> There is widespread support for a national initiative in long term preservation of digital content across a very broad range of stakeholder groups outside the traditional scholarly community (p.4) <warrant> <P2> Approaching persistent archiving from the perspective of infrastructure allows system designers to decouple the data storage from the various components that allow users to manage the data. Potentially, any component can be "swapped out" without affecting the rest of the system. Theoretically, many of the technical problems in archiving can be separated into their components, and as innovation occurs, those components can be updated so that the archive remains persistent in the context of rapid change. Similarly, as the storage media obsolesce, the data can be migrated without affecting the overall integrity of the system. (p.7)
Conclusions
RQ Scenario-planning exercises may help expose assumptions that could ultimately be limiting in the future.
Type
Electronic Journal
Title
Coming to TERM: Designing the Texas Repository E-mail
CA The sheer volume of e-mail makes it difficult to preserve. Because of recent problems in the management of government e-mail, the Texas Department of Information Resources has been exploring ways to centralize and store all Texas state government e-mail.
Phrases
<P1> The Open Archival Information System (OAIS) Reference Model has attracted wide attention as a workable model because it provides the elements that research indicates are necessary: a closely audited, well documented, and constantly maintained and updated system. These elements are especially attractive to government. This model has the advantage of being an ISO international standard. <P2> This preservation strategy respects the archival bond by providing evidence of the official business transactions of the state in an architecture that respects and maintains links and ties between the compound parts of the record as well as larger logical groupings of records. <P3> Hundreds of archival institutions exist in North America, and yet very few, public or private, are currently preserving electronic records for the long-term.
Conclusions
RQ How does one construct and manage a large repository of electronic data?
Type
Electronic Journal
Title
Primary Sources, Research, and the Internet: The Digital Scriptorium at Duke
First Monday, Peer Reviewed Journal on the Internet
Publication Year
1997
Volume
2
Issue
9
Critical Arguements
CA "As the digital revolution moves us ever closer to the idea of the 'virtual library,' repositories of primary sources and other archival materials have both a special opportunity and responsibility. Since the materials in their custody are, by definition, often unique, these institutions will need to work very carefully with scholars and other researchers to determine what is the most effective way of making this material accessible in a digital environment."
Phrases
<P1> The matter of Internet access to research materials and collections is not one of simply doing what we have always done -- except digitally. It represents instead an opportunity to rethink the fundamental triangular relationship between libraries and archives, their collections, and their users. <P2> Digital information as it exists on the Internet today requires more navigational, contextual, and descriptive data than is currently provided in traditional card catalogs or their more modern electronic equivalent. One simply cannot throw up vast amounts of textual or image-based data onto the World Wide Web and expect existing search engines to make much sense of it or users to be able to digest the results. ... Archivists and manuscript curators have for many years now been providing just that sort of contextual detail in the guides, finding aids, and indexes that they have traditionally prepared for their holdings. <P3> Those involved in the Berkeley project understood that HTML was essentially a presentational encoding scheme and lacked the formal structural and content-based encoding that SGML would offer. <P4> Encoded Archival Description is quickly moving towards become an internationally embraced standard for the encoding of archival metadata in a wide variety of archival repositories and special collections libraries. And the Digital Scriptorium at Duke has become one of the early implementors of this standard. <warrant>
Conclusions
RQ "Duke is currently involved in a project that is funded through NEH and also involves the libraries of Stanford, the University of Virginia, and the University of California-Berkeley. This project (dubbed the "American Heritage Virtual Digital Archives Project") will create a virtual archive of encoded finding aids from all four institutions. This archive will permit seamless searching of these finding aids -- at a highly granular level of detail -- through a single search engine on one site and will, it is hoped, provide a model for a more comprehensive national system in the near future."
Type
Electronic Journal
Title
The Warwick Framework: A container architecture for diverse sets of metadata
This paper is a abbreviated version of The Warwick Framework: A Container Architecture for Aggregating Sets of Metadata. It describes a container architecture for aggregating logically, and perhaps physically, distinct packages of metadata. This "Warwick Framework" is the result of the April 1996 Metadata II Workshop in Warwick U.K.
ISBN
1082-9873
Critical Arguements
CA Describes the Warwick Framework, a proposal for linking together the various metadata schemes that may be attached to a given information object by using a system of "packages" and "containers." "[Warwick Workshop] attendees concluded that ... the route to progress on the metadata issue lay in the formulation a higher-level context for the Dublin Core. This context should define how the Core can be combined with other sets of metadata in a manner that addresses the individual integrity, distinct audiences, and separate realms of responsibility of these distinct metadata sets. The result of the Warwick Workshop is a container architecture, known as the Warwick Framework. The framework is a mechanism for aggregating logically, and perhaps physically, distinct packages of metadata. This is a modularization of the metadata issue with a number of notable characteristics. It allows the designers of individual metadata sets to focus on their specific requirements, without concerns for generalization to ultimately unbounded scope. It allows the syntax of metadata sets to vary in conformance with semantic requirements, community practices, and functional (processing) requirements for the kind of metadata in question. It separates management of and responsibility for specific metadata sets among their respective "communities of expertise." It promotes interoperability by allowing tools and agents to selectively access and manipulate individual packages and ignore others. It permits access to the different metadata sets that are related to the same object to be separately controlled. It flexibly accommodates future metadata sets by not requiring changes to existing sets or the programs that make use of them."
Phrases
<P1> The range of metadata needed to describe and manage objects is likely to continue to expand as we become more sophisticated in the ways in which we characterize and retrieve objects and also more demanding in our requirements to control the use of networked information objects. The architecture must be sufficiently flexible to incorporate new semantics without requiring a rewrite of existing metadata sets. <warrant> <P2> Each logically distinct metadata set may represent the interests of and domain of expertise of a specific community. <P3> Just as there are disparate sources of metadata, different metadata sets are used by and may be restricted to distinct communities of users and agents. <P4> Strictly partitioning the information universe into data and metadata is misleading. <P5> If we allow for the fact that metadata for an object consists of logically distinct and separately administered components, then we should also provide for the distribution of these components among several servers or repositories. The references to distributed components should be via a reliable persistent name scheme, such as that proposed for Universal Resources Names (URNs) and Handles. <P6> [W]e emphasize that the existence of a reliable URN implementation is a necessary to avoid the problems of dangling references that plague the Web. <warrant> <P7> Anyone can, in fact, create descriptive data for a networked resource, without permission or knowledge of the owner or manager of that resource. This metadata is fundamentally different from that metadata that the owner of a resource chooses to link or embed with the resource. We, therefore, informally distinguish between two categories of metadata containers, which both have the same implementation [internally referenced and externally referenced metadata containers].
Conclusions
RQ "We run the danger, with the full expressiveness of the Warwick Framework, of creating such complexity that the metadata is effectively useless. Finding the appropriate balance is a central design problem. ... Definers of specific metadata sets should ensure that the set of operations and semantics of those operations will be strictly defined for a package of a given type. We expect that a limited set of metadata types will be widely used and 'understood' by browsers and agents. However, the type system must be extensible, and some method that allows existing clients and agents to process new types must be a part of a full implementation of the Framework. ... There is a need to agree on one or more syntaxes for the various metadata sets. Even in the context of the relatively simple World Wide Web, the Internet is often unbearably slow and unreliable. Connections often fail or time out due to high load, server failure, and the like. In a full implementation of the Warwick Framework, access to a "document" might require negotiation across distributed repositories. The performance of this distributed architecture is difficult to predict and is prone to multiple points of failure. ... It is clear that some protocol work will need to be done to support container and package interchange and retrieval. ... Some examination of the relationship between the Warwick Framework and ongoing work in repository architectures would likely be fruitful.
Type
Electronic Journal
Title
Collection-Based Persistent Digital Archives - Part 1
The preservation of digital information for long periods of time is becoming feasible through the integration of archival storage technology from supercomputer centers, data grid technology from the computer science community, information models from the digital library community, and preservation models from the archivistÔÇÖs community. The supercomputer centers provide the technology needed to store the immense amounts of digital data that are being created, while the digital library community provides the mechanisms to define the context needed to interpret the data. The coordination of these technologies with preservation and management policies defines the infrastructure for a collection-based persistent archive. This paper defines an approach for maintaining digital data for hundreds of years through development of an environment that supports migration of collections onto new software systems.
ISBN
1082-9873
Critical Arguements
CA "Supercomputer centers, digital libraries, and archival storage communities have common persistent archival storage requirements. Each of these communities is building software infrastructure to organize and store large collections of data. An emerging common requirement is the ability to maintain data collections for long periods of time. The challenge is to maintain the ability to discover, access, and display digital objects that are stored within an archive, while the technology used to manage the archive evolves. We have implemented an approach based upon the storage of the digital objects that comprise the collection, augmented with the meta-data attributes needed to dynamically recreate the data collection. This approach builds upon the technology needed to support extensible database schema, which in turn enables the creation of data handling systems that interconnect legacy storage systems."
Phrases
<P1> The ultimate goal is to preserve not only the bits associated with the original data, but also the context that permits the data to be interpreted. <warrant> <P2> We rely on the use of collections to define the context to associate with digital data. The context is defined through the creation of semi-structured representations for both the digital objects and the associated data collection. <P3>A collection-based persistent archive is therefore one in which the organization of the collection is archived simultaneously with the digital objects that comprise the collection. <P4> The goal is to preserve digital information for at least 400 years. This paper examines the technical issues that must be addressed and presents a prototype implementation. <P5>Digital object representation. Every digital object has attributes that define its structure, physical context, and provenance, and annotations that describe features of interest within the object. Since the set of attributes (such as annotations) will vary across all objects within a collection, a semi-structured representation is needed. Not all digital objects will have the same set of associated attributes. <P6> If possible, a common information model should be used to reference the attributes associated with the digital objects, the collection organization, and the presentation interface. An emerging standard for a uniform data exchange model is the eXtended Markup Language (XML). <P7> A particular example of an information model is the XML Document Type Definition (DTD) which provides a description for the allowed nesting structure of XML elements. Richer information models are emerging such as XSchema (which provides data types, inheritance, and more powerful linking mechanisms) and XMI (which provides models for multiple levels of data abstraction). <P8> Although XML DTDs were originally applied to documents only, they are now being applied to arbitrary digital objects, including the collections themselves. More generally, OSDs can be used to define the structure of digital objects, specify inheritance properties of digital objects, and define the collection organization and user interface structure. <P9> A persistent collection therefore needs the following components of an OSD to completely define the collection context: Data dictionary for collection semantics; Digital object structure; Collection structure; and User interface structure. <P10> The re-creation or instantiation of the data collection is done with a software program that uses the schema descriptions that define the digital object and collection structure to generate the collection. The goal is to build a generic program that works with any schema description. <P11> The information for which driver to use for access to a particular data set is maintained in the associated Meta-data Catalog (MCAT). The MCAT system is a database containing information about each data set that is stored in the data storage systems. <P12> The data handling infrastructure developed at SDSC has two components: the SDSC Storage Resource Broker (SRB) that provides federation and access to distributed and diverse storage resources in a heterogeneous computing environment, and the Meta-data Catalog (MCAT) that holds systemic and application or domain-dependent meta-data about the resources and data sets (and users) that are being brokered by the SRB. <P13> A client does not need to remember the physical mapping of a data set. It is stored as meta-data associated with the data set in the MCAT catalog. <P14> A characterization of a relational database requires a description of both the logical organization of attributes (the schema), and a description of the physical organization of attributes into tables. For the persistent archive prototype we used XML DTDs to describe the logical organization. <P15> A combination of the schema and physical organization can be used to define how queries can be decomposed across the multiple tables that are used to hold the meta-data attributes. <P16> By using an XML-based database, it is possible to avoid the need to map between semi-structured and relational organizations of the database attributes. This minimizes the amount of information needed to characterize a collection, and makes the re-creation of the database easier. <warrant> <P17> Digital object attributes are separated into two classes of information within the MCAT: System-level meta-data that provides operational information. These include information about resources (e.g., archival systems, database systems, etc., and their capabilities, protocols, etc.) and data objects (e.g., their formats or types, replication information, location, collection information, etc.); Application-dependent meta-data that provides information specific to particular data sets and their collections (e.g., Dublin Core values for text objects). <P18> Internally, MCAT keeps schema-level meta-data about all of the attributes that are defined. The schema-level attributes are used to define the context for a collection and enable the instantiation of the collection on new technology. <P19> The logical structure should not be confused with database schema and are more general than that. For example, we have implemented the Dublin Core database schema to organize attributes about digitized text. The attributes defined in the logical structure that is associated with the Dublin Core schema contains information about the subject, constraints, and presentation formats that are needed to display the schema along with information about its use and ownership. <P20> The MCAT system supports the publication of schemata associated with data collections, schema extension through the addition or deletion of new attributes, and the dynamic generation of the SQL that corresponds to joins across combinations of attributes. <P21> By adding routines to access the schema-level meta-data from an archive, it is possible to build a collection-based persistent archive. As technology evolves and the software infrastructure is replaced, the MCAT system can support the migration of the collection to the new technology.
Conclusions
RQ Collection-Based Persistent Digital Archives - Part 2
SOW
DC "The technology proposed by SDSC for implementing persistent archives builds upon interactions with many of these groups. Explicit interactions include collaborations with Federal planning groups, the Computational Grid, the digital library community, and individual federal agencies." ... "The data management technology has been developed through multiple federally sponsored projects, including the DARPA project F19628-95-C-0194 "Massive Data Analysis Systems," the DARPA/USPTO project F19628-96-C-0020 "Distributed Object Computation Testbed," the Data Intensive Computing thrust area of the NSF project ASC 96-19020 "National Partnership for Advanced Computational Infrastructure," the NASA Information Power Grid project, and the DOE ASCI/ASAP project "Data Visualization Corridor." Additional projects related to the NSF Digital Library Initiative Phase II and the California Digital Library at the University of California will also support the development of information management technology. This work was supported by a NARA extension to the DARPA/USPTO Distributed Object Computation Testbed, project F19628-96-C-0020."
Type
Electronic Journal
Title
Collection-Based Persistent Digital Archives - Part 2
"Collection-Based Persistent Digital Archives: Part 2" describes the creation of a one million message persistent E-mail collection. It discusses the four major components of a persistent archive system: support for ingestion, archival storage, information discovery, and presentation of the collection. The technology to support each of these processes is still rapidly evolving, and opportunities for further research are identified.
ISBN
1082-9873
Critical Arguements
CA "The multiple migration steps can be broadly classified into a definition phase and a loading phase. The definition phase is infrastructure independent, whereas the loading phase is geared towards materializing the processes needed for migrating the objects onto new technology. We illustrate these steps by providing a detailed description of the actual process used to ingest and load a million-record E-mail collection at the San Diego Supercomputer Center (SDSC). Note that the SDSC processes were written to use the available object-relational databases for organizing the meta-data. In the future, it may be possible to go directly to XML-based databases."
Phrases
<P1> The processes used to ingest a collection, transform it into an infrastructure independent form, and store the collection in an archive comprise the persistent storage steps of a persistent archive. The processes used to recreate the collection on new technology, optimize the database, and recreate the user interface comprise the retrieval steps of a persistent archive. <P2> In order to build a persistent collection, we consider a solution that "abstracts" all aspects of the data and its preservation. In this approach, data object and processes are codified by raising them above the machine/software dependent forms to an abstract format that can be used to recreate the object and the processes in any new desirable forms. <P3> The SDSC infrastructure uses object-relational databases to organize information. This makes data ingestion more complex by requiring the mapping of the XML DTD semi-structured representation onto a relational schema. <P4> The SDSC infrastructure uses object-relational databases to organize information. This makes data ingestion more complex by requiring the mapping of the XML DTD semi-structured representation onto a relational schema. <P5> The steps used to store the persistent archive were: (1) Define Digital Object: define meta-data, define object structure (OBJ-DTD) --- (A), define object DTD to object DDL mapping --- (B) (2) Define Collection: define meta-data, define collection structure (COLL-DTD) --- (C), define collection DTD structure to collection DDL mapping --- (D) (3) Define Containers: define packing format for encapsulating data and meta-data (examples are the AIP standard, Hierarchical Data Format, Document Type Definition) <P5> In the ingestion phase, the relational and semi-structured organization of the meta-data is defined. No database is actually created, only the mapping between the relational organization and the object DTD. <P6> Note that the collection relational organization does not have to encompass all of the attributes that are associated with a digital object. Separate information models are used to describe the objects and the collections. It is possible to take the same set of digital objects and form a new collection with a new relational organization. <P7> Multiple communities across academia, the federal government, and standards groups are exploring strategies for managing very large archives. The persistent archive community needs to maintain interactions with these communities to track development of new strategies for data management and storage. <warrant> <P8>
Conclusions
RQ "The four major components of the persistent archive system are support for ingestion, archival storage, information discovery, and presentation of the collection. The first two components focus on the ingestion of data into collections. The last two focus on access to the resulting collections. The technology to support each of these processes is still rapidly evolving. Hence consensus on standards has not been reached for many of the infrastructure components. At the same time, many of the components are active areas of research. To reach consensus on a feasible collection-based persistent archive, continued research and development is needed. Examples of the many related issues are listed below:
Type
Electronic Journal
Title
Buckets: A new digital technology for preserving NASA research
CA Buckets are information objects designed to reduce dependency on traditional archives and database systems thereby making them more resilent to the transient nature of information systems.
Phrases
Another focus of aggregation was including the metadata with data. Through experiences NASA researchers found that metadata tended to "drift" over time, becoming decoupled from the data it described or locked in specific DL systems and hard to extract or share with other systems. (p. 377) Buckets are designed to imbue the information objects with certain responsibilities, such as display, dissemination, protection, and maintenance of its contents. As such, buckets should be able to work with many DL systems simultaneously, and minimize or eliminate the necessary modification of DL systems to work with buckets. Ideally, buckets should work with everything and break nothing. This philosophy is formalized in the SODA DL model. the objects become "smarter" at the expense of the archives (that become "dumber"), as functionalities generally associated with archives are moved into the data objects themselves. (p. 390)
Conclusions
RQ The creation of high quality tools for bucket creation and administration is absolutely necessary. The extension of authentication and security measures is key to supporting more technologies. Many applications of this sort of information object independence remains to be explored.
Type
Electronic Journal
Title
Search for Tomorrow: The Electronic Records Research Program of the U.S. National Historical Publications and Records Commission
The National Historical Publications and Records Commission (NHPRC) is a small grant-making agency affiliated with the U.S. National Archives and Records Administration. The Commission is charged with promoting the preservation and dissemination of documentary source materials to ensure an understanding of U.S. history. Recognizing that the increasing use of computers created challenges for preserving the documentary record, the Commission adopted a research agenda in 1991 to promote research and development on the preservation and continued accessibility of documentary materials in electronic form. From 1991 to the present the Commission awarded 31 grants totaling $2,276,665 for electronic records research. Most of this research has focused on two issues of central concern to archivists: (1) electronic record keeping (tools and techniques to manage electronic records produced in an office environment, such as word processing documents and electronic mail), and (2) best practices for storing, describing, and providing access to all electronic records of long-term value. NHPRC grants have raised the visibility of electronic records issues among archivists. The grants have enabled numerous archives to begin to address electronic records problems, and, perhaps most importantly, they have stimulated discussion about electronic records among archivists and records managers.
Publisher
Elsevier Science Ltd
Critical Arguements
CA "The problem of maintaining electronic records over time is big, expensive, and growing. A task force on digital archives established by the Commission on Preservation and Access in 1994 commented that the life of electronic records could be characterized in the same words Thomas Hobbes once used to describe life: ÔÇ£nasty, brutish, and shortÔÇØ [1]. Every day, thousands of new electronic files are created on federal, state, and local government computers across the nation. A small but important portion of these records will be designated for permanent retention. Government agencies are increasingly relying on computers to maintain information such as census files, land titles, statistical data, and vital records. But how should electronic records with long-term value be maintained? Few government agencies have developed comprehensive policies for managing current electronic records, much less preserving those with continuing value for historians and other researchers. Because of this serious and growing problem, the National Historical Publications and Records Commission (NHPRC), a small grantmaking agency affiliated with the U.S. National Archives and Records Administration (NARA), has been making grants for research and development on the preservation and use of electronic documentary sources. The program is conducted in concert with NARA, which in 1996 issued a strategic plan that gives high priority to mastering electronic records problems in partnership with federal government agencies and the NHPRC.
Phrases
<P1> How can data dictionaries, information resource directory systems, and other metadata systems be used to support electronic records management and archival requirements? <P2> In spite of the number of projects the Commission has supported, only four questions from the research agenda have been addressed to date. Of these, the question relating to requirements for the development of data dictionaries and other metadata systems (question number four) has produced a single grant for a state information locator system in South Carolina, and the question relating to needs for archival education (question 10) has led to two grants to the Society of American Archivists for curricular materials. <P3> Information systems created without regard for these considerations may have deficiencies that limit the usefulness of the records contained on them. <warrant> <P4> The NHPRC has awarded major grants to four institutions over the past five years for projects to develop and test requirements for electronic record keeping: University of Pittsburgh (1993): A working set of functional requirements and metadata specifications for electronic record keeping systems; City of Philadelphia (1995, 1996, and 1997): A project to incorporate a subset of the Pittsburgh metadata specifications into a new human resources information system and other city systems as test cases and to develop comprehensive record keeping policies and standards for the cityÔÇÖs information technology systems; Indiana University (1995): A project to develop an assessment tool and methodology for analyzing existing electronic records systems, using the Pittsburgh functional requirements as a model and the student academic record system and a financial system as test cases; Research Foundation of the State University of New York-Albany, Center for Technology in Government (1996): A project to identify best practices for electronic record keeping, including work by the U.S. Department of Defense and the University of British Columbia in addition to the University of Pittsburgh. The Center is working with the stateÔÇÖs Adirondack Parks Agency in a pilot project to develop a system model for incorporating record keeping and archival considerations into the creation of networked computing and communications applications. <P5> No definitive solution has yet been identified for the problems posed by electronic records, although progress has been made in learning what will be needed to design functional electronic record keeping systems. <P6> With the proliferation of digital libraries, the need for long-term storage, migration and retrieval strategies for electronic information has become a priority for a wide variety of information providers. <warrant>
Conclusions
RQ "How best to preserve existing and future electronic formats and provide access to them over time has remained elusive. The answers cannot be found through theoretical research alone, or even through applied research, although both are needed. Answers can only emerge over time as some approaches prove able to stand the test of time and others do not. The problems are large because the costs of maintaining, migrating, and retrieving electronic information continue to be high." ... "Perhaps most importantly, these grants have stimulated widespread discussion of electronic records issues among archivists and record managers, and thus they have had an impact on the preservation of the electronic documentary record that goes far beyond the CommissionÔÇÖs investment."
SOW
DC The National Historic Publications and Records Commission (NHPRC) is the outreach arm of the National Archives and makes plans for and studies issues related to the preservation, use and publication of historical documents. The Commission also makes grants to non-Federal archives and other organizations to promote the preservation use of America's documentary heritage.
Type
Report
Title
Mapping of the Encoded Archival Description DTD Element Set to the CIDOC CRM
The CIDOC CRM is the first ontology designed to mediate contents in the area of material cultural heritage and beyond, and has been accepted by ISO TC46 as work item for an international standard. The EAD Document Type Definition (DTD) is a standard for encoding archival finding aids using the Standard Generalized Markup Language (SGML). Archival finding aids are detailed guides to primary source material which provide fuller information than that normally contained within cataloging records. 
Publisher
Institute of Computer Science, Foundation for Research and Technology - Hellas
Publication Location
Heraklion, Crete, Greece
Language
English
Critical Arguements
CA "This report describes the semantic mapping of the current EAD DTD Version 1.0 Element Set to the CIDOC CRM and its latest extension. This work represents a proof of concept for the functionality the CIDOC CRM is designed for." 
Conclusions
RQ "Actually, the CRM seems to do the job quite well ÔÇô problems in the mapping arise more from underspecification in the EAD rather than from too domain-specific notions. "┬á... "To our opinion, the archival community could benefit from the conceptualizations of the CRM to motivate more powerful metadata standards with wide interoperability in the future, to the benefit of museums and other disciplines as well."
SOW
DC "As a potential international standard, the EAD DTD is maintained in the Network Development and MARC Standards Office of the Library of Congress in partnership with the Society of American Archivists." ... "The CIDOC Conceptual Reference Model (see [CRM1999], [Doerr99]), in the following only referred to as ┬½CRM┬╗, is outcome of an effort of the Documentation Standards Group of the CIDOC Committee (see ┬½http:/www.cidoc.icom.org┬╗, ÔÇ£http://cidoc.ics.forth.grÔÇØ) of ICOM, the International Council of Museums beginning in 1996."
CA The archival profession has a brief window of opportunity to become stakeholders in the realm of electronic records. In order to accomplish that, they must answer not only the "what" but the "why" of recordkeeping in all of its implications.
Conclusions
RQ How will American archivists deal with the re-invention of professional roles that have traditionally been bifurcated by records on one side and archives on the other? Where does continuum thinking leave SAA and its primary constituency of historical archivists?
Type
Web Page
Title
Documenting Business: The Australian Recordkeeping Metadata Schema
In July 1999, the Australian Recordkeeping Metadata Schema (RKMS) was approved by its academic and industry steering group. This metadata set now joins other community specific sets in being available for use and implementation into workplace applications. The RKMS has inherited elements from and built on many other metadata standards associated with information management. It has also contributed to the development of subsequent sector specific recordkeeping metadata sets. The importance of the RKMS as a framework for 'mapping' or reading other sets and also as a standardised set of metadata available for adoption in diverse implementation environments is now emerging. This paper explores the context of the SPIRT Recordkeeping Metadata Project, and the conceptual models developed by the SPIRT Research Team as a framework for standardising and defining Recordkeeping Metadata. It then introduces the elements of the SPIRT Recordkeeping Metadata Schema and explores its functionality before discussing implementation issues with reference to document management and workflow technologies.
Critical Arguements
CA Much of the metadata work done so far has worked off the passive assumption of records as document-like objects. Instead, they need to be seen as active entities in business transactions.
Conclusions
RQ In order to decide which elements are to be used from the RKMS, organizations need to delineate the reach of specific implementations as far as how and when records need to be bound with metadata.
This portal page provides links to all EAD-related information as applicable to those institutional members of the U.K. Archives Hub. It provides links to Creating EAD records, More about EAD, Reference, and More resources.
Publisher
The Archives Hub
Publication Location
Manchester, England, U.K.
Language
English
Critical Arguements
CA "These pages have been designed to hold links and information which we hope will be useful to archivists and librarians working in the UK Higher and Further Education sectors."
SOW
DC The Archives Hub provides a single point of access to 17,598 descriptions of archives held in UK universities and colleges. At present these are primarily at collection-level, although complete catalogue descriptions are provided where they are available. The Archives Hub forms one part of the UK's National Archives Network, alongside related networking projects. A Steering Committee which includes representatives of contributing institutions, the National Archives and the other archive networks guides the progress of the project. There is also a Contributors' and Users' Forum which provides feedback to aid the development of the service. The service is hosted at MIMAS on behalf of the Consortium of University Research Libraries (CURL) and is funded by the Joint Information Systems Committee (JISC). Systems development work is undertaken at the University of Liverpool.
CA Databases have structure, and annotating them (with metadata for example) can be difficult. Work with semistructured data models in conjunction with the use of XML may solve this problem of accommodating unanticipated structures in databases.
Conclusions
RQ It's critical to develop tools to help data curators record and repeat the corrections they make.
Type
Web Page
Title
CDL Digital Object Standard: Metadata, Content and Encoding
This document addresses the standards for digital object collections for the California Digital Library 1. Adherence to these standards is required for all CDL contributors and may also serve University of California staff as guidelines for digital object creation and presentation. These standards are not intended to address all of the administrative, operational, and technical issues surrounding the creation of digital object collections.
Critical Arguements
CA These standards describe the file formats, storage and access standards for digital objects created by or incorporated into the CDL as part of the permanent collections. They attempt to balance adherence to industry standards, reproduction quality, access, potential longevity and cost.
Conclusions
RQ not applicable
SOW
DC "This is the first version of the CDL Digital Object Standard. This version is based upon the September 1, 1999 version of the CDL's Digital Image Standard, which included recommendations of the Museum Educational Site Licensing Project (MESL), the Library of Congress and the MOA II participants." ... "The Museum Educational Site Licensing Project (MESL) offered a framework for seven collecting institutions, primarily museums, and seven universities to experiment with new ways to distribute visual information--both images and related textual materials. " ... "The Making of America (MoA II) Testbed Project is a Digital Library Federation (DLF) coordinated, multi-phase endeavor to investigate important issues in the creation of an integrated, but distributed, digital library of archival materials (i.e., digitized surrogates of primary source materials found in archives and special collections). The participants include Cornell University, New York Public Library, Pennsylvania State University, Stanford University and UC Berkeley. The Library of Congress white papers and standards are based on the experience gained during the American Memory Pilot Project. The concepts discussed and the principles developed still guide the Library's digital conversion efforts, although they are under revision to accomodate the capabilities of new technologies and new digital formats." ... "The CDL Technical Architecture and Standards Workgroup includes the following members with extensive experience with digital object collection and management: Howard Besser, MESL and MOA II digital imaging testbed projects; Diane Bisom, University of California, Irvine; Bernie Hurley, MOA II, University of California, Berkeley; Greg Janee, Alexandria Digital Library; John Kunze, University of California, San Francisco; Reagan Moore and Chaitanya Baru, San Diego Supercomputer Center, ongoing research with the National Archives and Records Administration on the long term storage and retrieval of digital content; Terry Ryan, University of California, Los Angeles; David Walker, California Digital Library"
This document outlines the best practices guidelines for creation of EAD-encoded finding aids for submission to the Archives Hub in the U.K. It includes sections on Mandatory Fields, Access Points, Manual Encoding, Multilevel Descriptions, Saving and Submitting Files, and Links.
Notes
This is a downloadable .pdf file. Also available in Rich Text Format (.rtf).
Publisher
Archives Hub, U.K.
Publication Location
Manchester, England, U.K.
Language
English
Critical Arguements
CA "These pages have been designed to hold links and information which we hope will be useful to archivists and librarians working in the UK Higher and Further Education sectors."
Conclusions
RQ
SOW
DC The Archives Hub provides a single point of access to 17,598 descriptions of archives held in UK universities and colleges. At present these are primarily at collection-level, although complete catalogue descriptions are provided where they are available. The Archives Hub forms one part of the UK's National Archives Network, alongside related networking projects. A Steering Committee which includes representatives of contributing institutions, the National Archives and the other archive networks guides the progress of the project. There is also a Contributors' and Users' Forum which provides feedback to aid the development of the service. The service is hosted at MIMAS on behalf of the Consortium of University Research Libraries (CURL) and is funded by the Joint Information Systems Committee (JISC). Systems development work is undertaken at the University of Liverpool.
Type
Web Page
Title
Schema Registry: activityreports: Recordkeeping Metadata Standard for Commonwealth Agencies
CA "The Australian SPIRT Recordkeeping Metadata Project was initially a project funded under a programme known as the Strategic Partnership with Industry -- Research and Training (SPIRT) Support Grant -- partly funded by the Australian Research Council. The project was concerned with developing a framework for standardising and defining recordkeeping metadata and produced a metadata element set eventually known as the Australian Recordkeeping Metadata Schema (RKMS). The conceptual frame of reference in the project was based in Australian archival practice, including the Records Continuum Model and the Australian Series System. The RKMS also inherits part of the Australian Government Locator Service (AGLS) metadata set."
The creation and use of metadata is likely to become an important part of all digital preservation strategies whether they are based on hardware and software conservation, emulation or migration. The UK Cedars project aims to promote awareness of the importance of digital preservation, to produce strategic frameworks for digital collection management policies and to promote methods appropriate for long-term preservation - including the creation of appropriate metadata. Preservation metadata is a specialised form of administrative metadata that can be used as a means of storing the technical information that supports the preservation of digital objects. In addition, it can be used to record migration and emulation strategies, to help ensure authenticity, to note rights management and collection management data and also will need to interact with resource discovery metadata. The Cedars project is attempting to investigate some of these issues and will provide some demonstrator systems to test them.
Notes
This article was presented at the Joint RLG and NPO Preservation Conference: Guidelines for Digital Imaging, held September 28-30, 1998.
Critical Arguements
CA "Cedars is a project that aims to address strategic, methodological and practical issues relating to digital preservation (Day 1998a). A key outcome of the project will be to improve awareness of digital preservation issues, especially within the UK higher education sector. Attempts will be made to identify and disseminate: Strategies for collection management ; Strategies for long-term preservation. These strategies will need to be appropriate to a variety of resources in library collections. The project will also include the development of demonstrators to test the technical and organisational feasibility of the chosen preservation strategies. One strand of this work relates to the identification of preservation metadata and a metadata implementation that can be tested in the demonstrators." ... "The Cedars Access Issues Working Group has produced a preliminary study of preservation metadata and the issues that surround it (Day 1998b). This study describes some digital preservation initiatives and models with relation to the Cedars project and will be used as a basis for the development of a preservation metadata implementation in the project. The remainder of this paper will describe some of the metadata approaches found in these initiatives."
Conclusions
RQ "The Cedars project is interested in helping to develop suitable collection management policies for research libraries." ... "The definition and implementation of preservation metadata systems is going to be an important part of the work of custodial organisations in the digital environment."
SOW
DC "The Cedars (CURL exemplars in digital archives) project is funded by the Joint Information Systems Committee (JISC) of the UK higher education funding councils under Phase III of its Electronic Libraries (eLib) Programme. The project is administered through the Consortium of University Research Libraries (CURL) with lead sites based at the Universities of Cambridge, Leeds and Oxford."
Type
Web Page
Title
Metadata for preservation : CEDARS project document AIW01
This report is a review of metadata formats and initiatives in the specific area of digital preservation. It supplements the DESIRE Review of metadata (Dempsey et al. 1997). It is based on a literature review and information picked-up at a number of workshops and meetings and is an attempt to briefly describe the state of the art in the area of metadata for digital preservation.
Critical Arguements
CA "The projects, initiatives and formats reviewed in this report show that much work remains to be done. . . . The adoption of persistent and unique identifiers is vital, both in the CEDARS project and outside. Many of these initiatives mention "wrappers", "containers" and "frameworks". Some thought should be given to how metadata should be integrated with data content in CEDARS. Authenticity (or intellectual preservation) is going to be important. It will be interesting to investigate whether some archivists' concerns with custody or "distributed custody" will have relevance to CEDARS."
Conclusions
RQ Which standards and initiatives described in this document have proved viable preservation metadata models?
SOW
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.
This paper discusses how metadata standards can help organizations comply with the ISO 9000 standards for quality systems. It provides a brief overview of metadata, ISO 9000 and related records management standards. It then analyses in some depth the ISO 9000 requirements for quality records, and outlines the problems that some organizations have in complying with them. It also describes the metadata specifications developed by the University of Pittsburgh Electronic Recordkeeping project and the SPIRT Recordkeeping Metadata project in Australia and discusses the role of metadata in meeting ISO 9000 requirements for the creation and preservation of reliable, authentic and accessible records.
Publisher
Records Continuum Research Group
Critical Arguements
CA "During the last few years a number of research projects have studied the types of metadata needed to create, manage and make accessible quality records, i.e. reliable, authentic and useable records. This paper will briefly discuss the purposes of recordkeeping metadata, with reference to emerging records management standards, and the models presented by two projects, one in the United States and one in Australia. It will also briefly review the ISO 9000 requirements for records and illustrate how metadata can help an organization meet these requirements."
Conclusions
RQ "Quality records provide many advantages for organizations and can help companies meet the ISO 9000 certification. However, systems must be designed to create the appropriate metadata to ensure they comply with recordkeeping requirements, particularly those identified by records management standards like AS 4390 and the proposed international standard, which provide benchmarks for recordkeeping best practice. The Pittsburgh metadata model and the SPIRT framework provide organizations with standardized sets of metadata that would ensure the creation, preservation and accessibility of reliable, authentic and meaningful records for as long as they are of use. In deciding what metadata to capture, organisations should consider the cost of meeting the requirements of the ISO 9000 guidelines and any related records management best practice standards, and the possible risk of not meeting these requirements."
Type
Web Page
Title
The Gateway to Educational Materials: An Evaluation Study, Year 4: A Technical Report submitted to the US Department of Education
CA The Gateway to Educational Materials (GEM) is a Web site created through the efforts of several groups, including the US Department of Education, The National Library of Education, and a team from Syracuse University. The goal of the project is to provide teachers with a broad range of educational materials on the World Wide Web. This study evaluates The Gateway as an online source of educational information. The purpose of this evaluation is to provide developers of The Gateway with information about aspects of the system that might need improvement, and to display lessons learned through this process to developers of similar systems. It is the fourth in a series of annual studies, and focuses on effectiveness of The Gateway from the perspectives of end users and collection holders.
CA In March 2003, the intention of undertaking an international survey of LOM implementations was announced at the plenary meeting of the "Information Technology for Learning, Education and Training", ISO/IEC JTC1/SC36 sub-committee. The ISO/IEC JTC1/SC36 committee is international in both membership and emphasis, and has a working group, Working Group (WG) 4, "Management and Delivery for Learning, Education, and Training," which has been explicitly charged with the task of contributing to future standardization work on the LOM. <warrant> The international LOM Survey focuses on two questions: 1) "Which elements were selected for use or population?"; and 2) "How were these elements used, or what where the types of values assigned to them?" This report also attempts to draw a number of tentative suggestions and conclusions for further standardization work
Conclusions
RQ Based on its findings, the preliminary survey report was able to suggest a number of conclusions: First, fewer and better-defined elements may be more effective than the range of choice and interpretive possibilities currently allowed by the LOM. This seems to be especially the case regarding educational elements, which are surprisingly underutilized for metadata that it ostensibly and primarily educational. Second, clear and easily-supported means of working with local, customized vocabularies would also be very valuable. Third, it also seems useful to ensure that structures are provided to accommodate complex but more conventional aspects of resource description. These would include multiple title versions, as well as multilingual descriptions and values.
SOW
DC On June 12, 2002, 1484.12.1 - 2002 Learning Object Metadata (LOM) was approved by the IEEE-Standards Association.
Type
Web Page
Title
Towards a Digital Rights Expression Language Standard for Learning Technology
CA The Learning Technology Standards Committee (LTSC) of the Institute for Electrical and Electronic Engineers (IEEE) concentrated on making recommendations for standardizing a digital rights expression language (DREL) with the specific charge to (1) Investigate existing standards development efforts for DREL and digital rights. (2) Gather DREL requirements germane to the learning, education, and training industries. (3) Make recommendations as to how to proceed. (4) Feed requirements into ongoing DREL and digital rights standardization efforts, regardless of whether the LTSC decides to work with these efforts or embark on its own. This report represents the achievement of these goals in the form a of a white paper that can be used as reference for the LTSC, that reports on the current state of existing and proposed standardization efforts targeting digital rights expression languages and makes recommendations concerning future work.
Conclusions
RQ The recommendations of this report are: 1. Maintain appropriate liaisons between learning technology standards development organizations and those standards development organizations standardizing rights expression languages. The purpose of these liaisons is to continue to feed requirements into broader standardization efforts and to ensure that the voice of the learning, education and training community is heard. 2. Support the creation of application profiles or extensions of XrML and ODRL that include categories and vocabularies for roles common in educational and training settings. In the case of XrML, a name space for local context may be needed. (A name space is required for both XrML and ODRL for the ÔÇ£application profileÔÇØ or specifically the application ÔÇôLT application- extension) 3. Advocate the creation of a standard for expressing local policies in ways that can be mapped to rights expressions. This could be either through a data model or through the definition of an API or service. 4. Launch an initiative to identify models of rights enforcement in learning technology and to possibly abstract a common model for use by architecture and framework definition projects. 5. Further study the implications of patent claims, especially for educational and research purposes.
Type
Web Page
Title
METS : Metadata Encoding and Transmission Standard
CA "METS, although in its early stages, is already sufficiently established amongst key digital library players that it can reasonably be considered the only viable standard for digital library objects in the foreseeable future. Although METS may be an excellent framework, it is just that and only that. It does not prescribe the content of the metadata itself, and this is a continuing problem for METS and all other schema to contend with if they are to realize their full functionality and usefulness."
Conclusions
RQ The standardization (via some sort of cataloging rules) of the content held by metadata "containers" urgently needs to be addressed. If not, the full value of any metadata scheme, no matter how extensible or robust, will not be realized.
Type
Web Page
Title
National States Geographic Information Council (NSGIC) Metadata Primer -- A "How To" Guide on Metadata Implementation
The primer begins with a discussion of what metadata is and why metadata is important. This is followed by an overview of the Content Standards for Digital Geospatial Metadata (CSDGM) adopted by the Federal Geographic Data Committee (FGDC). Next, the primer focuses on the steps required to begin collecting and using metadata. The fourth section deals with how to select the proper metadata creation tool from the growing number being developed. Section five discusses the mechanics of documenting a data set, including strategies on reviewing the output to make sure it is in a useable form. The primer concludes with a discussion of other assorted metadata issues.
Critical Arguements
CA The Metadata Primer is one phase of a larger metadata research and education project undertaken by the National States Geographic Information Council and funded by the Federal Geographic Data Committee's Competetive Cooperative Agreements Program (CCAP). The primer is designed to provide a practical overview of the issues associated with developing and maintaining metadata for digital spatial data. It is targeted toward an audience of state, local, and tribal government personnel. The document provides a "cook book" approach to the creation of metadata. Because much of the most current information on metadata resides on the Internet, the primer summarizes relevant material available from other World Wide Web (WWW) home pages.
Conclusions
RQ To what extent could the NSGIC recommendations be used for non-geographic applications?
SOW
DC FGDC approved the Content Standard for Digital Geospatial Metadata (FGDC-STD-001-1998) in June 1998. FGDC is a 19-member interagency committee composed of representatives from the Executive Office of the President, Cabinet-level and independent agencies. The FGDC is developing the National Spatial Data Infrastructure (NSDI) in cooperation with organizations from State, local and tribal governments, the academic community, and the private sector. The NSDI encompasses policies, standards, and procedures for organizations to cooperatively produce and share geographic data.
Abstract The ability of investigators to share data is essential to the progress of integrative scientific research both within and across disciplines. This paper describes the main issues in achieving effective data sharing based on previous efforts in building scientific data networks and, particularly, recent efforts within the Earth sciences. This is presented in the context of a range of information architectures for effecting differing levels of standardization and centralization both from a technology perspective as well as a publishing protocol perspective. We propose a new Metadata Interchange Format (.mif) that can be used for more effective sharing of data and metadata across digital libraries, data archives and research projects.
Critical Arguements
CA "In this paper, we discuss two important information technology aspects of the electronic publication of data in the Earth sciences, metadata, and a variety of different concepts of electronic data publication. Metadata are the foundation of electronic data publications and they are determined by needs of archiving, the scientific analysis and reproducibility of a data set, and the interoperability of diverse data publication methods. We use metadata examples drawn from the companion paper by Staudigel et al. (this issue) to illustrate the issues involved in scaling-up the publication of data and metadata by individual scientists, disciplinary groups, the Earth science community-at-large and to libraries in general. We begin by reviewing current practices and considering a generalized alternative." ... 'For this reason, we will we first discuss different methods of data publishing via a scientific data network followed by an inventory of desirable characteristics of such a network. Then, we will introduce a method for generating a highly portable metadata interchange format we call .mif (pronounced dot-mif) and conclude with a discussion of how this metadata format can be scaled to support the diversity of interests within the Earth science community and other scientific communities." ... "We can borrow from the library community the methods by which to search for the existence and location of data (e.g., Dublin Core http://www.dublincore.org) but we must invent new ways to document the metadata needed within the Earth sciences and to comply with other metadata standards such as the Federal Geographic Data Committee (FGDC). To accomplish this, we propose a metadata interchange format that we call .mif that enables interoperability and an open architecture that is maximally independent of computer systems, data management approaches, proprietary software and file formats, while encouraging local autonomy and community cooperation. "
Conclusions
RQ "These scalable techniques are being used in the development of a project we call SIOExplorer that can found at http://sioexplorer.ucsd.edu although we have not discussed that project in any detail. The most recent contributions to this discussion and .mif applications and examples may be found at http:\\Earthref.org\metadata\GERM\."
SOW
DC This article was written by representatives of the San Diego Supercomputer Center and the Insititute of Geophysics and Planetary Physics under the auspices of the University of California, San Diego.
CA The role of archives and archivists is being fundamentally redefined in consideration of postcustodial theories and practice.
Conclusions
RQ Who is accountable? How explicit should the "imprint" of the archivist be in the shaping of the record? Who decides (and how) what we remember and what we keep?
Type
Web Page
Title
Metadata Reference Guide: ONIX ONline Information eXchange
CA According to Editeur, the group responsible for the maintenance of the ONIX standard, ONIX is the international standard for representing book, serial, and video product information in electronic form.
Type
Web Page
Title
Creating and Documenting Text: A Guide to Good Practice
CA "The aim of this Guide is to take users through the basic steps involved in creating and documenting an electronic text or similar digital resource. ... This Guide assumes that the creators of electronic texts have a number of common concerns. For example, that they wish their efforts to remain viable and usable in the long-term, and not to be unduly constrained by the limitations of current hardware and software. Similarly, that they wish others to be able to reuse their work, for the purposes of secondary analysis, extension, or adaptation. They also want the tools, techniques, and standards that they adopt to enable them to capture those aspects of any non-electronic sources which they consider to be significant -- whilst at the same time being practical and cost-effective to implement."
Conclusions
RQ "While a single metadata scheme, adopted and implemented wholescale would be the ideal, it is probable that a proliferation of metadata schemes will emerge and be used by different communities. This makes the current work centred on integrated services and interoperability all the more important. ... The Warwick Framework (http://www.ukoln.ac.uk/metadata/resources/wf.html) for example suggests the concept of a container architecture, which can support the coexistence of several independently developed and maintained metadata packages which may serve other functions (rights management, administrative metadata, etc.). Rather than attempt to provide a metadata scheme for all web resources, the Warwick Framework uses the Dublin Core as a starting point, but allows individual communities to extend this to fit their own subject-specific requirements. This movement towards a more decentralised, modular and community-based solution, where the 'communities of expertise' themselves create the metadata they need has much to offer. In the UK, various funded organisations such as the AHDS (http://ahds.ac.uk/), and projects like ROADS (http://www.ilrt.bris.ac.uk/roads/) and DESIRE (http://www.desire.org/) are all involved in assisting the development of subject-based information gateways that provide metadata-based services tailored to the needs of particular user communities."
This guide is optimized for creation of EAD-encoded finding aids for the collections of New York University and New York Historical Society. The links on the page list tools and files that may be downloaded and referenced for production of NYU-conformant finding aids.
Publisher
New York University
Critical Arguements
CA "This guide is optimized for creation of EAD-encoded finding aids for the collections of New York University and New York Historical Society. Instructions assume the use of NoteTab as the XML editor, utilizing template files that serve as base files for the different collections." 
Conclusions
RQ
SOW
DC This guide serves both New York University and the New York Historical Society.
Type
Web Page
Title
Capturing Electronic Transactional Evidence: The Future
To ensure that the digital collections submitted to RLG Cultural Materials can be discovered and understood, RLG has compiled these Descriptive Metadata Guidelines for contributors. While these guidelines reflect the needs of one particular service, they also represent a case study in information sharing across community and national boundaries. RLG Cultural Materials engages a wide range of contributors with different local practices and institutional priorities. Since it is impossible to find -- and impractical to impose -- one universally applicable standard as a submission format, RLG encourages contributors to follow the suite of standards applicable to their particular community (p.1).
Critical Arguements
CA "These guidelines . . . do not set a new standard for metadata submission, but rather support a baseline that can be met by any number of strategies, enabling participating institutions to leverage their local descriptions. These guidelines also highlight the types of metadata that enhance functionality for RLG Cultural Materials. After a contributor submits a collection, RLG maps that description into the RLG Cultural Materials database using the RLG Cultural Materials data model. This ensures that metadata from the various participant communities is integrated for efficient searching and retrieval" (p.1).
Conclusions
RQ Not applicable.
SOW
DC RLG comprises more than 150 research and cultural memory institutions, and RLG Cultural Materials elicits contributions from countless museums, archives, and libraries from around the world that, although they might retain local descriptive standards and metadata schemas, must conform to the baseline standards prescribed in this document in order to integrate into RLG Cultural Materials. Appendix A represents and evaluates the most common metadata standards with which RLG Cultural Materians is able to work.
Expanded version of the article "Ensuring the Longevity of Digital Documents" that appeared in the January 1995 edition of Scientific American (Vol. 272, Number 1, pp. 42-7).
Publisher
Council on Library and Information Resources
Critical Arguements
CA "It is widely accepted that information technology is revolutionizing our concepts of documents and records in an upheaval at least as great as the introduction of printing, if not of writing itself. The current generation of digital records therefore has unique historical significance; yet our digital documents are far more fragile than paper. In fact, the record of the entire present period of history is in jeopardy. The content and historical value of many governmental, organizational, legal, financial, and technical records, scientific databases, and personal documents may be irretrievably lost to future generations if we do not take steps to preserve them."
Conclusions
RQ "We must develop evolving standards for encoding explanatory annotations to bootstrap the interpretation of digital documents that are saved in nonstandard forms. We must develop techniques for saving the bit streams of software-dependent documents and their associated systems and application software. We must ensure that the hardware environments necessary to run this software are described in sufficient detail to allow their future emulation. We must save these specifications as digital documents, encoded using the bootstrap standards developed for saving annotations so that they can be read without special software (lest we be recursively forced to emulate one system in order to learn how to emulate another). We must associate contextual information with our digital documents to provide provenance as well as explanatory annotations in a form that can be translated into successive standards so as to remain easily readable. Finally, we must ensure the systematic and continual migration of digital documents onto new media, preserving document and program bit streams verbatim, while translating their contextual information as necessary."
DC OAIS emerged out of an initiative spearheaded by NASA's Consultative Committee for Space Data Systems. It has been shaped and promoted by the RLG and OCLC. Several international projects have played key roles in shaping the OAIS model and adapting it for use in libraries, archives and research repositories. OAIS-modeled repositories include the CEDARS Project, Harvard's Digital Repository, Koninklijke Bibliotheek (KB), the Library of Congress' Archival Information Package for audiovisual materials, MIT's D-Space, OCLC's Digital Archive and TERM: the Texas Email Repository Model.