Structural Metadata and the Social Limitation of Interoperability: A Sociotechnical View of XML and Digital Library Standards Development
Copyright © 2008 Jerome McDonough. Used by permission.
Table of Contents
Copyright © 2008 Jerome McDonough. Used by permission.
Table of Contents
Ten years after its endorsement by the World Wide Web Consortium, XML has achieved a high degree of adoption within numerous, disparate communities and in a vast range of application domains, from standards for electronic filing of federal income tax ([Internal Revenue Service, 2007]) to user interface design ([Goodger et al., 2001]). The digital library community has been an active and early adopter of XML, for use in structuring both content and metadata. The reasons for this rapid uptake of XML within the digital library community are familiar to anyone with experience in the world of markup languages:
XML helps ensure platform (and perhaps more critically vendor) independence;
XML provides the multilingual character support critical to the handling of library materials;
XML's extensibility and modularity allow libraries to customize its application within their own operating environments;
XML helps minimize software development costs by allowing libraries to leverage existing, open source development tools;
XML, through virtue of being an open standard which enables descriptive markup, may assist in the long-term preservation of electronic materials; and perhaps most importantly
XML provides a technological basis for interoperability of both content and metadata across library systems.
Despite its success, however, XML has not lived up to many librarians' expectations within one area, that of interoperability. Efforts to exchange information employing XML incarnations of descriptive metadata standards such as Dublin Core have fallen prey to a number of encoding and semantic inconsistencies ([Shreeves et al., 2005]). Perhaps more surprising is the failure of XML to ensure interoperability at a syntactic level. Digital library developers have expected that shared use of an XML standard for structuring of content and metadata (what is commonly called "structural metadata" within the digital library community) would ensure content interoperability and provide a clean division between content and higher level tools and services designed to work with standardized encodings of that content ([Hurley et al., 1999]). In practice, this goal has proved extraordinarily elusive. Experiments conducted by participants in the Library of Congress National Digital Infrastructure for Preservation Program (NDIIPP) to test the exchange of digital objects between repositories failed even when participants were using the same XML-based encoding format and producing valid XML instances to exchange ([DiLauro et al., 2005], [Shirky, 2005]).
While some of the failures experienced by the Library of Congress NDIIPP tests were the result of incompatible repository infrastructure, others resulted from mismatched expectations regarding the appropriate use of METS, the XML format employed for the test. DiLauro et al., discussing Johns Hopkins University's experience in the NDIIPP tests, state:
Stanford commented after their ingest of the JHU archive that they had expected one METS object for the entire archive. Because our approach resulted in many METS files — on the order of the number of items in the archive — the Stanford ingest programs experienced out-of-memory conditions. This situation may have been ameliorated had they used the reference code provided by JHU; however, this will be an area that we will look into for future archive ingest projects.
This matter points to a broader issue observed during the various import processes of this phase. Though three of the four non-LC participants (including JHU) used METS as part of their dissemination packages, each of our approaches was different. Clearly there would be some advantage to working toward at least some common elements for these processes ([DiLauro et al., 2005]).
To date, the digital library community has treated these interoperability issues surrounding structural metadata standards as a technical problem demanding a technical solution. Most efforts to solve the interoperability problems around structural metadata have focused on the use of a profiling mechanism to further constrain the creation of instance documents conforming with the XML schema, sometimes in conjunction with a validation mechanism (such as Schematron) to test instance documents conformance with the additional requirements established in the profile ([Littman, 2006], [Keith, 2005]). However, while these mechanisms may be successful in insuring interoperability within a narrowly defined local context, they are not in themselves any guarantee of interoperability at the scale envisioned by digital library projects such as Aquifer ([Kott et al., 2006]), which hope to promote the ready exchange and interoperability of digital library content among a multitude of institutions. The official METS profile registry already contains a variety of mutually incompatible profiles for similar types of objects. Allowing the specification of local profiles may formalize the problem, but it does not solve it.
If we are to deal with the issues of interoperability that continually manifest themselves in the realm of structural metadata standards for digital libraries, I believe we need to cease viewing this purely as a technical problem, and acknowledge that it is the result of the interplay of a variety of technical and social factors. The XML standards for structural metadata employed by the digital library community represent cases of sociotechnical systems, and only when we have analyzed the social, as well as the technical, components of these systems will we be in a position to suggest how they may be optimized to achieve the goals of their users.
All of the existing and developing standards for structural metadata within the digital library community are XML-based. Any sociotechnical examination of these standards therefore must start with at least some consideration of XML itself. ([Akrich, 1992]) argues that
...when technologists define the characteristics of their objects, they necessarily make hypotheses about the entities that make up the world into which the object is to be inserted. Designers thus define actors with specific tastes, competences, motives, aspirations, political prejudices, and the rest, and they assume that morality, technology, science, and economy will evolve in particular ways. A large part of the work of innovators is "inscribing" [emph. original] this vision of (or prediction about) the work in the technical content of the new object ([Akrich, 1992], pp. 207-208).
We can learn a great deal about the viewpoints of a particular technology's designers from the documents they author which either help define and implement the goals for the technology (e.g., use cases, user needs analysis, specification documents), or which attempt to explain or promote new technology to potential users. If we look at the original XML Recommendation, we can find a relatively clear set of goals for the technology in the "Origin and Goals" section:
XML shall be straightforwardly usable over the Internet.
XML shall support a wide variety of applications.
XML shall be compatible with SGML.
It shall be easy to write programs which process XML documents.
The number of optional features in XML is to be kept to the absolute minimum, ideally zero.
XML documents should be human-legible and reasonably clear.
The XML design should be prepared quickly.
The design of XML shall be formal and concise.
XML documents shall be easy to create.
Terseness in XML markup is of minimal importance ([World Wide Web Consortium, 1998]).
This relatively small set of goals for the XML language was further elaborated upon by members of the original W3C XML Working Group in a variety of papers they authored to introduce and clarify XML to its potential user community. ([Bosak, 1998]) further defines the goals of XML as supporting the user needs of "extensibility, to define new tags as needed," "structure, to model data to any level of complexity," "validation, to check data for structural correctness," "media independence, to publish content in multiple formats," and "vendor and platform independence, to process any conforming document using standard commercial software or even simple text tools." The benefits adhering to XML's providing a standardized format include "complete interoperability of both content and style across applications and platforms; freedom of content creators from vendor control of production tools; freedom of users to choose their own views into content; easy construction of powerful tools for manipulating content on a large scale; a level playing field for independent software developers; and true international publishing across all media." Emancipatory language is invoked repeatedly here through the use of the terms 'freedom' and 'independence,' particularly affording users the freedom "to define new tags" and in so doing "choose their own views into content." A key benefit to XML in this world view is the freedom it provides users to define their own structure for documents and data, using their own semantics, and to escape restrictions that software vendors (through their own inscriptions in their own products) might wish to impose on their users. Other articles by members of the XML Working Group (see, for example, [Bosak, 1997] and [Bosak and Bray, 1999]) reiterate a vision of XML as a technology allowing users to define their own structures while simultaneously supporting interoperability of documents and data on a global scale.
The XML 1.0 Recommendation can be seen as bearing the inscription of its designers' ideological stance towards appropriate mechanisms for data and document structuring as well as appropriate relationships between document creators and software and platform vendors. The effort to promote a meta-language over any specific markup language, the adoption of Unicode as a basic character set, and the elimination of SGML features such as CONCUR, OMITTAG and SUBDOC are some of the technological means through which XML's designers sought to normalize and reify a particular set of social and technological relationships. Nor did this process stop with the release of the XML 1.0 recommendation in 1998. The period between February 1998 and October 2001 saw the development and release of a plethora of additional XML specifications, including XML Namespaces, XSLT, XPath, XML Schema, XLink/XBase, XML Information Set and XSL-FO, as well as a variety of XML software tools including parsers, editors and stylesheet engines. All of these various technological objects presented their own opportunities for their designers to further refine the ideological inscription carried within the XML 1.0 Recommendation. One of these objects in particular, the Namespaces in XML Recommendation, deserves further examination due to its significant impact on structural metadata standards developed by the digital library community.
([World Wide Web Consortium, 1999a]) provides the following justification for the introduction of a formal namespace mechanism into XML:
We envision applications of Extensible Markup Language (XML) where a single XML document may contain elements and attributes (here referred to as a "markup vocabulary") that are defined for and used by multiple software modules. One motivation for this is modularity; if such a markup vocabulary exists which is well understood and for which there is useful software available, it is better to re-use this markup rather than re-invent it.
Such documents, containing multiple markup vocabularies, pose problems of recognition and collision. Software modules need to be able to recognize the tags and attributes which they are designed to process, even in the face of "collisions" occurring when markup intended for some other software package uses the same element type or attribute name.
While it was not widely recognized at the time, the combination of the XML namespaces and the XML Schema recommendations provided the digital library community with the precise combination of tools that it had been seeking for some two years, since the publishing of the Warwick Framework document ([Lagoze, Lynch and Daniel, 1996]). The Warwick Framework described a container architecture that would allow a digital library practitioner to bind together and encapsulate different forms of metadata describing a particular object, utilizing a modular system in which different forms of metadata could be plugged in as needed into the larger framework. The description of the Warwick Framework in the original proposal document bears at least some similarity to the list of benefits for XML itself included above:
It allows the designers of individual metadata sets to focus on their specific requirements and to work within their specific areas of expertise, without concerns for generalization to ultimately unbounded scope.
It allows the syntax of metadata sets to vary in conformance with semantic requirements, community practices, and functional (processing) requirements for the kind of metadata in question.
It distributes management of and responsibility for specific metadata sets among their respective "communities of expertise".
It promotes interoperability and extensibility by allowing tools and agents to selectively access and manipulate individual packages and ignore others.
It permits access to different metadata sets related to the same object to be separately controlled.
It flexibly accommodates future metadata sets by not requiring changes to existing sets or the programs that make use of them.
In the same year that the XML 1.0 Recommendation was released, a research project undertaken by U.C. Berkeley Library, Cornell University Library, New York Public Library, Penn State University Library and Stanford University Library sought "to develop best practices for the encoding of intellectual, structural and administrative data about primary resources housed in research libraries" ([U.C. Berkeley Library, 1997]). The project, known as the Making of America II (MOA2), sought to test its notions of "best practices" by developing a test bed repository of digital library materials that would be created by the participating institutions; the content and metadata for these digital library objects were to be transmitted to the central repository, housed at the U.C. Berkeley Library, using a transmission format to be developed by the project.
Given the participants' prior experience using SGML standards such as TEI and EAD, and the recent release of the XML 1.0 Recommendation, it was decided that the transmission format should be developed using XML. This format, known as the MOA2 DTD, was intended to convey descriptive metadata and administrative metadata about a digital library object, as well as providing a structural metadata framework which allowed MOA2 document authors to describe the logical or physical structure of a work (e.g., 'this is a book with chapters' or 'this is a bound volume with pages'), to associate portions of that structure with digital files representing the content of each portion, and to associate descriptive and administrative metadata with either segments of the work, or with the digital content files themselves. As the transmission format needed to support a variety of work genres (including still photographs, journals, ledgers and diaries), the elements for describing the structure of the object were highly abstract. Drawing upon previous work done by the Text Encoding Initiative, the MOA2 DTD represented the structure of any digital library object as a series of nested <div> elements. Child <fptr> elements underneath a <div> associated the <div> with digital content files, and a variety of XML ID/IDREF attributes were used to link <div> elements and content files with descriptive and administrative metadata.
While the MOA2 DTD was reasonably successful, experiments with the use of the format outside the original research project revealed several deficiencies, including a lack of support for time-based media, and what was seen by members of the digital library community as an excessively restrictive model with respect to the descriptive and administrative metadata allowed within a MOA2 document. In this latter regard, the MOA2 DTD was viewed as not living up to the promise of modular design and flexibility that should be upheld in XML schema design. A successor effort, METS, was launched in 2001 to try to improve on the MOA2 DTD. The METS format retained a very similar model for structuring a digital library object (with a few extensions to support time based media), but implemented a far more permissive structure with respect to descriptive and administrative metadata and was created using an XML Schema, rather than a DTD, in order to exploit XML namespaces and their further promise of modularity and schema reuse. The general structure for an object such as a digitized book in METS might look like this:
<?xml version="1.0" encoding="UTF-8"?> <mets xmlns="http://www.loc.gov/METS/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xmlns:xlink="http://www.w3.org/1999/xlink" xsi:schemaLocation="http://www.loc.gov/METS/ http://www.loc.gov/standards/mets/mets.xsd"> <dmdSec ID="DM1" > <mdWrap MDTYPE="OTHER"> <xmlData><meta>Descriptive Metadata for Book</meta> </xmlData> </mdWrap> </dmdSec> <amdSec> <techMD ID="AM1"> <mdWrap MDTYPE="OTHER"> <xmlData><meta>Administrative metadata applicable to TIFF files</meta></xmlData> </mdWrap></techMD> </amdSec> <fileSec> <fileGrp> <file ID="F1" ADMID="AM1"><FLocat LOCTYPE="URN" xlink:href="urn:x-mets:Location_of_Page_One_TIFF_Image"/></file> <file ID="F2" ADMID="AM1"><FLocat LOCTYPE="URN" xlink:href="urn:x-mets:Location_of_Page_Two_TIFF_Image"/></file> <file ID="F3" ADMID="AM1"><FLocat LOCTYPE="URN" xlink:href="urn:x-mets:Location_of_Page_Three_TIFF_Image"/></file> </fileGrp> </fileSec> <structMap> <div TYPE="book" DMDID="DM1"> <div TYPE="chapter"> <fptr FILEID="F1" /> <fptr FILEID="F2" /> <fptr FILEID="F3" /> </div> </div> </structMap> </mets>The goal of representing the structure of a work as a hierarchy of nested <div> elements with TYPE attributes was to have a relatively simple, abstract hierarchical structure that could be readily applied to a variety of materials. This was intended to promote the adoption of the standard (a single, simple standard is more likely to be adopted than a variety of complex ones), which in turn was seen as promoting interoperability. Having all of the digital library community using a single standard for structuring content and metadata would certainly be preferable to the community adopting a variety of disparate standards.
It should be noted that this move towards abstraction was a relatively significant break from the SGML design practices that many research libraries had been using to date. While it is true that the notion of using nested <div> elements was derived from the TEI text apparatus, TEI does not rely on pure abstraction; one does not expect to encounter encoding such as <div type="figure"> in a TEI document, when a <figure> element is available to use. Just as XML itself, the METS schema carries an inscription of its designers' worldview, that it was preferable to develop a single, simple, generalizable, highly abstract model and encoding mechanisms to structure content and metadata for digital library objects, rather than to pursue the development of a variety of highly specific schemas (one for photographs, one for journals, etc.), or a grand encompassing schema that contained elements appropriate to different genres that could be combined as needed (e.g., the TEI model). Despite its use of the abstract <div> element with a TYPE attribute to represent the structural components of a digital library object, however, the METS schema insisted on the use of more specific concrete elements to identify different forms of metadata, with the <dmdSec> element used for descriptive metadata and the <amdSec> element used for administrative metadata, along with a series of subelements for different forms of administrative metadata (technical, rights, provenance and source). This typification of different forms of metadata was itself an effort to promote modularity in further metadata schema development. By identifying specific subclasses of metadata within the METS schema, METS' designers hoped to encourage XML developers in the digital library world to create discrete, specialized metadata standards that would align with those subclasses, and that those encoding digital library objects could then select from a set of such modular XML metadata standards in composing a particular object. Through METS' design, its implementers consciously sought to encourage the adoption of modular schema design practices within the digital library community.
The digital library community was obviously not the only community in the late 1990s that recognized the potential benefits of XML, nor was it the only community to develop standards for structural metadata based on this notion of a highly abstract structure. In October of 2000, the Moving Picture Experts Group (MPEG) issued a call for proposals for "digital item declaration." This work culminated in the release of the MPEG-21 Digital Item Declaration (ISO/IEC 21000-2) standard in 2003. While MPEG-21 is similar to the METS standard in employing a rather abstract hierarchical structural mechanism for ordering content and metadata, it differs inasmuch as non-structural metadata (<Descriptor> elements in MPEG-21 parlance) are not typed, and structural metadata elements, while still rather abstract, are of three different types: <Container>, <Item> and <Component>. Our METS example above, if translated into MPEG-21 DIDL format, might take the following format:
<DIDL xmlns="urn:mpeg:mpeg21:2002:01-DIDL-NS" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="urn:mpeg:mpeg21:2002:01-DIDL-NS file:/Users/jeromemcdonough/Documents/GSLIS/Research/OAIORE/MPEG21.DIDL.OFFICIAL.xsd"> <Container> <Item> <Descriptor id="DM1"> <Statement mimeType="text/xml"><meta>Descriptive Metadata for Book</meta></Statement> </Descriptor> <Descriptor id="AM1"> <Statement mimeType="text/xml"><meta>Administrative Metadata Applicable to TIFF Files</meta></Statement> </Descriptor> <Item> <Component> <Descriptor><Reference target="#AM1" /></Descriptor> <Resource mimeType="image/tiff" ref="urn:x-mets:Location_of_Page_One_TIFF_Image"> </Resource> </Component> <Component> <Descriptor><Reference target="#AM1" /></Descriptor> <Resource mimeType="image/tiff" ref="urn:x-mets:Location_of_Page_Two_TIFF_Image" /> </Component> <Component> <Descriptor><Reference target="#AM1" /></Descriptor> <Resource mimeType="image/tiff" ref="urn:x-mets:Location_of_Page_Three_TIFF_Image" /> </Component> </Item> </Item> </Container> </DIDL>While it obviously uses somewhat different element names and linking mechanisms than METS, MPEG-21 DIDL follows a very similar model, and has been successfully employed by several digital libraries. The logical structure of the object is laid out via a hierarchical arrangement of <Container>, <Item> and <Component> elements, and various nodes in that tree structure can be associated with metadata via <Descriptor> elements, and content via <Resource> elements.
If we examine the MPEG-21 standard to determine what inscriptions its designers have placed within it regarding its use, we find messages very similar to that of METS. Structural metadata should be highly abstract, so that a very small set of elements can be employed to structure widely disparate content genres. While METS was perhaps more vocal in trying to push the message that further development of metadata schemas should try to create small, focused and modular metadata sets that could be drawn upon as needed to encode a particular object, MPEG-21's use of <Descriptor> elements that can be associated with other elements throughout an MPEG-21 document conveys the same message. Other structural metadata standards of interest to the digital library community employ similar mechanisms. The XFDU standard for data archiving uses hierarchies of <ContentUnit> elements that may be associated with <dataObjects> and <metadataObjects>. The IMS Content Packaging standard for learning objects uses hierarchies of <item> elements that may be associated with <resources> and <metadata>. While implementations differ in details, the pattern is similar and widespread across the various structural metadata standards of concern to the digital library community. Again and again we designers seeking to achieve wide adoption of their standard in order to promote interoperability between differing institutions; to secure this goal, they favor a highly abstract structural mechanism which can be applied to a wide variety of content, and mechanisms to allow a variety of additional metadata schema to 'plug and play' within the larger structural framework.
While of perhaps some limited interest to social researchers of technology, none of the preceding seems particularly surprising or problematic. That the designers of XML itself, and of encoding standards for digital library metadata and content, should favor flexibility, extensibility, modularity and the use of abstraction to support the generalizability of their standard, and hence promote its widespread adoption to help achieve interoperability, would not be a great shock to anyone who has spent more than five minutes in the company of technologists. These are all considered almost innate goods among software engineers in general and markup language enthusiasts in particular. Yet the NDIIPP tests cited previously would seem to indicate that flexibility, extensibility, modularity and abstraction are not in and of themselves sufficient to achieve interoperability. So what, specifically, is the problem that METS and other structural metadata standards are encountering?
One of the earliest discussion points in the development of the METS standard was which of the various elements within the schema should be declared mandatory and which optional. After some discussion among the members of the working group that established METS' original design, it was decided that the <structMap> element, which records the basic tree structure on to which content files and metadata are mapped in METS, would be the only required element. METS, in the group's opinion, was fundamentally a structural metadata standard; it existed to provide a framework into which other descriptive and administrative metadata, as well as content, could be placed. The <structMap> element provided the tree upon which all of the other structural components of METS where hung, where the logical and physical structure of a work could be delineated, and so was really the core of a METS file and the only section that needed to be mandatory. As the <structMap> was the only mandatory portion of a METS file, it was also expected that any structural description of a work should reside there; software that would process a METS file would expect to find logical or physical descriptions of the structure of a work residing within a structural map, and not elsewhere in the METS file.
It was a matter of some surprise for many in the METS community, then, when the Library of Congress, which serves as the maintenance agency for the METS standard, began to produce METS files for digital versions of certain kinds of audio recordings which placed the logical structure of the works in MODS records contained within the METS descriptive metadata section (<dmdSec>) rather than in the structural map, and registered a profile of METS establishing this as their formal internal practice for 'recorded events' ([Library of Congress, 2006]). The MODS record within a METS file would provide a logical structure for the work using a hierarchical arrangement of the MODS <relatedItem> element, while the METS <structMap> would contain the physical structure, with ID/IDREF links used to draw connections between the two structural descriptions. A recorded concert, for example, might have a MODS record containing a hierarchy such as this:
<mods:mods xmlns:mods="http://www.loc.gov/mods/v3" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://www.loc.gov/mods/v3 http://www.loc.gov/standards/mods/mods.xsd"> <mods:titleInfo> <mods:nonSort>The </mods:nonSort> <mods:title>1946 Library of Congress recital</mods:title> </mods:titleInfo> <mods:relatedItem type="constituent" ID="DMD_disc01_tr001"> <mods:titleInfo type="uniform"><mods:partName>Chaconne von Vitali</mods:partName></mods:titleInfo> </mods:relatedItem> <mods:relatedItem type="constituent" ID="DMD_disc01_tr002_005"> <mods:titleInfo><mods:title>Sonata in G minor, BWV 1 001</mods:title></mods:titleInfo> <mods:relatedItem type="constituent" ID="DMD_disc01_tr002"> <mods:titleInfo><mods:partName>Adagio</mods:partName> </mods:titleInfo> </mods:relatedItem> <mods:relatedItem type="constituent" ID="DMD_disc01_tr003"> <mods:titleInfo><mods:partName>Fuga : allegro</mods:partName> </mods:titleInfo> </mods:relatedItem> <mods:relatedItem type="constituent" ID="DMD_disc01_tr004"> <mods:titleInfo><mods:partName>Siciliano</mods:partName> </mods:titleInfo> </mods:relatedItem> <mods:relatedItem type="constituent" ID="DMD_disc01_tr005"> <mods:titleInfo><mods:partName>Presto</mods:partName> </mods:titleInfo> </mods:relatedItem> </mods:relatedItem> <mods:relatedItem type="constituent" ID="DMD_disc01_tr006"> <mods:titleInfo><mods:title>Paganiniana : variations</mods:title> </mods:titleInfo> </mods:relatedItem> <mods:identifier type="lccn">99594334</mods:identifier> </mods:mods>while the structural map would record the physical structure of the work as follows:
<div TYPE="cd:compactDiscObject" DMDID="MODS1"> <div TYPE="cd:disc"> <div DMDID="DMD_disc01_tr001" TYPE="cd:track"></div> <div DMDID="DMD_disc01_tr002" TYPE="cd:track"></div> <div DMDID="DMD_disc01_tr003" TYPE="cd:track"></div> <div DMDID="DMD_disc01_tr004" TYPE="cd:track"></div> <div DMDID="DMD_disc01_tr005" TYPE="cd:track"></div> <div DMDID="DMD_disc01_tr006" TYPE="cd:track"></div> </div> </div>
This represents a valid use of METS (in the technical XML sense), but is a departure from expected practice, which would be to include two structural maps within the METS file, one representing the physical structure and the second representing the logical structure of the work.
This example demonstrates two problems that have been impeding the development of interoperable content within the digital library community. The first is that the implementation of highly abstract elements for the definition of structure provides a tremendous amount of flexibility to document encoders; there is a vast number of potential encodings of any given object in METS, with variations possible in depth of structure (do I limit my structure to musical movements or do I provide structural information to the measure level?), labeling (you say TYPE="book", I say TYPE="monograph"), and arrangement (should the Lord of Rings film trilogy be encoded as a single METS file? Three METS files? Three METS files for the individual films and a fourth representing the abstract notion of the Trilogy?). This can lead to significant variation in encoding practices, even between two institutions dealing with remarkably similar material and using the same metadata standards, as noted by ([DiLauro et al., 2005]).
The second problem is what we might call the problem of standards independence. The various metadata schema standards which have been developed with the help of the digital library community during the past decade have all been created with the understanding that in order to ensure their usefulness in a variety of different organizations and environments, they should not contain inherent dependencies on other schemas; they need to be able to express all of the information necessary for their particular domain on their own. For many of these standards, the designers recognized a need to be able to account for relationships between various content objects being described by the standard, whether the description being applied was the more traditional form of intellectual description you would expect in a library catalog, or a technical description of the composition of a TIFF image. The result has been that a number of common metadata schemas within the digital library field contain elements for expressing structural metadata: Dublin Core has its <relation> element, MODS has its <RelatedItem> element, the PREMIS schema for preservation metadata has a <relation> element, even the MIX standard for still image technical metadata contains an element for referencing previous image metadata. As the standards' developers felt they should not make their efforts dependent on structural metadata mechanisms in other standards, they implemented their own. Unfortunately, with the addition of each new metadata standard containing structural metadata capabilities, the potential for difficulties with our first problem increases. Every new metadata standard within the digital library community seems to add another mechanism for describing the structural relationships between content objects, and hence greater potential for variation in object encoding practices.
The irony is that both these problems derive from the flexibility, extensibility, modularity and use of abstraction to create structural metadata elements of wide applicability that the designers of the metadata schemas hoped to promote. The potential range of variation in encoding structural metadata is the result of each of these factors. The use of abstraction in METS (i.e., the <div> element) was an attempt to make the standard flexible in application; however, it opens up a tremendous degree of play in encoding practice. If you ask two different individuals how many page breaks there are in a text, the likelihood that they will get the same answer is a good deal greater than if you ask them how many divisions there are in a text. The use of abstraction opens up encoding to a much greater degree of personal interpretation, and hence variation. The extensibility of METS and the hope to promote a modular system of metadata schema reuse that its authors inscribed within it open up the possibility of using other metadata schemas to encode structural metadata. And it was this same desire for flexibility and modularity that has led other metadata schema designers to include structural metadata components in their own schema; they wanted to ensure that their own efforts were flexible enough to be applied in a variety of settings, and with a variety of other schemas. But having to design their own schema without knowing the specific supporting capabilities to be found in other schemas with which their own might be used, they are inevitably forced to create structural metadata capacities of their own within their schema. The designers of metadata schemas (structural or otherwise) within the digital library community have sought to adhere to a particular set of design practices, seeking to create flexible, extensible, modular and generalized tools, and to promote like practice in others through inscription of their view of appropriate XML design within their technological artifacts. Unfortunately, promoting such good practices has been a death blow to one of the principle reasons for adopting XML in the first place: to ensure interoperability of digital library materials across systems. wide-scale interoperability requires wide-scale adoption, but the design practices of schema implementers intended to promote wide-scale adoption run directly counter to wide spread interoperability.
Hence XML's similarity to a rope. Like a rope, it is extraordinarily flexible; unfortunately, just as with rope, that flexibility makes it all too easy to hang yourself.
The digital library community seems to face a dilemma at this point. Through its pursuit of design goals of flexibility, extensibility, modularity and abstraction, and its promulgation of those goals as common practice through its implementation of XML metadata standards, it has managed to substantially impede progress towards another commonly held goal, interoperability of digital library content across a range of systems. How then, should the community respond?
One possible response to this situation would be to say that perhaps our community cares less about interoperability than we thought. Despite the existence of projects such as the DLF Aquifer venture, it may be that interoperability is in fact a lower priority for the digital library community than it likes to believe, and the adoption of metadata standards that impede interoperability is merely a reflection of that underlying reality, and not a major problem to resolve. There is at least some reason to suspect this may be the case. Most research libraries have a clearly defined local clientele, and while voices within the digital library community have been calling for the liberation of content from local silos for some time ([Seaman, 2003]), libraries' primary responsibility will always be to their local communities. If systems developed to deliver digital library content to the local community are successful in that context, and if the costs associated with achieving much more widespread interoperability are high, then many libraries may decide that interoperability, while desirable, is a goal which may have to wait.
If libraries do wish to make progress on the issue of interoperability of structural metadata, they will need to recognize that, as ([Renear and Golovchinksky, 2001]) observed, "every significant information processing standardization effort must skillfully negotiate competing and apparently irreconcilable objectives, [and] serve a wide variety of stakeholders with many different interests." In the case of structural metadata, the particular competing objectives that the digital library community does not seem to have successfully reconciled to date are what ([Kendall, 2007]), in a discussion of blogging practices, has labeled the problem of "control vs. connection." The structural metadata standards which have been developed to date, with their emphasis on flexibility, extensibility and modularity have sought to afford local institutions the greatest degree of control possible in their encoding practices. The standards are designed to allow any given institution to do what it wants. This has obvious benefits in terms of easing adoption of the standard in any given context, and as a result insuring the standards' widespread adoption (obviously a good thing in a standard). However, increasing the amount of local control over the ways in which a language is used and developed is fundamentally at odds with a language's ability to serve as a means for connection with others outside the local context. It is, in essence, promoting the development of regional dialects at the expense of mutual intelligibility. The particular case of structural metadata standards reveals that sufficient local variation in syntax, the ways in which people structure their objects using a markup language, can be as fatal to communication as variation in semantics.
Given this fundamental tradeoff between control and connection, libraries wishing to promote interoperability of digital library content have two possible strategies. The first, and most obvious, is to attempt to alter the balance currently struck between connection and control to more strongly favor connection. There are several mechanisms which the library community might employ in pursuit of this strategy, including the design and use of schema which more significantly restrict both the means for recording the structure of objects and the ability to employ arbitrary additional schema within instance documents (or developing profiles of existing schema to achieve the same ends), establishing formal rules of structural description (equivalent to rules of description used in cataloging for creating bibliographic records) dictating aspects of object encoding not susceptible to enforcement through XML's validation mechanisms, and mandating the use of particular controlled vocabularies and ontologies within document instances to record information such as a <div> element's TYPE attribute in METS. Decreasing the possibility for local variation in encoding of structural metadata will certainly help improve digital libraries' capability to interoperate with each other. However, removing local capacity for variation will also tend to reduce the number of institutions who are willing to use such a markup language. If the digital library community, for instance, was to revise the METS standard to forbid any use of a <relation> or <RelatedItem> element in a descriptive metadata section to express the logical structure of a work, it would assist in insuring interoperability of digital content, but it might also very well mean losing the Library of Congress's support for the standard. More importantly, however, such an approach seems to overlook one of the fundamental realities of the web environment: communities of practice no longer operate in splendid isolation from each other (if indeed, they ever did). Even if libraries could agree on a structural metadata standard that enabled a significantly greater degree of support for interoperability than we find with today's standards, libraries must now interact with a variety of other communities (publishers, museums, archives, educational technology companies, etc.) that are also creating their own structural metadata standards. The odds of our moving past this structural tower of Babel to achieve a single object encoding practice across all communities do not look promising, to say the least.
The second, and I believe more promising, strategy is to accept that the need for local community control over encoding practices is a valid one, that regional 'dialects' of markup languages are inevitable, and that we must find ways to facilitate information exchange across the boundaries of different communities' markup vernacular. However, this will require a significant shift in the digital library community's relationship to the notion of standards. Specifically, I believe the library community needs to shift from its current singular focus on schema development to a dual focus on both schema development and translation between schemas. It is worth noting here that the Library of Congress currently serves as maintenance agency for a variety of XML standards developed within the library community; if you examine the list of standards that they are maintaining ([Library of Congress, 2008]), however, you will find that while there are several metadata standards listed, standardized stylesheets to enable conversion between formats are not listed here.
Such stylesheets do exist in some cases. The Library of Congress has, for example, provided stylesheets to enable conversion of MODS descriptive metadata records into MARC/XML format and back. These efforts to formalize prior work that established cross-walks between different descriptive metadata standards are not, however, seen by the community as having the status and importance of standards, as exhibited by their omission from the 'Standards at the Library of Congress' web page. If the digital library community wishes to support interoperability while simultaneously affording institutions localized control over encoding practices, that situation needs to change. Enabling translation activity must be seen as important, if not more important, a standards activity than the creation of schema for metadata sets.
A heightened emphasis on standardizing translation between markup languages will obviously mean further work on formalizing translations between markup languages using XSLT, and treating those with the level of attention and care that the community has lavished on metadata schema. However, it might also be worth considering whether the notion of formal rules of structural description mentioned earlier might be of benefit in trying to achieve greater translatability between different markup languages. As an example of what this might mean, consider the example of the 1:1 principle in Dublin Core ([Hillmann, 2005]), that a single Dublin Core record should describe one and only one resource. The 1:1 principle provides guidance on the relationship between a metadata record and a described resource that is applicable outside the realm of Dublin Core; in fact, several other descriptive metadata standards developed since Dublin Core have made reference to the 1:1 principle as a guide to usage. We could easily envisage similar principles being developed for structural metadata that could guide usage of a variety of different structural metadata standards, and by working to insure similar use practices, would help insure ease of translation between different structural markup languages. We might, for instance, take as a working principle that any given structural metadata document should never contain more than two levels of structural hierarchy. Our METS example above passes muster with this rule; if, however, we modified it so that a third level of <div> elements was needed (of TYPE 'subchapter', for example), then we would be in violation of this principle. To fix this problem, we could employ METS' <mptr> element to allow the <div> elements for each chapter to reference separate METS files containing the structural descriptions for the individual chapters. Through the establishment of common principles of structural encoding and standardized stylesheets for translation, we might be able to improve our ability to interoperate while simultaneously retaining some flexibility for local encoding practice (although obviously adoption of common principles of structural encoding may impede local control in favor of connection to some degree).
The rise of the network information society is presenting libraries with a variety of new challenges. Perhaps the most significant of these is the heightened degree of interaction with communities of practice that do not share libraries' standards, practices or values. If libraries are to survive and thrive in this new information society, they must alter their own value structure to prioritize communication with other communities to an equal, if not greater, extent than internal communication between libraries. If they pursue this course, they may find that issues of internal interoperability of library systems are more tractable than they have appeared to date.
[Akrich, 1992] Akrich, M. (1992). The De-Scription of Technical Objects. In Weibe E. Bijker &John Law (Eds.) Shaping Technology/Building Society: Studies in Sociotechnical Change. Cambridge, MA: MIT Press. Retrieved from http://hdl.handle.net.proxy2.library.uiuc.edu/2027/heb.01128.0001.001
[Bosak, 1998] Bosak, J. (October, 1998). Media-Independent Publishing: Four Myths about XML. Computer 31(10). IEEE Computer Society. Retrieved from http://ieeexplore.ieee.org/iel4/2/15590/00722303.pdf
[Bosak and Bray, 1999] Bosak, J., and Bray, T. (May 1999). XML and the Second-Generation Web. Scientific American 280(5). Retrieved from http://search.ebscohost.com/login.aspx?direct=true&db=aph&AN=1775198&site=ehost-live
[Daniel and Lagoze, 1997] Daniel, Jr., R. & Lagoze, C. (November 1997). Extending the Warwick Framework: From Metadata Containers to Active Digital Objects. D-Lib Magazine 3(10). Retrieved from http://www.dlib.org/dlib/november97/daniel/11daniel.html
[DiLauro et al., 2005] DiLauro, T., Patton, M., Reynolds, D. & Choudhury, G. S. (Dec. 2005). The Archive Ingest and Handling Test. D-Lib Magazine 11(12). Retrieved from http://www.dlib.org/dlib/december05/choudhury/12choudhury.html
[Goodger et al., 2001] Goodger, B., Hickson, I., Hyatt, D. & Waterson, C. (2001). XML User Interface Language (XUL) 1.0 [online]. David Hyatt, ed. Mozilla Foundation. Retrieved from http://www.mozilla.org/projects/xul/xul.html
[Hillmann, 2005] Hillmann, D. (2005). Using Dublin Core [online]. Dublin Core Metadata Initiative. Retrieved from http://dublincore.org/documents/usageguide/index.shtml
[Hurley et al., 1999] Hurley, B. J., Price-Wilken, J., Proffitt, M., & Besser, H. (1999). The Making of America II test bed Project: A Digital Library Service Model. Washington, DC: Digital Library Federation. Retrieved from http://www.clir.org/pubs/reports/pub87/pub87.pdf
[Internal Revenue Service, 2007] Internal Revenue Service (2008). Current Valid XML Schemas for 11/20/1120-F/1120S/7004 Modernized e-File (MeF) [online]. Washington, DC: United States Department of the Treasure. Retrieved from http://www.irs.gov/efile/article/0,,id=128360,00.html
[Keith, 2005] Keith, C. (2005). Making METS Profiles Machine-Actionable to Support Validation and Facilitate Interoperability. Digital Library Federation Spring Forum 2005, April 13-15, 2005, San Diego, CA. Retrieved from http://www.diglib.org/forums/Spring2005/presentations/keith-2005-04.pdf
[Kendall, 2007] Kendall, L. (September 2007). Shout Into the Wind, and It Shouts Back: Identity and interactional tensions on LiveJournal. First Monday 12(9). Retrieved from http://www.firstmonday.org/issues/issue12_9/kendall/index.html
[Kott et al., 2006] Kott, K., Dunn, J., Halbert, M., Johnston, L., Milewicz, L. & Shreeves, S. (May 2006). Digital Library Federation (DLF) Aquifer Project. D-Lib Magazine 12(5). Retrieved from http://www.dlib.org/dlib/may06/kott/05kott.html
[Lagoze, Lynch and Daniel, 1996] Lagoze, C., Lynch, C. & Daniel, Jr., R. (June 28, 1996). The Warwick Framework: A Container Architecture for Aggregating Sets of Metadata. Retrieved from http://www.ifla.org/documents/libraries/cataloging/metadata/tr961593.pdf
[Library of Congress, 2006] Library of Congress (Jan. 12, 2006). Library of Congress METS Profile for Recorded Events. Retrieved from http://www.loc.gov/standards/mets/profiles/00000009.html
[Littman, 2006] Littman, J. (May 2006). A Technical Approach and Distributed Model for Validation of Digital Objects. D-Lib Magazine 12(5). Retrieved from http://www.dlib.org/dlib/may06/littman/05littman.html
[Nelson et al., 2005] Nelson, M. L., Bollen, J., Manepalli, G. & Haq, R. (Dec. 2005). Archive Ingest and Handling Test: The Old Dominion University Approach. D-Lib Magazine 11(12). Retrieved from http://www.dlib.org/dlib/december05/nelson/12nelson.html
[Renear and Golovchinksky, 2001] Renear, A. & Golovchinksy, G. (2001). Content Standards for Electronic Books: The OEBF Publication Structure and the Role of Public Interest Participation. Journal of Library Administration 35(1/2). Retrieved from http://www.haworthpress.com/store/E-Text/View_EText.asp?sid=VNKBQXQJNSCX9GMAS5QPA2M1RQGEF1C7&a=3&s=J111&v=35&i=1%2F2&fn=J111v35n01%5F07
[Seaman, 2003] Seaman, D. (July/August 2003). Deep Sharing: A Case for the Federated Digital Library. Educause Review 38(4). Retrieved from http://www.google.com/url?sa=t&ct=res&cd=9&url=http%3A%2F%2Fwww.educause.edu%2Fir%2Flibrary%2Fpdf%2Ferm0348.pdf&ei=V_kISImMN6jkigGTkLCQDA&usg=AFQjCNH1Dwk2C8eEk9vNFXGg4IwUm9YoZw&sig2=sJnhZs0JwbGwGZaybhbZOA
[Shirky, 2005] Shirky, C. (Dec. 2005). AIHT: Conceptual Issues from Practical Tests. D-Lib Magazine 11(12). Retrieved from http://www.dlib.org/dlib/december05/shirky/12shirky.html
[Shreeves et al., 2005] Shreeves, S. L., Knutson, E. M., Stvilia, B., Palmer, C. L., Twidale, M. B. & Cole, T. W. (2005). Is 'Quality' Metadata 'Shareable' Metadata? The Implications of Local Metadata Practices for Federated Collections. In H. A. Thompson (Ed.) Proceedings of the Twelfth National Conference of the Association of College and Research Libraries, April 7-10 2005, Minneapolis, MN. Chicago, IL: Association of College and Research Libraries, pp. 223-237. Retrieved from http://hdl.handle.net/2142/145
[U.C. Berkeley Library, 1997] The Library, University of California at Berkeley (1997). The Making of America II test bed Project Proposal. Retrieved from http://sunsite.berkeley.edu/moa2/moaproposal.html
[World Wide Web Consortium, 1998] World Wide Web Consortium (Feb. 10, 1998). Extensible Markup Language (XML) 1.0: W3C Recommendation 10-February-1998. Tim Bray, Jean Paoli & C. M. Sperberg-McQueen (Eds.). Retrieved from http://www.w3.org/TR/1998/REC-xml-19980210
[World Wide Web Consortium, 1999a] World Wide Web Consortium (Jan. 14, 1999). Namespaces in XML. Tim Bray, Dave Hollander & Andrew Layman (Eds.). Retrieved from http://www.w3.org/TR/1999/REC-xml-names-19990114/
[World Wide Web Consortium, 1999b] World Wide Web Consortium (Jan. 14, 1999). The World Wide Web Consortium Issues "Namespaces in XML" as a W3C Recommendation [Press Release]. Retrieved from http://www.w3.org/Press/1999/XML-Namespaces-REC
 See http://www.loc.gov/standards/mets/mets-registered-profiles.html for the set of formally registered METS profiles.
 I am including within this set the Metadata Encoding & Transmission Standard (METS, see http://www.loc.gov/standards/mets/), the MPEG-21 Digital Item Declaration Language (DIDL, see http://www.chiariglione.org/mpeg/standards/mpeg-21/mpeg-21.htm), the Open Archives Initiative Object Reuse & Exchange standard (OAI-ORE, see http://www.openarchives.org/ore/) and the Fedora Object XML specification (FOXML, see http://www.fedora-commons.org/documentation/2.2.2/userdocs/digitalobjects/introFOXML.html). There are a variety of other structural metadata standards that are of particular interest to the digital library community, although not being developed within it, including standards for data archiving such as the XML Formatted Data Unit specification (XFDU, see http://sindbad.gsfc.nasa.gov/xfdu/) being developed by the Consultative Committee on Space Data Systems and standards for structuring content for e-learning systems such as the IMS Content Packaging specification (IMS-CP, see http://www.imsglobal.org/content/packaging/) developed by the IMS Global Learning Consortium and the Shareable Content Object Reference Model (SCORM, see http://www.adlnet.gov/scorm/index.aspx) developed by the United States Department of Defense.
 See http://sunsite.berkeley.edu/moa2/papers/moa2dtd2.htm to see an HTML version of the DTD.
 Complete MODS & METS records for this example can be found at the Library of Congress webpage for the 1946 Library of Congress recital at http://lcweb2.loc.gov/diglib/ihas/loc.natlib.ihas.200003790/default.html