12.327, FYI: Workshop: Web-Based Lang Documentation & OLAC
The LINGUIST Network
linguist at linguistlist.org
Thu Feb 8 23:03:53 UTC 2001
LINGUIST List: Vol-12-327. Thu Feb 8 2001. ISSN: 1068-4875.
Subject: 12.327, FYI: Workshop: Web-Based Lang Documentation & OLAC
Moderators: Anthony Aristar, Wayne State U.<aristar at linguistlist.org>
Helen Dry, Eastern Michigan U. <hdry at linguistlist.org>
Andrew Carnie, U. of Arizona <carnie at linguistlist.org>
Reviews (reviews at linguistlist.org):
Simin Karimi, U. of Arizona
Terence Langendoen, U. of Arizona
Editors (linguist at linguistlist.org):
Karen Milligan, WSU Naomi Ogasawara, EMU
Lydia Grebenyova, EMU Jody Huellmantel, WSU
James Yuells, WSU Michael Appleby, EMU
Marie Klopfenstein, WSU Ljuba Veselinova, Stockholm U.
Software: John Remmers, E. Michigan U. <remmers at emunix.emich.edu>
Gayathri Sriram, E. Michigan U. <gayatri at linguistlist.org>
Home Page: http://linguistlist.org/
The LINGUIST List is funded by Eastern Michigan University, Wayne
State University, and donations from subscribers and publishers.
Editor for this issue: Lydia Grebenyova <lydia at linguistlist.org>
=================================Directory=================================
1)
Date: Thu, 01 Feb 2001 12:57:25 EST
From: "J. Albert Bickford" <albert_bickford at sil.org>,
Steven Bird <sb at ldc.upenn.edu>
Subject: Workshop Report: Web-Based Lang Documentation & Description, & OLAC
-------------------------------- Message 1 -------------------------------
Date: Thu, 01 Feb 2001 12:57:25 EST
From: "J. Albert Bickford" <albert_bickford at sil.org>,
Steven Bird <sb at ldc.upenn.edu>
Subject: Workshop Report: Web-Based Lang Documentation & Description, & OLAC
Workshop Report
Workshop on Web-Based Language Documentation and Description,
and the Open Language Archives Community
J. Albert Bickford
SIL-Mexico and University of North Dakota
albert_bickford at sil.org
The Workshop on Web-Based Language Documentation and Description (December
12-15, 2000, University of Pennsylvania) brought together linguists,
archivists, software developers, publishers and funding agencies to discuss
how best to publish information about language on the internet. This
workshop, together with the Open Language Archives Community which is
developing out of it, seem especially important in providing useful
information about linguistics and less-commonly studied languages for both
scholars and the wide general audience that can be found on the web. I hope
that this report will be useful in understanding these new developments in
the linguistics publishing and archiving field.
The aim of the workshop was to establish an infrastructure for electronic
publishing that simultaneously addresses the needs of users (including
scholars, language communities, and the general public), creators,
archivists, software developers, and funding agencies. Such an
infrastructure would ideally meet a number of requirements important to
these different stakeholders, such as:
* provide a single entry point on the internet through which all materials
can be easily located, regardless of where they are stored (on the
internet or in a traditional archive). Essentially, this would be a
massive union catalog of the whole internet and beyond.
* identify every language uniquely and precisely, so that all materials
relevant to a particular language can be located
* make available software for creating, using, and archiving data
(especially data in special formats); this includes software to help
convert data from older formats to newer ones
* serve as a forum for giving and receiving advice about software, archiving
practices, and related matters
* provide opportunity for comments and reviews of materials published within
the system
The workshop was organized by Steven Bird (University of Pennsylvania) and
Gary Simons (SIL International).[1] It included approximately 40
presentations and several working sessions on a variety of topics.
There was general agreement among the participants that a system for
organizing the wealth of language-related material on the internet is
needed, and that an appropriate way to establish one is by following the
guidelines of the Open Archives Initiative (OAI)
[http://www.openarchives.org/]. (These guidelines provide a general
framework for creating systems like this for specific scholarly
communities.) An OAI publishing and archiving system contains the following
elements:
* data providers, which house the materials that are indexed in the system
* a standardized set of cataloguing information for describing each of the
materials, also known as "metadata" (i.e., data about data)
* service providers, which collect the metadata from all the data providers
and allow users to search it in various ways so as to locate materials of
interest to them
In the case of linguistics, the system will be known as the Open Language
Archives Community (OLAC). The Linguist list [http://www.linguistlist.org/]
has agreed to serve the system as its primary service provider. It will be
the main source that people will use to find materials through the
system. Further information about OLAC can be found at
[http://www.language-archives.org/]. The agreement to establish OLAC is
probably the most important accomplishment of the workshop.
This agreement was solidified through working sessions which met during the
workshop and started the process of working through the details in various
areas, such as:
* Character encoding: Unicode, fonts, character sets, etc.
* Data structure for different types of data (lexicons, annotated text, etc.)
* Metadata (cataloguing information that should be common to the whole
community and how it should be represented in the computer) and other
concerns of archivists
* Ethics, especially the responsibilities that archivists and publishers
have to language communities
* Expectations of users, creators (e.g. authors), software developers
These and other issues will continue to be discussed on email lists in the
coming months, ultimately culminating in recommendations for "best practice" in
each area, together with a preliminary launch of the whole system, hopefully
within a year. (Prototypes of the system are available now at the OLAC address
above, along with various planning documents.)
There were also a number of conference papers, which provided a foundation for
making the working sessions productive. Rather than list or review all the
presentations here, I will summarize them, since they are all available on the
conference website [http://www.ldc.upenn.edu/exploration/expl2000/]. The topics
covered included the following:
* Proposals for various aspects of the OLAC system
* Concerns of various stakeholders, such as archivists, sponsors, language
communities
* Descriptions and demonstrations of specific software, research projects,
and web publishing systems
* Metadata and metadata standards
* Technical issues, such as Unicode, the OAI, sorting, data formats for
different types of language materials (e.g. dictionaries, annotated text,
example sentences in linguistic papers, and audio)
One insight that I gleaned from these presentations was a better understanding
of glossed interlinear text. Interlinear text is not a type of data, but rather
just one possible way of displaying an annotated text. The annotations on a
text can consist of many types of information: alternate transcriptions,
morpheme glosses, word glosses, free translations, syntactic structure (and
possibly several alternative tree structures for the same text), discourse
structure, audio and video recordings, footnotes and commentary on various
issues, etc. What ties them all together is a "timeline" that proceeds from the
beginning to the end of a text, to which different types of information are
anchored. Aligned interlinear glosses are one way of displaying some of this
information, but not the only way, and not even the most appropriate way for
some types of information. The traditional arrangement of Talmudic material,
for example, with the core text in the center of the page and commentary around
the edges, is another possible display of annotated text, in which the
annotations are associated more with whole sentences and paragraphs than with
individual morphemes. There are also some sophisticated examples available for
presenting audio alongside interlinear text. (For example, check out the LACITO
archive [ http://lacito.archivage.vjf.cnrs.fr/ ]!)
Throughout, it was very clear that those at the conference had a great deal in
common with each other:
* a primary concern for descriptive (as distinguished from theoretical) linguistics
* a desire to make language materials available, to communities of speakers
and the general public as well as scholars
* an interest in taking advantage of the Internet, which provides a means
of publishing such materials that by-passes the limitations of
traditional publication (since the costs are so much lower, and thus
appropriate for materials that have smaller audiences)
* awareness that many materials may be less than fully-polished yet still
valuable to some people and worth archiving
* a sense of frustration with the currently confused state of the art in
data formats, especially fonts and character encoding, and the lack of
good information about how best to archive and publish on the web
* awareness of the large amount of data that is in data formats which will
be obsolete in a few years (and thus a willingness to accept data in
whatever form it is in, while also seeing a need for software to help
convert data to newer formats)
* a strong suspicion toward and distrust of rigid requirements, yet a
willingness to adopt standards voluntarily when their usefulness has been
demonstrated
Finally, the conference pointed out several trends that will be increasingly
important in future years.
* The speakers of lesser-known languages will be more actively involved the
production and use of materials in and about their languages, and their
concerns will increasingly have to be considered by scholars. These
include carefully documenting permissions and levels of access to
materials, making sure that language materials are available to the
communities themselves, and being careful that scholars do not
inadvertently aid commercial interests in exploiting native
knowledge-systems (such as medicinal use of plants) without appropriate
compensation.
* The boundary between publishing, libraries, and archiving is being
blurred by the shift to the digital world. Materials can be "archived" on
the web, which is a type of publication. Electronic "libraries" are
springing up in many places. Published and unpublished works from around
the world can be listed together in one common catalog. The same
technology is important in both spheres of activity. In short, these
activities are merging under a new umbrella that could be called
"scholarly information management". A corollary to this trend is that
archiving is not just something done at the end of a research project;
it's part of the ongoing process of managing the information that the
project produces.
* In such a world, and with huge numbers of resources available to sift
through, metadata becomes increasingly important. A freeform paragraph
description in a publications catalog is no longer good enough. It is the
metadata that users will consult in order to find materials of interest
to them, so the metadata must be carefully structured, accurate and
current. More and more, scholars will have to think not just about
producing materials but also about how to describe them so as to make
them accessible to others.
* Unicode [http://www.unicode.org/] is the way of the future for
representation of special characters in computers. The days of special
fonts for each language project are numbered. Instead, Unicode will make
possible a single set of fonts that meets virtually everyone's needs in
the same package. Over the next few years, most people will be switching
their computers over to using Unicode almost exclusively (that is, if
they want to take advantage of newer software).
* Language data will increasingly need to be structured carefully so that
not only can people view it and use it, but machines will be able to
understand and manipulate it in various ways. This will most likely be
done using XML (Extensible Markup Language) which is already
widely-supported in the computer industry, with more support becoming
available regularly.[2]
All in all, it was a workshop that was both stimulating and practical, one
which will have an unusual amount of influence in months and years to come.
Footnotes:
[1] Funding was provided by the Institute for Research in Cognitive Science
(IRCS) of the University of Pennsylvania, the International Standards in
Language Engineering Spoken Language Group (ISLE), and Talkbank.
[2] Since XML's development has been closely-associated with the World Wide
Web consortium [http://www.w3.org/XML/], it has been widely regarded as
the successor to HTML for web pages. However, this is just a small part
of its usefulness; it is a general-purpose system for representing the
structure of information in a document or database, which can be
customized for myriads of purposes. Many software tools are currently
available for creating and manipulating data in XML, with more being
created all the time. One, Extensible Stylesheet Language
Transformations [http://www.w3.org/TR/xslt], can do complex
restructuring of XML data.
(This report will be published in Notes on Linguistics,
http://www.sil.org/linguistics/NOL.htm)
---------------------------------------------------------------------------
LINGUIST List: Vol-12-327
More information about the LINGUIST
mailing list