17.65, Review: Computational Ling: Nass & Brave (2005)
LINGUIST List
linguist at LINGUISTLIST.ORG
Wed Jan 11 22:13:01 UTC 2006
LINGUIST List: Vol-17-65. Wed Jan 11 2006. ISSN: 1068 - 4875.
Subject: 17.65, Review: Computational Ling: Nass & Brave (2005)
Moderators: Anthony Aristar, Wayne State U <aristar at linguistlist.org>
Helen Aristar-Dry, Eastern Michigan U <hdry at linguistlist.org>
Reviews (reviews at linguistlist.org)
Sheila Dooley, U of Arizona
Terry Langendoen, U of Arizona
Homepage: http://linguistlist.org/
The LINGUIST List is funded by Eastern Michigan University, Wayne
State University, and donations from subscribers and publishers.
Editor for this issue: Lindsay Butler <lindsay at linguistlist.org>
================================================================
What follows is a review or discussion note contributed to our
Book Discussion Forum. We expect discussions to be informal and
interactive; and the author of the book discussed is cordially
invited to join in. If you are interested in leading a book
discussion, look for books announced on LINGUIST as "available
for review." Then contact Sheila Dooley at dooley at linguistlist.org.
===========================Directory==============================
1)
Date: 09-Jan-2006
From: Richard Sproat < rws at uiuc.edu >
Subject: Wired for Speech
-------------------------Message 1 ----------------------------------
Date: Wed, 11 Jan 2006 17:03:21
From: Richard Sproat < rws at uiuc.edu >
Subject: Wired for Speech
AUTHOR: Nass, Clifford; Brave, Scott
TITLE: Wired for Speech
SUBTITLE: How Voice Activates and Advances the Human-Computer
Relationship
PUBLISHER: MIT Press
YEAR: 2005
Announced at http://linguistlist.org/issues/16/16-2705.html
Richard Sproat, Departments of Linguistics and ECE, University of
Illinois at Urbana-Champaign
OVERVIEW
The topic of this book is voice user interfaces, an example of which is
the automated system that one interacts with when one calls United
Airlines and wishes to check on the arrival or departure time for a
flight. Other examples include systems where one speaks to a
graphical avatar (a ''talking head'') that serves as an automated
information kiosk; or the National Oceanic and Atmospheric
Administration's Weather Radio, which presents the weather using a
text-to-speech (TTS) synthesizer. In short, a voice user interface is
any automated system that allows a user to access information,
possibly with automatic speech recognition (ASR) for voice input, and
with either prerecorded prompts or TTS technology to produce output.
The book is not about the technology underlying voice user interfaces.
Rather, it is about how humans react to them and interact with them in
controlled experiments, and how this information should guide the
design of the ''persona'' that the interface presents to the world.
The book is divided into fourteen main chapters, and a two-page
summary chapter. After a brief stage-setting first chapter, the authors
turn in chapters 2-3 to their first topic, namely the gender of voices.
How do people react to synthetic male versus female voices? How
does gender stereotyping affect people's perception of the quality or
believability of a voice user interface system? Can one get around
user prejudices by having a ''gender neutral'' voice?
Chapters 4-5 turn to the issue of voice ''personality''. People infer
many aspects of other people's personality by the way they talk, and
the kinds of words they use. And, as it turns out, people
impute ''personalities'' to voice user interfaces. As with gender,
preconceived notions about personalities have a strong effect on the
user's perception of a voice user interface.
Chapter 6 deals with the issue of regional or foreign accents and
perceived ethnicity. Once again people's prejudices about accent and
race carry over to machines, even though the notion that a machine
has a geographical or ethnic background is obviously absurd.
Chapters 7-8 discuss emotion and how that should be expressed, or
not expressed, in voice user interfaces. One of the clear suggestions
of this section of the book is that, where possible, it is important for a
voice user interface to match its emotion to the (expected) emotional
state of the user.
Chapter 9 asks when and how a voice user interface should use
multiple voices. A couple of conclusions are drawn: first, if multiple
voices are used, they should be matched to the tasks being
performed. For example, the authors suggest using an officious
sounding voice to guide users through a complex menu system, and a
warm friendly sounding voice to reassure users that they are being
guided to the right place. Second, despite the common notion
of ''voice fonts'' (e.g. Raman 2004), users do not treat different voices
the same way as they treat different textual fonts since a change of
voice has social implications that a change of font does not.
Chapter 10 deals with the question of whether voice interfaces should
say ''I'', and thereby make perceived claims to being human. From the
authors' experiments, it seems that systems that use synthetic (TTS)
voices should not say ''I''.
Chapter 11 deals with recorded speech versus TTS, and real faces
versus synthetic faces, and concludes that people react better to a
system that has either a synthetic face speaking with a synthetic
voice, or a real face speaking with a real voice; users do not like it
when the conditions are crossed.
Chapter 12 argues that it is generally bad to mix obviously recorded
speech with obviously synthetic speech: for example, it would be a
bad design choice to have a system produce a canned phrase
like ''Good Morning Ms.'' using a prerecorded voice, and then finish
the utterance with an obviously synthetic voice saying the name of the
user. This, at least, is one section of the book that will seem obvious
to anyone who has worked on the technology of speech synthesis: we
have known for a long time that it is not a good idea to mix high quality
prerecorded speech with poor quality synthesis. The chapter also
contains a discussion of humor, though it is not obvious how this
relates to the main topic of the chapter.
The final two chapters, 13-14, shift the ground from voice (and video)
output to voice input. Chapter 13 deals with the issue of how
comfortably people will interact when they know, or are constantly
reminded that they are being recorded. A set of experiments with
various kinds of attached microphones versus unobtrusive array
microphones, showed that users who used the less obtrusive
microphones were more creative in their responses and more willing
to disclose sensitive information. Finally, chapter 14 discusses what
systems should do when they misrecognize a user: what are the
relative costs and benefits of the system accepting blame (''I'm sorry, I
did not understand you'') versus implicating blame on the user (''You
are speaking too quickly, please slow down.'').
Three main themes run through this book.
The first is simply this: we are ''wired'' for speech. Even though users
know they are dealing with an automated system, if the system takes
speech as input, or produces speech as output, users cannot help but
treat the system as if it were another human, and will apply the same
beliefs and prejudices to the automated system as they would to a
human that had the same behavior. That is, if a person feels more
comfortable with a male human explaining how to operate a complex
piece of equipment than with a female, then that prejudice will carry
over to an automated assistant that has a female voice. This first
point is consistent with previous work from Nass's lab: in general,
people seem to treat computers as if they were people, even though
they know full well that they are not (Nass, Steuer & Tauber, 1994).
The second theme is that there is no way around the first theme: for
instance, we cannot solve the problem of speakers' inherent gender
biases by making a system with a voice of ambiguous gender. Users
will just think that the system is weird and will react to it worse than if
the system clearly indicates that it is ''female'' or ''male''.
Finally, user perceptions of voice user interfaces have direct
implications for users' views of whatever service or product the system
is trying to sell. Just as a skilled salesman can make a product seem
more desirable than it might otherwise seem, so a well-designed voice
interface can make claims of a product's value seem more believable.
DETAILED CRITIQUE
To place the current research in some historical perspective it is worth
noting that Nass's research on ''Computers as Social Actors'' was the
inspiration for Microsoft ''Bob'' which, after its demise, led eventually
to ''Clippy'', the Microsoft Office automated assistant. Neither of these
products have been well received and there has been much
discussion of why (e.g., Schwartz, 2003), a topic that would take us
beyond the scope of this review. The authors are evidently very
proud of their long experience at providing user-interface design
advice to corporations: the preface to the book is highly self-laudatory,
and contains a fairly long list of consulting contracts that Nass's lab
has had with various companies over the years, including such varied
companies as BMW, Charles Schwab, General Magic, Macromedia,
NTT, Philips and US West.
In the overview above, reference was made to experiments conducted
by the authors to validate their claims about design issues for voice
user interfaces, and it is worth summarizing one of those experiments
just to give a flavor of the kind of research the authors performed. For
example, in assessing the importance of gender stereotyping, the
authors conducted an experiment, where participants were directed to
an online auction site which offered a set of stereotypically male and
stereotypically female merchandise, with descriptions from eBay.
Descriptions were read to the listeners either with a female voice or a
male voice generated with the Festival TTS system (Taylor, Black &
Caley, 1998). Subjects were then asked to rate how credible the
description they heard was. The results of this experiment (reported
on pages 25-27) were that speakers rated the product descriptions as
more credible if the gender of the voice matched the ''gender'' of the
product.
While the focus of the book is on the use of technology, rather than
the technology itself, one cannot forget that any use of a technology
presumes some understanding of the technology that is being used.
>From the technological perspective there are a couple of points of
interest about this book. First, I personally found it noteworthy that
the majority of the discussion focusses on synthesis rather than
recognition. This focus is the opposite of the focus in the speech
technology community, where synthesis has long taken a back seat to
recognition, and where synthesis has traditionally been regarded as
much easier than recognition. But the focus of the current book on
synthesis is, after all, natural: although speech recognition is an
important part of many voice user interfaces, it is the voice with which
the system speaks that gives it its ''personality'' and its apparent
human-like qualities.
Second, it is unfortunately the case that the authors do not always
seem to understand the technology that they are evaluating. On
several occasions they imply that changing voices, changing the
emotions of voices, and changing the gender of voices is a
straightforward process. This is misleading at a number of levels.
First, consider emotion. While Nass and Brave are correct that many
of the acoustic correlates of emotion are known, and while it is true
that rendering emotion in synthetic speech has been a research topic
since Cahn's work (Cahn, 1989), it is still not possible to produce
convincing renditions of all emotions. Second, while it might seem as if
it should be easy in general to change the voice or the gender used
by the system, in practice there are limitations. To understand this, it is
necessary to briefly remind the reader of the various methods used to
produce speech output in TTS systems. The oldest approach,
exemplified by the Klatt synthesizer (Klatt 1980) and its commercial
offspring DECTalk, is a fully parametric system where all parameters
of the voice, including pitch, formant values, spectral tilt, and many
others, are controllable. In such a system it is indeed in principle easy
to produce new voices --- but at a cost: the quality of the resulting
speech sounds distinctly mechanical, largely because we do not have
good models of how to control the parameters over time. Such
limitations in our understanding have been sidestepped in much of the
recent work on ''unit selection'' based methods. These methods,
pioneered in work on the CHATR system (Hunt & Black, 1996), and
exemplified in commercial systems such as AT&T's ''Natural Voices'',
depends upon a huge database of speech from one speaker. During
synthesis, a set of units as closely as possible matching the intended
utterance is selected on the fly from the database. The resulting
speech can sound very good in the best case --- and downright silly in
the worst. But one of the practical discoveries of this work is that the
less one fiddles with the speech, the better and more natural the
resulting synthesis sounds. This means that modification of speech
such as changing the pitch is to be eschewed. The result: if you want
a different voice, you have to record a different speaker, and analyze
their speech. If you want a different emotion, you have to record your
speaker performing speech with that emotion. This is certainly a lot
easier to do than it used to be, but at a minimum one is looking at
recording an hour's worth of a speech. This clearly involves more than
turning a few knobs, which is all that Nass and Brave seem to imply is
needed.
Turning away from technological issues, there are problems with the
design of the book itself. As the authors state at the outset, they use
endnotes extensively for background information that can be freely
skipped by the casual reader. For example, the data and statistical
analyses of all the experiments are presented in endnotes, not in the
body of the text, which merely summarizes the results. Also,
bibliographic references are all given in the endnotes. This design
choice has both a good and a bad aspect. It surely helps the non-
specialist reader, who will not necessarily be inclined to look at the
authors' data in detail, but will be satisfied with the authors' synopsis
of the results. But it is annoying for someone who has a technical
background in the field, since following up any single point
necessitates thumbing to the back of the book. The lack of a standard
bibliography is also an extremely bothersome feature of the book.
Negatives aside, this is a book worth reading by anyone interested in
speech technology. Those of us who have worked on developing the
technology underlying voice user interfaces have traditionally not
thought much about the actual design of the end product. Nass and
Brave have clearly thought about these issues more than anyone
else.
Still, while it is useful to understand what features of speech work best
for which applications, we should not lose sight of the fact that the
underlying technology is itself immature, and that just building a
system that can communicate effectively with inexperienced users is
still a challenge. In the ''Restaurant at the End of the Universe'', the
second book in Douglas Adams' ''Hitchhiker'' series, Ford Prefect (the
Betelgeusian companion of the hero, Arthur Dent) berates the
Golgafrincham colonizers of prehistoric Earth for not having made
much progress on the invention of the wheel. A marketing consultant
fires back at Ford and asks him, if he is so smart, what color it should
be.
REFERENCES
Cahn, J. 1989. ''Generating Expression in Synthesized Speech.''
Master's thesis. Massachusetts Institute of Technology.
Hunt, A. and Black, A. 1996. ''Unit selection in a concatenative speech
synthesis system using a large speech database.'' Proceedings of
ICASSP 96, vol 1, pp 373-376, Atlanta, Georgia.
Klatt, D. 1980. ''Software for a cascade/parallel formant synthesizer'',
Journal of the Acoustical Society of America, 67.3, 971-995.
Nass, C., Steuer, J. S., & Tauber, E. (1994). ''Computers are social
actors.'' Proceeding of the CHI Conference, 72-77. Boston, MA
Raman, TV. 2004. ''Emacsspeak -- The Complete Audio Desktop''.
http://emacspeak.sourceforge.net/
Schwartz, L. 2003. ''Why people hate the paperclip: Labels,
appearance, behavior and social responses to user interface agents.''
Master's Thesis, Stanford University.
Taylor, P., Black, A. and Caley, R. 1998 ''The architecture of the
Festival Speech Synthesis System'' 3rd ESCA Workshop on Speech
Synthesis, pp. 147-151, Jenolan Caves, Australia.
ABOUT THE REVIEWER
Richard Sproat is professor in the departments of Linguistics and
Electrical & Computer Engineering at the University of Illinois at
Urbana-Champaign. His interests include multilingual text processing
and speech technology. Prior to coming to the University of Illinois,
Sproat worked in industrial research at AT&T Bell Laboratories, with
his primary area of research being text-to-speech synthesis. Sproat
was one of the main architects of the Bell Labs multilingual text-to-
speech synthesizer. He was also involved in the design of the SABLE
text-to-speech markup language, a precursor to the W3C's SSML.
-----------------------------------------------------------
LINGUIST List: Vol-17-65
More information about the LINGUIST
mailing list