19.1514, Diss: Comp Ling: Eisenstein: 'Gesture in Automatic Discourse Proces...'
LINGUIST Network
linguist at LINGUISTLIST.ORG
Thu May 8 13:05:51 UTC 2008
LINGUIST List: Vol-19-1514. Thu May 08 2008. ISSN: 1068 - 4875.
Subject: 19.1514, Diss: Comp Ling: Eisenstein: 'Gesture in Automatic Discourse Proces...'
Moderators: Anthony Aristar, Eastern Michigan U <aristar at linguistlist.org>
Helen Aristar-Dry, Eastern Michigan U <hdry at linguistlist.org>
Reviews: Randall Eggert, U of Utah
<reviews at linguistlist.org>
Homepage: http://linguistlist.org/
The LINGUIST List is funded by Eastern Michigan University,
and donations from subscribers and publishers.
Editor for this issue: Evelyn Richter <evelyn at linguistlist.org>
================================================================
To post to LINGUIST, use our convenient web form at
http://linguistlist.org/LL/posttolinguist.html.
===========================Directory==============================
1)
Date: 07-May-2008
From: Jacob Eisenstein < jacobe at csail.mit.edu >
Subject: Gesture in Automatic Discourse Processing
-------------------------Message 1 ----------------------------------
Date: Thu, 08 May 2008 09:03:59
From: Jacob Eisenstein [jacobe at csail.mit.edu]
Subject: Gesture in Automatic Discourse Processing
E-mail this message to a friend:
http://linguistlist.org/issues/emailmessage/verification.cfm?iss=19-1514.html&submissionid=177682&topicid=14&msgnumber=1
Institution: Massachusetts Institute of Technology
Program: Computer Science and Artificial Intelligence Laboratory
Dissertation Status: Completed
Degree Date: 2008
Author: Jacob Eisenstein
Dissertation Title: Gesture in Automatic Discourse Processing
Dissertation URL: http://people.csail.mit.edu/jacobe/diss.html
Linguistic Field(s): Computational Linguistics
Dissertation Director(s):
Randall Davis
Regina Barzilay
Dissertation Abstract:
Computers cannot fully understand spoken language without access to the
wide range of modalities that accompany speech. This thesis addresses the
particularly expressive modality of hand gesture, and focuses on building
structured statistical models at the intersection of speech, vision, and
meaning.
My approach is distinguished in two key respects. First, gestural patterns
are leveraged to discover parallel structures in the meaning of the
associated speech. This differs from prior work that attempted to interpret
individual gestures directly, an approach that was prone to a lack of
generality across speakers. Second, I present novel, structured statistical
models for multimodal language processing, which enable learning about
gesture in its linguistic context, rather than in the abstract.
These ideas find successful application in a variety of language processing
tasks: resolving ambiguous noun phrases, segmenting speech into topics, and
producing keyframe summaries of spoken language. In all three cases, the
addition of gestural features -- extracted automatically from video --
yields significantly improved performance over a state-of-the-art text-only
alternative. This marks the first demonstration that hand gesture improves
automatic discourse processing.
-----------------------------------------------------------
LINGUIST List: Vol-19-1514
More information about the LINGUIST
mailing list