Temporal relationship between auditory and visual prosodic cues

Erin Cvejic, Jeesun Kim, Chris Wayne Davis

Research output: Contribution to conferencePaperpeer-review

3 Citations (Scopus)

Abstract

It has been reported that non-articulatory visual cues to prosody tend to align with auditory cues, emphasizing auditory events that are in close alignment (visual alignment hypothesis). We investigated the temporal relationship between visual and auditory prosodic cues in a large corpus of utterances to determine the extent to which non-articulatory visual prosodic cues align with auditory ones. Six speakers saying 30 sentences in three prosodic conditions (×2 repetitions) were recorded in a dialogue exchange task, to measure how often eyebrow movements and rigid head tilts aligned with auditory prosodic cues, the temporal distribution of such movements, and the variation across prosodic conditions. The timing of brow raises and head tilts were not aligned with auditory cues, and the occurrence of visual cues was inconsistent, lending little support for the visual alignment hypothesis. Different types of visual cues may combine with auditory cues in different ways to signal prosody.

Original languageEnglish
Pages981-984
Number of pages4
Publication statusPublished - 2011
Event12th Annual Conference of the International Speech Communication Association, INTERSPEECH 2011 -
Duration: 27 Aug 2011 → …

Conference

Conference12th Annual Conference of the International Speech Communication Association, INTERSPEECH 2011
Period27/08/11 → …

Fingerprint

Dive into the research topics of 'Temporal relationship between auditory and visual prosodic cues'. Together they form a unique fingerprint.

Cite this