Rethinking multimodal integration: The importance of event unity
Abstract
Our perceptual system continuously processes information from multiple modalities. Some of these derive from common sources – such hearing a speaker’s voice while watching their lip movements. Others arrive independently – for example watching the same speaker’s lips when a passing train masks their voice. Deciding which sights and sounds “go together” is essential in reconstructing an accurate internal representation of our environment. The challenges of this process are illustrated vividly whenever clever ventriloquists trick audiences into “hearing” puppets speak.
This “ventriloquist effect” is explored in laboratories through tasks involving localization of disembodied tone beeps and light flashes. Divorced from real world referents, these stimuli provide context-free tools thought to illuminate generalized principals of perceptual processing. A large body of research suggests a pattern of “optimal integration” in which the highest quality information is weighted most strongly, accurately describing both audio-visual (Alais & Burr, 2004) and visual-haptic tasks (Ernst & Banks 2002). However, the degree to which these models generalize beyond artificial stimuli is a matter of some debate.
Here I will discuss new findings from my group arguing for a broader understanding of audio-visual integration using naturalistic stimuli. Broadly construed, this approach explores trade-offs between three parameters. The first two, coincidence in space and time, are well understood by virtue of previous work. However the third– event congruity – also plays a crucial albeit overlooked role. Although more difficult to quantify than spatial agreement or temporal alignment, event congruity plays an important role in the audio-visual binding of speech (Vatakis & Spence, 2007) impact events (Vatakis & Spence, 2008), and musical notes (Chuen & Schutz, in press; Schutz & Kubovy, 2009). Together this research raises intriguing questions about the best practices for exploring the perceptual system’s ability to integrate realistic, time-varying stimuli more complex than flashes and beeps.
Published
How to Cite
Issue
Section
License
Copyright on articles is held by the author(s). The corresponding author has the right to grant on behalf of all authors and does grant on behalf of all authors, a worldwide exclusive licence (or non-exclusive license for government employees) to the Publishers and its licensees in perpetuity, in all forms, formats and media (whether known now or created in the future)
i) to publish, reproduce, distribute, display and store the Contribution;
ii) to translate the Contribution into other languages, create adaptations, reprints, include within collections and create summaries, extracts and/or, abstracts of the Contribution;
iii) to exploit all subsidiary rights in the Contribution,
iv) to provide the inclusion of electronic links from the Contribution to third party material where-ever it may be located;
v) to licence any third party to do any or all of the above.