Audio-visual Influence on Speech Perception

Audio-visual Influence on Speech Perception

Author: Lena Quinto

Publisher:

Published: 2007

Total Pages: 60

ISBN-13: 9780494403983

DOWNLOAD EBOOK

The importance of visual cues in speech perception is illustrated by the McGurk effect, whereby incongruent visual cues affect the perception speech sounds. It is unclear whether similar effects occur for sung materials. In Experiment 1, participants heard sequences of syllables (la-la-la-ba, la-la-la-ga) that were spoken or sung. Sung stimuli were ascending triads (do-mi-so) that returned to the tonic (do). Incongruent stimuli were created by combining an auditory /ba/ with a visual /ga/. Participants reported the final syllable. Results revealed overwhelming auditory dominance for spoken and for sung conditions. In Experiment 2, background noise was added to increase attention to visual cues. Auditory dominance prevailed in quiet but visual dominance prevailed in noise. In Experiment 3 the target syllable was isolated. As before, participants exhibited auditory dominance, but they had greater difficulty detecting sung syllables than spoken syllables that were presented in isolation. The contributions of visual and auditory cues from the preceding context are discussed.


Hearing Eye II

Hearing Eye II

Author: Douglas Burnham

Publisher: Psychology Press

Published: 2013-10-28

Total Pages: 338

ISBN-13: 1135471959

DOWNLOAD EBOOK

This volume outlines some of the developments in practical and theoretical research into speechreading lipreading that have taken place since the publication of the original "Hearing by Eye". It comprises 15 chapters by international researchers in psychology, psycholinguistics, experimental and clinical speech science, and computer engineering. It answers theoretical questions what are the mechanisms by which heard and seen speech combine? and practical ones what makes a good speechreader? Can machines be programmed to recognize seen and seen-and-heard speech?. The book is written in a non-technical way and starts to articulate a behaviourally-based but cross-disciplinary programme of research in understanding how natural language can be delivered by different modalities.


Audiovisual Speech Processing

Audiovisual Speech Processing

Author: Gérard Bailly

Publisher: Cambridge University Press

Published: 2012-04-26

Total Pages: 507

ISBN-13: 1107006821

DOWNLOAD EBOOK

This book presents a complete overview of all aspects of audiovisual speech including perception, production, brain processing and technology.


Toward a Unified Theory of Audiovisual Integration in Speech Perception

Toward a Unified Theory of Audiovisual Integration in Speech Perception

Author: Nicholas Altieri

Publisher: Universal-Publishers

Published: 2010-09-09

Total Pages:

ISBN-13: 1599423618

DOWNLOAD EBOOK

Auditory and visual speech recognition unfolds in real time and occurs effortlessly for normal hearing listeners. However, model theoretic descriptions of the systems level cognitive processes responsible for integrating auditory and visual speech information are currently lacking, primarily because they rely too heavily on accuracy rather than reaction time predictions. Speech and language researchers have argued about whether audiovisual integration occurs in a parallel or in coactive fashion, and also the extent to which audiovisual occurs in an efficient manner. The Double Factorial Paradigm introduced in Section 1 is an experimental paradigm that is equipped to address dynamical processing issues related to architecture (parallel vs. coactive processing) as well as efficiency (capacity). Experiment 1 employed a simple word discrimination task to assess both architecture and capacity in high accuracy settings. Experiments 2 and 3 assessed these same issues using auditory and visual distractors in Divided Attention and Focused Attention tasks respectively. Experiment 4 investigated audiovisual integration efficiency across different auditory signal-to-noise ratios. The results can be summarized as follows: Integration typically occurs in parallel with an efficient stopping rule, integration occurs automatically in both focused and divided attention versions of the task, and audiovisual integration is only efficient (in the time domain) when the clarity of the auditory signal is relatively poor--although considerable individual differences were observed. In Section 3, these results were captured within the milieu of parallel linear dynamic processing models with cross channel interactions. Finally, in Section 4, I discussed broader implications for this research, including applications for clinical research and neural-biological models of audiovisual convergence.


Hearing by Eye II

Hearing by Eye II

Author: Ruth Campbell

Publisher: Psychology Press

Published: 1998

Total Pages: 338

ISBN-13: 9780863775024

DOWNLOAD EBOOK

This volume outlines developments in practical and theoretical research into speechreading lipreading.


Communication Acoustics

Communication Acoustics

Author: Jens Blauert

Publisher: Springer Science & Business Media

Published: 2005-05-20

Total Pages: 404

ISBN-13: 9783540221623

DOWNLOAD EBOOK

- Speech Generation: Acoustics, Models and Applications (Arild Lacroix). - The Evolution of Digital Audio Technology (John Mourjopoulos). - Audio-Visual Interaction (Armin Kohlrausch) . - Speech and Audio Coding (Ulrich Heute) . - Binaural Technique (Dorte Hammerhoei, Henrik Moeller). - Auditory Virtual Environment (Pedro Novo). - Evolutionary Adaptions for Auditory Communication (Georg Klump). - A Functional View on the Human Hearing Organ (Herbert Hudde). - Modeling of Binaural Hearing (Jonas Braasch). - Psychoacoustics and Sound Quality (Hugo Fastl). - Semiotics for Engineers (Ute Jekosch). - Quality of Transmitted Speech for Humans and Machines (Sebastian Möller).


The Impact of Auditory-visual Speech Perception on Working Memory

The Impact of Auditory-visual Speech Perception on Working Memory

Author: Jana B. Frtusova

Publisher:

Published: 2010

Total Pages: 0

ISBN-13:

DOWNLOAD EBOOK

Adding visual speech information (i.e. lip movements) to auditory speech information (i.e. voice) can enhance speech comprehension in younger and older adults while at the same time it reduces electrical brain responses, as measured by event-related potentials (ERPs). Thus, the brain seems to allocate fewer resources to speech comprehension when audio-visual (AV) speech information is available. This study examined whether the brain resources saved at the perceptual level during AV presentation allow younger and older adults to perform better on a working memory task, and whether older adults benefit to the same extent as younger adults. Twenty older adults and 23 younger adults completed an n-back working memory task (0-, 1-,2-, 3-back) under visual-only (V-only), auditory-only (A-only), and AV condition while ERPs were recorded. The results showed a decrease in reaction time across all memory loads and an improvement in accuracy for 2back and 3-back during AV compared to the V-only and A-only conditions. In addition, ERP analysis from a sample of 12 younger and 12 older adults showed a smaller N1 amplitude for the older group during AV compared to A-only presentation. The attenuation of N1, however, did not correlate with behavioural data. Nor did it show a relationship with changes either in the latency or the amplitude of P3, an ERP that reflects working memory processes. Thus, despite clear behavioural improvements on the working memory task during AV speech presentation, a more direct relationship between facilitation of sensory processing and working memory improvement was not identified.


The Oxford Handbook of Psycholinguistics

The Oxford Handbook of Psycholinguistics

Author: M. Gareth Gaskell

Publisher: Oxford University Press, USA

Published: 2007

Total Pages: 880

ISBN-13: 0198568975

DOWNLOAD EBOOK

The ability to communicate through spoken and written language is one of the defining characteristics of the human race, yet it remains a deeply mysterious process. The young science of psycholinguistics attempts to uncover the mechanisms and representations underlying human language. This interdisciplinary field has seen massive developments over the past decade, with a broad expansion of the research base, and the incorporation of new experimental techniques such as brain imaging and computational modelling. The result is that real progress is being made in the understanding of the key components of language in the mind. The Oxford Handbook of Psycholinguistics brings together the views of 75 leading researchers in psycholinguistics to provide a comprehensive and authoritative review of the current state of the art in psycholinguistics. With almost 50 chapters written by experts in the field, the range and depth of coverage is unequalled. The contributors are eminent in a wide range of fields, including psychology, linguistics, human memory, cognitive neuroscience, bilingualism, genetics, development and neuropsychology. Their contributions are organised into six themed sections, covering word recognition, the mental lexicon, comprehension and discourse, language production, language development, and perspectives on psycholinguistics. The breadth of coverage, coupled with the accessibility of the short chapter format should make the handbook essential reading for both students and researchers in the fields of psychology, linguistics and neuroscience.


Audiovisual Speech Processing

Audiovisual Speech Processing

Author: Gérard Bailly

Publisher: Cambridge University Press

Published: 2012-04-26

Total Pages: 507

ISBN-13: 110737815X

DOWNLOAD EBOOK

When we speak, we configure the vocal tract which shapes the visible motions of the face and the patterning of the audible speech acoustics. Similarly, we use these visible and audible behaviors to perceive speech. This book showcases a broad range of research investigating how these two types of signals are used in spoken communication, how they interact, and how they can be used to enhance the realistic synthesis and recognition of audible and visible speech. The volume begins by addressing two important questions about human audiovisual performance: how auditory and visual signals combine to access the mental lexicon and where in the brain this and related processes take place. It then turns to the production and perception of multimodal speech and how structures are coordinated within and across the two modalities. Finally, the book presents overviews and recent developments in machine-based speech recognition and synthesis of AV speech.