Audio-visual Influence on Speech Perception PDF Download

Are you looking for read ebook online? Search for your book and save it on your Kindle device, PC, phones or tablets. Download Audio-visual Influence on Speech Perception PDF full book. Access full book title Audio-visual Influence on Speech Perception by Lena Quinto. Download full books in PDF and EPUB format.

Audio-visual Influence on Speech Perception

Audio-visual Influence on Speech Perception PDF Author: Lena Quinto
Publisher:
ISBN: 9780494403983
Category :
Languages : en
Pages : 60

Book Description
The importance of visual cues in speech perception is illustrated by the McGurk effect, whereby incongruent visual cues affect the perception speech sounds. It is unclear whether similar effects occur for sung materials. In Experiment 1, participants heard sequences of syllables (la-la-la-ba, la-la-la-ga) that were spoken or sung. Sung stimuli were ascending triads (do-mi-so) that returned to the tonic (do). Incongruent stimuli were created by combining an auditory /ba/ with a visual /ga/. Participants reported the final syllable. Results revealed overwhelming auditory dominance for spoken and for sung conditions. In Experiment 2, background noise was added to increase attention to visual cues. Auditory dominance prevailed in quiet but visual dominance prevailed in noise. In Experiment 3 the target syllable was isolated. As before, participants exhibited auditory dominance, but they had greater difficulty detecting sung syllables than spoken syllables that were presented in isolation. The contributions of visual and auditory cues from the preceding context are discussed.

Audio-visual Influence on Speech Perception

Audio-visual Influence on Speech Perception PDF Author: Lena Quinto
Publisher:
ISBN: 9780494403983
Category :
Languages : en
Pages : 60

Book Description
The importance of visual cues in speech perception is illustrated by the McGurk effect, whereby incongruent visual cues affect the perception speech sounds. It is unclear whether similar effects occur for sung materials. In Experiment 1, participants heard sequences of syllables (la-la-la-ba, la-la-la-ga) that were spoken or sung. Sung stimuli were ascending triads (do-mi-so) that returned to the tonic (do). Incongruent stimuli were created by combining an auditory /ba/ with a visual /ga/. Participants reported the final syllable. Results revealed overwhelming auditory dominance for spoken and for sung conditions. In Experiment 2, background noise was added to increase attention to visual cues. Auditory dominance prevailed in quiet but visual dominance prevailed in noise. In Experiment 3 the target syllable was isolated. As before, participants exhibited auditory dominance, but they had greater difficulty detecting sung syllables than spoken syllables that were presented in isolation. The contributions of visual and auditory cues from the preceding context are discussed.

Hearing Eye II

Hearing Eye II PDF Author: Douglas Burnham
Publisher: Psychology Press
ISBN: 1135471959
Category : Psychology
Languages : en
Pages : 338

Book Description
This volume outlines some of the developments in practical and theoretical research into speechreading lipreading that have taken place since the publication of the original "Hearing by Eye". It comprises 15 chapters by international researchers in psychology, psycholinguistics, experimental and clinical speech science, and computer engineering. It answers theoretical questions what are the mechanisms by which heard and seen speech combine? and practical ones what makes a good speechreader? Can machines be programmed to recognize seen and seen-and-heard speech?. The book is written in a non-technical way and starts to articulate a behaviourally-based but cross-disciplinary programme of research in understanding how natural language can be delivered by different modalities.

Audiovisual Speech Processing

Audiovisual Speech Processing PDF Author: Gérard Bailly
Publisher: Cambridge University Press
ISBN: 1107006821
Category : Computers
Languages : en
Pages : 507

Book Description
This book presents a complete overview of all aspects of audiovisual speech including perception, production, brain processing and technology.

The Effect on Audiovisual Speech Perception of Auditory and Visual Source Separation

The Effect on Audiovisual Speech Perception of Auditory and Visual Source Separation PDF Author: Stuart Matthew Leech
Publisher:
ISBN:
Category :
Languages : en
Pages : 0

Book Description


Toward a Unified Theory of Audiovisual Integration in Speech Perception

Toward a Unified Theory of Audiovisual Integration in Speech Perception PDF Author: Nicholas Altieri
Publisher: Universal-Publishers
ISBN: 1599423618
Category :
Languages : en
Pages :

Book Description
Auditory and visual speech recognition unfolds in real time and occurs effortlessly for normal hearing listeners. However, model theoretic descriptions of the systems level cognitive processes responsible for integrating auditory and visual speech information are currently lacking, primarily because they rely too heavily on accuracy rather than reaction time predictions. Speech and language researchers have argued about whether audiovisual integration occurs in a parallel or in coactive fashion, and also the extent to which audiovisual occurs in an efficient manner. The Double Factorial Paradigm introduced in Section 1 is an experimental paradigm that is equipped to address dynamical processing issues related to architecture (parallel vs. coactive processing) as well as efficiency (capacity). Experiment 1 employed a simple word discrimination task to assess both architecture and capacity in high accuracy settings. Experiments 2 and 3 assessed these same issues using auditory and visual distractors in Divided Attention and Focused Attention tasks respectively. Experiment 4 investigated audiovisual integration efficiency across different auditory signal-to-noise ratios. The results can be summarized as follows: Integration typically occurs in parallel with an efficient stopping rule, integration occurs automatically in both focused and divided attention versions of the task, and audiovisual integration is only efficient (in the time domain) when the clarity of the auditory signal is relatively poor--although considerable individual differences were observed. In Section 3, these results were captured within the milieu of parallel linear dynamic processing models with cross channel interactions. Finally, in Section 4, I discussed broader implications for this research, including applications for clinical research and neural-biological models of audiovisual convergence.

Hearing by Eye II

Hearing by Eye II PDF Author: Ruth Campbell
Publisher: Psychology Press
ISBN: 9780863775024
Category : Computers
Languages : en
Pages : 338

Book Description
This volume outlines developments in practical and theoretical research into speechreading lipreading.

The Impact of Auditory-visual Speech Perception on Working Memory

The Impact of Auditory-visual Speech Perception on Working Memory PDF Author: Jana B. Frtusova
Publisher:
ISBN:
Category :
Languages : en
Pages : 0

Book Description
Adding visual speech information (i.e. lip movements) to auditory speech information (i.e. voice) can enhance speech comprehension in younger and older adults while at the same time it reduces electrical brain responses, as measured by event-related potentials (ERPs). Thus, the brain seems to allocate fewer resources to speech comprehension when audio-visual (AV) speech information is available. This study examined whether the brain resources saved at the perceptual level during AV presentation allow younger and older adults to perform better on a working memory task, and whether older adults benefit to the same extent as younger adults. Twenty older adults and 23 younger adults completed an n-back working memory task (0-, 1-,2-, 3-back) under visual-only (V-only), auditory-only (A-only), and AV condition while ERPs were recorded. The results showed a decrease in reaction time across all memory loads and an improvement in accuracy for 2back and 3-back during AV compared to the V-only and A-only conditions. In addition, ERP analysis from a sample of 12 younger and 12 older adults showed a smaller N1 amplitude for the older group during AV compared to A-only presentation. The attenuation of N1, however, did not correlate with behavioural data. Nor did it show a relationship with changes either in the latency or the amplitude of P3, an ERP that reflects working memory processes. Thus, despite clear behavioural improvements on the working memory task during AV speech presentation, a more direct relationship between facilitation of sensory processing and working memory improvement was not identified.

Communication Acoustics

Communication Acoustics PDF Author: Jens Blauert
Publisher: Springer Science & Business Media
ISBN: 9783540221623
Category : Computers
Languages : en
Pages : 404

Book Description
- Speech Generation: Acoustics, Models and Applications (Arild Lacroix). - The Evolution of Digital Audio Technology (John Mourjopoulos). - Audio-Visual Interaction (Armin Kohlrausch) . - Speech and Audio Coding (Ulrich Heute) . - Binaural Technique (Dorte Hammerhoei, Henrik Moeller). - Auditory Virtual Environment (Pedro Novo). - Evolutionary Adaptions for Auditory Communication (Georg Klump). - A Functional View on the Human Hearing Organ (Herbert Hudde). - Modeling of Binaural Hearing (Jonas Braasch). - Psychoacoustics and Sound Quality (Hugo Fastl). - Semiotics for Engineers (Ute Jekosch). - Quality of Transmitted Speech for Humans and Machines (Sebastian Möller).

Audiovisual Speech Processing

Audiovisual Speech Processing PDF Author: Gérard Bailly
Publisher: Cambridge University Press
ISBN: 110737815X
Category : Language Arts & Disciplines
Languages : en
Pages : 507

Book Description
When we speak, we configure the vocal tract which shapes the visible motions of the face and the patterning of the audible speech acoustics. Similarly, we use these visible and audible behaviors to perceive speech. This book showcases a broad range of research investigating how these two types of signals are used in spoken communication, how they interact, and how they can be used to enhance the realistic synthesis and recognition of audible and visible speech. The volume begins by addressing two important questions about human audiovisual performance: how auditory and visual signals combine to access the mental lexicon and where in the brain this and related processes take place. It then turns to the production and perception of multimodal speech and how structures are coordinated within and across the two modalities. Finally, the book presents overviews and recent developments in machine-based speech recognition and synthesis of AV speech.

How Does Feedback Impact Training in Audio-visual Speech Perception?

How Does Feedback Impact Training in Audio-visual Speech Perception? PDF Author: Amy Ranta
Publisher:
ISBN:
Category :
Languages : en
Pages : 29

Book Description
Abstract: Integration of visual and auditory speech cues is a process used by listeners in compromised listening situations, as well as in normal environments, as exemplified by the McGurk effect (McGurk and McDonald, 1976). Audio-visual integration of speech appears to be a skill independent of the ability to process auditory or visual speech cues alone. Grant and Seitz (1998) argued for independence of this process based on their findings that integration abilities could not be predicted from auditory-only or visual-only performance. Gariety (2009) and James (2009) further supported this argument by training listeners in the auditory-only modality with degraded speech syllables, then testing those listeners in the auditory-only, visual-only, and audio-visual conditions. Their results showed an increase in auditory-only performance, but no improvement in integration. Recently, DiStefano (2010) conducted a training study in which listeners were trained in the audio-visual modality with degraded speech syllables. Results showed that the performance increased only for the audio-visual conditions, and did not increase in the auditory-only or visual-only conditions. Interestingly, performance improved only for stimulus pairs that were "congruent" (i.e. the auditory and visual inputs were the same syllable) and did not increase for "discrepant" stimuli (i.e. the auditory and visual inputs were different syllables). It is possible that the feedback provided in DiStefano's study impacted their pattern of results. However, the question remains as to whether integration of discrepant stimuli can be trained. In the present study, five listeners received ten hours of training sessions in the audio-visual condition with degraded speech signals similar to those used by Shannon et al. (1995). The feedback given during training was designed to encourage McGurk-type combination and fusion responses, in contrast to DiStefano's study, in which feedback was given to encourage responses that matched the auditory signal. A comparison of pre-training and post-training scores showed little to no improvement in auditory-only performance, and a slight decrease in visual-only performance for congruent stimuli. Further, a substantial increase in McGurk-type responses was seen from pre- to post-test for discrepant stimuli. These results provide further support that integration is an independent process, and that the feedback provided strongly influences response patterns. This strong lack of generalization from training also should be incorporated into designing effective integration training programs for aural rehabilitation.