Attention and stimulus-driven processes in auditory object perception

Project: Research project

Project Details

Description

DESCRIPTION (provided by applicant): In natural situations, the sound environment is dynamically and rapidly changing, with multiple sources overlapping in time and competing for attention. The ability to listen to your friend talking while walking down a noisy city street requires brain mechanisms that disentangle the sound mixture, separating your friend's voice from the sounds of the cars and other passing conversations. Auditory scene analysis, the ability to parse and organize the mixture of sound input, is a fundamental auditory process. Yet, the neural mechanisms that subserve perceptual sound organization are still poorly understood, and often rely on theories developed primarily to describe mechanisms of the visual system. Deficits in the ability to select relevant information when there are multiple competing sound sources is a common complaint in aging and in individuals with hearing loss, and can greatly hinder communication ability. There is no computer algorithm or prosthetic device that can mimic what the brain does when there is competing background noise. The overall goal of this proposal is to characterize how the auditory system adapts to dynamically changing multi-stream environments, allowing rapid and flexible shifting to different sound events in one's surroundings. Specific Aim 1 determines how ambiguous input is physiologically stored. Specific Aim 2 characterizes how attention modifies neural activity to support behavior in ambiguous situations. Specific Aim 3 identifies how neural representations of auditory input accommodate to changing multi-stream environments. The aims will be accomplished by obtaining behavioral and multiple electrophysiological indices of sound organization when the input is perceptually ambiguous, thus providing a novel model in normal hearing adults for characterizing how the brain maintains stable sound events in noisy environments. A key strength of the current project is the ability to neurophysiologically assess sound organization in auditory cortex for both attended and unattended sounds. The results of the proposed experiments will elucidate how dynamically changing environments are maintained by brain systems; characterizing how automatic and attentive mechanisms of scene analysis interact in the perception of one among many streams. This will fill a profound gap in our understanding of the neural mechanisms contributing to the perception of stable auditory events in complex and dynamically changing sound environments.
StatusFinished
Effective start/end date4/15/018/31/18

Funding

  • National Institute on Deafness and Other Communication Disorders: $344,684.00
  • National Institute on Deafness and Other Communication Disorders: $352,750.00
  • National Institute on Deafness and Other Communication Disorders: $277,638.00
  • National Institute on Deafness and Other Communication Disorders: $247,083.00
  • National Institute on Deafness and Other Communication Disorders: $351,326.00
  • National Institute on Deafness and Other Communication Disorders: $337,131.00
  • National Institute on Deafness and Other Communication Disorders: $333,653.00
  • National Institute on Deafness and Other Communication Disorders: $354,875.00
  • National Institute on Deafness and Other Communication Disorders: $343,485.00
  • National Institute on Deafness and Other Communication Disorders: $309,338.00
  • National Institute on Deafness and Other Communication Disorders: $348,165.00
  • National Institute on Deafness and Other Communication Disorders: $277,638.00
  • National Institute on Deafness and Other Communication Disorders: $348,165.00
  • National Institute on Deafness and Other Communication Disorders: $354,875.00
  • National Institute on Deafness and Other Communication Disorders: $277,638.00

Fingerprint

Explore the research topics touched on by this project. These labels are generated based on the underlying awards/grants. Together they form a unique fingerprint.