Human communication and language skills rely heavily on the ability to detect and process auditory inputs. This paper reviews possible applications of the event-related potential (ERP) technique to the study of cortical mechanisms supporting human auditory processing, including speech stimuli. Following a brief introduction to the ERP methodology, the remaining sections focus on demonstrating how ERPs can be used in humans to address research questions related to cortical organization, maturation and plasticity, as well as the effects of sensory deprivation, and multisensory interactions. The review is intended to serve as a primer for researchers interested in using ERPs for the study of the human auditory system.
Citation: Alexandra P. Key. Human Auditory Processing: Insights from Cortical Event-related Potentials[J]. AIMS Neuroscience, 2016, 3(2): 141-162. doi: 10.3934/Neuroscience.2016.2.141
[1] | Shu Hui Yau, Fabrice Bardy, Paul F Sowman, Jon Brock . The Magnetic Acoustic Change Complex and Mismatch Field: A Comparison of Neurophysiological Measures of Auditory Discrimination. AIMS Neuroscience, 2017, 4(1): 14-27. doi: 10.3934/Neuroscience.2017.1.14 |
[2] | Mercedes Fernandez, Juliana Acosta, Kevin Douglass, Nikita Doshi, Jaime L. Tartar . Speaking Two Languages Enhances an Auditory but Not a Visual Neural Marker of Cognitive Inhibition. AIMS Neuroscience, 2014, 1(2): 145-157. doi: 10.3934/Neuroscience.2014.2.145 |
[3] | Byron Bernal, Alfredo Ardila . From Hearing Sounds to Recognizing Phonemes: Primary Auditory Cortex is A Truly Perceptual Language Area. AIMS Neuroscience, 2016, 3(4): 454-473. doi: 10.3934/Neuroscience.2016.4.454 |
[4] | Eduardo Mercado III . Relating Cortical Wave Dynamics to Learning and Remembering. AIMS Neuroscience, 2014, 1(3): 185-209. doi: 10.3934/Neuroscience.2014.3.185 |
[5] | Matthew S. Sherwood, Jason G. Parker, Emily E. Diller, Subhashini Ganapathy, Kevin Bennett, Jeremy T. Nelson . Volitional down-regulation of the primary auditory cortex via directed attention mediated by real-time fMRI neurofeedback. AIMS Neuroscience, 2018, 5(3): 179-199. doi: 10.3934/Neuroscience.2018.3.179 |
[6] | Charles R Larson, Donald A Robin . Sensory Processing: Advances in Understanding Structure and Function of Pitch-Shifted Auditory Feedback in Voice Control. AIMS Neuroscience, 2016, 3(1): 22-39. doi: 10.3934/Neuroscience.2016.1.22 |
[7] | Zoha Deldar, Carlos Gevers-Montoro, Ali Khatibi, Ladan Ghazi-Saidi . The interaction between language and working memory: a systematic review of fMRI studies in the past two decades. AIMS Neuroscience, 2021, 8(1): 1-32. doi: 10.3934/Neuroscience.2021001 |
[8] | Joan Belo, Maureen Clerc, Daniele Schön . The effect of familiarity on neural tracking of music stimuli is modulated by mind wandering. AIMS Neuroscience, 2023, 10(4): 319-331. doi: 10.3934/Neuroscience.2023025 |
[9] | Pascale Voelker, Ashley N Parker, Phan Luu, Colin Davey, Mary K Rothbart, Michael I Posner . Increasing the amplitude of intrinsic theta in the human brain. AIMS Neuroscience, 2020, 7(4): 418-437. doi: 10.3934/Neuroscience.2020026 |
[10] | Iliyan Ivanov, Kristin Whiteside . Dyadic Brain - A Biological Model for Deliberative Inference. AIMS Neuroscience, 2017, 4(4): 169-188. doi: 10.3934/Neuroscience.2017.4.169 |
Human communication and language skills rely heavily on the ability to detect and process auditory inputs. This paper reviews possible applications of the event-related potential (ERP) technique to the study of cortical mechanisms supporting human auditory processing, including speech stimuli. Following a brief introduction to the ERP methodology, the remaining sections focus on demonstrating how ERPs can be used in humans to address research questions related to cortical organization, maturation and plasticity, as well as the effects of sensory deprivation, and multisensory interactions. The review is intended to serve as a primer for researchers interested in using ERPs for the study of the human auditory system.
The ability to detect and process auditory inputs plays a major role in human communication. Therefore, identification and assessment of the brain mechanisms supporting this ability is critical for understanding typical and atypical communication skills, including speech and language. Over the years, animal studies provided valuable data regarding the general structure and function of the primate auditory cortex. However, the human brain is not identical to that of any animal model, and the human auditory functioning includes aspects not present in other primates (e.g., spoken language). Therefore, studies of human auditory cortex structure and function are critically important. The main constraint is the fact that many of the effective techniques used to examine animal brains cannot be systematically applied to humans due to the invasive nature (e.g., intracranial recordings) and/or ethical concerns (e.g., experimental auditory deprivation). Advancements in non-invasive neuroimaging techniques (e.g., functional magnetic resonance imaging, fMRI; electroencephalography, EEG; magnetoencephalography, MEG; optical imaging) provide an opportunity to examine various aspects of human auditory cortex functions. These methods have yielded valuable results replicating and extending findings from the animal models.
This review focuses on the uses of scalp-recorded cortical event-related potentials (ERPs) for investigation of human auditory function. For readers interested in the evoked responses reflecting subcortical neural mechanisms of auditory processing, [1] and [2] provide relevant information. Following a brief introduction to the ERPs, the remaining sections demonstrate the success of using ERPs to address questions related to cortical maturation, functional organization, and plasticity, which have been previously studied using more invasive methods. A separate section examines studies of human auditory development and functioning following deprivation as well as restoration of auditory input. The final section describes ERP studies of multisensory interactions in the auditory cortex. From this review it should be clear that ERPs can be a valuable tool for investigating a wide range of questions concerning neural mechanisms supporting human auditory processing and provide novel insights into brain-behavior connections in the area of speech and language.
The ERP is a portion of the EEG that is time-locked to a stimulus event (see Figure 1). It reflects a momentary change in the ongoing electrical brain activity in response to an external or internal stimulus. This temporal synchronization with the evoking event allows to separate stimulus-related brain activity from the continuous background EEG when averaged across multiple trials [3]. It is important to note that the ERPs reflect oscillatory brain activity that is time- and phase-locked to stimulus onset. Examining time- but not phase-locked activity (known as the induced signal) by means of a time-frequency analysis could provide additional information about the neural mechanisms of interest (e.g., [4,5]), but is beyond the scope of this paper.
ERPs waveforms are described as a series of positive and negative peaks characterized in terms of size, or amplitude (µV height of the wave at different points), latency (time from stimulus onset), scalp distribution, and more recently, brain sources of measured activity. The peaks are labeled by polarity (‘P’ for positive and ‘N’ for negative deflections) and by latency (e.g., ‘N100’ indicates a negative peak occurring at 100 ms after the stimulus event) or sequential number (‘N1’ reflects the first negative peak after stimulus onset). ERPs offer a very high temporal resolution measuring voltage changes in milliseconds and show reliable sensitivity in detecting functional changes of brain activity [6,7].
The millisecond-level resolution is the major strength of the ERPs and represents an advantage over other imaging techniques (e.g., fMRI) as it tracks information processing at the speed comparable to that of the brain processes. This is particularly important for measures of auditory function because it allows for the fine-grained level of analysis needed to accurately characterize brain’s ability to process auditory inputs, including rapidly changing speech sounds [8].
The auditory ERP studies have used a wide variety of auditory stimuli (from simple clicks and tones to syllables, words, and complex sentences) and can be broadly categorized into two types: (1) those assessing individual stimulus detection and processing and (2) those measuring differentiation of contrasting stimuli. Stimulus processing studies may include a single stimulus or multiple stimuli of interest and focus on ERP characteristics associated with the individual sounds. Conversely, auditory differentiation studies use at least two stimuli that differ on a particular dimension of interest, such as basic physical characteristics (e.g., frequency, duration), articulatory features (e.g., voice onset timing, place of articulation), or semantic aspects (e.g., real vs. nonsense words). The analyses then examine the difference in ERPs elicited by the contrasting stimuli. For a more detailed introduction to the ERP technique, including experimental design and data interpretation issues, see [9,10].
Generally, variations in ERP peaks are interpreted to reflect speed (latency) or extent (amplitude) of processing. Shorter latencies indicate more rapid processing while larger amplitudes might suggest increased brain activity (i.e., a greater number of neurons firing in synchrony). The ERP responses thought to reflect activity of the primary and secondary auditory cortex include P1, N1, P2, and mismatch negativity (MMN); however, their specific latencies and scalp distribution vary with age (see [11,12,13], for further details).
The passive nature of many ERP paradigms is an additional benefit of this technique. Behavioral measures of auditory processing typically rely on participant’s overt reaction to auditory stimulation, such as moving (e.g., turning head, raising hand, or pushing a button) or speaking. Unfortunately, the developmental level of the participant limits the use of such procedures, and the results can be confounded by inattention, poor motor control, lack of motivation, and many other variables besides the auditory processing ability. On the other hand, ERPs can be obtained without participant’s active involvement in the test because cortical brain responses to auditory stimulation are detectable even in the absence of overt behavioral reactions.
One limitation of the ERP technique is that it does not offer detailed spatial resolution needed to conclusively link the electrical activity recorded on the scalp with the specific sources within the brain. Even with high-density (e.g., 128 or 256 channels) electrode arrays, not all brain electrical activity is captured due to variations in signal strength, distance from various cortical areas to the scalp, and differences in the orientation of cortical columns generating the signal. Nevertheless, studies using convergent methodologies (e.g., ERP and fMRI, lesion studies, MEG) provided consistent results identifying most likely brain sources of the commonly observed ERP responses (see [14] for a review), and therefore allow ERP data to be interpreted with regard to the anatomical aspects of auditory processing.
A vast number of ERP studies have successfully examined a wide range of questions regarding the auditory function in humans, such as cortical maturation changes, topographic organization, and sound discrimination ability. Capitalizing on the benefits of ERP methodology, these studies were able to document auditory processing in general, and speech/language processing in particular, from the earliest developmental stages (e.g., preterm infants) and throughout the lifespan.
While the basic elements of the human auditory system are already mature at birth (e.g., the cochlea [15]), and the main cortical structures (e.g., Heschl’s gyrus, superior temporal gyrus) are in place by 37 weeks gestation [16], the auditory cortex undergoes many structural and functional changes during the first few years of life [17,18]. Although cortical layer I resembles adult structure at birth, the deep layers (lower layer III and layer IV) develop between 6 months and 5 years of age, followed by maturation of the superficial layers (upper layer III and layer II) that continues until 10-12 years [17,19].
Many of these developmental changes are reflected in the maturational time course of various ERPs. Cortical auditory responses begin to emerge as early as 25 weeks conceptional age as an ERP consisting of a negative peak followed by a positive peak [20], and an auditory change discrimination could be observed in MEG data at 27 weeks gestation [21]. Because the early negative potential is typically reduced or absent in waveforms of term infants and young children, it has been interpreted to reflect an unstimulated auditory system or the activity of non-primary auditory areas [22,23].
Precursors of all major auditory ERP peaks present at 12 months of age are evident at birth, with maturation reflected in the increase of peak amplitudes and decrease in latencies [24,25,26]. However, such changes are not uniform across the peaks and do not always follow a linear trajectory [24,27]. The time course of these developmental changes may also vary by gender [28] and by the developmental stage at birth, as infants born prematurely take longer to demonstrate auditory processing comparable to that of the infants born full-term [29].
The early positive peak (often labeled P1), typically seen even in newborns and remaining the most prominent ERP response through the early childhood [24,30], has been frequently used as an index of maturation of the thalamo-cortical auditory pathway [31,32]. This peak is thought to reflect activation of cortical layers III and IV, and consistent with their rapid maturation, P1peak latency quickly shortens with age [19].
On the other hand, the N1 peak, the most pronounced auditory ERP in adult waveforms, is not always detected in infants and young children. An auditory N1 can be recorded from 3 years of age using a slow stimulation rate (1:1.2-2.4 sec; [33,34]), but its characteristics continue to mature into adolescence and young adulthood [19,35]. This delayed emergence and protracted development of the N1 response is attributed to slower maturation of the axons in cortical layers II and upper III [19].
Animal studies have demonstrated that the auditory cortex is highly organized such that various properties of the auditory stimuli (e.g., loudness, frequency, location in space) are processed by dedicated neurons (e.g., [36,37]) arranged in spatially contiguous groups. For example, intracranial recordings in awake monkeys established that auditory stimulus intensity is processed by at least three different groups of neurons, those sensitive to a specific stimulus level, and those responding to low or to high intensities [38]. Furthermore, while some neurons showed a linear increase in firing rate with the increase in stimulus intensity, others demonstrated a nonlinear pattern of response. Human ERP studies documented that increases in sound levels were associated with increased cortical activity [39,40]. Increase in click or tone level from 0 to 40 dB SL resulted in a systematic increase of P1-N1 and N1-P2 amplitudes [41] and decrease in latencies [42], with both of these measures reaching a plateau at intensities above 50-60 dB SL. More recent evidence suggests that there may not be an asymptote level for the ERP changes in response to increased stimulus intensity, as ERP amplitudes continued to increase as tone levels shifted from 30 to 90 dB SPL when stimuli were delivered with slow presentation rate [43]. Similar shortened latency of P1, N1, and P2 peaks in response to increased loudness was reported for consonant-vowel syllables as their level increased by 0, 20, or 40 dB SL above the participants’ sound discrimination threshold needed to achieve > 95% accuracy [44]. However, under these conditions, the ERP amplitudes decreased with increased sound intensity, possibly due to reduced attentional demands for discrimination of louder stimuli.
In the stimulus frequency domain, marmoset studies demonstrated that stimulus frequency is processed in the auditory cortex by bands of neurons with a distinct spatial distribution: those tuned to low frequencies located parallel to the lateral sulcus, while high-frequency contours placed perpendicular to that sulcus [45]. Human ERP measures are also sensitive to stimulus frequency. Picton et al. [46] reported a larger N1 peak in response to low- compared to high-frequency signals. Increase in signal frequency also shortened N1 peak latency [23,47,48], see also [49] for frequency effects on N1 and MMN properties.
Development of brain source analysis procedures for ERPs allowed to take scalp data another step further and to demonstrate tonotopic organization within the auditory cortex. Using current source density mapping and spatio-temporal dipole modeling, Bertrand et al. [50] demonstrated that observed differences in N1 amplitude and peak latency in response to tones of different frequency (250, 500, 1000, 2000 and 4000 Hz) were due to changes in the orientation of the brain sources in the auditory cortex. Similar topography was reported for stimulus harmonics of complex tones [47]. There, brain sources of the N1 became more frontally oriented as spectral profiles of the stimulus harmonics increased regardless of the fundamental frequencies (i.e., pitch) of the stimuli.
Using magnetic rather than electrical signal (MEG), which is primarily sensitive to location rather than orientation of the brain source, Pantev et al. [51] demonstrated that neurons activated by higher frequencies were located deeper in the supratemporal plane (see also [52]), with the depth of the Nlm (magnetic equivalent of N1) increasing linearly with the logarithm of the stimulus frequency (500, 1000 and 4000 Hz). Although not manifested in N1 peak latency changes, possibly due to largeacoustical variation of the stimuli, similar tonotopic organization was also documented for vowels, with the maps distributed perpendicular to those for pure tones [53].
In addition to the use of ERPs to assess cortical maturation and topography, a great number of studies focused on documenting specific auditory abilities, such as sound differentiation. The mismatch negativity response (MMN) has been viewed by many as a “gold standard” ERP discrimination measure. The MMN can be evoked by any perceivable physical deviance between auditory stimuli, such as differences in duration, frequency, or intensity, as well as by more complex characteristics such as voice onset time or place of articulation of speech sounds (for reviews, see [54,55,56]).
Infant MMN studies demonstrated that even an immature auditory system is capable of carrying out complex sound discriminations. Premature infants born at 30-35 weeks gestational age are capable of discriminating vowels, suggesting that the human fetus may learn to discriminate some sounds while still in the womb [57]. In addition to vowel discrimination, full-term newborns can discriminate relatively small changes in tone frequency (e.g., 1000 vs. 1100 Hz [58,59]; 500 vs. 750Hz [60]) and in duration (60-100 ms [61]; 160 ms [62]; 200 ms [60]), and do so in a manner similar to adults as reflected in the similar time course and scalp distributions of the MMN response. Interestingly, some discrimination effects were evident even when newborns were sleeping [63]. In addition to the sound duration discrimination, a developing auditory system is also sensitive to differences in the timing between the sounds, as unexpected changes in the length of the regular inter-tone interval (500 ms vs. 1500 ms) elicited an MMN response in 10-month-old infants [64].
Sensitivity to timing differences is also relevant to speech processing. Friederici et al. [65] tested the ability of 8-week-old infants in the awake and asleep state to discriminate vowel duration in natural speech syllables /baa/ and /ba/. All infants were able to detect the longer stimulus presented infrequently among the short syllables. However, when the shorter vowel was the rare stimulus, only the awake infants produced a mismatch response. The differences in outcomes were attributed to the shorter vowel being less salient than the longer stimulus and thus requiring a certain level of alertness to be detected [65]. Using similar stimuli, Pihko et al., [66] demonstrated that the ability to detect the shorter stimulus improves with age: while only some newborns could detect the short deviant (possibly due to individual differences in sleep/wake state), by 6 months of age, all infants in their study differentiated the vowels of different length.
A separate line of research demonstrated sensitivity to temporal differences as the important aspect of categorical speech perception. Intracranial recordings from primary and secondary auditory cortex in adults reflected differences in evoked responses to syllables with the voice onset time (VOT) between 40-80 ms compared to 0-20 ms [67]. While responses to longer VOT included markers of consonant release and voicing onset, the latter was greatly reduced for short VOT stimuli. The temporal distribution of such differences matched behavioral sound discrimination data indicating a categorical shift in consonant identification from a voiced to an unvoiced stop consonant as the VOT increased from 20 to 40 ms [67]. Examining N1 amplitude in response to two-tone non-speech stimuli resembling the temporal delay of voice-onset time (VOT), Simos and Molfese [68] demonstrated that similar sensitivity to the temporal lag in sound elements is already present within 48 hours from birth. The categorical boundaries for consonant-vowel speech syllables are similar in infants, children and adults [69,70,71].
Human auditory system is also adept at discriminating other phonetic characteristics such as formant transitions associated with differences in the place of articulation (POA; [72]). Similar to the VOT findings, even young infants can discriminate consonants varying in POA (newborns—[70]; 2-4 month olds—[73], [74]). Amplitude of the P2 varies in response to POA contrasts and can be predictive of later language outcomes [75]. MMN responses in children (7-11 years) and adults are also sensitive to POA features and were present in response to synthesized speech variants of the syllable /da/ presented among standard /da/ stimuli, even though these sounds appeared to be indiscriminable using behavioral paradigms [76].
Examination of ERPs to speech sounds also highlighted hemisphere differences in auditory processing. While both VOT and POA differences affect ERPs over both hemispheres (e.g., the N1 response in adults), they also elicit lateralized ERP changes. Timing differences are more likely to be seen over the right hemisphere, while POA effects are typically observed over the left hemisphere (see [77] for a review).
Animal studies in kittens and rats have shown that repetitive exposure to stimuli with specific spectral and temporal properties (e.g., modulated tones) during early development can reorganize the primary auditory cortex (e.g., expand representation of specific frequencies [78,79,80]. ERP evidence has also demonstrated the impact of postnatal auditory stimulation on human auditory development.
While basic sound detection abilities appear to be independent of auditory experience as reflected in the absence of group differences in the P2 response to a non-speech chime, P2 responses to speech sounds in premature infants (35-38 weeks gestation) were less well developed than those of full-term newborns and 1-week old infants [81] (see also [29] and [82] for effects of extreme prematurity). Most pronounced experience-related differences were evident in the increased P2 response to their mother’s voice in 1-week-old infants compared to that of newborns, while ERPs responses to the stranger’s voice did not differ between these groups [81]. However, ERP studies in very premature infants born prior to 30 weeks gestation demonstrated that environmental auditory exposure might not be sufficient for proper auditory development [82]). Brain responses (infant P1-N2 and the mismatch response) of such infants recorded at 35 weeks postconceptual age reflected auditory change detection in pure tone stimulibut had lower amplitudes compared to those of infants born closer to term, likely due to disruptions in corticogenesis processes and/or atypical auditory stimulation associated with neonatal intensive care environment (see also [83]).
ERP findings also indicate that even a brief exposure (2.5-5 hours) to auditory stimuli can alter sound processing. Cheour [84] demonstrated that passive exposure to vowels (/i/ and /y/i/ vs. /y/) could result in improved sound discrimination in newborn infants as reflected in increased MMN response. The effects of such exposure lasted at least 24 hours and generalized to similar stimuli of different pitch. Experience-dependent plasticity is also apparent in the time course of cerebral specialization for native language processing. Cheour et al. [85] assessed sound discrimination ability in a sample of Finnish and Estonian infants using vowel targets that were present in both Finnish and Estonian languages or specific to Estonian only. At 6 months of age, stimuli from both languages elicited similar MMN responses in Finnish infants. However, at 12 months of age, MMN responses were larger for the Finnish language vowels suggesting that infants became more specialized in processing sounds of their native language. Estonian infants at 12 months did not show significant differences in MMN response to the target vowels (both were common in their language) and generated a larger MMN to the Estonian vowel than the Finnish group. Similar increases in specialization of auditory processing of the native sounds were present for consonant contrasts in American infants: at 7 months of age, infants were able to discriminate non-native contrasts but the effect disappeared by 11 months, while responsiveness to native language consonant contrasts increased over time [86]. Interestingly, ERPs data also indicate that the decline in nonnative speech perception reflects reorganization rather than loss of ability [87] as ERP responses to the nonnative contrast do not disappear but occur at later latencies or different scalp areas than those typically reflecting native sound discrimination [88,89,90].
Auditory cortex retains some plasticity beyond the first year of life as experience-related changes were observed in older children and adults. For 3-6 year old Finnish children, 2 months of attendance at a school or a daycare center where French was spoken 50-90% of the time, were sufficient to develop a MMN response for a French speech contrast [91]. In adults, auditory training to discriminate VOT differences increased amplitudes of the N1-P2 responses reflecting increased neural synchrony and heightened sensitivity to previously subtle acoustic cues [92,93].
Changes in auditory processing can also be induced in a mature system following intense and targeted exposure, both in animals [94] and humans. Reinke et al. [95] examined vowel discrimination abilities in two groups of adults during two ERP sessions (600 trials per session) separated by 1 week. During that week, one group received four training sessions on vowel identification task (1080 trials). Training resulted in a decrease of the N1 and P2 latencies. A larger P2 amplitude was observed for all participants in the post-training ERP session, but the amount of increase was greater for the trained group. In a different study, the initial ERP assessment (5400 trials) was followed by 5 daily sessions (600 trails) of consonant discrimination training [96]. At the post-training ERP session, there were no training effects on the P2 amplitude, as both trained and untrained participants exhibited a larger P2 (with slight group differences in topographic distributions). One explanation for the lack of clear training effects suggests that the two groups received a very similar number of stimulus exposures and the training sessions might have been too brief to alter ERP responses beyond the change induced by the initial exposure to the stimuli. Indeed, a significant increase in P2 was observed between the first and second halves of the pre-training ERP session, suggesting that even exposure to a speech sound without explicit training goals can alter cortical processing.
The auditory system may also lose its ability to process certain auditory cues, possibly due to decreased neural synchrony [93] and/or increased refractory time [97]. Studies in rats demonstrated that with the increasing age, the auditory cortex neurons were most likely to respond to slower-changing stimuli [98]. In humans, ERP studies noted similar patterns of slowing: in older adults with and without hearing loss N1 and P2 latencies were prolonged in response to stimuli with increased VOT durations [90,93]. Alterations in the N1 response suggest that older auditory systems are less able to time-lock to the onset of voicing when there is a gap between the onset of the consonant burst and the onset of voicing or there may be a slower recovery processes from the initial excitation response to the consonant burst before neurons are able to fire again in response to the onset of voicing [97,99]. Delayed P2 latencies suggest impairments in neural mechanisms responsible for detecting complex acoustic signals associated with speech.
Animal and human studies have demonstrated that the development of the central auditory system depends on both internal (genetic) mechanisms and external (sensory) inputs [100,101]. Atypical auditory development due to auditory deprivation (such as congenital deafness or loss of hearing early in life) provides a unique opportunity to examine the role of individual experience with sound for auditory development and the extent of cortical plasticity following restoration of auditory inputs (e.g., via hearing aides or cochlear implants) at various ages.
ERP studies in persons with hearing loss demonstrate that extended sound deprivation affects auditory development. A large negativity preceding the P1 response has been seen in congenitally deaf children and in profoundly hearing-impaired children prior to activation of cochlear implants or acquisition of hearing aids [102]. This early negativity resembled ERPs of preterm infants [103] suggesting delayed cortical maturation in the absence of auditory input.
Acquired hearing loss in adults also leads to altered cortical functioning. Moderate to severe sensorineural hearing loss was associated with prolonged N1, N2, and P3 latencies as well as reduced N2 amplitudes relative to normal hearing peers [104]. Mild to moderate sensorineural hearing loss did not result in the same increase in latencies, but was associated with reduction in the amplitude of N1 [105]. Oates et al. [106] conducted a more extensive investigation into the impact of sensorineural hearing loss on the amplitude and latencies of auditory ERPs (N1, MMN, N2, P3) in response to speech syllables. They demonstrated that as degree of hearing loss increased, latencies of ERP components were prolonged and the amplitudes were reduced or absent, with the specific ERP peak alterations dependent on stimulus intensity and the degree of hearing loss. Furthermore, changes in peak latency occurred prior to changes in amplitude. Even mild (25 to 49 dB HL) threshold elevations resulted in delayed ERPs, while the amplitude reduction was evident in participants with hearing loss of at least 60 dB HL for the 65 dB SPL stimuli and 75 dB HL for the 80 dB SPL stimuli. Additionally, although reduced amplitude and delayed peak latency of auditory ERPs were also observed in normal hearing individuals with simulated hearing loss, in the participants with sensorineural hearing loss, ERP alterations were present at better signal-to-noise ratios (i.e., for more audible sounds; [106]). Several recent studies have replicated the reported peak latency delays, but also observed an increase in the amplitudes of the auditory P2 response in adults with mild-to-moderate hearing loss, suggesting increased effortful listening and the use of compensatory information processing strategies [107,108].
Evidence from studies of cats with noise-induced hearing loss suggests that enriched auditory environment can counteract some of the detrimental cortical reorganization associated with the impaired auditory processing [109]. The use of hearing aids in humans increases the range of audible inputs and can improve morphology of cortical auditory ERPs. However, the amount of change depends on the degree of sensorineural hearing loss, the intensity of the stimuli, and is not uniform across ERP components [110]. The largest improvements associated with sound amplification occurred for individuals in the severe-to-profound hearing-impaired group. At 65 dB SPL stimulus intensity, the use of hearing aids resulted in the increased amplitudes (by 50%) and decreased latencies (by 30 ms for N1), suggesting synchronous activation of a larger pool of cortical neurons compared to the unaided condition. However, for the 80 dB SPL stimuli, amplification improved only amplitudes of the N1, possibly due to the output limiting system of the hearing aids. ERP changes associated with the use of hearing aids may be due not only to the general increase in intensity of the stimuli, but to additional alterations in sound characteristics. In a sample of normal hearing adults listening to sounds via hearing aids, increases in N1-P2 amplitudes in the aided condition were related to improvement in the signal-to-noise ratio between the stimulus and low-level environmental and/or circuit noise, as both types of auditory input were amplified by the hearing aid [43].
Cochlear implants are often used to restore auditory input in patients with hearing loss for whom hearing aids do not provide sufficient benefits. ERP evidence suggests that within 6-8 months post-implantation auditory cortex undergoes significant reorganization [111] as indicated by changes in several ERP components, and some changes in stimulus processing and discrimination can be seen within days after the initial processor fitting [112,113].As the earliest cortical ERP response elicited by auditory stimuli, P1 peak latency has been used for assessment of auditory cortex development in children with hearing impairments and cochlear implant users [23,114,115]. The presence of P1 following activation of the cochlear implants supports the assumption that the deep layers of auditory cortex may develop to some extent even in the absence of auditory stimulation [116], however that development is incomplete. ERPs elicited by brief (100-200µs) clicks or pulses in pediatric implant users were similar to those of younger controls [31]. Following activation of the implant, P1 latencies for prelingually deaf children implanted by 3.5 years of age decreased at a faster rate compared to typical development and were within normal limits after 6 months of implant use [114,117], see also [118] for a review.
Longer periods of auditory deprivation may lead to more extensive alterations in the auditory cortex [119]. Several studies suggest that cross-modal reorganization may begin around 7 years of age (e.g., [120]) and consequently limit cortical ability to process inputs from the auditory system. P1 latencies for prelingually deaf children implanted after 7 years of age showed less pronounced reduction and remained outside of the normal range even after several years of implant use [111]. The lag in P1 latency remained proportionate to the length of deafness [121].
Even greater alterations are observed for the N1 response that reflects activity of the upper layers of the auditory cortex. Unlike the deeper structures, maturation of these layers is experience-dependent and does not occur if adequate auditory stimulation was not available early in life [117]. Consequently, the N1 peak is often missing or appears immature in cochlear implant users with the history of prolonged periods of deafness during childhood [116,122]. The absence of a clear N1 response may also suggest increased likelihood of difficulties with processing speech in noise (see also [123]), because better perception of degraded speech is associated with greater maturity of the upper cortical layers [121].
N1 latencies in adults with unilateral cochlear implants are usually delayed for all auditory stimuli (e.g., tones, vowels, syllables), and N1 amplitudes are reduced for speech [124]. N1 amplitude and peak latency in implant users correlate with behavioral discrimination of speech sounds (spondee identification; [125]). Opposite of normal hearing persons, N1 amplitudes in implant users are larger for low frequency sounds and reduce with the increase in stimulus frequency [126]. These findings suggest that a prolonged period of poor high frequency stimulation common in profound deafness results in cortical reorganization favoring low frequencies. Furthermore, restoration of high frequency inputs following cochlear implantation may not be sufficient to establish typical cortical organization [126]. Over 6 months after implantation, N1 latency may shorten toward normal values, but in the prelingually deaf patients such changes are less pronounced and more variable over time. Also, a more diffuse topography was noted for N1 in cochlear implant users compared to a normal hearing person [112]. A recent longitudinal study of postlingually deaf adults demonstrated increased amplitude and decreased peak latency of the N1 response after 8 weeks of cochlear implant use [127]. However, while the N1 amplitudes became comparable to those of the normal hearing group after one year of experience with the cochlear implant, the latencies remained delayed. The magnitude of the delay was related to the duration of deafness prior to the implantation.
The degree of alteration in ERPs may be stimulus specific. In postlingually deaf adults using cochlear implants, N1-P2 latencies were delayed for frequency, syllable, and vowel contrast stimuli, N1 amplitudes were reduced for all speech stimuli, but the P2 was smaller for consonant-vowel syllables only [124]. Furthermore, ERP alterations were related to the degree of cochlear implant benefit: N1-P2 peak amplitudes and latencies in successful implant users were similar to those of normal hearing controls, and better performing users elicited larger P2 amplitudes than moderate performers [128,129]. Shorter P2 latencies were associated with shorter durations of deafness [126] and better speech perception scores [125]. Degree of benefit from the cochlear implants can also be reflected in the MMN response as its amplitude and latency are similar to those of controls in successful implant users, while in poor users, this response is typically absent [130].
Recent increase in the number of bilateral cochlear implantations further extended our understanding of the maturational processes following restoration of auditory stimulation. Infants who received bilateral cochlear implants simultaneously showed P1 latencies similar to those of age-matched hearing peers by 1 month post-activation [131]. P1 development in infants receiving two cochlear implants sequentially by 12 to 24 months of age was similar to that of unilateral cochlear implant users with latencies reaching normal limits by 3 to 6 months following the first implant acquisitions. However, after receiving the second cochlear implant, P1 latencies for that ear were within normal limits within 1 month following implantation [131]. Similar fast normalization of P1 was reported for a child receiving sequential implants prior to the age of 3.5 years [113]. In a child receiving the second implant after 7 years of age, ERPs remained similar to those observed in late-implanted unilateral implant users, suggesting that stimulation from the first implant provided minimal impact on the ipsilateral auditory cortex development [113] .
Animal studies have consistently demonstrated that the auditory cortex could be a multisensory structure, and sound processing may be affected by inputs from other modalities [132]. Parallel findings have been obtained in humans (see [133,134] for detailed reviews). Sams et al. [135] were among the first to use ERPs to demonstrate that when auditory inputs (e.g., a spoken syllable) were paired with incongruent visual stimuli (e.g., lip movements corresponding to a different syllable), the listeners tended to misidentify the auditory stimulus, and the resulting illusory percept was treated by the auditory cortex as a novel acoustic event (i.e., the McGurk effect [136]), eliciting an MMN response (see also [137,138]).
Other findings indicate that multisensory effects may be present even earlier in time as reflected in the superadditive effects where an ERP response to a multimodal stimulus is greater than the sum of the responses to its unimodal components. When the auditory, visual, and audiovisual objects were attended, the P1 (P50 in the author’s notation) to the audiovisual stimuli was larger than the sum of its amplitudes for the auditory and visual stimuli presented alone [139]. However, in the unattended condition, the P1 response to multisensory trials was smaller, similar to the sensory gating effect filtering out irrelevant inputs [140]. Giard and Peronnet [141] reported increased amplitude of the auditory N1-P2 complex in multisensory trials (tones presented simultaneously with circles). This increase exceeded the sum of the auditory and visual ERP responses. The observed multisensory effects varied among participants. For those better in visual object recognition, adding a visual cue to the auditory stimulus increased the activity in the auditory cortex. Conversely, for participants performing better in the auditory task, adding a visual cue to the auditory stimulus did not result in the change of activity in the auditory cortex. The results suggest that at an early stage of sensory analysis, multisensory integration enhanced neural activity for the nondominant modality [141]. However, a number of recent studies report evidence of such early multisensory integration effects only when auditory and visual stimuli are both attended to [139] and provide redundant information. When visual stimuli were task-irrelevant, multisensory enhancement was observed only for later ERP responses thought to reflect cognitive rather than predominantly sensory processes [142].
In addition to the superadditive response described above, ERP studies also documented the existence of a multisensory suppression phenomenon. Van Wassenhove and colleagues [143] demonstrated that audiovisual speech stimuli resulted in reduced N1-P2 responses compared to the auditory speech alone. Interestingly, amplitude reduction was independent of the audiovisual speech congruency, participants’ expectations, or attended modality. Similar evidence for suppressive, speech-specific audiovisual integration mechanisms was provided by Besle and colleagues [144,145] who reported a decrease of N1 in response to audiovisual syllables, possibly reflecting attempts to minimize the processing of redundant cross-modal information.
Visual speech also appears to speed up cortical processing of auditory signals as indicated by the reduction in the auditory N1 latency. This effect depended on visual saliency of the stimulus and was not found during object recognition [141] or discrimination of verbal material presented in spoken and written forms [146]. Given the interpretation of the auditory N1 as reflecting stimulus feature analysis [147], these findings suggest that lip movements may facilitate feature analysis of the syllables in the auditory cortex, an interpretation consistent with the behavioral evidence of faster identification of multimodal compared to auditory syllables [144,148].
The differences in the direction of the multisensory effects could be attributed to the more complex nature of the speech stimuli. However, other findings suggest that these effects are not speech-specific as the multisensory suppression occurred for both congruent and incongruent speech [148] and non-speech [149] multimodal stimuli. The more likely source of the differences is the physical arrangement of the stimuli. In most cases of audiovisual speech, the onset of the visual stimuli precedes that of the sound, while in studies showing superadditive effects, the onsets of stimuli in both modalities were simultaneous. Asynchronous stimulus onset violates the temporal congruence requirement of the superadditive responses and can reflect activity of the subcortical structures sensitive to audiovisual asynchrony [150], which in turn may lead to response suppression in the sensory specific cortices. A study by Stekelenburg and Vroomen [149] further supported and extended the role of temporal asynchrony in the generation of multisensory effects. Both N1 and P2 suppression effects were observed for speech and non-speech stimuli. However, the reduction in amplitude and peak latency of the N1 depended mainly on the presence of anticipatory motion in the visual stimulus rather than its congruence with the auditory stimuli. The N1 suppression effect did not occur in the absence of the anticipatory motion. On the other hand, the amplitude reduction of the P2 was larger for incongruent than congruent multimodal stimuli [149].
Studies focused on understanding human auditory function and its cortical mechanisms achieved great success in using measures of brain activity, such as ERPs, to address a wide range of research questions. As demonstrated in the present review, the ERP has proved to be a reliable correlate of sound detection, speech processing, and multisensory integration in the auditory cortex. The noninvasive nature and ease of acquisition of ERPs allowed researchers to compare auditory function between infants, children, and adults, track maturational changes across the lifespan, document the role of postnatal experience (or lack thereof) in auditory cortex development, and understand the contribution of auditory processing to general cognitive functioning.
Methodological advances, including the development of novel analysis techniques for modeling brain sources of scalp activity, allowed ERPs to be used not only for general detection of brain responses to sounds but for examination of more advanced questions concerning specific functional organization and plasticity of the auditory cortex previously studied using more invasive methods. Paired with new clinical developments, such as cochlear implantation that restores auditory input for persons with hearing loss, ERPs provided a unique opportunity to track cortical development and document critical periods for various auditory functions. Auditory ERP measures have also demonstrated good correspondence with the results of standardized behavioral assessments, increasing ERP utility as a potential assessment tool, especially in populations with limited behavioral repertoire (e.g., young infants) or cognitive impairments precluding them from complying with instructions and providing valid overt responses.
The next frontier for ERP applications to the study of auditory processing would be to move from their nearly exclusive use as a research tool for group-level analyses (e.g., comparing groups with and without hearing loss) to interpreting single person data in a diagnostic way, similar to the clinical use of auditory brainstem responses (ABR). ERPs are highly sensitive to developmental and individual differences. Establishment of age-specific normative criteria for ERP peak characteristics (e.g., amplitude, latency, and topographic distribution) as well as the identification of stimuli and testing paradigms with the best diagnostic value would lead to further advancements in understanding of typical and atypical auditory processes. Our improved ability to conclusively interpret individual ERPs would provide an important bridge between psychophysiological research and clinical or educational practice, leading to new developments in prevention, early detection, and effective treatment of deficits in auditory function.
This work was supported in part by NICHD grants P30HD15052 and U54 HD083211 to the Vanderbilt Kennedy Center for Research on Human Development. The figure was created by Kylie Muccilli.
The author declares no conflicts of interest in this paper.
[1] |
Skoe E, Kraus N. (2010) Auditory brain stem response to complex sounds: A tutorial. Ear Hear 31: 302-324. doi: 10.1097/AUD.0b013e3181cdb272
![]() |
[2] | Bidelman GM, Moreno S, Alain C (2013) Tracing the emergence of categorical speech perception in the human auditory system. Hum Brain Map J 79: 201-212. |
[3] | Rockstroh B, Elbert T, Birbaumer N, et al. (1982) Slow brain potentials and behavior. Baltimore, MD: Urban-Schwarzenberg; 1982. |
[4] |
Giraud A-L, Poeppel D (2012) Cortical oscillations and speech processing: emerging computational principles and operations. Nat Neurosci 15: 511-517. doi: 10.1038/nn.3063
![]() |
[5] | Jochaut D, Lehongre K, Saitovitch A, et al. (2015) Atypical coordination of cortical oscillations in response to speech in autism. Front Hum Neurosci 27: 9. |
[6] | Rüsseler J, Nager W, Möbes J, et al. (2005) Cognitive adaptations and neuroplasticity: Lessons from event-related brain potentials. In: Heil P, Schleich H, Budinger E, Konig R, editors. The auditory cortex: A synthesis of human and animal research. Mahwah, NJ: Lawrence Erlbaum Associates, Inc; 2005. pp. 467-485. |
[7] | Zani A, Proverbio AM (2003) Cognitive electrophysiology of mind and brain. In: Zani A, Proverbio AM, editors. The Cognitive Electrophysiology of Mind and Brain. San Diego, CA: Academic Press; 2003. pp. 3–36. |
[8] |
Yoder PJ, Camarata S, Camarata M, et al. (2006) Association between differentiated processing of syllables and comprehension of grammatical morphology in children with Down syndrome. Am J Ment Retard 111: 138-152. doi: 10.1352/0895-8017(2006)111[138:ABDPOS]2.0.CO;2
![]() |
[9] | Luck SJ (2005) An Introduction to the Event-Related Potential Technique. The MIT Press. |
[10] | Luck SJ (2014) An Introduction to the event-related potential technique. 2nd ed. The MIT Press. |
[11] |
Eggermont JJ, Ponton CW (2002) The neurophysiology of auditory perception: from single units to evoked potentials. Audiol Neurootol 7: 71-99. doi: 10.1159/000057656
![]() |
[12] | Fishman YI (2013) The Mechanisms and Meaning of the Mismatch Negativity. Brain Topogr 27: 500-526. |
[13] |
Tomé D, Barbosa F, Nowak K, et al. (2015) The development of the N1 and N2 components in auditory oddball paradigms: a systematic review with narrative analysis and suggested normative values. J Neural Transmiss 122: 375-391. doi: 10.1007/s00702-014-1258-3
![]() |
[14] |
Key APF, Dove GO, Maguire MJ (2005) Linking Brainwaves to the Brain: An ERP Primer. Develop Neuropsychol 27: 183-215. doi: 10.1207/s15326942dn2702_1
![]() |
[15] |
Granier-Deferre C, Lecanuet JP, Cohen H, et al. (1985) Feasibility of Prenatal Hearing Test. Acta Oto-Laryngologica 99: 93-101. doi: 10.3109/00016488509121762
![]() |
[16] |
Moore JK, Guan YL (2001) Cytoarchitectural and axonal maturation in human auditory cortex. J Assoc Res Otolaryngol 2: 297-311. doi: 10.1007/s101620010052
![]() |
[17] | Ponton C, Eggermont JJ, Khosla D, et al. (2013) Maturation of human central auditory system activity: separating auditory evoked potentials by dipole source modeling. Clin Neurophysiol 113: 407-420. |
[18] |
Moore JK, Niparko JK, Perazzo LM, et al. (1997) Effect of adult-onset deafness on the human central auditory system. Ann Otol Rhinol Laryngol 106: 385-390. doi: 10.1177/000348949710600505
![]() |
[19] |
Ponton CW, Eggermont JJ, Kwong B, et al. (2000) Maturation of human central auditory system activity: evidence from multi-channel evoked potentials. Clin Neurophysiol 111: 220-236. doi: 10.1016/S1388-2457(99)00236-9
![]() |
[20] | Rotteveel JJ, de Graaf R, Stegeman DF, et al. (1987) The maturation of the central auditory conduction in preterm infants until three months post term. V. The auditory cortical response (ACR). Hear Res 27: 95-110. |
[21] |
Holst M, Eswaran H, Lowery C, et al. (2005) Development of auditory evoked fields in human fetuses and newborns: A longitudinal MEG study. Clin Neurophysiol 116: 1949-1955. doi: 10.1016/j.clinph.2005.04.008
![]() |
[22] |
Moller AR, Rollins PR (2002) The non-classical auditory pathways are involved in hearing in children but not in adults. Neurosci Letter 319: 41-44. doi: 10.1016/S0304-3940(01)02516-2
![]() |
[23] |
Gordon KA, Tanaka S, Wong DDE, et al. (2008) Characterizing responses from auditory cortex in young people with several years of cochlear implant experience. Clin Neurophysiol 119: 2347-2362. doi: 10.1016/j.clinph.2008.06.013
![]() |
[24] |
Kushnerenko E, Čeponienė R, Balan P, et al. (2002) Maturation of the auditory change detection response in infants: a longitudinal ERP study. NeuroReport 13: 1843–1848. doi: 10.1097/00001756-200210280-00002
![]() |
[25] |
Trainor L, McFadden M, Hodgson L, et al. (2003) Changes in auditory cortex and the development of mismatch negativity between 2 and 6 months of age. Int J Psychophysiol 51: 5-15. doi: 10.1016/S0167-8760(03)00148-X
![]() |
[26] |
Wunderlich JL, Cone-Wesson BK (2006) Maturation of CAEP in infants and children: A review. Hear Res 212: 212-223. doi: 10.1016/j.heares.2005.11.008
![]() |
[27] | Molfese DL, Molfese VJ (2000) The continuum of language development during infancy and early childhood: Electrophysiological correlates. In: Rovee-Collier C, Lipsitt L, Hayne H, editors. Progress in infancy research. Lawrence Erlbaum Associates. pp. 251-287. |
[28] | Nanova P, Lyamova L, Hadjigeorgieva M, et al. (2008) Gender-specific development of auditory information processing in children: An ERP study. Clin Neurophysiol 119: 1992-2003. |
[29] | Key APF, Lambert EW, Aschner JL, et al. (2012) Influence of gestational age and postnatal age on speech sound processing in NICU infants. Psychophysiology 49: 720-731. |
[30] |
Pang EW, Taylor MJ (2000) Tracking the development of the N1 from age 3 to adulthood: an examination of speech and non-speech stimuli. Clin Neurophysiol 111: 388-397. doi: 10.1016/S1388-2457(99)00259-X
![]() |
[31] |
Ponton CW, Don M, Eggermont JJ, et al. (1996) Auditory system plasticity in children after long periods of complete deafness. NeuroReport 8: 61. doi: 10.1097/00001756-199612200-00013
![]() |
[32] |
Ponton CW, Don M, Eggermont JJ, et al. (1996) Maturation of human cortical auditory function: differences between normal-hearing children and children with cochlear implants. Ear Hear 17: 430-437. doi: 10.1097/00003446-199610000-00009
![]() |
[33] |
Paetau R, Ahonen A, Salonen O, et al. (1995) Auditory evoked magnetic fields to tones and pseudowords in healthy children and adults. J Clin Neurophysiol 12: 177-185. doi: 10.1097/00004691-199503000-00008
![]() |
[34] |
Sharma A, Kraus N, McGee TJ, et al. (1997) Developmental changes in P1 and N1 central auditory responses elicited by consonant-vowel syllables. Electroencephalography clin Neurophysiol 104: 540-545. doi: 10.1016/S0168-5597(97)00050-6
![]() |
[35] |
Gilley PM, Sharma A, Dorman M, et al. (2005) Developmental changes in refractoriness of the cortical auditory evoked potential. Clin Neurophysiol 116: 648-657. doi: 10.1016/j.clinph.2004.09.009
![]() |
[36] |
Kaas JH, Hackett TA, Tramo MJ, et al. (1999) Auditory processing in primate cerebral cortex. Current Opinion Neurobiol 9: 164-170. doi: 10.1016/S0959-4388(99)80022-1
![]() |
[37] |
Morel A, Garraghty PE, Kaas JH (1993) Tonotopic organization, architectonic fields, and connections of auditory cortex in macaque monkeys. J Comp Neurol 335: 437-459. doi: 10.1002/cne.903350312
![]() |
[38] | Recanzone GH, Guard DC, Phan ML (2000) Frequency and intensity response properties of single neurons in the auditory cortex of the behaving macaque monkey. J Neurophysiol 83: 2315-2331. |
[39] |
Rapin I, Schimmel H, Tourk LM, et al. (1966) Evoked responses to clicks and tones of varying intensity in waking adults. Electroencephalography Clin Neurophysiol 21: 335-344. doi: 10.1016/0013-4694(66)90039-3
![]() |
[40] |
Beagley HA, Knight JJ (1967) Changes in Auditory Evoked Response with Intensity. J Laryngology Otology 81: 861-873. doi: 10.1017/S0022215100067815
![]() |
[41] |
Madell JR, Goldstein R (1972) Relation between loudness and the amplitude of the early components of the averaged electroencephalic response. J Speech Hear Res 15: 134-141. doi: 10.1044/jshr.1501.134
![]() |
[42] |
Picton TW, Hillyard SA, Krausz HI, et al. (1974) Human auditory evoked potentials. I. Evaluation of components. Electroencephalography Clin Neurophysiol 36: 179-190. doi: 10.1016/0013-4694(74)90155-2
![]() |
[43] |
Billings CJ, Tremblay KL, Souza PE, et al. (2007) Effects of hearing aid amplification and stimulus intensity on cortical auditory evoked potentials. Audiol Neurotol 12: 234-246. doi: 10.1159/000101331
![]() |
[44] |
Garinis AC, Cone-Wesson BK (2007) Effects of stimulus level on cortical auditory event-related potentials evoked by speech. J Am Aca Audiology 18: 107-116. doi: 10.3766/jaaa.18.2.3
![]() |
[45] | Aitkin LM, Merzenich MM, Irvine DR, et al. (2013) Frequency representation in auditory cortex of the common marmoset (Callithrix jacchus jacchus). J Comp Neurol 252: 175-185. |
[46] |
Picton TW, Woods DL, Proulx GB (1978) Human auditory sustained potentials. II. Stimulus relationships. Electroencephalography Clin Neurophysiol 45: 198-210. doi: 10.1016/0013-4694(78)90004-4
![]() |
[47] |
Crottaz-Herbette S, Ragot R (2000) Perception of complex sounds: N1 latency codes pitch and topography codes spectra. Clin Neurophysiol 111: 1759-1766. doi: 10.1016/S1388-2457(00)00422-3
![]() |
[48] |
Verkindt C, Bertrand O, Perrin F, et al. (1995) Tonotopic organization of the human auditory cortex: N100 topography and multiple dipole model analysis. Electroencephalography Clin Neurophysiol 96:143–56. doi: 10.1016/0168-5597(94)00242-7
![]() |
[49] |
Wunderlich JL, Cone-Wesson BK (2001) Effects of stimulus frequency and complexity on the mismatch negativity and other components of the cortical auditory-evoked potential. J Acoust Soc Am 109: 1526-1537. doi: 10.1121/1.1349184
![]() |
[50] | Bertrand O, Perrin F, Pernier J (1991) Evidence for a tonotopic organization of the auditory cortex observed with auditory evoked potentials. Acta Oto-Laryngologica 111: 116-123. |
[51] |
Pantev C, Bertrand O, Eulitz C, et al. (1995) Specific tonotopic organizations of different areas of the human auditory cortex revealed by simultaneous magnetic and electric recordings. Electroencephalography Clin Neurophysiol 94: 26-40. doi: 10.1016/0013-4694(94)00209-4
![]() |
[52] |
Lütkenhöner B, Steinsträter O (1998) High-precision neuromagnetic study of the functional organization of the human auditory cortex. Audiol Neurootol 3: 191-213. doi: 10.1159/000013790
![]() |
[53] |
Shestakova A, Brattico E, Soloviev A, et al. (2004) Orderly cortical representation of vowel categories presented by multiple exemplars. Brain Res Cogn Brain Res 21: 342-350. doi: 10.1016/j.cogbrainres.2004.06.011
![]() |
[54] |
Näätänen R, Paavilainen P, Rinne T, et al. (2007) The mismatch negativity (MMN) in basic research of central auditory processing: A review. Clin Neurophysiol 118: 2544-2590. doi: 10.1016/j.clinph.2007.04.026
![]() |
[55] | May PJC, Tiitinen H (2010) Mismatch negativity (MMN), the deviance-elicited auditory deflection, explained. Psychophysiology 47: 66-122. |
[56] |
Garrido MI, Kilner JM, Stephan KE, et al. (2009) The mismatch negativity: A review of underlying mechanisms. Clin Neurophysiol 120: 453-463. doi: 10.1016/j.clinph.2008.11.029
![]() |
[57] |
Cheour-Luhtanen M, Alho K, Sainio K, et al. (1996) The ontogenetically earliest discriminative response of the human brain. Psychophysiology 33: 478-481. doi: 10.1111/j.1469-8986.1996.tb01074.x
![]() |
[58] | Leppänen PHT, Richardson U, Pihko E, et al. (2002) Brain responses to changes in speech sound durations differ between infants with and without familial risk for dyslexia. Develop Neuropsychol 22: 407-422. |
[59] |
Alho K, Sainio K, Sajaniemi N, et al. (1990) Event-related brain potential of human newborns to pitch change of an acoustic stimulus. Electroencephalography Clin Neurophysiol 77: 151-155. doi: 10.1016/0168-5597(90)90031-8
![]() |
[60] |
Čeponienė R, Kushnerenko E, Fellman V, et al. (2002) Event-related potential features indexing central auditory discrimination by newborns. Brain Res Cogn Brain Res 13: 101-113. doi: 10.1016/S0926-6410(01)00093-3
![]() |
[61] | Cheour M, Kushnerenko E, Čeponienė R, et al. (2002) Electric brain responses obtained from newborn infants to changes in duration in complex harmonic tones. Develop Neuropsychol 22: 471-479. |
[62] |
Kushnerenko E, Čeponienė R, Fellman V, et al. (2001) Event-related potential correlates of sound duration: similar pattern from birth to adulthood. NeuroReport 12: 3777-3781. doi: 10.1097/00001756-200112040-00035
![]() |
[63] |
Martynova O, Kirjavainen J, Cheour M (2003) Mismatch negativity and late discriminative negativity in sleeping human newborns. Neurosci Letter 340: 75-78. doi: 10.1016/S0304-3940(02)01401-5
![]() |
[64] |
Brannon EM, Roussel LW, Meck WH, et al. (2004) Timing in the baby brain. Brain Res Cogn Brain Res 21: 227-233. doi: 10.1016/j.cogbrainres.2004.04.007
![]() |
[65] |
Friederici AD, Friedrich M, Weber C (2002) Neural manifestation of cognitive and precognitive mismatch detection in early infancy. NeuroReport 13: 1251-1254. doi: 10.1097/00001756-200207190-00006
![]() |
[66] | Pihko E, Leppänen PH, Eklund KM, et al. (1999) Cortical responses of infants with and without a genetic risk for dyslexia: I. Age effects. NeuroReport 10: 901-905. |
[67] | Steinschneider M, Volkov IO, Noh MD, et al. (1999) Temporal encoding of the voice onset time phonetic parameter by field potentials recorded directly from human auditory cortex. J Neurophysiol 82: 2346-2357. |
[68] |
Simos PG, Molfese DL, Brenden RA (1997) Behavioral and Electrophysiological Indices of Voicing-Cue Discrimination: Laterality Patterns and Development. Brain Language 57: 122-150. doi: 10.1006/brln.1997.1836
![]() |
[69] | Molfese DL, Papanicolaou A, Hess TM, et al. (1979) Neuroelectrical correlates of semantic processes. Evoked brain potentials and behavior. Plenum Pub Cor: 89-106. |
[70] | Molfese DL, Molfese VJ (1979) Hemisphere and stimulus differences as reflected in the cortical responses of newborn infants to speech stimuli. Develop Psychol 15: 505-511. |
[71] |
Segalowitz SJ, Cohen H (1989) Right hemisphere EEG sensitivity to speech. Brain Language 37: 220-231. doi: 10.1016/0093-934X(89)90016-3
![]() |
[72] | Dehaene-Lambertz G, Baillet S (1998) A phonological representation in the infant brain. NeuroReport 9: 1885. |
[73] |
Dehaene-Lambertz G, Dehaene S (1994) Speed and cerebral correlates of syllable discrimination in infants. Nature 370: 292-295. doi: 10.1038/370292a0
![]() |
[74] | Dehaene-Lambertz G (2000) Cerebral specialization for speech and non-speech stimuli in infants. J Cogn Neurosci 12: 449-460. |
[75] | Molfese DL (2000) Predicting dyslexia at 8 years of age using neonatal brain responses. Brain Language 72: 238-245. |
[76] |
Kraus N, McGee T, Micco A, et al. (1993) Mismatch negativity in school-age children to speech stimuli that are just perceptibly different. Electroencephalography Clin Neurophysiol 88: 123-130. doi: 10.1016/0168-5597(93)90063-U
![]() |
[77] | Molfese D, Key A, Maguire M, et al. (2005) Event-related evoked potentials (ERPs) in speech perception. In: Pisoni D, Remez R, editors. The Handbook of Speech Perception. Malden, MA: pp. 99-121. |
[78] |
Zhang LI, Bao S, Merzenich MM (2002) Disruption of primary auditory cortex by synchronous auditory inputs during a critical period. Proc Natl Acad Sci USA 99: 2309-2314. doi: 10.1073/pnas.261707398
![]() |
[79] | Stanton SG, Harrison RV (1996) Abnormal cochleotopic organization in the auditory cortex of cats reared in a frequency augmented environment. Aud Neurosci 2: 97-107. |
[80] |
Nakahara H, Zhang LI, Merzenich MM (2004) Specialization of primary auditory cortex processing by sound exposure in the “critical period.” Proceed Nation Aca Sci 101: 7170-7174. doi: 10.1073/pnas.0401196101
![]() |
[81] |
deRegnier R-A, Wewerka S, Georgieff MK, et al. (2002) Influences of postconceptional age and postnatal experience on the development of auditory recognition memory in the newborn infant. Dev Psychobiol 41: 216-225. doi: 10.1002/dev.10070
![]() |
[82] | Bisiacchi PS, Mento G, Suppiej A (2009) Cortical auditory processing in preterm newborns: An ERP study. Biolog Psychol 82: 176-185. |
[83] |
Fellman V, Kushnerenko E, Mikkola K, et al. (2004) Atypical auditory event-related potentials in preterm infants during the first year of life: A possible sign of cognitive dysfunction? Pediatric Res 56: 291-297. doi: 10.1203/01.PDR.0000132750.97066.B9
![]() |
[84] | Cheour M, Martynova O, Näätänen R, et al. (2002) Speech sounds learned by sleeping newborns. Nature 415: 599-600. |
[85] |
Cheour M, Čeponienė R, Lehtokoski A, et al. (1998) Development of language-specific phoneme representations in the infant brain. Nat Neurosci 1: 351-353. doi: 10.1038/1561
![]() |
[86] |
Rivera-Gaxiola M, Silva Pereyra J, Kuhl PK (2005) Brain potentials to native and non-native speech contrasts in 7-and 11-month-old American infants. Develop Sci 8: 162-172. doi: 10.1111/j.1467-7687.2005.00403.x
![]() |
[87] | Aaltonen O, Niemi P, Nyrke T, et al. (1987) Event-related brain potentials and the perception of a phonetic continuum. Biolog Psychol 24: 197-207. |
[88] |
Dehaene-Lambertz G (1997) Electrophysiological correlates of categorical phoneme perception in adults. NeuroReport 8: 919-924. doi: 10.1097/00001756-199703030-00021
![]() |
[89] |
Rivera-Gaxiola M, Csibra G, Johnson MH (2000) Electrophysiological correlates of cross-linguistic speech perception in native English speakers. Behav Brain Res 111: 13-23. doi: 10.1016/S0166-4328(00)00139-X
![]() |
[90] |
Sharma A, Dorman MF (2000) Neurophysiologic correlates of cross-language phonetic perception. J Acoust Soc Am 107: 2697. doi: 10.1121/1.428655
![]() |
[91] |
Cheour M, Shestakova A, Alku P, et al. (2002) Mismatch negativity shows that 3-6-year-old children can learn to discriminate non-native speech sounds within two months. Neurosci Letter 325: 187-190. doi: 10.1016/S0304-3940(02)00269-0
![]() |
[92] | Tremblay K, Kraus N, McGee T, et al. (2001) Central auditory plasticity: changes in the N1-P2 complex after speech-sound training. Ear Hear 22: 79-90. |
[93] |
Tremblay KL, Kraus N et al. (2002) Auditory training induces asymmetrical changes in cortical neural activity. J Speech Lang Hear Re 45: 564-572. doi: 10.1044/1092-4388(2002/045)
![]() |
[94] |
Beitel RE, Schreiner CE, Cheung SW, et al. (2003) Reward-dependent plasticity in the primary auditory cortex of adult monkeys trained to discriminate temporally modulated signals. Proc Natl Acad Sci USA 100: 11070-11075. doi: 10.1073/pnas.1334187100
![]() |
[95] |
Reinke KS, He Y, Wang C, et al. (2003) Perceptual learning modulates sensory evoked response during vowel segregation. Cogn Brain Res 17: 781-791. doi: 10.1016/S0926-6410(03)00202-7
![]() |
[96] |
Sheehan KA, McArthur GM, Bishop DVM (2005) Is discrimination training necessary to cause changes in the P2 auditory event-related brain potential to speech sounds? Cogn Brain Res 25: 547-553. doi: 10.1016/j.cogbrainres.2005.08.007
![]() |
[97] |
Tremblay K, Ross B (2007) Effects of age and age-related hearing loss on the brain. J Commu Disorder 40: 305-312. doi: 10.1016/j.jcomdis.2007.03.008
![]() |
[98] |
Mendelson JR, Ricketts C (2001) Age-related temporal processing speed deterioration in auditory cortex. Hear Res 158: 84-94. doi: 10.1016/S0378-5955(01)00294-5
![]() |
[99] |
Papanicolaou AC, Loring DW, Eisenberg HM (1984) Age-related differences in recovery cycle of auditory evoked potentials. Neurobiol Aging 5: 291-295. doi: 10.1016/0197-4580(84)90005-8
![]() |
[100] | Illing R-B (2004) Maturation and plasticity of the central auditory system. Acta Oto-Laryngologica 124: 6-10. |
[101] |
Kral A (2004) Postnatal cortical development in congenital auditory deprivation. Cerebral Cortex 15: 552-562. doi: 10.1093/cercor/bhh156
![]() |
[102] | Sharma A, Tobey E, Dorman M, et al. (2004) Central auditory maturation and babbling development in infants with cochlear implants. Arch Otolaryngol Head Neck Surg 130: 511-516. |
[103] | Weitzman L, Graziani L, Duhamel L (1967) Maturation and topography of the auditory evoked response of the prematurely born infant. Electroencephalography Clin Neurophysiol 23: 82-83. |
[104] |
Polen SB (1984) Auditory event related potentials. Seminars Hear 5: 127-141. doi: 10.1055/s-0028-1095227
![]() |
[105] |
Wall LG, Dalebout SD, Davidson SA, et al. (1991) Effect of Hearing Impairment on Event-Related Potentials for Tone and Speech Distinctions. Folia Phoniatr Logop 43: 265-274. doi: 10.1159/000266137
![]() |
[106] | Oates PA, Kurtzberg D, Stapells DR (2002) Effects of sensorineural hearing loss on cortical event-related potential and behavioral measures of speech-sound processing. Ear Hear 23:399-415. |
[107] | Campbell J, Sharma A (2013) Compensatory changes in cortical resource allocation in adults with hearing loss. Front Syst Neurosci 7: 71. |
[108] |
Bertoli S, Probst R, Bodmer D (2011) Late auditory evoked potentials in elderly long-term hearing-aid users with unilateral or bilateral fittings. Hear Res 280: 58-69. doi: 10.1016/j.heares.2011.04.013
![]() |
[109] |
Noreña AJ, Eggermont JJ (2005) Enriched acoustic environment after noise trauma reduces hearing loss and prevents cortical map reorganization. J Neurosci 25: 699-705. doi: 10.1523/JNEUROSCI.2226-04.2005
![]() |
[110] |
Korczak PA, Kurtzberg D, Stapells DR (2005) Effects of sensorineural hearing loss and personal hearing aids on cortical event-related potential and behavioral measures of speech-sound processing. Ear Hear 26: 165-185. doi: 10.1097/00003446-200504000-00005
![]() |
[111] |
Sharma A, Dorman MF, Spahr AJ (2002) A sensitive period for the development of the central auditory system in children with cochlear implants: Implications for age of implantation. Ear Hear 23: 532-539. doi: 10.1097/00003446-200212000-00004
![]() |
[112] |
Jordan K, Schmidt A, Plotz K, et al. (1997) Auditory event-related potentials in post- and prelingually deaf cochlear implant recipients. Am J Otology 18: S116-117. doi: 10.1016/S0196-0709(97)90099-5
![]() |
[113] | Sharma A, Dorman MF, Kral A (2005) The influence of a sensitive period on central auditory development in children with unilateral and bilateral cochlear implants. Hear Res 203: 134-143. |
[114] |
Sharma A, Dorman MF, Spahr AJ (2002) Rapid development of cortical auditory evoked potentials after early cochlear implantation. NeuroReport 13: 1365-1368. doi: 10.1097/00001756-200207190-00030
![]() |
[115] |
Dorman MF, Sharma A, Gilley P, et al. (2007) Central auditory development: Evidence from CAEP measurements in children fit with cochlear implants. J Commu Disorders 40: 284-294. doi: 10.1016/j.jcomdis.2007.03.007
![]() |
[116] |
Ponton CW, Eggermont JJ (2001) Of kittens and kids: altered cortical maturation following profound deafness and cochlear implant use. Audiol Neurootol 6: 363-380. doi: 10.1159/000046846
![]() |
[117] | Sharma A, Dorman M, Spahr A, et al. (2002) Early cochlear implantation in children allows normal development of central auditory pathways. Ann Otol Rhinol Laryngol Suppl 189: 38-41. |
[118] |
Johnson JM (2009) Late Auditory Event-Related Potentials in Children With Cochlear Implants: A Review. Develop Neuropsychol 34: 701-720. doi: 10.1080/87565640903265152
![]() |
[119] |
Finney EM, Fine I, Dobkins KR (2001) Visual stimuli activate auditory cortex in the deaf. Nat Neurosci 4: 1171-1173. doi: 10.1038/nn763
![]() |
[120] | Lee DS, Lee JS, Oh SH, et al. (2001) Cross-modal plasticity and cochlear implants. Nature 409: 149-150. |
[121] |
Eggermont JJ, Ponton CW (2003) Auditory-evoked potential studies of cortical maturation in normal hearing and implanted children: Correlations with changes in structure and speech perception. Acta Oto-Laryngologica 123: 249-252. doi: 10.1080/0036554021000028098
![]() |
[122] | Singh S, Liasis A, Rajput K, et al. (2004) Event-related potentials in pediatric cochlear implant patients. Ear Hear 25: 598-610. |
[123] |
Martin BA, Kurtzberg D, Stapells DR (1999) The effects of decreased audibility produced by high-pass noise masking on N1 and the mismatch negativity to speech sounds /ba/and/da. J Speech Language Hear Res 42: 271-286. doi: 10.1044/jslhr.4202.271
![]() |
[124] |
Groenen PAP, Beynon AJ, Snik AFM, et al. (2001) Speech-evoked cortical potentials recognition in cochlear implant users and speech. Scand Audiol 30: 31-40. doi: 10.1080/010503901750069554
![]() |
[125] | Makhdoum MJ, Groenen PAP, Snik AFM, et al. (1998) Intra-and interindividual correlations between auditory evoked potentials and speech perception in cochlear implant users. Scand Audiol 27: 13-20. |
[126] | Kelly AS, Purdy SC, Thorne PR (2005) Electrophysiological and speech perception measures of auditory processing in experienced adult cochlear implant users. Clin Neurophysiol 116: 1235-1246. |
[127] | Sandmann P, Plotz K, Hauthal N, et al. (2015) Rapid bilateral improvement in auditory cortex activity in postlingually deafened adults following cochlear implantation. Clin Neurophysiol 126: 594-607. |
[128] |
Groenen PA, Makhdoum M, van den Brink JL, et al. (1996) The relation between electric auditory brain stem and cognitive responses and speech perception in cochlear implant users. Acta Oto-Laryngologica 116: 785-790. doi: 10.3109/00016489609137926
![]() |
[129] | Micco AG, Kraus N, Koch DB, et al. (1995) Speech-evoked cognitive P300 potentials in cochlear implant recipients. Am J Otology 16: 514-520. |
[130] |
Kraus N, Micco AG, Koch DB, et al. (1993) The mismatch negativity cortical evoked potential elicited by speech in cochlear-implant users. Hear Res 65: 118-124. doi: 10.1016/0378-5955(93)90206-G
![]() |
[131] |
Bauer PW, Sharma A, Martin K, et al. (2006) Central auditory development in children with bilateral cochlear implants. Arch Otolaryngol Head Neck Surg 132: 1133-1136. doi: 10.1001/archotol.132.10.1133
![]() |
[132] |
Wallace MT, Ramachandran R, Stein BE (2004) A revised view of sensory cortical parcellation. Proc Natl Acad Sci USA 101: 2167-2172. doi: 10.1073/pnas.0305697101
![]() |
[133] | Murray MM, Spierer L (2009) Auditory spatio-temporal brain dynamics and their consequences for multisensory interactions in humans. Hear Res 258: 121-133. |
[134] | Musacchia G, Schroeder CE (2009) Neuronal mechanisms, response dynamics and perceptual functions of multisensory interactions in auditory cortex. Hear Res 258: 72-79. |
[135] |
Sams M, Aulanko R, Hämäläinen M, et al. (1991) Seeing speech: visual information from lip movements modifies activity in the human auditory cortex. Neurosci Letters 127: 141-145. doi: 10.1016/0304-3940(91)90914-F
![]() |
[136] |
McGurk H, MacDonald J (1976) Hearing lips and seeing voices. Nature 264: 746-748. doi: 10.1038/264746a0
![]() |
[137] |
Colin C, Radeau M, Soquet A, et al. (2002) Mismatch negativity evoked by the McGurk-MacDonald effect: a phonetic representation within short-term memory. Clin Neurophysiol 113: 495-506. doi: 10.1016/S1388-2457(02)00024-X
![]() |
[138] |
Möttönen R, Krause CM, Tiippana K, et al. (2002) Processing of changes in visual speech in the human auditory cortex. Brain Res Cogn Brain Res 13: 417-425. doi: 10.1016/S0926-6410(02)00053-8
![]() |
[139] |
Talsma D, Doty TJ, Woldorff MG (2006) Selective Attention and audiovisual integration: Is attending to both modalities a prerequisite for early integration? Cerebral Cortex 17: 679-690. doi: 10.1093/cercor/bhk016
![]() |
[140] |
Boutros NN, Belger A (1999) Midlatency evoked potentials attenuation and augmentation reflect different aspects of sensory gating. Biolog Psychiatry 45: 917-922. doi: 10.1016/S0006-3223(98)00253-4
![]() |
[141] |
Giard M, Peronnet F (1999) Auditory-visual integration during multimodal object recognition in humans: A behavioral and electrophysiological study. J Cogn Neurosci 11: 473-490. doi: 10.1162/089892999563544
![]() |
[142] | Li Q, Wu J, Touge T (1991) Audiovisual interaction enhances auditory detection in late stage: an event-related potential study. NeuroReport 21: 173-178. |
[143] |
Van Wassenhove V, Grant KW, Poeppel D (2005) Visual speech speeds up the neural processing of auditory speech. Proc Natl Acad Sci USA 102: 1181-1186. doi: 10.1073/pnas.0408949102
![]() |
[144] | Besle J, Fort A, Delpuech C, et al. (2004) Bimodal speech: early suppressive visual effects in human auditory cortex. Europ J Neurosci 20: 2225-2234. |
[145] |
Besle J, Fischer C, Bidet-Caulet A, et al. (2008) Visual Activation and audiovisual interactions in the auditory cortex during speech perception: Intracranial recordings in humans. J Neurosci 28: 14301-14310. doi: 10.1523/JNEUROSCI.2875-08.2008
![]() |
[146] | Raij T, Uutela K, Hari R (2000) Audiovisual integration of letters in the human brain. Neuron 28: 617-625. |
[147] |
Näätänen R, Picton T (1987) The N1 wave of the human electric and magnetic response to sound: a review and an analysis of the component structure. Psychophysiology 24: 375-425. doi: 10.1111/j.1469-8986.1987.tb00311.x
![]() |
[148] |
Klucharev V (2003) Electrophysiological indicators of phonetic and non-phonetic multisensory interactions during audiovisual speech perception. Cogn Brain Res 18: 65-75. doi: 10.1016/j.cogbrainres.2003.09.004
![]() |
[149] |
Stekelenburg JJ, Vroomen J (2007) Neural correlates of multisensory integration of ecologically valid audiovisual events. J Cogn Neurosci 19: 1964-1973. doi: 10.1162/jocn.2007.19.12.1964
![]() |
[150] | Bushara KO, Grafman J, Hallett M (2001) Neural correlates of auditory–visual stimulus onset asynchrony detection. J Neurosci 21: 300-304. |
1. | Catherine Morgan, Linda Fetters, Lars Adde, Nadia Badawi, Ada Bancale, Roslyn N. Boyd, Olena Chorna, Giovanni Cioni, Diane L. Damiano, Johanna Darrah, Linda S. de Vries, Stacey Dusing, Christa Einspieler, Ann-Christin Eliasson, Donna Ferriero, Darcy Fehlings, Hans Forssberg, Andrew M. Gordon, Susan Greaves, Andrea Guzzetta, Mijna Hadders-Algra, Regina Harbourne, Petra Karlsson, Lena Krumlinde-Sundholm, Beatrice Latal, Alison Loughran-Fowlds, Catherine Mak, Nathalie Maitre, Sarah McIntyre, Cristina Mei, Angela Morgan, Angelina Kakooza-Mwesige, Domenico M. Romeo, Katherine Sanchez, Alicia Spittle, Roberta Shepherd, Marelle Thornton, Jane Valentine, Roslyn Ward, Koa Whittingham, Alieh Zamany, Iona Novak, Early Intervention for Children Aged 0 to 2 Years With or at High Risk of Cerebral Palsy, 2021, 175, 2168-6203, 846, 10.1001/jamapediatrics.2021.0878 |