PMCCPMCCPMCC

Search tips
Search criteria 

Advanced

 
Logo of frontagineuroLink to Publisher's site
 
Front Aging Neurosci. 2012; 4: 30.
Published online 2012 November 23. Prepublished online 2012 October 8. doi:  10.3389/fnagi.2012.00030
PMCID: PMC3504955

Musical experience strengthens the neural representation of sounds important for communication in middle-aged adults

Abstract

Older adults frequently complain that while they can hear a person talking, they cannot understand what is being said; this difficulty is exacerbated by background noise. Peripheral hearing loss cannot fully account for this age-related decline in speech-in-noise ability, as declines in central processing also contribute to this problem. Given that musicians have enhanced speech-in-noise perception, we aimed to define the effects of musical experience on subcortical responses to speech and speech-in-noise perception in middle-aged adults. Results reveal that musicians have enhanced neural encoding of speech in quiet and noisy settings. Enhancements include faster neural response timing, higher neural response consistency, more robust encoding of speech harmonics, and greater neural precision. Taken together, we suggest that musical experience provides perceptual benefits in an aging population by strengthening the underlying neural pathways necessary for the accurate representation of important temporal and spectral features of sound.

Keywords: auditory, brainstem, musical experience, speech in noise, aging, musicians

Introduction

Hearing speech in a noisy environment is difficult for everyone, yet older adults are particularly vulnerable to the effects of background noise (Gordon-Salant and Fitzgibbons, 1995). Given that everyday activities often occur in noisy environments, speech-in-noise perception is an important aspect of daily communication. Indeed, difficulty hearing in noise is one of the top complaints of older adults (Tremblay et al., 2003; Yueh et al., 2003). Additionally, their reduced ability to hear in noise may lead to the avoidance of social situations where noise is present, resulting in social isolation and decreased quality of life (Heine and Browning, 2002). With widespread population aging (Vincent and Velkoff, 2010), it is becoming increasingly pressing to understand the age-related changes in communication skills as well as the underlying biology that contributes to these communication problems.

Aging has a pervasive impact on the neural encoding of sound, with delayed neural responses and decreased neural precision (Walton et al., 1998; Burkard and Sims, 2001; Finlayson, 2002; Tremblay et al., 2003; Lister et al., 2011; Parthasarathy and Bartlett, 2011; Recanzone et al., 2011; Vander Werff and Burns, 2011; Wang et al., 2011; Anderson et al., 2012; Konrad-Martin et al., 2012; Parbery-Clark et al., 2012). While it was once thought that these effects were an obligatory trajectory of aging, an increasing body of work contradicts this notion (Thomas and Baker, 2012). Instead, studies using animal models have suggested that windows of critical period plasticity can be reopened for learning (Zhou et al., 2011) and that age-related declines are reversed with training (de Villers-Sidani et al., 2010). Recently, we demonstrated that lifelong musical training similarly prevents such declines (Parbery-Clark et al., 2012), suggesting that intensive auditory experience may act in some capacity as an “aging antidote.” The study of aging musicians may therefore inform what constitutes “optimal aging,” fostering the development of remediation strategies.

Intensive auditory experience, such as that offered by musical training, enhances brain systems underlying the neural encoding of communication sounds (Pantev et al., 2003; Fujioka et al., 2004; Schon et al., 2004; Shahin et al., 2005; Magne et al., 2006; Moreno and Besson, 2006; Marques et al., 2007; Musacchia et al., 2007; Lee et al., 2009; Tervaniemi et al., 2009; as reviewed in Kraus and Chandrasekaran, 2010; Besson et al., 2011; Bidelman et al., 2011a,b; Chobert et al., 2011; Marie et al., 2011; Shahin, 2011), including those aspects of neural encoding that are crucial for hearing in noise in young adults and children (Parbery-Clark et al., 2009a, 2011b; Bidelman and Krishnan, 2010; Strait et al., in press). Despite evidence for a speech-in-noise advantage in older adult musicians (Parbery-Clark et al., 2011a; Zendel and Alain, 2011), the mechanism through which musical experience impacts the neural encoding of speech in noise in an older population is poorly understood. Here, we aimed to delineate the effects of musical experience on the neural encoding of speech in noise by assessing speech-evoked auditory brainstem responses (ABRs) in quiet and noise in a middle-aged population of musicians and nonmusicians.

We focused our analyses on neural response timing, spectral encoding, and phase-locking to the stimulus, both in terms of the temporal envelope and higher-frequency components, because these elements decline with age (Anderson et al., 2012; Parbery-Clark et al., 2012; Ruggles et al., 2012), yet are enhanced in young musicians (Musacchia et al., 2007; Parbery-Clark et al., 2009a; Strait et al., 2012). Additionally these particular neural response components are important contributors to speech-in-noise perception in young adults and children. For example, there is a well-defined relationship between neural response timing and hearing in noise, with earlier response timing relating with improved speech-in-noise perception (Parbery-Clark et al., 2009a; Anderson et al., 2010). We also know that accurately perceiving and encoding the timbral structure unique to an individual's voice facilitates the creation of an auditory object (Griffiths and Warren, 2004; Shinn-Cunningham and Best, 2008) and its subsequent segregation from competing auditory streams (Iverson, 1995). Timbre perception is driven by both envelope and harmonic encoding (Krimphoff et al., 1994; McAdams et al., 1995) with both of these components known to play a role in hearing in noise (Swaminathan and Heinz, 2012; Strait et al., in press). As such, we hypothesized that musicians have enhanced neural encoding of the spectral and temporal components of the speech stimulus, resulting in a more precise neural representation of this signal. We were also interested in defining the relationship between these neural measures and indices of speech-in-noise perception in middle-aged listeners. To this aim we administered standardized (i.e., Hearing in Noise Test (HINT); Nilsson et al., 1994), and subjective (i.e., self-report questionnaire of perceived difficulties hearing in noise; Gatehouse and Noble, 2004) measures of hearing in noise, predicting that the neural measures would relate to speech-in-noise performance, providing at least a partial explanation for the middle-aged musician advantage for hearing in noise.

Methods

Participants

Forty-eight middle-aged adults (45–65 years, mean age 56 ± 5 years) participated. All subjects had normal hearing for octave frequencies from 0.125–4 kHz bilaterally ≤20 dB HL, pure-tone average ≤10 dB HL. Participants had no history of neurological or learning disorders, did not have asymmetric pure-tone thresholds (defined as ≥15 dB difference at two or more frequencies between ears) and demonstrated normal click-evoked ABRs (wave V latency ≤6.8 ms at 80 dB SPL). No participant reported a history of chemotherapy, taking ototoxic medications, major surgeries, or head trauma. In addition, all participants were native English speakers and had normal non-verbal IQ: Abbreviated Wechsler's Adult Scale of Intelligence's matrix reasoning subtest, (Wechsler, 1999). All experimental procedures were approved by the Northwestern University Institutional Review Board and participants provided informed consent.

Twenty-three subjects were categorized as musicians, having started musical training before the age of nine and consistently engaged in musical activities a minimum of three times a week throughout their lifetimes. For information relating to participants' music practice histories, see Table Table1.1. Twenty-five subjects were categorized as nonmusicians with 17 having had no musical training and eight having fewer than 4 years of musical experience. The groups did not differ in age, hearing thresholds, sex, or non-verbal IQ (all P > 0.1; Table Table2).2). Participants were also matched on measures of physical activity [F(1, 47) = 0.032, p = 0.858], assessed by asking participants to describe the type and quantity of physical activity they engaged in each week. To account for varying types of physical activity, “walking” and “biking” were given half values while “running,” “weight training,” and other more vigorous activities were given full values. From these values, the total hours of physical activity per week was calculated for each participant. Participants were then assigned a value based on their overall activity level: 0 (<1 h/week), 1 (1–2 h/week), 2 (2–3 h/week), 3 (3–4 h/week), or 4 (>4 h/week); Table Table2.2. Two musicians and three nonmusicians were left-handed. In terms of alcohol consumption, 4 musicians and 6 nonmusicians reported never drinking.

Table 1
Participants' musical practice history.
Table 2
Participant characteristics: means (with SDs) for the musician and nonmusician groups are listed for age, pure-tone averages (0.5–4 kHz HL), click wave V latencies, non-verbal IQ percentiles (WASI Matrix Reasoning Subtest), and physical activity ...

Electrophysiology

Stimulus

The speech stimulus was a 170 ms six-formant speech syllable /da/ synthesized at a 20 kHz sampling rate. This syllable has a steady fundamental frequency (F0 = 100 Hz) except for an initial 5 ms (onset) burst. During the first 50 ms (transition from the stop burst /d/ to the vowel /a/) the lower three formants change over time (F1, 400–720 Hz; F2, 1700–1240 Hz; F3, 2580–2500 Hz) but stabilize for the 120 ms steady-state vowel. The upper three formants are constant throughout (F4, 3300 Hz; F5, 3750 Hz; F6, 4900 Hz; See Figure Figure1).1). The /da/ was chosen because it combines a transient (the /d/) and periodic (the /a/) segment, two acoustic features which have been extensively studied using auditory brainstem responses (ABRs) (Skoe and Kraus, 2010). Additionally, stop consonants pose perceptual challenges to both young and older listeners (Miller and Nicely, 1955; Ohde and Abou-Khalil, 2001).

Figure 1
Stimulus waveform (A), spectrogram (B), and group average response (C) for the speech syllable /da/. The group average response plotted is the older musician response in quiet.

Electrophysiologic recording parameters and procedure. ABRs were differentially recorded at a 20 kHz sampling rate using Ag-AgCl electrodes in a vertical montage (Cz active, FPz ground and linked-earlobe references) in Neuroscan Acquire 4.3 (Compumedics, Inc., Charlotte, NC). Contact impedance was 2 kΩ or less across all electrodes. Stimuli were presented binaurally in alternating polarities at 80 dB SPL with an 83 ms inter-stimulus interval (Scan 2, Compumedics, Inc.) through ER-3 insert earphones (Etymotic Research, Inc., Elk Grove Village, IL). During the recording session (26 ± 2 min) subjects watched a silent, captioned movie of their choice to facilitate a restful state.

Data reduction. Responses were band-pass filtered offline from 70 to 2000 Hz in MATLAB (12 dB/octave, zero phase-shift; The Mathworks, Inc., Natick, MA) and epoched using a −40 to 213 ms time window referenced to stimulus onset. Any sweep with an amplitude beyond ±35 μV was considered artifact and rejected, resulting in a total of 6000 response trials for each subject. The responses of the two polarities were added to minimize the influence of cochlear microphonic and stimulus artifact on the response (Aiken and Picton, 2008). Response amplitudes were baseline corrected to the prestimulus period.

Timing

We manually identified peaks in the subcortical responses generated by synchronous neural firing to the speech syllable /da/. The identification provides each peak's latency and amplitude. Peaks were labeled according to stimulus onset at time 0 ms such that a peak occurring at ~33–34 ms after onset would be called Peak 33. The first major peak, in response to the onset of the sound, was identified as Peak 9, those that correspond to the transition were peaks 33, 43, and 53, and to the vowel were peaks 63–163 at 10 ms intervals (Figure (Figure2).2). Two individuals who were blind to participant group independently identified each peak. An additional peak-picker confirmed peak identification and resolved disagreement between the two.

Figure 2
Average brainstem responses to /da/ in musician (red) and nonmusician (black) middle-aged adults in quiet (A) and noise (B). In quiet, musicians had earlier neural response timing for the onset and transition portion; in noise, musicians had earlier neural ...

All participants had distinct transition and vowel peaks for both the quiet and noise conditions, but onsets were absent for one participant (a nonmusician) in the quiet condition and for four participants (1 musician, 3 nonmusicians) in noise. Statistical analyses for onset peak latencies only included participants with discernible peaks in both quiet and noise (n = 44). For correlational analyses between peak timing and speech-in-noise perception, composite peak timing scores were created for the transition and the vowel regions. These composite scores were calculated by taking the average latency of peaks 33–53 for the transition and 63–163 for the vowel which when reported are denoted as transitionmean and vowelmean.

Spectral representation: fundamental frequency and harmonics

The neural encoding of the stimulus spectrum was calculated using a fast Fourier transform in MATLAB. The average spectral amplitudes relating to the transition (20–60 ms) and the vowel (60–170 ms) regions were determined by using 20-Hz bins centered around the frequencies of interest which included the fundamental frequency (F0) and its subsequent integer harmonics H2–H10 (200–1000 Hz, whole integer multiples of the F0). These values were used for all statistical analyses except for correlations for which we created a composite harmonic score by averaging the H2–H10 bins, representing the strength of overall harmonic encoding.

Stimulus-to-response—envelope analyses and waveform correlation

To measure the effect of noise on the neural response we employed two types of stimulus-to-response correlations. The first was to assess the effect of noise on the global envelope encoding by calculating the degree of correlation between the envelope of the stimulus and each participant's neural envelope encoding in the quiet and noise conditions. The second was to assess the effect of noise on neural response morphology by calculating the degree of similarity between the stimulus waveform and each participant's neural response in both the quiet and noise conditions. For this second analysis, two time ranges were chosen corresponding to the transition and the vowel. In both cases, we band-pass filtered the stimulus to match the brainstem response characteristics (70–2000 Hz). For the envelope analyses we obtained the broadband amplitude envelopes by performing a Hilbert transform on the stimulus and response waveforms and low-pass filtering at 200 Hz. To calculate the correlations between the stimulus and the responses we used the xcorr function in MATLAB (Skoe and Kraus, 2010). In both cases, the degree of similarity was calculated by shifting the stimulus waveform over 7–12 ms range relative to the regions of interest until a maximum correlation value was found. The 7–12 ms time lag was chosen because it encompasses the stimulus transmission delay (from the ER-3 transducer and ear insert ~1.1 ms) and the neural lag between the cochlea and the rostral brainstem. Average r-values were Fisher transformed for statistical analysis. Higher r-values indicate greater degrees of correlation.

Response consistency. Response consistency was calculated across trials over the length of the recording period (i.e., 6000 sweeps) by creating a composite response consistency score for each subject. Specifically, we created 300 randomly selected pairs of 3000, non-overlapping sweep sub-averages. To determine the degree of similarity between the individual pair sets, each pair of sub-averages was cross-correlated in MATLAB to generate a Pearson's correlation coefficient. This process was performed for each of the 300 pairs and the final value represents the average of the 300 correlation values. Response consistency was computed for the two time regions of interest: the transition and the vowel. Average r-values were Fisher transformed for statistical analysis. Higher r-values indicate greater degrees of correlation.

Hearing in noise ability. We used the Hearing in Noise Test (HINT; Bio-logic Systems Corp; Mundelein, IL) (Nilsson et al., 1994) to assess speech perception in noise. HINT is an adaptive test of speech recognition that measures speech perception ability in noise. During this test participants repeated short and semantically and syntactically simple sentences (e.g., she stood near the window) that were presented in speech-shaped background noise. Speech stimuli consisted of Bamford-Kowal-Bench (Bench et al., 1979) sentences (12 lists of 20 sentences) spoken by a male and were presented in free field. Participants sat one meter from the loudspeaker from which the target sentences and the noise originated at a 0° azimuth. The noise presentation level was fixed at 65 dB SPL and the program adjusted perceptual difficulty by increasing or decreasing the intensity level of the target sentences until the threshold signal-to-noise ratio (SNR) was determined. Perceptual speech-in-noise thresholds were defined as the level difference (in dB) between the speech and the noise presentation levels at which 50% of sentences are correctly repeated. A lower SNR indicates better performance.

Self-reported hearing in noise ability. We administered the Speech subscale of the Speech, Spatial, and Qualities Questionnaire (Gatehouse and Noble, 2004) to gauge an individual's perception of their hearing in noise. This questionnaire consists of 14 questions about hearing performance in various environments using a 10-point Likert scale. See Table Table33 for a complete list of the questions.

Table 3
Means, standard deviations, and significance values for the musician and nonmusicians groups' self-assessment of their speech perception and speech-in-noise abilities.

Statistical analyses

All statistical analyses were conducted in SPSS Version 18.0 (SPSS Inc., Chicago, IL). Repeated measure analyses of variance (RMANOVA) were used for group (musician vs. nonmusician) × condition (quiet vs. noise) comparisons for latency, spectral representation, stimulus-to-response correlations, envelope encoding, and response consistency. Univariate analyses of variance were used for behavioral measures. Post-hoc tests were used when appropriate. To assess relationships among variables, Pearson r correlations were used. Levene's test was used to ensure homogeneity of variance for all measures and the Shapiro-Wilk test was used to ensure that all variables were normally distributed. Bonferroni corrections for multiple comparisons were applied as appropriate; p-values reflect two-tailed tests. The SSQ (self-reported hearing in noise ability) was the only test that violated the assumption of normality. Neither log nor reciprocal transforms rendered these data normal. As such, we only used these data to quantify group differences using the non-parametric Mann-Whitney test; correlations with other variables were not explored.

Results

Summary of results

Musicians demonstrated greater speech-in-noise perception [HINT: F(1, 47) = 20.276, p < 0.005; musicians mean: −3.16, SD 0.61; nonmusicians mean: −2.34, SD 0.63] and rated themselves as having less difficulty hearing in noise than nonmusicians as assessed by the SSQ (Table (Table3).3). Musicians exhibited more robust neural encoding of speech in both quiet and noise. Musicians had earlier neural response timing, greater neural representation of the stimulus harmonics as well as more precise phase-locking to the stimulus both in terms of temporal envelope and stimulus-to-response correlations. Musicians also demonstrated less neural response degradation in noise evidenced by smaller neural timing shifts and smaller decreases in neural response consistency. We also found that specific neural measures such as earlier neural response timing and more robust brainstem responses to speech correlated with better speech-in-noise performance as measured by HINT.

Timing

Musicians demonstrated enhanced onset and transition timing in quiet and limited degradative effects of background noise for all aspects of neural timing. To quantify effects of musicianship and noise on neural response timing, we divided the neural response into three time regions: onset, transition, and vowel. We performed a mixed-model repeated-measures ANOVA (RMANOVA) 2 group (musician/nonmusician) × 2 condition (quiet/noise) with latencies in the three distinct time regions entered as dependent variables. Noise delayed peak timing across all time regions [onset: F(1, 42) = 98.008, p < 0.001; transition: F(1, 46) = 19.113, p < 0.001; vowel F(1, 46) = 2.375, p = 0.025]. Musicians demonstrated earlier neural response timing for both the onset [F(1, 42) = 11.080, p = 0.002] and the transition [F(1, 46) = 13.219, p < 0.001] but not for the vowel [F(1, 46) = 1.471, p = 0.185]. A significant group-by-condition interaction was found for all three time regions [onset: F(1, 42) = 4.822, p = 0.034; transition: F(1, 46) = 3.668, p < 0.019; vowel F(1, 46) = 2.053, p = 0.050]. Post-hoc tests revealed that musicians had significantly earlier responses in both quiet and noise conditions for the onset and transition [Onsetquiet: F(1, 42) = 4.521, p = 0.039; Onsetnoise: F(1, 42) = 12.720, p = 0.001; Transitionquiet: F(1, 46) = 10.459, p < 0.001; Transitionnoise: F(1, 46) = 11.786, p < 0.001], whereas for the steady-state musicians and nonmusicians were equated in quiet but musicians were earlier in noise [Vowelquiet: F(1, 46) = 1.423, p = 0.205; Vowelnoise: F(1, 46) = 1.912, p = 0.071]. In summary, musicians demonstrate earlier response timing in quiet for the onset and the transition but not the vowel. We also find that the addition of background noise delays neural responses for both groups, but that musicians' responses shifted less than those of nonmusicians (Figure (Figure22).

Spectral representation

Harmonics

For the vowel, in both quiet and noise, musicians demonstrated more robust auditory brainstem representation of the harmonics than nonmusicians; no musician advantage was found for the neural encoding of the harmonics in the transition. A 2 (musician/nonmusician) × 2 condition (quiet/noise) × 9 harmonicsH2−H10 RMANOVA revealed a main effect of noise and musicianship on responses to the vowel, with noise reducing spectral amplitudes, [F(1, 46) = 4.655, p < 0.001] and musicians having greater spectral amplitudes than the nonmusicians [F(1, 46) = 2.831, p = 0.012] but no noise × musicianship interaction [F(1, 46) = 1.476, p = 0.192]. For the transition, again noise resulted in a reduction in harmonic amplitude [F(1, 46) = 7.418, p < 0.001] but there was no musician advantage [F(1, 46) = 1.046, p = 0.423] nor a significant noise × musicianship interaction [F(1, 46) = 1.001, p = 0.456; Figure Figure33].

Figure 3
Spectral encoding for the transition (A and C) and vowel (B and D) in quiet (A and B) and noise (C and D). Musicians (red) demonstrated enhanced spectral encoding for the vowel in both quiet and noise; nonmusicians (black) had greater F0 encoding in the ...

Fundamental frequency (F0)

For the vowel, in both quiet and noise, musicians demonstrated a trend toward a greater representation of the fundamental frequency. This was not observed for responses to the transition. A 2 group (musician/nonmusician) × 2 condition (quiet/noise) RMANOVA revealed a weak trend for musicianship [F(1, 46) = 2.900, p = 0.095] but no main effect of noise [F(1, 46) = 0.089, p = 0.767] nor noise × musicianship interaction [F(1, 46) = 1.404, p = 0.242]. For the transition, there was a main effect of noise [F(1, 46) = 48.977, p < 0.001], no main effect of musicianship [F(1, 46) = 0.004, p = 0.300] but a significant interaction [F(1, 46) = 7.063, p = 0.011]. Post-hoc tests revealed that nonmusicians had greater representation of the F0 in quiet [F(1, 46) = 4.103, p = 0.049] but not in noise [F(1, 46) = 0.070, p = 0.792; Figure Figure33].

Stimulus to response

Envelope analyses

In both quiet and noise, musicians had better neural representation of the stimulus envelope [Figure [Figure4;4; F(1, 46) = 23.893, p < 0.001; Table Table4].4]. Noise had a significant effect on envelope encoding, in that for both groups, envelope encoding got stronger in noise [F(1, 46) = 4.665, p = 0.036; Table Table5].5]. No significant noise × musicianship interaction was found [F(1, 46) = 0.071, p = 0.792].

Figure 4
Envelope correlations between the stimulus (A) and the responses from the two conditions: quiet (B and C) and noise (D and E). The neural encoding of the stimulus envelope was greater in musicians (red) than nonmusicians (black) for both quiet and noise. ...
Table 4
Stimulus-to-response (envelope and waveform) correlation values (Pearson r): means (with SDs) for the musicians and nonmusicians across the relevant time regions.
Table 5
Response consistency scores (Pearson r-values): means (with SDs) for the musicians and nonmusicians across the transition and the vowel.

Waveform correlation

Musicians demonstrated more precise neural representation of the vowel in both quiet and noise [F(1, 46) = 20.290, p < 0.001; Table Table4].4]. The addition of background noise degraded neural response morphology [F(1, 46) = 5.492, p = 0.023], but no significant interaction was present [F(1, 46) = 0.504, p = 0.481]. For the transition, no effect of noise, [F(1, 46) = 5.492, p = 0.429], musicianship, [F(1, 46) = 1.584, p = 0.215], or a significant interaction [F(1, 46) = 0.504, p = 0.522] was found, suggesting that this particular analytical measure did not capture the degradation caused by noise in this time region.

Response consistency

Musicians had greater neural response consistency in both quiet and noise for the vowel [F(1, 46) = 13.488, p = 0.001], despite the addition of noise resulting in a decline in response consistency for both groups [F(1, 46) = 5.795, p < 0.020]. No significant noise × group interaction was present [F(1, 46) = 0.022, p = 0.882]. For the transition, noise reduced response consistency [F(1, 46) = 67.884, p < 0.001]; yet musicians did not demonstrate an overall enhancement in both quiet and noise conditions [F(1, 46) = 0.803, p = 0.375]. Rather, there was a trending interaction [F(1, 46) = 3.072, p = 0.086] with musicians and nonmusicians having equivalent response consistency in quiet [F(1, 46) = 0.033, p = 0.856] but musicians having marginally greater response consistency in noise [F(1, 46) = 3.133, p = 0.083; Table Table55].

Brainstem-hearing in noise relationships

Brainstem measures in both quiet and noise related to speech-in-noise perception as measured by HINT. The accuracy with which the ABR represented the envelope of the speech sound related to HINT (envelopequiet: r = −0.278, p = 0.05; envelopenoise: r = −0.346, p = 0.016). In all cases, earlier neural response latencies (Figure (Figure5,5, Table Table6)6) and greater SRvowel correlations (Figure (Figure6)6) were associated with better HINT scores. SRtransition correlations in quiet and noise were not related to speech-in-noise perception (all p > 0.1).

Figure 5
Relationships between speech-in-noise performance and brainstem response timing. Earlier neural response timing in the transition for both the quiet (A) and noise (B) conditions is associated with better hearing in noise. Similar relationships (not plotted ...
Table 6
Correlations (with significance levels) between peak latency for the onset, transition, and vowel peaks for the two conditions (i.e., Quiet and Noise) and HINT.
Figure 6
Relationships between speech-in-noise performance and stimulus-to-response waveform (i.e., vowel) correlations. Better hearing in noise was associated with higher stimulus-to-response correlations in quiet (A) and noise (B), suggesting that greater precision ...

Response consistency also related to HINT. In quiet, the RCvowel related with speech-in-noise perception (r = −0.307, p = 0.034) but not the RCtransition (r = −0.202, p = 0.169). In noise, the RCtransition related with hearing in noise (r = −0.291, p = 0.045) but not the vowel (r = −0.185, p = 0.208). Lastly, neither the representation of the F0 nor the harmonics directly related to speech-in-noise performance (all p > 0.1).

Discussion

Here we show that middle-aged musicians have greater neural fidelity of the stimulus with faster neural response timing, better envelope encoding, greater neural representation of the stimulus harmonics as well as less neural degradation with the addition of background noise. These subcortical measures are all associated with better speech perception in noise. Furthermore, we reveal that middle-aged musicians rate their speech-in-noises abilities higher than nonmusicians, suggesting that musicians' communication skills are higher than nonmusicians in real-world listening environments. Taken together, these results indicate that musical experience in an older adult population is associated with more precise neural responses and greater resistance to the deleterious effects of background noise.

More precise neural encoding relates with speech-in-noise perception

Hearing in noise relies on the ability to distinguish and track the target voice from the background noise, and recognizing the distinct timbral signature of a person is a key way to achieve this. Envelope and harmonic cues contribute to timbre (Krimphoff et al., 1994; McAdams et al., 1995), making them an important component of the neural code.

Envelope encoding, like stimulus-to-response correlations, is thought to represent the neural encoding of mid-to-high frequency neurons (Dau, 2003; Parbery-Clark et al., 2009a; Ruggles et al., 2012), thus providing a direct link between envelope encoding, the neural representation of higher harmonics and timbre. Furthermore, robust envelope and stimulus-to-response correlations are behaviorally relevant in that they facilitate listening in complex environments such as in background noise (Parbery-Clark et al., 2009a; Swaminathan and Heinz, 2012) or reverberant environments (Ruggles et al., 2012). Our results indicate that middle-aged musicians have stronger representation of envelope, stimulus-to-response and harmonic encoding than nonmusicians and we believe that the strengthened encoding of these spectral features may afford musicians the ability to better discern and segregate voices, giving them an advantage for speech-in-noise perception (Parbery-Clark et al., 2009a,b, 2011a; Zendel and Alain, 2011; Strait et al., in press). Throughout their training and subsequent musical experience, musicians spend countless hours attending to spectrally rich musical sounds, learning to use subtle differences in acoustic cues to discriminate instruments. Spectral information is of great behavioral relevance for musicians, with young adult musicians detecting slight harmonic differences as well as having a greater neural representation of harmonics (Koelsch et al., 1999; Shahin et al., 2005; Musacchia et al., 2008; Lee et al., 2009; Parbery-Clark et al., 2009a; Zendel and Alain, 2009). Our results indicate that older musicians also have a greater neural representation of the harmonics than nonmusicians suggesting that musical experience maintains spectral encoding despite the general trajectory of decline in the ability of the nervous system to represent spectral cues across the lifespan (Clinard et al., 2010; Ruggles et al., 2011, 2012; Anderson et al., 2012).

Middle-aged musicians demonstrate enhanced neural timing of speech in both quiet and noise—as has been found in child musicians (Strait et al., in press), whereas young adult musicians (Parbery-Clark et al., 2009a) only exhibit these enhancements in the more challenging of the two conditions—in noise. In explaining the developmental trajectory between child musicians to young adults, we propose that musical training during childhood accelerates the developmental trajectory of neural mechanisms underpinning the neural encoding of sound, as demonstrated by earlier neural response timing in child musicians, whether it be in the presence or absence of background noise (Strait et al., in press). By young adulthood, we suggest that nonmusicians have “caught up” with the musicians in that both groups are equated for response timing in quiet, even though musicians are still earlier in noise (Parbery-Clark et al., 2009a; Strait et al., in press). Here we extend this proposal to suggest that on the other side of the life cycle—that of aging—musical experience prevents declines in neural mechanisms that underlie neural encoding irrespective of the listening environment.

Our results highlight faster response timing in middle-aged musicians for the onset and transition—two parts of the response that decline with both age (Anderson et al., 2012; Parbery-Clark et al., 2012) and the introduction of noise (Cunningham et al., 2002; Parbery-Clark et al., 2009a; Anderson et al., 2010), are the most challenging in terms of perception (Miller and Nicely, 1955) and neural encoding (Anderson et al., 2010). Importantly, in quiet, there were no group differences for the vowel, indicating that the middle-aged nonmusician's neural responses are not globally delayed for response timing; rather, these effects were exclusively found in the response to the most complex portions of the sound (Parbery-Clark et al., 2012). The addition of background noise did result in a general delay for both groups; still, musicians' responses were delayed to a lesser extent. Musicians' decreased neural response degradation in noise was further evidenced by more consistent neural responses. Taken together, our results provide evidence for musical training across the life span having a pervasive effect on sensory and neural processing, maintaining neural function both in quiet and noisy conditions.

Musicians: model of aging

To date, the majority of research supporting the use of musicians as a model of plasticity has focused on child or young adult populations (for review see: Münte et al., 2002; Zatorre and McGill, 2005; Habib and Besson, 2009; Kraus and Chandrasekaran, 2010). While this work has increased our understanding of the effects of music on the nervous system, the role of musical training in the older normal hearing adult population remains largely unexplored. Given that musical training strengthens those skills that decline with age, we argue that the musician's brain provides an optimal model for studying the effects of age on the nervous system. Aging declines are thought to start as early as middle age (Salthouse et al., 1996; Helfer and Vargo, 2009; Ruggles et al., 2011, 2012; Parbery-Clark et al., 2012) and are accompanied by a decrease in central nervous system function, which holds important implications for perceptual and cognitive skills (Craik and Salthouse, 2000). Given that aging musicians maintain an advantage over nonmusicians in terms of neural processing (Parbery-Clark et al., 2012), auditory perception (Parbery-Clark et al., 2011a; Zendel and Alain, 2011) and cognitive abilities (Hanna-Pladdy and MacKay, 2011; Parbery-Clark et al., 2011a; Hanna-Pladdy and Gajewski, 2012), older musicians may provide a means to better understand what contributes to successful aging.

The application of musical experience to the study of aging requires knowledge of the effects of aging on the nervous system. One of the major neurophysiological hallmarks of aging is delayed neural timing and decreased temporal processing (Walton et al., 1998; Burkard and Sims, 2001; Frisina, 2001; Finlayson, 2002; Tremblay et al., 2003; Frisina and Walton, 2006; Lister et al., 2011; Parthasarathy and Bartlett, 2011; Recanzone et al., 2011; Vander Werff and Burns, 2011; Wang et al., 2011; Anderson et al., 2012; Konrad-Martin et al., 2012; Parbery-Clark et al., 2012). These age-related deficits are caused, at least in part, by a decrease in inhibitory mechanisms. With aging, the inhibitory neurotransmitters that facilitate the accurate neural encoding of temporally dynamic and complex sounds (Walton et al., 1998; Caspary et al., 2002, 2008) as well as response consistency (Pichora-Fuller and Schneider, 1992) are reduced throughout the auditory pathway (Caspary et al., 1995, 2005; Wang et al., 2009; de Villers-Sidani et al., 2010; Hughes et al., 2010; Juarez-Salinas et al., 2010). Because the ABR requires a high-degree of neural synchronicity (Kraus et al., 2000), decreased neural consistency such as that caused by temporal jitter (Pichora-Fuller et al., 2007) or neural response variability (Turner et al., 2005; Yang et al., 2009) associated with aging can also contribute to delayed neural response timing and reduced spectral encoding (Anderson et al., 2012). Here we present musician advantages for neural response timing, spectral encoding, and neural response consistency—all factors known to decline with age. For these reasons, we propose that the study of the older musician may be beneficial in elucidating the specific neural components that are enhanced relative to nonmusicians or impervious to age-related declines—highlighting which aspects may be amenable to rehabilititation.

Future directions

We document enhanced neural encoding in a normal hearing, middle-aged adult musician population. Because aging also results in a higher prevalence of hearing loss, it will be important to define how musical experience interacts in an older adult population with sensory hearing loss. Additionally, our earlier work demonstrated that young adult musicians (19–30 years) had minimal neural differences in quiet (Parbery-Clark et al., 2009a), yet the present results show striking group differences in a middle-aged group (45–65 years) for the same condition. Determining the time course of the neural changes that occur between these two age groups (i.e., young and middle-aged adults) will further our understanding of the effects of aging on the nervous system, as well as the role musicianship plays to offset these declines.

Conclusions

We reveal strengthened neural encoding of the important acoustic ingredients for speech perception in noise for middle-aged musicians, potentially providing a neural basis for their behavioral advantage for hearing in noise.

Conflict of interest statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

The authors wish to thank the participants who participated in this study as well as Dana Strait, Trent Nicol, and Travis White–Schwoch who provided comments on an earlier version of this manuscript. This work was supported by NSF 0842376 to Nina Kraus.

References

  • Aiken S. J., Picton T. W. (2008). Envelope and spectral frequency-following responses to vowel sounds. Hear. Res. 245, 35–47 10.1016/j.heares.2008.08.004 [PubMed] [Cross Ref]
  • Anderson S., Parbery-Clark A., White-Schwoch T., Kraus N. (2012). Aging affects neural precision of speech encoding. J. Neurosci. 32, 14156–14164 10.1523/JNEUROSCI.2176-12.2012 [PMC free article] [PubMed] [Cross Ref]
  • Anderson S., Skoe E., Chandrasekaran B., Kraus N. (2010). Neural timing is linked to speech perception in noise. J. Neurosci. 30, 4922–4926 10.1523/JNEUROSCI.0107-10.2010 [PMC free article] [PubMed] [Cross Ref]
  • Bench J., Kowal A., Bamford J. (1979). The BKB (Bamford-Kowal-Bench) sentence lists for partially-hearing children. Br. J. Audiol. 13, 108–112 [PubMed]
  • Besson M., Chobert J., Marie C. (2011). Transfer of training between music and speech: common processing, attention, and memory. Front. Psychology 2:94 10.3389/fpsyg.2011.00094 [PMC free article] [PubMed] [Cross Ref]
  • Bidelman G., Krishnan A. (2010). Effects of reverberation on brainstem representation of speech in musicians and non-musicians. Brain Res. 1355, 112–125 10.1016/j.brainres.2010.07.100 [PMC free article] [PubMed] [Cross Ref]
  • Bidelman G. M., Gandour J. T., Krishnan A. (2011a). Musicians demonstrate experience-dependent brainstem enhancement of musical scale features within continuously gliding pitch. Neurosci. Lett. 503, 203–207 10.1016/j.neulet.2011.08.036 [PMC free article] [PubMed] [Cross Ref]
  • Bidelman G. M., Krishnan A., Gandour J. T. (2011b). Enhanced brainstem encoding predicts musicians' perceptual advantages with pitch. Eur. J. Neurosci. 33, 530–538 10.1111/j.1460-9568.2010.07527.x [PMC free article] [PubMed] [Cross Ref]
  • Burkard R. F., Sims D. (2001). The human suditory brainstem tesponse to high click rates: aging effects. Am. J. Audiol. 10, 53–61 [PubMed]
  • Caspary D. M., Ling L., Turner J. G., Hughes L. F. (2008). Inhibitory neurotransmission, plasticity and aging in the mammalian central auditory system. J. Exp. Biol. 211, 1781–1791 10.1242/jeb.013581 [PMC free article] [PubMed] [Cross Ref]
  • Caspary D. M., Milbrandt J. C., Helfert R. H. (1995). Central auditory aging: GABA changes in the inferior colliculus. Exp. Gerentol. 30, 349–360 [PubMed]
  • Caspary D. M., Palombi P. S., Hughes L. F. (2002). GABAergic inputs shape responses to amplitude modulated stimuli in the inferior colliculus. Hear. Res. 168, 163–173 10.1016/S0378-5955(02)00363-5 [PubMed] [Cross Ref]
  • Caspary D. M., Schatteman T. A., Hughes L. F. (2005). Age-related changes in the inhibitory response properties of dorsal cochlear nucleus output neurons: role of inhibitory inputs. J. Neurosci. 25, 10952–10959 10.1523/JNEUROSCI.2451-05.2005 [PubMed] [Cross Ref]
  • Chobert J., Marie C., François C., Schön D., Besson M. (2011). Enhanced passive and active processing of syllables in musician children. J. Cogn. Neurosci. 23, 3874–3887 10.1162/jocn_a_00088 [PubMed] [Cross Ref]
  • Clinard C. G., Tremblay K. L., Krishnan A. R. (2010). Aging alters the perception and physiological representation of frequency: evidence from human frequency-following response recordings. Hear. Res. 264, 48–55 10.1016/j.heares.2009.11.010 [PMC free article] [PubMed] [Cross Ref]
  • Craik F. I. M., Salthouse T. A. (2000). The Handbook of Aging and Cognition II. Mahwah, NJ: Lawrence Erlbaum Associates
  • Cunningham J., Nicol T., King C., Zecker S. G., Kraus N. (2002). Effects of noise and cue enhancement on neural responses to speech in auditory midbrain, thalamus and cortex. Hear. Res. 169, 97–111 10.1016/S0378-5955(02)00344-1 [PubMed] [Cross Ref]
  • Dau T. (2003). The importance of cochlear processing for the formation of auditory brainstem and frequency following responses. J. Acoust. Soc. Am. 113, 936 10.1121/1.1534833 [PubMed] [Cross Ref]
  • de Villers-Sidani E., Alzghoul L., Zhou X., Simpson K. L., Lin R. C. S., Merzenich M. M. (2010). Recovery of functional and structural age-related changes in the rat primary auditory cortex with operant training. Proc. Natl. Acad. Sci. U.S.A. 107, 13900–13905 10.1073/pnas.1007885107 [PubMed] [Cross Ref]
  • Finlayson P. G. (2002). Paired-tone stimuli reveal reductions and alterations in temporal processing in inferior colliculus neurons of aged animals. J. Assoc. Res. Otolaryngol. 3, 321–331 10.1007/s101620020038 [PMC free article] [PubMed] [Cross Ref]
  • Frisina R. D. (2001). Subcortical neural coding mechanisms for auditory temporal processing. Hear. Res. 158, 1–27 10.1016/S0378-5955(01)00296-9 [PubMed] [Cross Ref]
  • Frisina R. D., Walton J. P. (2006). Age-related structural and functional changes in the cochlear nucleus. Hear. Res. 217, 216–233 10.1016/j.heares.2006.02.003 [PubMed] [Cross Ref]
  • Fujioka T., Trainor L. J., Ross B., Kakigi R., Pantev C. (2004). Musical training enhances automatic encoding of melodic contour and interval structure. J. Cogn. Neurosci. 16, 1010–1021 10.1162/0898929041502706 [PubMed] [Cross Ref]
  • Gatehouse S., Noble W. (2004). The speech, spatial and qualities of hearing scale (SSQ). Int. J. Audiol. 43, 85–99 [PubMed]
  • Gordon-Salant S., Fitzgibbons P. J. (1995). Comparing recognition of distorted speech using an equivalent signal-to-noise ratio index. J. Speech Hear. Res. 38, 706–713 [PubMed]
  • Griffiths T. D., Warren J. D. (2004). What is an auditory object? Nat. Rev. Neurosci. 5, 887–892 10.1038/nrn1538 [PubMed] [Cross Ref]
  • Habib M., Besson M. (2009). What do music training and musical experience teach us about brain plasticity? Music Percept. 26, 279–285
  • Hanna-Pladdy B., Gajewski B. (2012). Recent and past musical activity predicts cognitive aging variability: direct comparison with general lifestyle activities. Front. Hum. Neurosci. 6:198 10.3389/fnhum.2012.00198 [PMC free article] [PubMed] [Cross Ref]
  • Hanna-Pladdy B., MacKay A. (2011). The relation between instrumental musical activity and cognitive aging. Neuropsychology 25, 378 10.1037/a0021895 [PubMed] [Cross Ref]
  • Heine C., Browning C. J. (2002). Communication and psychosocial consequences of sensory loss in older adults: overview and rehabilitation directions. Disabil. Rehabil. 24, 763–773 10.1080/09638280210129162 [PubMed] [Cross Ref]
  • Helfer K., Vargo M. (2009). Speech recognition and temporal processing in middle-aged women. J. Am. Acad. Audiol. 20, 264–271 [PMC free article] [PubMed]
  • Hughes L. F., Turner J. G., Parrish J. L., Caspary D. M. (2010). Processing of broadband stimuli across A1 layers in young and aged rats. Hear. Res. 264, 79–85 10.1016/j.heares.2009.09.005 [PMC free article] [PubMed] [Cross Ref]
  • Iverson P. (1995). Auditory Stream Segmentation by musical timbre: effects of static and dynamic acoustic attributes. J. Exp. Psychol. Hum. Percept. Perform. 21, 751–763 [PubMed]
  • Juarez-Salinas D. L., Engle J. R., Navarro X. O., Recanzone G. H. (2010). Hierarchical and serial processing in the spatial auditory cortical pathway is degraded by natural aging. J. Neurosci. 30, 14795–14804 10.1523/JNEUROSCI.3393-10.2010 [PMC free article] [PubMed] [Cross Ref]
  • Koelsch S., Schröger E., Tervaniemi M. (1999). Superior attentive and pre-attentive auditory processing in musicians. Neuroreport 10, 1309–1313 [PubMed]
  • Konrad-Martin D., Dille M. F., McMillan G., Griest S., McDermott D., Fausti S. A., et al. (2012). Age-related changes in the auditory brainstem response. J. Am. Acad. Audiol. 23, 18–35 10.3766/jaaa.23.1.3 [PubMed] [Cross Ref]
  • Kraus N., Bradlow M. A., Cunningham C. J., King C. D., Koch D. B., Nicol T. G., et al. (2000). Consequences of neural asynchrony: a case of AN. J. Assoc. Res. Otolaryngol. 1, 33–45 [PMC free article] [PubMed]
  • Kraus N., Chandrasekaran B. (2010). Music training for the development of auditory skills. Nat. Rev. Neurosci. 11, 599–605 10.1038/nrn2882 [PubMed] [Cross Ref]
  • Krimphoff J., McAdams S., Winsberg S. (1994). Caractérisation du timbre des sons complexes. II: analyses acoustiques et quantification psychophysique. J. Phys. 4, 625–628
  • Lee K. M., Skoe E., Kraus N., Ashley R. (2009). Selective subcortical enhancement of musical intervals in musicians. J. Neurosci. 29, 5832–5840 10.1523/JNEUROSCI.6133-08.2009 [PubMed] [Cross Ref]
  • Lister J. J., Roberts R. A., Lister F. L. (2011). An adaptive clinical test of temporal resolution: age effects. Int. J. Audiol. 50, 367–374 10.3109/14992027.2010.551218 [PubMed] [Cross Ref]
  • Magne C., Schön D., Besson M. (2006). Musician children detect pitch violations in both music and language better than nonmusician children: behavioral and electrophysiological approaches. J. Cogn. Neurosci. 18, 199–211 10.1162/089892906775783660 [PubMed] [Cross Ref]
  • Marie C., Magne C., Besson M. (2011). Musicians and the metric structure of words. J. Cogn. Neurosci. 23, 294–305 10.1162/jocn.2010.21413 [PubMed] [Cross Ref]
  • Marques C., Moreno S., Castro S. L., Besson M. (2007). Musicians detect pitch violation in a foreign language better than nonmusicians: behavioral and electrophysiological evidence. J. Cogn. Neurosci. 19, 1453–1463 10.1162/jocn.2007.19.9.1453 [PubMed] [Cross Ref]
  • McAdams S., Winsberg S., Donnadieu S., de Soete G., Krimphoff J. (1995). Perceptual scaling of synthesized musical timbres: common dimensions, specificities, and latent subject classes. Psychol. Res. 58, 177–192 [PubMed]
  • Miller G. A., Nicely P. E. (1955). An analysis of perceptual confusions among some English consonants. J. Acoust. Soc. Am. 27, 338–352
  • Moreno S., Besson M. (2006). Musical training and language related brain electrical activity in children. Psychophysiology 43, 287–291 10.1111/j.1469-8986.2006.00401.x [PubMed] [Cross Ref]
  • Münte T. F., Altenmuller E., Jancke L. (2002). The musician's brain as a model of neuroplasticity. Nat. Rev. Neurosci. 3, 473–478 10.1038/nrn843 [PubMed] [Cross Ref]
  • Musacchia G., Sams M., Skoe E., Kraus N. (2007). Musicians have enhanced subcortical auditory and audiovisual processing of speech and music. Proc. Natl. Acad. Sci. U.S.A. 104, 15894–15898 10.1073/pnas.0701498104 [PubMed] [Cross Ref]
  • Musacchia G., Strait D., Kraus N. (2008). Relationships between behavior, brainstem and cortical encoding of seen and heard speech in musicians. Hear. Res. 241, 34–42 10.1016/j.heares.2008.04.013 [PMC free article] [PubMed] [Cross Ref]
  • Nilsson M., Soli S., Sullivan J. (1994). Development of the hearing in noise test for the measurement of speech reception thresholds in quiet and in noise. J. Acoust. Soc. Am. 95, 1085–1099 10.1121/1.408469 [PubMed] [Cross Ref]
  • Ohde R. N., Abou-Khalil R. (2001). Age differences for stop-consonant and vowel perception in adults. J. Acoust. Soc. Am. 110, 2156–2166 10.1121/1.1399047 [PubMed] [Cross Ref]
  • Pantev C., Ross B., Fujioka T., Trainor L. J., Schulte M., Schulz M. (2003). Music and learning-induced cortical plasticity. Ann. N.Y. Acad. Sci. 999, 438–450 10.1196/annals.1284.054 [PubMed] [Cross Ref]
  • Parbery-Clark A., Anderson S., Hittner E., Kraus N. (2012). Musical experience offsets age-related delays in neural timing. Neurobiol. Aging 33, 1483 10.1016/j.neurobiolaging.2011.12.015 [PubMed] [Cross Ref]
  • Parbery-Clark A., Skoe E., Kraus N. (2009a). Musical experience limits the degradative effects of background noise on the neural processing of sound. J. Neurosci. 29, 14100–14107 10.1523/JNEUROSCI.3256-09.2009 [PubMed] [Cross Ref]
  • Parbery-Clark A., Skoe E., Lam C., Kraus N. (2009b). Musician enhancement for speech in noise. Ear Hear. 30, 653 10.1097/AUD.0b013e3181b412e9 [PubMed] [Cross Ref]
  • Parbery-Clark A., Strait D. L., Anderson S., Hittner E., Kraus N. (2011a). Musical experience and the aging auditory system: implications for cognitive abilities and hearing speech in noise. PLoS ONE 6:e18082 10.1371/journal.pone.0018082 [PMC free article] [PubMed] [Cross Ref]
  • Parbery-Clark A., Strait D. L., Kraus N. (2011b). Context-dependent encoding in the auditory brainstem subserves enhanced speech-in-noise perception in musicians. Neuropsychologia 49, 3338–3345 10.1016/j.neuropsychologia.2011.08.007 [PMC free article] [PubMed] [Cross Ref]
  • Parthasarathy A., Bartlett E. L. (2011). Age-related auditory deficits in temporal processing in F-344 rats. Neuroscience 192, 619–630 10.1016/j.neuroscience.2011.06.042 [PubMed] [Cross Ref]
  • Pichora-Fuller M. K., Schneider B. A. (1992). The effect of interaural delay of the masker on masking-level differences in young and old adults. J. Acoust. Soc. Am. 91, 2129–2135 10.1121/1.403673 [PubMed] [Cross Ref]
  • Pichora-Fuller M. K., Schneider B. A., MacDonald E., Pass H. E., Brown S. (2007). Temporal jitter disrupts speech intelligibility: a simulation of auditory aging. Hear. Res. 223, 114–121 10.1016/j.heares.2006.10.009 [PubMed] [Cross Ref]
  • Recanzone G. H., Engle J. R., Juarez-Salinas D. L. (2011). Spatial and temporal processing of single auditory cortical neurons and populations of neurons in the macaque monkey. Hear. Res. 271, 115–122 10.1016/j.heares.2010.03.084 [PMC free article] [PubMed] [Cross Ref]
  • Ruggles D., Bharadwaj H., Shinn-Cunningham B. G. (2011). Normal hearing is not enough to guarantee robust encoding of suprathreshold features important in everyday communication. Proc. Natl. Acad. Sci. 108, 15516–15521 10.1073/pnas.1108912108 [PubMed] [Cross Ref]
  • Ruggles D., Bharadwaj H., Shinn-Cunningham B. G. (2012). Why middle-aged listeners have trouble hearing in everyday settings. Curr. Biol. 22, 1417–1422 10.1016/j.cub.2012.05.025 [PubMed] [Cross Ref]
  • Salthouse T. A., Hancock H. E., Meinz E. J., Hambrick D. Z. (1996). Interrelations of age, visual acuity, and cognitive functioning. J. Gerontol. B Psychol. Sci. 51, P317–P330 [PubMed]
  • Schon D., Magne C., Besson M. (2004). The music of speech: music training facilitates pitch processing in both music and language. Psychophysiology 41, 341–349 10.1162/jocn.2007.19.9.1453 [PubMed] [Cross Ref]
  • Shahin A. J. (2011). Neurophysiological influence of musical training on speech perception. Front. Psychology 2:126 10.3389/fpsyg.2011.00126 [PMC free article] [PubMed] [Cross Ref]
  • Shahin A. J., Roberts L. E., Pantev C., Trainor L. J., Ross B. (2005). Modulation of P2 auditory-evoked responses by the spectral complexity of musical sounds. Neuroreport 16, 1781–1785 [PubMed]
  • Shinn-Cunningham B., Best V. (2008). Selective attention in normal and impaired hearing. Trends Amplif. 12, 283 10.1177/1084713808325306 [PMC free article] [PubMed] [Cross Ref]
  • Skoe E., Kraus N. (2010). Auditory brain stem response to complex sounds: a tutorial. Ear Hear. 31, 302–324 10.1097/AUD.0b013e3181cdb272 [PMC free article] [PubMed] [Cross Ref]
  • Strait D., Parbery-Clark A., Hittner E., Kraus N. (in press). Musical training during early childhood enhances the neural encoding of speech in noise. Brain Lang. [Epub ahead of print]. 10.1016/j.bandl.2012.09.001 [PMC free article] [PubMed] [Cross Ref]
  • Strait D. L., Chan K., Ashley R., Kraus N. (2012). Specialization among the specialized: Auditory brainstem function is tuned in to timbre. Cortex 48, 360 10.1016/j.cortex.2011.03.015 [PubMed] [Cross Ref]
  • Swaminathan J., Heinz M. G. (2012). Psychophysiological analyses demonstrate the importance of neural envelope coding for speech perception in noise. J. Neurosci. 32, 1747–1756 10.1523/JNEUROSCI.4493-11.2012 [PMC free article] [PubMed] [Cross Ref]
  • Tervaniemi M., Kruck S., de Baene W., Schroger E., Alter K., Friederici A. D. (2009). Top-down modulation of auditory processing: effects of sound context, musical expertise and attentional focus. Eur. J. Neurosci. 30, 1636–1642 10.1111/j.1460-9568.2009.06955.x [PubMed] [Cross Ref]
  • Thomas C., Baker C. I. (2012). Teaching an adult brain new tricks: a critical review of evidence for training-dependent structural plasticity in humans. Neuroimage. [Epub ahead print]. 10.1016/j.neuroimage.2012.03.069 [PubMed] [Cross Ref]
  • Tremblay K., Piskosz M., Souza P. (2003). Effects of age and age-related hearing loss on the neural representation of speech cues. Clin. Neurophysiol. 114, 1332–1343 10.1016/S1388-2457(03)00114-7 [PubMed] [Cross Ref]
  • Turner J. G., Hughes L. F., Caspary D. M. (2005). Affects of aging on receptive fields in rat primary auditory cortex layer V neurons. J. Neurophysiol. 94, 2738–2747 10.1152/jn.00362.2005 [PubMed] [Cross Ref]
  • Vander Werff K. R., Burns K. S. (2011). Brain stem responses to speech in younger and older adults. Ear Hear. 32, 168–180 10.1097/AUD.0b013e3181f534b5 [PubMed] [Cross Ref]
  • Vincent G. K., Velkoff V. A. (2010). The Next Four Decades. The Older Population in the United States: 2010 to 2050. Washington, DC: US D.O.C.-E.A.S. Administration
  • Walton J. P., Frisina R. D., O'Neill W. E. (1998). Age-related alteration in processing of temporal sound features in the auditory midbrain of the CBA mouse. J. Neurosci. 18, 2764–2776 [PubMed]
  • Wang H., Turner J. G., Ling L., Parrish J. L., Hughes L. F., Caspary D. M. (2009). Age-related changes in glycine receptor subunit composition and binding in dorsal cochlear nucleus. Neuroscience 160, 227–239 10.1016/j.neuroscience.2009.01.079 [PMC free article] [PubMed] [Cross Ref]
  • Wang M., Wu X., Li L., Schneider B. A. (2011). The effects of age and interaural delay on detecting a change in interaural correlation: the role of temporal jitter. Hear. Res. 275, 139–149 10.1016/j.heares.2010.12.013 [PubMed] [Cross Ref]
  • Wechsler D. (1999). Wechsler Abbreviated Scale of Intelligence (WASI). San Antonio, TX: The Psychological Corporation; 10.1207/s15324826an1004_8 [Cross Ref]
  • Yang Y., Liang Z., Li G., Wang Y., Zhou Y. (2009). Aging affects response variability of V1 and MT neurons in rhesus monkeys. Brain Res. 1274, 21–27 10.1016/j.brainres.2009.04.015 [PubMed] [Cross Ref]
  • Yueh B., Shapiro N., MacLean C. H., Shekelle P. G. (2003). Screening and management of adult hearing loss in primary care: scientific review. J. Am. Med. Assoc. 289, 1976–1985 10.1001/jama.289.15.1976 [PubMed] [Cross Ref]
  • Zatorre R., McGill J. (2005). Music, the food of neuroscience? Nature 434, 312–315 10.1038/434312a [PubMed] [Cross Ref]
  • Zendel B. R., Alain C. (2009). Concurrent sound segregation is enhanced in musicians. J. Cogn. Neurosci. 21, 1488–1498 10.1162/jocn.2009.21140 [PubMed] [Cross Ref]
  • Zendel B. R., Alain C. (2011). Musicians experience less age-related decline in central auditory processing. Psychol. Aging 27, 410–417 10.1037/a0024816 [PubMed] [Cross Ref]
  • Zhou X., Panizzutti R., de Villers-Sidani É., Madeira C., Merzenich M. M. (2011). Natural restoration of critical period plasticity in the juvenile and adult primary auditory cortex. J. Neurosci. 31, 5625–5634 10.1523/JNEUROSCI.6470-10.2011 [PubMed] [Cross Ref]

Articles from Frontiers in Aging Neuroscience are provided here courtesy of Frontiers Media SA