PMCCPMCCPMCC

Search tips
Search criteria 

Advanced

 
Logo of nihpaAbout Author manuscriptsSubmit a manuscriptHHS Public Access; Author Manuscript; Accepted for publication in peer reviewed journal;
 
Epilepsia. Author manuscript; available in PMC 2009 March 1.
Published in final edited form as:
PMCID: PMC2645716
NIHMSID: NIHMS78304

A Comparison of Five FMRI Protocols for Mapping Speech Comprehension Systems

Abstract

Aims

Many fMRI protocols for localizing speech comprehension have been described, but there has been little quantitative comparison of these methods. We compared five such protocols in terms of areas activated, extent of activation, and lateralization.

Methods

FMRI BOLD signals were measured in 26 healthy adults during passive listening and active tasks using words and tones. Contrasts were designed to identify speech perception and semantic processing systems. Activation extent and lateralization were quantified by counting activated voxels in each hemisphere for each participant.

Results

Passive listening to words produced bilateral superior temporal activation. After controlling for pre-linguistic auditory processing, only a small area in the left superior temporal sulcus responded selectively to speech. Active tasks engaged an extensive, bilateral attention and executive processing network. Optimal results (consistent activation and strongly lateralized pattern) were obtained by contrasting an active semantic decision task with a tone decision task. There was striking similarity between the network of brain regions activated by the semantic task and the network of brain regions that showed task-induced deactivation, suggesting that semantic processing occurs during the resting state.

Conclusions

FMRI protocols for mapping speech comprehension systems differ dramatically in pattern, extent, and lateralization of activation. Brain regions involved in semantic processing were identified only when an active, non-linguistic task was used as a baseline, supporting the notion that semantic processing occurs whenever attentional resources are not controlled. Identification of these lexical-semantic regions is particularly important for predicting language outcome in patients undergoing temporal lobe surgery.

Keywords: language, fMRI, speech, comprehension, semantics, temporal lobe

Introduction

Localization of language areas prior to brain surgery can help determine the risk of post-operative aphasia and may be useful for modifying surgical procedures to minimize such risk. Anterior temporal lobe resection is a common and highly effective treatment for intractable epilepsy (Wiebe et al., 2001; Tellez-Zenteno et al., 2005), but carries a 30–50% risk of decline in naming ability when performed on the left temporal lobe (Hermann et al., 1994; Langfitt & Rausch, 1996; Bell et al., 2000; Sabsevitz et al., 2003). In addition to retrieval of names, the left temporal lobe is classically associated with speech comprehension (Wernicke, 1874). These seemingly different language functions both depend on common systems for processing speech sounds (phonology) and word meanings (lexical semantics), both of which are located largely in the temporal lobe (Indefrey & Levelt, 2004; Awad et al., 2007). Identification of these phonological and lexical-semantic systems is therefore an important goal in the presurgical mapping of language functions.

Functional magnetic resonance imaging (fMRI) is used increasingly for this purpose (Binder, 2006). FMRI is a safe, non-invasive procedure for localizing hemodynamic changes associated with neural activity. Many fMRI studies conducted on healthy adults have investigated the brain correlates of speech comprehension, though with a variety of activation procedures and widely varying results. There has been little systematic, quantitative comparison of these activation protocols. There is at present little agreement, for example, on which type of procedure produces the strongest activation, which is most specific for detecting processes of interest, and which is associated with the greatest degree of hemispheric lateralization.

Speech comprehension protocols can be categorized in general terms according to stimulus and task factors (Table 1). Speech is an acoustically complex stimulus that engages much of the auditory cortex bilaterally in pre-linguistic processing of spectral and temporal information (Binder et al., 2000; Poeppel, 2001). Most models of speech perception posit a late stage of auditory perception in which this complex acoustic information activates long-term representations for consonant and vowel phonemes (Pisoni, 1973; Stevens & Blumstein, 1981; Klatt, 1989). Many fMRI studies of speech perception have aimed to isolate this phonemic stage of the speech perception process by contrasting spoken words with non-speech auditory sounds that presumably 'subtract out' earlier stages of auditory processing. These non-speech control sounds have varied in terms of their similarity to speech, ranging from steady state 'noise' with no spectral or temporal information (Binder et al., 2000), to 'tones' possessing relatively simple spectral and temporal information (Démonet et al., 1992; Binder et al., 2000; Desai et al., 2005), to various synthetic speech-like sounds with spectrotemporal complexity comparable to speech (Scott et al., 2000; Dehaene-Lambertz et al., 2005; Liebenthal et al., 2005; Mottonen et al., 2006). As expected, the more similar the control sounds are to the speech sounds in terms of acoustic complexity, the less activation occurs in primary and association auditory areas in the superior temporal lobe (Binder et al., 2000; Scott et al., 2000; Davis & Johnsrude, 2003; Specht & Reul, 2003; Uppenkamp et al., 2006).

Table 1
Five types of protocols for mapping speech comprehension areas.

Another principle useful for categorizing speech comprehension studies is whether or not an active task is requested of the participants. Many studies employed passive listening to speech, whereas others required participants to respond to the speech sounds according to particular criteria. Active tasks focus participants' attention on a specific aspect of a stimulus, such as its form or meaning, which is assumed to cause 'top-down' activation of the neural systems relevant for processing the attended information. For example, some prior studies of speech comprehension sought to identify the brain regions specifically involved in processing word meanings (lexical-semantic system) by contrasting a semantic task using speech sounds with a phonological task using speech sounds (Démonet et al., 1992; Mummery et al., 1996; Binder et al., 1999).

A final factor to consider in categorizing speech comprehension studies is whether or not an active task is used for the baseline condition. Performing a task of any kind requires a variety of general functions such as focusing attention on the relevant aspects of the stimulus, holding the task instructions in mind, making a decision, and generating a motor response. Active control tasks are used to 'subtract out' these and other general processes that are considered non-linguistic in nature and therefore irrelevant for the purpose of language mapping. Another potential benefit of such tasks is that they provide a better-controlled baseline state than resting (Démonet et al., 1992). Evidence suggests that the conscious resting state is characterized by ongoing mental activity experienced as 'daydreams', 'mental imagery', 'inner speech' and the like, which is interrupted when an overt task is performed (Antrobus et al., 1966; Pope & Singer, 1976; Singer, 1993; Teasdale et al., 1993; Binder et al., 1999; McKiernan et al., 2006). It has been proposed that this 'task unrelated thought' depends on the same conceptual knowledge systems that underlie language comprehension and production of propositional language (Binder et al., 1999). Thus, resting states and states in which stimuli are presented with no specific task demands may actually be conditions in which there is continuous processing of conceptual knowledge and mental production of meaningful 'inner speech'.

Table 1 illustrates five types of contrasts that have been used to map speech comprehension systems. The first four are obtained by crossing either a passive or active task state with a control condition using either silence or a non-speech auditory stimulus. The last contrast uses speech sounds in both conditions while contrasting an active semantic task with an active phonological task. Though there are other possible contrasts not listed in the table (e.g., Active Speech vs. Passive Non-Speech), most prior imaging studies on this topic can be classified into one of these five general types. Within each type, of course, are many possible variations on both stimulus content (e.g., relative concreteness, grammatical class, or semantic category of words; sentences vs. single words) and specific task requirements.

Our aim in the current study was to provide a meaningful comparison between these five types of protocols through controlled manipulation of the factors listed in Table 1. Interpreting differences between any two of these types requires that the same or comparable word stimuli be used in each case. We used single words for each protocol, as these lend themselves readily to use in semantic tasks. Similarly, we used the same non-speech tone stimuli for the passive and active control conditions in protocols 2 and 4. Finally, because fMRI results are known to be highly variable across individuals and scanning sessions, we scanned the same 26 individuals on all five protocols in the same imaging session. The results were characterized quantitatively in terms of extent and magnitude of activation, specific brain regions activated, and lateralization of activation. The results provide the first clear picture of the relative differences and similarities, advantages and disadvantages of these speech comprehension mapping protocols.

Methods

Participants

Participants in the study were 26 healthy adults (13 men, 13 women), ranging in age from 18 to 29 years, with no history of neurologic, psychiatric, or auditory symptoms. All participants indicated strong right-hand preferences (laterality quotient > 50) on the Edinburgh Handedness Inventory (Oldfield, 1971). Participants gave written informed consent and were paid a small hourly stipend. The study received prior approval by the Medical College of Wisconsin Human Research Review Committee.

Task Conditions and Behavioral Measures

Task conditions during scanning included a resting state, a passive tone listening task, an active tone decision task, a passive word listening task, a semantic decision task, and a phoneme decision task (Table 2). Auditory stimuli were presented with a computer and a pneumatic audio system, as previously described (Binder et al., 1997). Participants kept their eyes closed during all conditions. For the resting condition, participants were instructed to remain relaxed and motionless. No auditory stimulus was presented other than the baseline scanner noise. Stimuli in the passive and active tone tasks were digitally-synthesized 500-Hz and 750-Hz sinewave tones of 150 ms duration each, separated by 250-ms inter-tone intervals. These were presented as sequences of 3 to 7 tones. In the passive tone task, participants were asked simply to listen to these sequences. In the tone decision task, participants were required to respond by pressing a button with the left hand for any sequence containing two 750 Hz tones. The left hand was used in this and all other active tasks to minimize any leftward bias due to activation of the left hemisphere motor system.

Table 2
Conditions used in the study.

Stimuli in the passive word and semantic decision tasks were 192 spoken English nouns designating animals (e.g., turtle), presented at the rate of one word every 3 seconds. In the passive word task, participants were asked simply to listen passively to these words. In the semantic decision task, participants were required to respond by a left hand button press for animals they considered to be both "found in the United States" and "used by people." No animal word was used more than once in the same task. The tone and word stimuli were matched on average intensity, average sound duration (750 ms), average trial duration (3 s), and frequency of targets (37.5% of trials). Characteristics of the tone and word stimuli and the rationale for the tone and semantic decision tasks are described elsewhere in greater detail (Binder et al., 1995; Binder et al., 1997).

Stimuli in the phoneme decision task were spoken consonant-vowel (CV) syllables, including all combinations of the consonants b, d, f, g, h, j, k, l, m, n, p, r, s, t, v, w, y, and z with the five vowels /æ /, /i/, /a/, /o/, and /u/. All syllables were edited to a duration of 400 ms. Each trial presented three CV syllables in rapid sequence, e.g., /pa dæ su/. Subjects were required to respond by a left hand button press when a triplet of CV syllables included both of the consonants /b/ and /d/.

Seven functional scans were acquired (Table 3). In each scan, one of the conditions alternated eight times with one of the other conditions in a standard block design. Each of these sixteen blocks lasted 24 s and included 8 stimulus presentations (i.e., 8 tone trains, 8 words, or 8 CV triplets). To avoid the possibility that participants might automatically perform the active decision tasks during the passive conditions, all passive scans were acquired before the participants received any training on the decision tasks or had knowledge that tasks were to be performed. Within this group of scans (scans 1–3), the order of scans was randomized and counterbalanced across participants.

Table 3
Composition and order of the functional scans.

Following the passive scans, participants were instructed in the Tone Decision task and performed a Tone Decision vs. Rest scan to identify brain regions potentially involved in task-unrelated conceptual processing during the resting state. To avoid the possibility that participants might rehearse or review the semantic task during these resting blocks, the Tone Decision vs. Rest scan was always acquired first after the passive scans and before participants had any knowledge of the semantic task. The remaining three scans were acquired after training on the Semantic Decision task. These included a Semantic Decision vs. Rest scan, a Semantic Decision vs. Tone Decision scan, and a Semantic Decision vs. Phoneme Decision scan. The order of these three scans was randomized and counterbalanced across participants, except that the Phoneme Decision task was restricted to one of the final two scans. This restriction was to avoid the mental burden of having to learn and perform two new tasks simultaneously (Semantic Decision and Phoneme Decision) for scan 5.

Performances on the Tone Decision and Phoneme Decision tasks were scored as the proportion of correct responses. Responses on the semantic decision task were scored using response data from a group of 50 normal right-handed controls on the same stimulus sets. Items responded to with a probability greater than .75 by controls (e.g., lamb, salmon, horse, goose) were categorized as targets, and items responded to with a probability less than .25 by controls (e.g., ape, cockroach, lion, stork) were categorized as distractors. Performance by each subject was then scored as the proportion of correct discriminations between targets and distractors.

Image Acquisition

Scanning was conducted on a 1.5 Tesla General Electric Signa scanner (GE Medical Systems, Milwaukee, WI) using a 3-axis local gradient coil with an insertable transmit/receive birdcage radiofrequency coil. Padding was placed behind the neck and around the head as needed to relax the cervical spine and to fill the space between the head and inner surface of the coil. Functional imaging employed a gradient-echo echoplanar sequence with the following parameters: 40 ms echo time, 4 s repetition time, 24 cm field of view, 64×64 pixel matrix, and 3.75×3.75×7.0 mm voxel dimensions. Seventeen to 19 contiguous sagittal slice locations were imaged, encompassing the entire brain. One hundred sequential image volumes were collected in each functional run, giving a total duration of 6 min, 40 s for each run. Each 100-image functional run began with 4 baseline images (16 s) to allow MR signal to reach equilibrium, followed by 96 images during which two comparison conditions were alternated for eight cycles. High resolution, T1-weighted anatomical reference images were obtained as a set of 124 contiguous sagittal slices using a 3D spoiled-gradient-echo ("SPGR") sequence.

Image Processing and Subtraction Analysis

Image analysis was done with AFNI software (available at http://afni.nimh.nih.gov/afni) (Cox, 1996). Motion artifacts were minimized by registration of the raw echoplanar image volumes in each run to the first steady-state volume (fifth volume) in the run. Estimates of the three translation and three rotation movements at each point in each time-series were computed during registration and saved. The first four images of each run, during which spin relaxation reaches an equilibrium state, were discarded, and the mean, linear trend, and second-order trend were removed on a voxel-wise basis from the remaining 96 image volumes of each run.

Multiple regression analysis of each run in each subject was performed to identify voxels showing task-associated changes in BOLD signal. An idealized BOLD response was derived by convolving the 24-sec on/off task alternation function with a canonical hemodynamic response modeled using a gamma function (Cohen, 1997). Movement vectors computed during image registration were included in the model to remove residual variance associated with motion-related changes in MRI signal. This analysis generated, for each functional run, a map of beta coefficients representing the magnitude of the response at each voxel and a map of correlation coefficients representing the statistical fit of the observed data to the idealized BOLD response function.

Group activation maps were created with a random-effects model treating subject as a random factor. The beta coefficient map from each subject was spatially smoothed with a 6-mm full-width-half-maximum Gaussian kernel to compensate for inter-subject variance in anatomical structure. These maps were then resized to fit standard stereotaxic space (Talairach & Tournoux, 1988) using piece-wise affine transformation and linear interpolation to a 1-mm3 voxel grid. A single-sample, two-tailed t test was then conducted at each voxel for each run to identify voxels with mean beta coefficients that differed from zero. These group maps were thresholded using a voxel-wise 2-tailed probability of P < 0.0001 (|t-deviate| ≥ 4.55) and minimum cluster size of 200 mm3, resulting in a whole-brain corrected, 2-tailed probability threshold of P < 0.05 for each group map, as determined by Monte-Carlo simulation.

The final analysis determined the mean number of significantly activated voxels in each hemisphere for each task contrast. Individual masks of the supratentorial brain volume were created for each subject by thresholding the first echoplanar image volume to exclude voxels outside of the brain, followed by manual editing to remove the cerebellum and brainstem. The resulting mask was aligned to standard stereotaxic space and divided at the midline to produce separate left and right whole-hemisphere regions of interest (ROIs). The correlation coefficient maps from each subject were then thresholded at a whole-brain corrected P < 0.05 (voxel-wise P < 0.001 and minimum cluster size of 295 mm3) and converted to standard stereotaxic space. Activated voxels were then automatically counted in the left and right hemisphere ROIs for each subject and task contrast. A laterality index (LI) was computed for each subject and task contrast using the formula (L − R)/(L + R), where L and R are the number of voxels in the left and right hemisphere ROIs. LIs computed in this way are known to vary as a function of the significance threshold, becoming more symmetrical as the threshold is lowered (Adcock et al., 2003). LIs for some tasks have been shown to vary substantially depending on the brain region in which the voxels are counted (Lehéricy et al., 2000; Spreer et al., 2002). To permit a meaningful comparison between activation protocols, we therefore used the same threshold and whole-hemisphere ROI for all protocols.

Results

Task Performance

All participants learned the tasks easily and tolerated the scanning procedure well. Performance on the Tone Decision task was uniformly good, with participants attaining a mean score of 98.4% correct (SD = 1.9, range 89–100%). A paired t-test showed no difference in Tone Decision performance when the task was paired with rest compared to when it was paired with the Semantic Decision task (P = 0.256). Participants also performed well in discriminating targets from distractors on the Semantic Decision task, with a mean score of 92.3% correct (SD = 4.4, range 72–100%). It should be noted that judgments in the Semantic Decision task are subjective and depend on participants' personal experiences, hence there are no strictly correct or incorrect responses. Accuracy scores reflect the similarity between a participant's responses and those of a group of participants, and are intended merely to demonstrate compliance with the task. Repeated-measures ANOVA showed no difference in Semantic Decision performance when the task was paired with rest, with the Tone Decision task, or with the Phoneme Decision task (P = 0.745). Accuracy on the Phoneme Decision task averaged 92.4% correct (SD = 4.7, range 77– 100%).

FMRI Results

Results for each of the main speech comprehension contrasts are described below, as well as several relevant contrasts between the control conditions. Peak activation coordinates for each contrast are given in the Appendix.

Passive Words vs. Rest

Activation during passive listening to words, compared to a resting state, occurred mainly in the superior temporal gyrus (STG) bilaterally, including Heschl's gyrus (HG) and surrounding auditory association cortex in the planum temporale (PT), lateral STG, and upper bank of the superior temporal sulcus (STS) (Figure 1, top panel). Smaller foci of activation were observed in two left hemisphere regions, including the inferior precentral sulcus (junction of Brodmann areas (BA) 6, 44, and 8) and the posterior inferior temporal gyrus (BA 37). Stronger activation during the resting state (blue areas in Figure 1, top) was observed in the posterior cingulate gyrus and precuneus bilaterally.

Figure 1
Group fMRI activation maps from three passive listening experiments. Top: Passive listening to words relative to resting. Middle: Passive listening to tones relative to resting. Bottom: Passive listening to words relative to tones. Data are displayed ...

Passive Words vs. Passive Tones

Much of the STG activation to words could be due to pre-linguistic processing of auditory information, and therefore not specific to speech or language. As shown in the middle panel of Figure 1, passive listening to tone sequences elicited a very similar pattern of bilateral activation in STG, HG, and PT, demonstrating that activation in these regions is not specific to speech.

The lower panel of Figure 1 shows a direct contrast between passive words and passive tones. Contrasting words with this non-speech auditory control condition should eliminate activation related to pre-linguistic auditory processing. A comparison of the top and bottom panels of Figure 1 confirms that activation in HG, PT, and surrounding regions of the STG was greatly reduced by incorporating this control condition. Activation for words relative to tones is restricted to ventral regions of the STG lying in the STS. This STS activation is clearly lateralized to the left hemisphere, consistent with the idea that additional activation for words over tones reflects language-related phoneme perception processes. No other areas showed greater activation for words over tones. Stronger activation for tones was noted in the posterior STG bilaterally, and in the right posterior cingulate gyrus.

Semantic Decision vs. Rest

In contrast to the passive word listening condition, the Semantic Decision task requires participants to focus attention on the words, retrieve specific semantic knowledge, make a decision, and generate a motor response. Like the passive listening condition, this task produced bilateral activation of the STG due to auditory processing, though this activation was somewhat more extensive than in the passive listening condition (Figure 2, upper panel). In addition, the Semantic Decision task activated a complex, bilateral network of frontal, limbic, and subcortical structures. Activation in the lateral frontal lobe was strongly left-lateralized, involving cortex in the posterior IFG (pars opercularis) and adjacent MFG, and distinct regions of ventral and dorsal premotor cortex (BA 6, frontal eye field). Strong, symmetric activation occurred in the supplementary motor area (SMA), anterior cingulate gyrus, and anterior insula. Smaller cortical activations involved the left intraparietal sulcus (IPS) and the right central sulcus, the latter consistent with use of a left hand response for the task. Bilateral activation also occurred in several subcortical regions, including the putamen (stronger on the right), anterior thalamus, medial geniculate nuclei, and paramedian mesencephalon. Finally, there was activation in the cerebellum bilaterally, involving both medial and lateral structures, and stronger on the left. Regions of task-induced deactivation (i.e., relatively higher BOLD signal in the resting condition) were observed in the posterior cingulate gyrus and adjacent precuneus bilaterally, the rostral and subgenual cingulate gyrus bilaterally, and the left postcentral gyrus.

Figure 2
Group fMRI activation maps from three active listening experiments. Top: Semantic Decision relative to resting. Middle: Tone Decision relative to resting. Bottom: Semantic Decision relative to Tone Decision. Display conventions as described for Figure ...

Semantic Decision vs. Tone Decision

Much of the activation observed in the Semantic Decision – Rest contrast could be explained by general executive, attention, working memory, and motor processes that are not specific to language tasks. As shown in the middle panel of Figure 2, the Tone Decision task elicited a similar pattern of bilateral activation in premotor, SMA, anterior cingulate, anterior insula, and subcortical regions. In addition, the Tone Decision task activated cortex in the right hemisphere that was not engaged by the Semantic task, including the posterior IFG, dorsolateral prefrontal cortex (inferior frontal sulcus), SMG (BA 40), and mid-MTG. Regions of task-induced deactivation were much more extensive with this task, involving large regions of the posterior cingulate gyrus and precuneus bilaterally, rostral/subgenual cingulate gyrus bilaterally, left orbital and medial frontal lobe, dorsal prefrontal cortex in the SFG and adjacent MFG (mainly on the left), angular gyrus (mainly left), ventral temporal lobe (parahippocampus, mainly on the left), and the left anterior temporal pole.

The lower panel of Figure 2 shows the contrast between Semantic Decision and Tone Decision tasks. Using the Tone Decision task as a control condition should eliminate activation in general executive systems as well as in low-level auditory and motor areas. A comparison of the top and bottom panels of Figure 2 confirms these predictions, showing subtraction of the bilateral activation in dorsal STG, premotor cortex and SMA, anterior insula, and deep nuclei that is common to both tasks. Compared to the Tone task, the Semantic task produced relative BOLD signal enhancement in many left hemisphere association and heteromodal regions, including much of the prefrontal cortex (anterior and mid-IFG, SFG, and portions of MFG), several regions in the lateral and ventral temporal lobe (MTG and ITG, anterior fusiform gyrus, parahippocampal gyrus, anterior hippocampus), the angular gyrus, and the posterior cingulate gyrus. Many of these regions (dorsal prefrontal cortex, angular gyrus, posterior cingulate) were not prominently activated when the Semantic Decision task was contrasted with Rest, suggesting that these regions are also active during the resting state. Their appearance in the Semantic Decision – Tone Decision contrast is due to their relative deactivation (or lack of tonic activation) during the Tone task. Other areas activated by the Semantic task relative to the Tone task included the right cerebellum and smaller foci in the pars orbitalis of the right IFG, right SFG, right angular gyrus, right posterior cingulate gyrus, and left anterior thalamus.

Several areas showed relatively higher BOLD signals during the Tone Decision task, including the PT bilaterally, the right SMG and anterior IPS, and scattered regions of premotor cortex bilaterally.

Semantic Decision vs. Phoneme Decision

Stimuli used in the Tone Decision task are acoustically much simpler than the speech sounds used in the Semantic Decision task and contain no phonemic information. The Phoneme Decision task, which requires participants to process meaningless speech sounds (pseudowords), provides a control for phonemic processing, allowing more specific identification of brain regions involved in semantic processing.

The following regions showed stronger activation for the Semantic compared to the Phoneme task (Figure 3): left dorsal prefrontal cortex (SFG and adjacent MFG), pars orbitalis (BA 47) of the left IFG, left orbital frontal cortex, left angular gyrus, bilateral ventromedial temporal lobe (parahippocampus, fusiform gyrus, and anterior hippocampus, more extensive on the left), bilateral posterior cingulate gyrus, and right posterior cerebellum. Small activations were observed in the anterior left STS and the right pars orbitalis. In contrast to the Semantic Decision – Tone Decision contrast (see Figure 2, bottom), there was little or no activation of dorsal regions of the left IFG or adjacent MFG, or of the lateral temporal lobe (MTG, ITG). These latter regions must have been activated in common during both the Semantic and Phoneme Decision tasks, and are therefore likely to be involved in pre-semantic phonological processes, such as phoneme recognition. Posterior regions of the left IFG and adjacent premotor cortex (BA 44/6) were in fact activated more strongly by the Phoneme task than the Semantic task. Other regions showing this pattern included the right posterior IFG and premotor cortex, extensive regions of the STG bilaterally, the SMG and anterior IPS bilaterally, and the SMA bilaterally.

Figure 3
Group fMRI activation map from the Semantic Decision – Phoneme Decision contrast.

A notable aspect of the Semantic Decision activation pattern for this contrast is how closely it resembles the network of brain areas showing task-induced deactivation (i.e., stronger activation during the resting state) in the contrast between Tone Decision and Rest (blue areas in Figure 2, middle panel). Figure 4 shows these regions of stronger activation for the resting state, duplicated from Figure 2, together with the areas activated by the Semantic relative to the Phonemic task. In both the Semantic Decision – Phoneme Decision contrast and the Rest – Tone Decision contrast, stronger BOLD signals are observed in left angular gyrus, left dorsal prefrontal cortex (SFG and adjacent MFG), left orbital frontal cortex, left pars opercularis, posterior cingulate gyrus, bilateral ventromedial temporal lobe (more extensive on the left), and left temporal pole.

Figure 4
A comparison of left hemisphere areas activated by the Semantic Decision task relative to the Phoneme Decision task (top) and areas active during Rest relative to the Tone Decision task (bottom).

Activation Extent and Degree of Lateralization

For clinical applications, it is important not only that a language mapping protocol identify targeted linguistic systems, but also that it produce consistent activation at the single subject level and a left-lateralized pattern useful for determining language dominance. We quantified the extent and lateralization of activation for each task protocol by counting the number of voxels that exceeded a whole-brain corrected significance threshold in each participant (Table 4). Repeated-measures ANOVA showed effects of task protocol on total activation volume (F(4,100) = 19.528, P < 0.001), left hemisphere activation volume (F(4,100) = 23.070, P < 0.001), right hemisphere activation volume (F(4,100) = 15.133, P < 0.001), and laterality index (F(4,100) = 21.045, P < 0.001). Total activation volume was largest for the Semantic Decision – Tone Decision protocol, and was greater for this protocol than for all others (all pair-wise P < 0.05, Bonferroni corrected for multiple comparisons) except the Semantic Decision – Rest protocol. Total activation volume was greater for all of the active task protocols than for any of the passive protocols (all pair-wise P < 0.05, Bonferroni corrected). Left hemisphere activation volume was largest for the Semantic Decision – Tone Decision protocol, which produced significantly more activated left hemisphere voxels than any of the other four protocols (all pair-wise P < 0.05, Bonferroni corrected). The laterality index was greatest for the Semantic Decision – Tone Decision and Semantic Decision – Phoneme Decision protocols. LIs for these tasks did not differ, but both were greater than the LIs for Passive Words – Rest and Semantic Decision – Rest (all pair-wise P < 0.05, Bonferroni corrected). Single-sample t-tests showed that LIs for the Passive Words – Rest (P = 0.23) and Semantic Decision – Rest (P = 0.07) protocols did not differ from zero, whereas LIs for the other three protocols were all significantly greater than zero (all P < 0.0001). Finally, LIs for the Semantic Decision – Tone Decision and Semantic Decision – Phoneme Decision protocols showed much less variation across the group (smaller SD) compared to the protocols using passive conditions (all F ratios > 4.4, all P < 0.001).

Table 4
Mean extent and lateralization of language-related activation for 5 fMRI contrasts. Numbers in parentheses are standard deviations. Voxel counts have been converted to normalized volumes of activation, expressed in ml.

In summary, the active task protocols (Semantic Decision – Rest, Semantic Decision – Tone Decision, and Semantic Decision – Phoneme Decision) produced much more activation than the passive protocols, the Semantic Decision – Tone Decision protocol produced by far the largest activation volume in the left hemisphere, and the protocols that included a non-resting control (Passive Words – Tones, Semantic Decision – Tone Decision, and Semantic Decision – Phoneme Decision) were associated with stronger left-lateralization of activation than the protocols that used a resting baseline. Of the five protocols, the Semantic Decision – Tone Decision protocol showed the optimal combination of activation volume, leftward lateralization, and consistency of lateralization.

For clinical applications, it is important to know the probability of detecting significant activation in targeted ROIs in individual patients. We previously constructed left frontal, temporal, and angular gyrus ROIs using activation maps from the Semantic Decision – Tone Decision contrast in a group of 80 right-handed adults (Frost et al., 1999; Szaflarski et al., 2002; Sabsevitz et al., 2003). With these three ROIs as targets, activated voxels (whole-brain corrected P < 0.05) were detected in 100% of the 26 participants in the current study in all three ROIs using the Semantic Decision – Tone Decision protocol.

Discussion

Our aim in this study was to compare five types of functional imaging contrasts used to examine speech comprehension networks. The contrasts produced markedly different patterns of activation and lateralization. These differences have important implications for the selection and interpretation of clinical fMRI language mapping protocols.

Passive Words – Rest

Many researchers have attempted to identify comprehension networks by contrasting listening to words with a resting state. Our Passive Words – Rest contrast confirms similar prior studies showing bilateral, symmetric activation of the STG, including primary auditory areas in Heschl's gyrus and planum temporale as well as surrounding association cortex, during passive word listening (Petersen et al., 1988; Wise et al., 1991; Mazoyer et al., 1993; Price et al., 1996; Binder et al., 2000; Specht & Reul, 2003). Interpretations of this STG activation vary, with some authors equating it to the 'receptive language area of Wernicke' and others arguing that it represents a pre-linguistic auditory stage of processing (Binder et al., 1996a; Binder et al., 2000). The latter account arises from the fact, often neglected in traditional models of language processing, that speech sounds are complex acoustic events. Speech phonemes (consonants and vowels), prosodic intonation, and speaker identity are all encoded in subtle spectral (frequency) and temporal patterns that must be recognized quickly and efficiently by the auditory system (Klatt, 1989). Analogous to the monkey STG, which is comprised largely of neurons coding such auditory information (Baylis et al., 1987; Rauschecker et al., 1995), the human STG (including the classical Wernicke area) appears to be specialized for processing complex auditory information. Thus, much of the activation observed in contrasts between word listening and resting can be attributed to auditory perceptual processes rather than to recognition of specific words. According to this model, activation should occur in the same STG regions during listening to spoken nonwords (e.g., "slithy toves") and to complex sounds that are not speech. Many imaging studies have confirmed these predictions (Wise et al., 1991; Démonet et al., 1992; Price et al., 1996; Binder et al., 2000; Scott et al., 2000; Davis & Johnsrude, 2003; Specht & Reul, 2003; Uppenkamp et al., 2006). Because auditory perceptual processes are represented in both left and right STG, this model also accounts for why the activation observed with passive listening is bilateral, and why lateralization measures obtained with this type of contrast are not correlated with language dominance as measured by the Wada test (Lehericy et al., 2000).

Passive Words – Passive Tones

We used simple tone sequences to 'subtract out' activation in the STG due to auditory perceptual processes with the expectation that only relatively early auditory processing would be removed using this control. The Passive Words – Passive Tones contrast confirmed similar prior studies, showing activation favoring speech in the mid-portion of the STS, with strong left lateralization (Mummery et al., 1999; Binder et al., 2000; Scott et al., 2000; Ahmad et al., 2003; Desai et al., 2005; Liebenthal et al., 2005; Benson et al., 2006). Nearly all of the activation observed in the dorsal STG in the Passive Words – Rest contrast was removed by incorporating this simple non-speech control, confirming that this more dorsal activation is not specific to speech or to words. The resulting activation, though strongly left-lateralized, was much less extensive than with other protocols. Six participants had little or no measurable activation (<0.1 ml) with the Passive Words – Passive Tones protocol. Among the other 20 participants, the total activation volume for the Semantic Decision – Tone Decision protocol was, on average, 44 times greater than for Passive Words – Passive Tones.

Semantic Decision – Rest

Active tasks that require participants to consciously process specific information about a stimulus are often used to enhance activation in brain regions associated with such processing. We designed a semantic decision task that required participants to retrieve specific factual information about a concept and use that information to make an explicit decision. Like the passive listening condition, this task produced bilateral activation of the STG due to auditory processing. This activation was somewhat more extensive than in the passive listening condition, consistent with previous reports that attention enhances auditory cortex activation (O'Leary et al., 1996; Grady et al., 1997; Jancke et al., 1999; Petkov et al., 2004; Johnson & Zatorre, 2005; Sabri et al., 2008). The Semantic Decision – Rest contrast also activated widespread prefrontal, anterior cingulate, anterior insula, and subcortical structures bilaterally. Some of these activations can be attributed to general task performance processes that are not specific to language. For example, any task that requires a decision about a stimulus must engage attentional systems that enable the participant to attend to and maintain attention on the stimulus. Similarly, any such task must involve maintenance of the task instructions and response procedure in working memory, a mechanism for making a decision based on the instructions, and a mechanism for executing a particular response. Dorsolateral prefrontal and inferior frontal cortex, premotor cortex, SMA, anterior cingulate, anterior insula, IPS, and subcortical nuclei have all been linked with these general attention and executive processes in prior studies (Paulesu et al., 1993; Braver et al., 1997; Smith et al., 1998; Honey et al., 2000; Adler et al., 2001; Braver et al., 2001; Ullsperger & von Cramon, 2001; Corbetta & Shulman, 2002; Krawczyk, 2002; Binder et al., 2004). Although there is modest leftward lateralization of the activation in some of these areas, most notably in the left dorsolateral prefrontal cortex, much of it is bilateral and symmetric, consistent with prior studies of attention and working memory.

Semantic Decision – Tone Decision

Active control tasks are used to subtract activation related to general task processes (Démonet et al., 1992). The aim is to design a control task that activates these systems to roughly the same degree as the language task while making minimal demands on language-specific processes. The Tone Decision task used in the present study requires subjects to maintain attention on a series of non-linguistic stimuli, hold these in working memory, generate a decision consistent with task instructions, and produce an appropriate motor response. Because many of these processes are common to both the Semantic Decision and Tone Decision tasks, activations associated with general executive and attentional demands of the Semantic Decision task are not observed in the Semantic Decision – Tone Decision contrast. As with the Passive Words – Passive Tones contrast, the tone stimuli used in the control task also cancel out activation in the dorsal STG bilaterally and even produce relatively greater activation of the planum temporale compared to words (Binder et al., 1996a). Other areas with relatively higher BOLD signals during the Tone Decision task included scattered regions of premotor cortex bilaterally, the right SMG, and right anterior IPS. These activations probably reflect the greater demands made by the Tone task on auditory short-term memory (Crottaz-Herbette et al., 2004; Arnott et al., 2005; Gaab et al., 2006; Brechmann et al., 2007; Sabri et al., 2008).

Most striking about the Semantic Decision – Tone Decision contrast, however, are the extensive, left-lateralized activations in the angular gyrus, dorsal prefrontal cortex, and ventral temporal lobe that were not visible in the Semantic Decision – Rest map. These areas have been linked with lexical-semantic processes in many prior imaging studies (Démonet et al., 1992; Price et al., 1997; Cappa et al., 1998; Binder et al., 1999; Roskies et al., 2001; Binder et al., 2003; Devlin et al., 2003; Scott et al., 2003; Spitsyna et al., 2006). Lesions in these sites produce deficits of language comprehension and concept retrieval in patients with Wernicke aphasia, transcortical aphasia, Alzheimer disease, semantic dementia, herpes encephalitis, and other syndromes (Alexander et al., 1989; Damasio, 1989; Gainotti et al., 1995; Dronkers et al., 2004; Nestor et al., 2006; Noppeney et al., 2007). These regions form a widely distributed, left-lateralized network of higher-order, supramodal cortical areas distinct from early sensory and motor systems. We propose that this network is responsible for storing and retrieving the conceptual knowledge that underlies word meaning. Processing of such conceptual knowledge is the foundation for both language comprehension and propositional language production (Levelt, 1989; Awad et al., 2007). It is these brain regions, in other words, that represent the 'language comprehension areas' in the human brain, as opposed to the early auditory areas, attentional networks, and working memory systems highlighted by the Semantic Decision – Rest contrast.

Why is activation in these regions not visible in the Semantic Decision – Rest contrast? The most likely explanation is that these regions are also active during the conscious resting state (Binder et al., 1999). These activation patterns suggest, in particular, that people retrieve and use conceptual knowledge and process word meanings even when they are outwardly 'resting'. Though counterintuitive to many behavioral neuroscientists, this notion has a long history in cognitive psychology, where it has been discussed under such labels as 'stream of consciousness', 'inner speech', and 'task-unrelated thoughts' (James, 1890; Hebb, 1954; Pope & Singer, 1976). Far from being a trivial curiosity, this ongoing conceptual processing may be the mechanism underlying our unique ability as humans to plan the future, interpret past experience, and invent useful artifacts (Binder et al., 1999). The existence of ongoing conceptual processing is supported not only by everyday introspection and a body of behavioral research (Antrobus et al., 1966; Singer, 1993; Teasdale et al., 1993; Giambra, 1995), but also by functional imaging studies (Binder et al., 1999; McKiernan et al., 2006; Mason et al., 2007). In particular, a recent study demonstrated a correlation between the occurrence of unsolicited thoughts and fMRI BOLD signals in the left angular gyrus and ventral temporal lobe (McKiernan et al., 2006). A key finding from both the behavioral and imaging studies is that ongoing conceptual processes are interrupted when subjects must attend and respond to an external stimulus. This observation allows us to explain why activation in this semantic network is observed when the Semantic Decision task is contrasted with the Tone Decision task but not when it is contrasted with a resting state. Unlike the resting state, the tone task interrupts semantic processing, thus a difference in level of activation of the semantic system occurs only when the tone task is used as a baseline.

This account also explains why this semantic network was not visible in either of the passive listening contrasts. Passive listening makes no demands on attention or decision processes and is therefore similar to resting. According to the model presented here, conceptual processes continue unabated during passive listening regardless of the type of stimuli presented. The semantic network therefore remains equally active through all of these conditions and is not visible in contrasts between them. Based on these findings, we disagree with authors who advocate the use of passive listening paradigms for mapping language comprehension systems. This position was articulated strongly by Crinion et al. (Crinion et al., 2003), who compared active and passive contrasts using speech and reversed speech stimuli. Similar activations were observed with both paradigms, which the authors interpreted as evidence that active suppression of default semantic processing is not necessary. Two aspects of the Crinion et al. study are noteworthy, however. First, the active task required participants to detect 2 or 3 changes from a male to a female speaker during a story that lasted several minutes. This task was likely very easy and did not continuously engage participants' attention. Second, the activation observed with these contrasts was primarily in the left STS, resembling the passive Words–Tones contrast in the current study. There was relatively little activation of ventral temporal regions such as those activated here in the Semantic Decision – Tone Decision contrast. Thus, we interpret the activations reported by Crinion et al. as occurring mainly at the level of phoneme perception, though their sentence materials likely also activated dorsolateral temporal lobe regions involved in syntactic parsing (Humphries et al., 2006; Caplan et al., 2008). In contrast to these systems, mapping conceptual/semantic systems in the ventral temporal lobe requires active suppression of ongoing conceptual processes.

The model just developed is outlined in schematic form in Table 5. The table indicates, for each of the experimental conditions, whether or not the condition engages any of the following six processes: auditory perception, phoneme perception, retrieval of concept knowledge, attention, working memory, and response production. A comparison of the + and − entries for any two of the conditions provides a prediction for which processes are likely to be represented in the contrast between those conditions. Reviewing the contrasts discussed above, for example, the Passive Words – Rest contrast is predicted to reveal activation related to auditory and phoneme perception, and the Passive Words – Passive Tones contrast is predicted to show activation related more specifically to phoneme perception. Semantic Decision – Rest is predicted to show activation in auditory and phoneme perception, attention, working memory, and response production systems. Semantic Decision – Tone Decision is predicted to show activation in phoneme perception and concept knowledge systems.

Table 5
Hypothesized neural systems engaged by the six conditions used in the study.

Semantic Decision – Phoneme Decision

The last contrast we examined aims to isolate comprehension processes related to retrieval of word meaning. Because the Tone Decision task uses non-speech stimuli, activation observed in the Semantic Decision – Tone Decision contrast represents both phoneme perception and lexical-semantic stages of comprehension. As illustrated in Table 5, the Phoneme Decision task, which incorporates nonword speech stimuli, is designed to activate the same auditory and phoneme perception processes engaged by the Semantic Decision task, but with minimal activation of conceptual knowledge. Areas activated in the Semantic Decision – Phoneme Decision contrast included the angular gyrus, ventral temporal lobe, dorsal prefrontal cortex, pars orbitalis of the IFG, orbital frontal cortex and posterior cingulate gyrus, all with strong leftward lateralization. These results are consistent with prior studies using similar contrasts (Démonet et al., 1992; Price et al., 1997; Cappa et al., 1998; Binder et al., 1999; Roskies et al., 2001; Devlin et al., 2003; Scott et al., 2003). Compared to the Semantic Decision – Tone Decision contrast, the Semantic Decision – Phoneme Decision contrast produces less extensive activation, consistent with the hypothesis that some of the activation observed in the former contrast is due to pre-semantic speech perception processes. It is also possible that the phoneme decision stimuli, though not words, were sufficiently word-like to partially or transiently activate word codes (Luce & Pisoni, 1998), resulting in partial masking of the lexical-semantic system.

The Phoneme Decision task requires participants to identify individual phonemes in the speech input and hold these in memory for several seconds. Thus this task makes greater demands on auditory analysis and phonological working memory than the Semantic Decision task. Posterior regions of the left IFG and adjacent premotor cortex (BA 44/6), and the SMG bilaterally, were activated more strongly by the Phoneme task than the Semantic task, supporting previous claims for involvement of these regions in phonological processes (Démonet et al., 1992; Paulesu et al., 1993; Buckner et al., 1995; Fiez, 1997; Devlin et al., 2003). Other areas activated more strongly by the Phoneme task included large regions of the STG bilaterally, the SMA bilaterally, the right posterior IFG and premotor cortex, and the anterior IPS bilaterally.

A close look at Table 5 reveals another contrast that could be used to identify activation related specifically to conceptual processing. In the contrast Rest – Tone Decision, the only system predicted to be more active during resting than during the tone task is the conceptual system. Figure 4 shows a side-by-side comparison of the regions activated by the Semantic Decision – Phoneme Decision contrast and the Rest – Tone Decision contrast. In both cases, activation is observed in the left angular gyrus, ventral temporal lobe, dorsal prefrontal cortex, orbital frontal lobe, and posterior cingulate gyrus. The similarity between these activation maps is striking, and particularly so because they were generated using such different task contrasts. The same brain areas are activated by (a) resting compared to a tone decision task, and (b) a semantic decision task compared to a phoneme decision task. This outcome is highly counterintuitive and can only be explained by a model, such as the one in Table 5, that includes activation of conceptual processes during the resting state (Binder et al., 1999). Note that this resting state activation is the same phenomenon often called 'the default state' in research on task-induced deactivations (Raichle et al., 2001). Our model provides an explicit account of the close similarity between semantic and task-induced deactivation networks, which has not been addressed in previous accounts of the default state.

Implications for Language Mapping

These observations have several implications for the design of clinical language mapping protocols. We found that the Semantic Decision – Tone Decision contrast produced an optimal combination of consistent activation and a strongly left-lateralized pattern. These results are not due to any unique characteristics of this particular semantic task; many similar tasks could be designed to focus participants' attention on word concepts. Moreover, as should be clear from this study, activation patterns are not determined by a single task condition, but rather by differences in processing demands between two (or more) conditions. In designing protocols for mapping language comprehension areas, it is critically important to incorporate a contrasting condition that interrupts ongoing conceptual processes by engaging the subject in an attentionally-demanding task. The Tone Decision task accomplishes this by requiring continuous perceptual analysis of meaningless tone sequences. The Semantic Decision – Tone Decision contrast thus identifies not only differences related to pre-semantic phoneme perception but also differences in the degree of semantic processing. It is this strong contrast, not the Semantic Decision task alone, that accounts for the extensive activation.

Consistent and extensive activation, however, is not the only goal of language mapping. More important is that the activation represents linguistic processes of interest. All effortful cognitive tasks require general executive and attentional processes that may not be relevant for the purpose of language mapping. The Tone Decision task also engages these processes and thus 'subtracts' activation due to them, resulting in a map that more specifically identifies language-related activation. As with the semantic task, these characteristics of the tone task are not unique, and there are many possible variations on this task that could accomplish the same goals, perhaps more effectively.

The activation patterns we observed reinforce current views regarding the neuroanatomical representation of speech comprehension processes. In contrast to the traditional view that localizes comprehension processes in the posterior STG (Geschwind, 1971), modern neuroimaging and lesion correlation data suggest that comprehension depends on a widely distributed cortical network involving many regions outside the posterior STG and planum temporale. These areas include distributed semantic knowledge stores located in ventral temporal (MTG, ITG, fusiform gyrus, temporal pole) and inferior parietal (angular gyrus) cortices, as well as prefrontal regions involved in retrieval and selection of semantic information.

FMRI methods for mapping lexical-semantic processes have particular relevance for the presurgical evaluation of patients with intractable epilepsy. The most common surgical procedure for intractable epilepsy is resection of the anterior temporal lobe, and the most consistently reported language deficit after anterior temporal lobe resection is anomia (Hermann et al., 1994; Langfitt & Rausch, 1996; Bell et al., 2000; Sabsevitz et al., 2003). Although these patients have difficulty producing names, the deficit is not caused by a problem with speech articulation or articulatory planning, but by a lexical-semantic retrieval impairment. Difficulty retrieving names in such cases is a manifestation of partial damage to the semantic system that stores knowledge about the concept being named or to the connections between the concept and its phonological representation (Levelt, 1989; Lambon Ralph et al., 2001). Efforts to use fMRI to predict and prevent anomia from anterior temporal lobe resection should therefore focus on identification of this lexical-semantic retrieval system rather than on auditory processes or motor aspects of speech articulation, neither of which are affected by temporal lobe surgery or play a role in language outcome.

Consistent with this view is the fact that anterior temporal lobe resections commonly involve ventral regions of the temporal lobe while typically sparing most of the STG. Figure 5 shows an overlap map of 23 left anterior temporal lobe resections performed at our center, computed by digital subtraction of preoperative and postoperative anatomical scans in each patient. As shown in the figure, the typical resection overlaps ventral lexical-semantic areas activated in the Semantic Decision – Tones contrast (green), but not STG areas activated in the Semantic Decision – Rest contrast. From a clinical standpoint, it is critical to detect these language zones that lie within the region to be resected, particularly since it is damage to these lexical-semantic systems that underlies the language deficits observed in these patients. In support of this model, lateralization of temporal lobe activation elicited with the Semantic Decision – Tone Decision contrast described here has been shown to predict naming decline after left anterior temporal lobe resection (Sabsevitz et al., 2003).

Figure 5
Overlap of the Semantic Decision – Rest and Semantic Decision – Tone Decision activations with a map of left anterior temporal lobe (ATL) resections in 23 epilepsy patients. ATL resection volumes were mapped in each patient by registration ...

The current study focused on cognitive processes elicited by various tasks and did not attempt to resolve all issues regarding language mapping methods. For example, no attempt was made to compare lateralization in different ROIs. Previous studies showed that placement of ROIs in brain regions with lateralized activation (e.g., IFG) can circumvent the problem of nonspecific bilateral activation in other regions (Léhericy et al., 2000). Furthermore, we restricted the current investigation to protocols involving single words, yet recent data suggest that spoken sentences may provide a more potent stimulus for eliciting temporal lobe activation (Vandenberghe et al., 2002; Humphries et al., 2006; Spitsyna et al., 2006; Awad et al., 2007). Development of more sensitive methods for identifying semantic networks in the anterior ventral temporal lobe is a particularly important goal for future research. We also did not investigate the relative ease with which these protocols can be applied to patients with neurological conditions. The Semantic Decision – Tone Decision protocol, as conducted here with a regular alternation between active tasks, has been used successfully by our group in over 200 epilepsy patients with full-scale IQ ranging as low as 70 (Binder et al., 1996b; Springer et al., 1999; Sabsevitz et al., 2003; Binder et al., 2008). Personnel with expertise in cognitive testing, such as a neuropsychologist or cognitive neurologist, are needed to train the patient to understand and perform the tasks, as is standard with Wada testing and other quantitative tests of brain function. Finally, the current study does not address whether removal of the temporal lobe regions activated in these protocols reliably causes language deficits. Does sparing or removal of these regions account for variation in language outcome across a group of patients? This question is beyond the scope of the current study but will be critical to address in future research.

Acknowledgments

Thanks to Julie Frost, William Gross, Edward Possing, Thomas Prieto, and B. Douglas Ward for technical assistance. Supported by National Institute of Neurological Diseases and Stroke grants R01 NS33576 and R01 NS35929, and by National Institutes of Health General Clinical Research Center grant M01 RR00058.

APPENDIX

Location of activation peaks in the atlas of Talairach and Tournoux (1988).

Passive Words > Rest
LocationBAxyzz-score
Superior Temporal
    L planum temporale42/22−52−2566.16
    L planum temporale42/22−43−3176.04
    L Heschl's g.41/42−30−29105.30
    L ant. STG22−486−44.99
    L ant. STS21/22−54−9−44.99
    R STG2253−2045.47
    R ant. STG2258−6−25.01
    R ant. STG22/38445−114.56
    R ant. STG22/38355−174.52
Other
    L precentral g.6−47−3295.27
    L SFG6−620444.86
    L post. ITG37−42−54−104.60
Passive Words > Passive Tones
LocationBAxyzz-score
Superior Temporal
    L ant. STS21/22−54−10−64.79
    L STS21/22−50−2914.36
Semantic Decision > Rest
LocationBAxyzz-score
Medial Frontal
    L SMA6−15467.03
    L SMA6−3−6604.43
    L ant. cingulate g.24−1010406.24
    L ant. cingulate g.24−527284.40
    R ant. cingulate g.32316416.42
    R SMA652614.73
Lateral Frontal
    L pars operc./IFS44/8−4413266.30
    L precentral s.6−394315.72
    L precentral s.6/44−382185.42
    L post. MFG6−304564.90
    L post. MFG8−38−3584.79
    L precentral g.6−38−9504.48
    R pars triang./IFS45/464423304.71
Anterior Insula
    L ant. insula-−331925.74
    L ant. insula-−3417125.73
    R ant. insula-3318−36.19
    R ant. insula-3419115.33
Superior Temporal
    L HG41−49−1517.00
    L STG22−60−2366.35
    L ant. STG22−48−1−55.63
    L planum parietale22/40−35−38194.54
    R HG4246−20126.47
    R STG2258−12−16.45
    R post. STS2153−3565.25
Other Cortical
    L IPS7−28−61334.96
    L cingulate isthmus30−6−4945.23
    R lingual g.1718−9004.61
    R lingual g.1719−7764.49
Subcortical
    L caudate-−13−6185.78
    L putamen-−20154.98
    L globus pallidus-−15−8−34.78
    L med. geniculate body-−17−21−55.27
    L med. thalamus-−6−18125.26
    L midbrain-−5−21−75.55
    R putamen-201166.48
    R putamen-20846.42
    R corona radiata-22−15195.65
    R ant. thalamus-6−15135.29
    R ant. thalamus-7−1034.73
    R ventral thalamus-9−24−14.99
    R cerebral peduncle-19−21−35.20
    midbrain tectum-−1−2705.20
Cerebellum
    L lat. cerebellum-−37−54−255.66
    L med. cerebellum-−9−65−175.34
    R med. cerebellum-9−67−235.55
    R cerebellum-30−50−254.43
Semantic Decision > Tone Decision
LocationBAxyzz-score
Dorsal Prefrontal
    L SFS6−2514537.49
    L SFG8−831466.02
    L SFS10−235365.85
    L SFG6−518575.84
    L SFG10−860165.58
    L SFG9−1149375.30
    L post. MFG6−324595.17
    L SFG8−1935484.92
    SFG6018455.95
Lateral Frontal
    L ant. MFG10−334746.24
    L ant. IFS46/10−3936−16.20
    L IFS44/6−418306.06
    L IFS46−3516265.16
    L pars triang.45−522655.86
    L pars triang.45−4617245.61
    L pars triang.45−451995.00
    L pars orbitalis47−4422−45.83
    L pars orbitalis47−2921−55.66
    L pars orbitalis47/45−4934−15.17
    L pars orbitalis47−2816−184.92
    R pars orbitalis473925−75.18
Medial Frontal
    L gyrus rectus12−228−155.13
    L ant. cingulate g.32−1035274.91
    L orbital frontal pole10/11−1759−74.48
Ventral Temporal
    L ant. STS21/22−57−9−85.67
    L ant. STS21/22−48−9−84.84
    L STS21/22−39−15−155.20
    L MTG21−56−39−64.91
    L post. ITG37−45−51−145.48
    L post. ITS37−45−45−24.61
    L fusiform g.20/36−32−31−165.94
    L ant. fusiform g.36−30−17−255.89
    L ant. fusiform g.20/36/38−29−9−324.93
Angular Gyrus
    L angular g.39−37−66506.36
    L angular g.39−45−68255.96
    L angular g.39−42−75325.31
    L angular g.39−30−59295.14
    R angular g.3946−67344.80
Posterior Medial
    L post. cingulate g.30−4−53105.92
    L precuneus31−12−63274.87
    L cingulate isthmus27−9−31−34.73
Subcortical
    L ant. thalamus-−4−3135.17
    L cerebral peduncle-−11−15−94.65
    R caudate head-143114.48
Cerebellum
    L med. cerebellum-−6−77−305.57
    R med. cerebellum-14−76−266.64
    R lat. cerebellum-38−72−396.04
    R lat. cerebellum-33−59−365.71
    R post. cerebellum-23−83−345.53
    R cerebellum-25−68−284.75
Semantic Decision > Phoneme Decision
LocationBAxyzz-score
Dorsal Prefrontal
    L SFG6/8−1518506.65
    L SFG8−828456.23
    L SFS6−2516556.19
    L SFG9−1251375.63
    L SFG8−734545.55
    L SFG9−1055185.04
    L SFG9−1643304.98
    L SFG9−1241444.94
Ventral Prefrontal
    L pars orbitalis47−3328−86.76
    L pars orbitalis47/11−2122−174.57
    L susorbital s.12/32−838−95.19
    L med. orbital g.11−1123−135.09
    L ant. MFG10−4441−64.54
    L ant. MFG10−305104.42
    L frontal pole10−186384.36
    R pars orbitalis473029−85.39
Ventral Temporal
    L hippocampus-−30−26−126.25
    L hippocampus-−28−36−45.79
    L hippocampus-−11−3575.28
    L parahippocampus28/35/36−31−20−206.19
    L ant. fusiform g.36−28−13−284.86
    L ant. MTG21−53−5−185.00
    L ant. MTG21−62−6−124.95
    R collateral s.28/35/3625−30−155.40
    R fusiform g.20/3732−38−185.15
Angular Gyrus
    L angular g.39−45−68266.27
    L angular g.39−38−77285.31
    L angular g.39−53−65344.75
Posterior Medial
    L post. cingulate g.30/23−4−51115.60
    L post. cingulate g.23−20−61144.75
    L post. cingulateg.23−8−52284.66
    L precuneus31−10−63255.29
    L cingulate isthmus30−13−4204.68
    R post. cingulate g.232−37315.28
Occipital Pole
    L occipital pole17/18−19−103−55.48
    R occipital pole1829−99114.56
    R occipital pole17/1815−103104.53
Cerebellum
    R lat. cerebellum-44−73−345.83
    R lat. cerebellum-32−70−365.67
    R post. cerebellum-17−84−344.68
    R post. cerebellum-31−76−254.56

Abbreviations: ant. = anterior, BA = approximate Brodmann area, g. = gyrus, med. = medial, IFS = inferior frontal sulcus, IPS = intraparietal sulcus, ITG = inferior temporal gyrus, L = left, lat. = lateral, MFG = middle frontal gyrus, MTG = middle temporal gyrus, operc. = opercularis, post. = posterior, R = right, SFG = superior frontal gyrus, SMA = supplementary motor area, STG = superior temporal gyrus, STS = superior temporal sulcus, triang. = triangularis

References

  • Adcock JE, Wise RG, Oxbury JM, Oxbury SM, Matthews PM. Quantitative fMRI assessment of the differences in lateralization of language-related brain activation in patients with temporal lobe epilepsy. Neuroimage. 2003;18:423–438. [PubMed]
  • Adler CM, Sax KW, Holland SK, Schmithorst V, Rosenberg L, Strakowski SM. Changes in neuronal activation with increasing attention demand in healthy volunteers: An fMRI study. Synapse. 2001;42:266–272. [PubMed]
  • Ahmad Z, Balsamo LM, Sachs BC, Xu B, Gaillard WD. Auditory comprehsnion of language in young children: Neural networks identified with fMRI. Neurology. 2003;60:1598–1605. [PubMed]
  • Alexander MP, Hiltbrunner B, Fischer RS. Distributed anatomy of transcortical sensory aphasia. Arch. Neurol. 1989;46:885–892. [PubMed]
  • Antrobus JS, Singer JL, Greenberg S. Studies in the stream of consciousness: Experimental enhancement and suppression of spontaneous cognitive processes. Perceptual and Motor Skills. 1966;23:399–417.
  • Arnott SR, Grady CL, Hevenor SJ, Graham S, Alain C. The functional organization of auditory working memory as revealed by fMRI. J. Cogn. Neurosci. 2005;17:819–831. [PubMed]
  • Awad M, Warren JE, Scott SK, Turkheimer FE, Wise RJS. A common system for the comprehension and production of narrative speech. J. Neurosci. 2007;27:11455–11464. [PubMed]
  • Baylis GC, Rolls ET, Leonard CM. Functional subdivisions of the temporal lobe neocortex. J. Neurosci. 1987;7:330–342. [PubMed]
  • Bell BD, Davies KG, Hermann BP, Walters G. Confrontation naming after anterior temporal lobectomy is related to age of acquisition of the object names. Neuropsychologia. 2000;38:83–92. [PubMed]
  • Benson RR, Richardson M, Whalen DH, Lai S. Phonetic processing areas revealed by sinewave speech and acoustically similar non-speech. Neuroimage. 2006;31:342–353. [PubMed]
  • Binder JR, Rao SM, Hammeke TA, Frost JA, Bandettini PA, Jesmanowicz A, Hyde JS. Lateralized human brain language systems demonstrated by task subtraction functional magnetic resonance imaging. Arch. Neurol. 1995;52:593–601. [PubMed]
  • Binder JR, Frost JA, Hammeke TA, Rao SM, Cox RW. Function of the left planum temporale in auditory and linguistic processing. Brain. 1996a;119:1239–1247. [PubMed]
  • Binder JR, Swanson SJ, Hammeke TA, Morris GL, Mueller WM, Fischer M, Benbadis S, Frost JA, Rao SM, Haughton VM. Determination of language dominance using functional MRI: A comparison with the Wada test. Neurology. 1996b;46:978–984. [PubMed]
  • Binder JR, Frost JA, Hammeke TA, Cox RW, Rao SM, Prieto T. Human brain language areas identified by functional MRI. J. Neurosci. 1997;17:353–362. [PubMed]
  • Binder JR, Frost JA, Hammeke TA, Bellgowan PSF, Rao SM, Cox RW. Conceptual processing during the conscious resting state: a functional MRI study. J. Cogn. Neurosci. 1999;11:80–93. [PubMed]
  • Binder JR, Frost JA, Hammeke TA, Bellgowan PSF, Springer JA, Kaufman JN, Possing ET. Human temporal lobe activation by speech and nonspeech sounds. Cereb. Cortex. 2000;10:512–528. [PubMed]
  • Binder JR, McKiernan KA, Parsons M, Westbury CF, Possing ET, Kaufman JN, Buchanan L. Neural correlates of lexical access during visual word recognition. J. Cogn. Neurosci. 2003;15:372–393. [PubMed]
  • Binder JR, Liebenthal E, Possing ET, Medler DA, Ward BD. Neural correlates of sensory and decision processes in auditory object identification. Nature Neurosci. 2004;7:295–301. [PubMed]
  • Binder JR. FMRI of language systems: methods and applications. In: Faro SH, Mohammed FB, editors. Functional MRI: Basic Principles and Clinical Applications. New York: Springer-Verlag; 2006. pp. 245–277.
  • Binder JR, Sabsevitz DS, Swanson SJ, Hammeke TA, Raghavan M, Mueller WM. Use of preoperative functional MRI to predict verbal memory decline after temporal lobe epilepsy surgery. Epilepsia. 2008 in press. [PMC free article] [PubMed]
  • Braver TS, Cohen JD, Nystrom LE, Jonides J, Smith EE, Noll DC. A parametric study of prefrontal cortex involvement in human working memory. Neuroimage. 1997;5:49–62. [PubMed]
  • Braver TS, Barch DM, Gray JR, Molfese DL, Snyder A. Anterior cingulate cortex and response conflict: Effects of frequency, inhibition and errors. Cereb. Cortex. 2001;11:825–836. [PubMed]
  • Brechmann A, Gaschler-Markefski B, Sohr M, Yoneda K, Kaulisch T, Scheich H. Working memory specific activity in auditory cortex: potential correlates of sequential processing and maintenance. Cereb. Cortex. 2007;17:2544–2552. [PubMed]
  • Buckner RL, Raichle ME, Petersen SE. Dissociation of human prefrontal cortical areas across different speech production tasks and gender groups. J. Neurosci. 1995;74:2163–2173. [PubMed]
  • Caplan D, Stanczak L, Waters G. Syntactic and thematic constraint effects on blood oxygenation level dependent signal correlates of comprehension of relative clauses. J. Cogn. Neurosci. 2008;20:643–656. [PMC free article] [PubMed]
  • Cappa SF, Perani D, Schnur T, Tettamanti M, Fazio F. The effects of semantic category and knowledge type on lexical-semantic access: A PET study. Neuroimage. 1998;8:350–359. [PubMed]
  • Cohen MS. Parametric analysis of fMRI data using linear systems methods. Neuroimage. 1997;6:93–103. [PubMed]
  • Corbetta M, Shulman GL. Control of goal-directed and stimulus-driven attention in the brain. Nature Rev. Neurosci. 2002;3:201–215. [PubMed]
  • Cox RW. AFNI: Software for analysis and visualization of functional magnetic resonance neuroimages. Comput. Biomed. Res. 1996;29:162–173. [PubMed]
  • Crinion JT, Lambon-Ralph MA, Warburton EA, Howard D, Wise RJS. Temporal lobe regions engaged during normal speech comprehension. Brain. 2003;126:1193–1201. [PubMed]
  • Crottaz-Herbette S, Anagnoson RT, Menon V. Modality effects in verbal working memory: differential prefrontal and parietal responses to auditory and visual stimuli. Neuroimage. 2004;21:340–351. [PubMed]
  • Damasio H. Neuroimaging contributions to the understanding of aphasia. In: Boller F, Grafman J, editors. Handbook of neuropsychology. Amsterdam: Elsevier; 1989. pp. 3–46.
  • Davis MH, Johnsrude IS. Hierarchical processing in spoken language comprehension. J. Neurosci. 2003;23:3423–3431. [PubMed]
  • Dehaene-Lambertz G, Pallier C, Serniclaes W, Sprenger-Charolles L, Jobert A, Dehaene S. Neural correlates of switching from auditory to speech perception. Neuroimage. 2005;24:21–33. [PubMed]
  • Démonet J-F, Chollet F, Ramsay S, Cardebat D, Nespoulous J-L, Wise R, Rascol A, Frackowiak R. The anatomy of phonological and semantic processing in normal subjects. Brain. 1992;115:1753–1768. [PubMed]
  • Desai R, Liebenthal E, Possing ET, Waldron E, Binder JR. Volumetric vs. surface-based alignment for localization of auditory cortex activation. Neuroimage. 2005;26:1019–1029. [PubMed]
  • Devlin JT, Matthews PM, Rushworth MFS. Semantic processing in the left inferior prefrontal cortex: A combined functional magnetic resonance imaging and transcranial magnetic stimulation study. J. Cogn. Neurosci. 2003;15:71–84. [PubMed]
  • Dronkers NF, Wilkins DP, Van Valin RD, Redfern BB, Jaeger JJ. Lesion analysis of the brain areas involved in language comprehension. Cognition. 2004;92:145–177. [PubMed]
  • Fiez JA. Phonology, semantics and the role of the left inferior prefrontal cortex. Hum. Brain Mapp. 1997;5:79–83. [PubMed]
  • Frost JA, Binder JR, Springer JA, Hammeke TA, Bellgowan PSF, Rao SM, Cox RW. Language processing is strongly left lateralized in both sexes: Evidence from FMRI. Brain. 1999;122:199–208. [PubMed]
  • Gaab N, Gaser C, Schlaug G. Improvement-related functional plasticity following pitch memory training. Neuroimage. 2006;31:255–263. [PubMed]
  • Gainotti G, Silveri MC, Daniele A, Giustoli L. Neuroanatomical correlates of category-specific semantic disorders: A critical survey. Memory. 1995;3:247–264. [PubMed]
  • Geschwind N. Aphasia. NEJM. 1971;284:654–656. [PubMed]
  • Giambra LM. A laboratory method for investigating influences on switching attention to task-unrelated imagery and thought. Conscious. Cogn. 1995;4:1–21. [PubMed]
  • Grady CL, Van Meter JW, Maisog JM, Pietrini P, Krasuski J, Rauschecker JP. Attention-related modulation of activity in primary and secondary auditory cortex. Neuroreport. 1997;8:2511–2516. [PubMed]
  • Hebb DO. The problem of consciousness and introspection. In: Adrian ED, Bremer F, Jasper HH, editors. Brain mechanisms and consciousness. A symposium. Springfield, IL: Charles C Thomas; 1954. pp. 402–421.
  • Hermann BP, Wyler AR, Somes G, Clement L. Dysnomia after left anterior temporal lobectomy without functional mapping: frequency and correlates. Neurosurgery. 1994;35:52–57. [PubMed]
  • Honey GD, Bullmore ET, Sharma T. Prolonged reaction time to a verbal working memory task predicts increased power of posterior parietal cortical activation. Neuroimage. 2000;12:495–503. [PubMed]
  • Humphries C, Binder JR, Medler DA, Liebenthal E. Syntactic and semantic modulation of neural activity during auditory sentence comprehension. J. Cogn. Neurosci. 2006;18:665–679. [PMC free article] [PubMed]
  • Indefrey P, Levelt WJM. The spatial and temporal signatures of word production components. Cognition. 2004;92:101–144. [PubMed]
  • James W. Principles of psychology. vol.1. New York: Dover Publications; 1890. The stream of consciousness; pp. 224–290.
  • Jancke L, Mirzazade S, Shah NJ. Attention modulates activity in the primary and the secondary auditory cortex: A functional magnetic resonance imaging study in human subjects. Neurosci Lett. 1999;266:125–128. [PubMed]
  • Johnson JA, Zatorre RJ. Attention to simultaneous unrelated auditory and visual events: Behavioral and neural correlates. Cereb. Cortex. 2005;15:1609–1620. [PubMed]
  • Klatt DH. Review of selected models of speech perception. In: Marslen-Wilson WD, editor. Lexical representation and process. Cambridge, MA: MIT Press; 1989. pp. 169–226.
  • Krawczyk DC. Contributions of the prefrontal cortex to the neural basis of human decision making. Neurosci. Biobehav. Rev. 2002;26:631–664. [PubMed]
  • Lambon Ralph MA, McClelland J, Patterson K, Galton CJ, Hodges JR. No right to speak? The relationship between object naming and semantic impairment: Neuropsychological evidence and a computational model. J. Cogn. Neurosci. 2001;13:341–356. [PubMed]
  • Langfitt JT, Rausch R. Word-finding deficits persist after left anterotemporal lobectomy. Arch. Neurol. 1996;53:72–76. [PubMed]
  • Lehéricy S, Cohen L, Bazin B, Samson S, Giacomini E, Rougetet R, Hertz-Pannier L, LeBihan D, Marsault C, Baulac M. Functional MR evaluation of temporal and frontal language dominance compared with the Wada test. Neurology. 2000;54:1625–1633. [PubMed]
  • Levelt WJM. Speaking: From intention to articulation. Cambridge, MA: MIT Press; 1989.
  • Liebenthal E, Binder JR, Spitzer SM, Possing ET, Medler DA. Neural substrates of phonetic perception. Cereb. Cortex. 2005;15:1621–1631. [PubMed]
  • Luce PA, Pisoni DB. Recognizing spoken words: The neighborhood activation model. Ear Hear. 1998;19:1–36. [PMC free article] [PubMed]
  • Mason MF, Norton MI, Van Horn JD, Wegner DM, Grafton ST, Macrae CN. Wandering minds: the default network and stimulus-independent thought. Science. 2007;315:393–395. [PMC free article] [PubMed]
  • Mazoyer BM, Tzourio N, Frak V, Syrota A, Murayama N, Levrier O, Salamon G, Dehaene S, Cohen L, Mehler J. The cortical representation of speech. J. Cogn. Neurosci. 1993;5:467–479. [PubMed]
  • McKiernan KA, D'Angelo BR, Kaufman JN, Binder JR. Interrupting the "stream of consciousness": An fMRI investigation. Neuroimage. 2006;29:1185–1191. [PMC free article] [PubMed]
  • Mottonen R, Calvert GA, Jaaskelainen IP, Matthews PM, Thesen T, Tuomainen J, Sams M. Perceiving identical sounds as speech or non-speech modulates activity in the left posterior superior temporal sulcus. Neuroimage. 2006;30:563–569. [PubMed]
  • Mummery CJ, Patterson K, Hodges JR, Wise RJS. Generating 'tiger' as an animal name or a word beginning with T: differences in brain activation. Proc Royal Soc Lond B. 1996;263:989–995. [PubMed]
  • Mummery CJ, Ashburner J, Scott SK, Wise RJS. Functional neuroimaging of speech perception in six normal and two aphasic subjects. J. Acoust. Soc. Am. 1999;106:449–457. [PubMed]
  • Nestor PJ, Fryer TD, Hodges JR. Declarative memory impairments in Alzheimer's disease and semantic dementia. Neuroimage. 2006;30:1010–1020. [PubMed]
  • Noppeney U, Patterson K, Tyler LK, Moss HE, Stamatakis EA, Bright P, Mummery CJ, Price CJ. Temporal lobe lesions and semantic impairment: a comparison of herpes simplex virus encephalitis and semantic dementia. Brain. 2007;130:1138–1147. [PubMed]
  • O'Leary DS, Andreasen NC, Hurtig RR, Hichwa RD, Watkins GL, Boles-Ponto LL, Rogers M, Kirchner PT. A positron emission tomography study of binaurally and dichotically presented stimuli: effects of level of language and directed attention. Brain Lang. 1996;53:20–39. [PubMed]
  • Oldfield RC. The assessment and analysis of handedness: The Edinburgh inventory. Neuropsychologia. 1971;9:97–113. [PubMed]
  • Paulesu E, Frith CD, Frackowiak RSJ. The neural correlates of the verbal component of working memory. Nature. 1993;362:342–345. [PubMed]
  • Petersen SE, Fox PT, Posner MI, Mintun M, Raichle ME. Positron emission tomographic studies of the cortical anatomy of single-word processing. Nature. 1988;331:585–589. [PubMed]
  • Petkov CI, Kang X, Alho K, Bertrand O, Yund EW, Woods DL. Attentional modulation of human auditory cortex. Nature Neurosci. 2004;7:658–663. [PubMed]
  • Pisoni DB. Auditory and phonetic memory codes in the discrimination of consonants and vowels. Perception and Psychophysics. 1973;13:253–260. [PMC free article] [PubMed]
  • Poeppel D. Pure word deafness and the bilateral processing of the speech code. Cognitive Sci. 2001;25:679–693.
  • Pope KS, Singer JL. Regulation of the stream of consciousness: Toward a theory of ongoing thought. In: Schwartz GE, Shapiro D, editors. Consciousness and self-regulation. New York: Plenum Press; 1976. pp. 101–135.
  • Price CJ, Wise RJS, Warburton EA, Moore CJ, Howard D, Patterson K, Frackowiak RSJ, Friston KJ. Hearing and saying. The functional neuro-anatomy of auditory word processing. Brain. 1996;119:919–931. [PubMed]
  • Price CJ, Moore CJ, Humphreys GW, Wise RJS. Segregating semantic from phonological processes during reading. J. Cogn. Neurosci. 1997;9:727–733. [PubMed]
  • Raichle ME, McLeod AM, Snyder AZ, Powers WJ, Gusnard DA, Shulman GL. A default mode of brain function. PNAS. 2001;98:676–682. [PubMed]
  • Rauschecker JP, Tian B, Hauser M. Processing of complex sounds in the Macaque nonprimary auditory cortex. Science. 1995;268:111–114. [PubMed]
  • Roskies AL, Fiez JA, Balota DA, Raichle ME, Petersen SE. Task-dependent modulation of regions in the left inferior frontal cortex during semantic processing. J. Cogn. Neurosci. 2001;13:829–843. [PubMed]
  • Sabri M, Binder JR, Desai R, Medler DA, Leitl MD, Liebenthal E. Attentional and linguistic interactions in speech perception. Neuroimage. 2008 in press. [PMC free article] [PubMed]
  • Sabsevitz DS, Swanson SJ, Hammeke TA, Spanaki MV, Possing ET, Morris GL, Mueller WM, Binder JR. Use of preoperative functional neuroimaging to predict language deficits from epilepsy surgery. Neurology. 2003;60:1788–1792. [PubMed]
  • Scott SK, Blank C, Rosen S, Wise RJS. Identification of a pathway for intelligible speech in the left temporal lobe. Brain. 2000;123:2400–2406. [PubMed]
  • Scott SK, Leff AP, Wise RJS. Going beyond the information given: a neural system supporting semantic interpretation. Neuroimage. 2003;19:870–876. [PubMed]
  • Singer JL. Experimental and theoretical studies of consciousness. Chichester: John Wiley & Sons; 1993. Experimental studies of ongoing conscious experience; pp. 100–122.
  • Smith EE, Jonides J, Marshuetz C, Koeppe RA. Components of verbal working memory: Evidence from neuroimaging. PNAS. 1998;95:876–882. [PubMed]
  • Specht K, Reul J. Functional segregation of the temporal lobes into highly differentiated subsystems for auditory perception: an auditory rapid event-related fMRI task. Neuroimage. 2003;20:1944–1954. [PubMed]
  • Spitsyna G, Warren JE, Scott SK, Turkheimer FE, Wise RJS. Converging language streams in the human temporal lobe. J. Neurosci. 2006;26:7328–7336. [PubMed]
  • Spreer J, Arnold S, Quiske A, Ziyeh S, Altenmüller DM, Herpers M, Kassubek J, Klisch J, Steinhoff BJ, Honegger J, Schulze-Bonhage A, Schumacher M. Determination of hemisphere dominance for language: comparison of frontal and temporal fMRI activation with intracarotid amytal testing. Neuroradiology. 2002;44:467–474. [PubMed]
  • Springer JA, Binder JR, Hammeke TA, Swanson SJ, Frost JA, Bellgowan PSF, Brewer CC, Perry HM, Morris GL, Mueller WM. Language dominance in neurologically normal and epilepsy subjects: a functional MRI study. Brain. 1999;122:2033–2045. [PubMed]
  • Stevens KN, Blumstein SE. The search for invariant acoustic correlates of phonetic features. In: Eimas PD, Miller JL, editors. Perspectives on the study of speech. Hillsdale, NJ: Erlbaum; 1981. pp. 1–38.
  • Szaflarski JP, Binder JR, Possing ET, McKiernan KA, Ward DB, Hammeke TA. Language lateralization in left-handed and ambidextrous people: fMRI data. Neurology. 2002;59:238–244. [PubMed]
  • Talairach J, Tournoux P. Co-Planar Stereotaxic Atlas of the Human Brain. New York: Thieme Medical Publishers; 1988.
  • Teasdale JD, Proctor L, Lloyd CA, Baddeley AD. Working memory and stimulus-independent thought: Effects of memory load and presentation rate. European Journal of Cognitive Psychology. 1993;5:417–433.
  • Tellez-Zenteno JF, Dhar R, Wiebe S. Long-term seizure outcomes following epilepsy surgery: a systematic review and meta-analysis. Brain. 2005;128:1188–1198. [PubMed]
  • Ullsperger M, von Cramon DY. Subprocesses of performance monitoring: A dissociation of error processing and response competition revealed by event-related fMRI and ERPs. Neuroimage. 2001;14:1387–1401. [PubMed]
  • Uppenkamp S, Johnsrude IS, Norris D, Marslen-Wilson W, Patterson RD. Locating the initial stages of speech-sound processing in human temporal cortex. Neuroimage. 2006;31:1284–1296. [PubMed]
  • Vandenberghe R, Nobre AC, Price CJ. The response of left temporal cortex to sentences. J. Cogn. Neurosci. 2002;14:550–560. [PubMed]
  • Wernicke C. Der aphasische Symptomenkomplex. Breslau: Cohn & Weigert; 1874.
  • Wiebe S, Blume W, Girvin JP, Eliasziw M. A randomized, controlled trial of surgery for temporal-lobe epilepsy. NEJM. 2001;345:311–318. [PubMed]
  • Wise R, Chollet F, Hadar U, Friston K, Hoffner E, Frackowiak R. Distribution of cortical neural networks involved in word comprehension and word retrieval. Brain. 1991;114:1803–1817. [PubMed]