|Home | About | Journals | Submit | Contact Us | Français|
Semantic memory includes all acquired knowledge about the world and is the basis for nearly all human activity, yet its neurobiological foundation is only now becoming clear. Recent neuroimaging studies demonstrate two striking results: the participation of modality-specific sensory, motor, and emotion systems in language comprehension, and the existence of large brain regions that participate in comprehension tasks but are not modality-specific. These latter regions, which include the inferior parietal lobe and much of the temporal lobe, lie at convergences of multiple perceptual processing streams. These convergences enable increasingly abstract, supramodal representations of perceptual experience that support a variety of conceptual functions including object recognition, social cognition, language, and the remarkable human capacity to remember the past and imagine the future.
Human brains acquire and use concepts with such apparent ease that the neurobiology of this complex process seems almost to have been taken for granted. Although philosophers have puzzled for centuries over the nature of concepts , semantic memory (see Glossary) became a topic of formal study in cognitive science only relatively recently . This history is remarkable, given that semantic memory is one of our most defining human traits, encompassing all the declarative knowledge we acquire about the world. A short list of examples includes the names and physical attributes of all objects, the origin and history of objects, the names and attributes of actions, all abstract concepts and their names, knowledge of how people behave and why, opinions and beliefs, knowledge of historical events, knowledge of causes and effects, associations between concepts, categories and their bases, and on and on.
Also remarkable is the variety of everyday cognitive activities that depend on this extensive store of knowledge. A common example is the recognition and use of objects, which has been the focus of much theoretical and empirical work on semantic memory [3–7]. Recognition and use of objects, however, is a capacity shared by many non-human animals that interact with food sources, build simple structures, or use simple tools. More uniquely human is the ability to represent concepts in the form of language, which allows not only the spread of conceptual knowledge in an abstract symbolic form, but also a cognitive mechanism for the fluid and flexible manipulation, association, and combination of concepts [8, 9]. Thus humans use conceptual knowledge for much more than merely interacting with objects. All of human culture, including science, literature, social institutions, religion, and art, is constructed from conceptual knowledge. We do not reason, plan the future or remember the past without conceptual content – all of these activities depend on activation of concepts stored in semantic memory.
Scientific study of human semantic memory processes has been limited in the past both by a relatively restricted focus on object knowledge and by an experimental tradition emphasizing stimulus-driven brain activity. Human brains are occupied much of the day with reasoning, planning, and remembering. This highly conceptual activity need not be triggered by stimuli in the immediate environment – all of it can be done, and usually is, in the privacy of one's own mind. Together with recent insights gained from studies of patients with semantic memory loss, functional imaging data are rapidly converging on a new anatomical model of the brain systems involved in these processes. Given the centrality of semantic memory to human behavior and human culture, the significance of these discoveries can hardly be overstated.
In this article we propose a large-scale neural model of semantic processing that synthesizes multiple lines of empirical and theoretical work. Our core argument is that semantic memory consists of both modality-specific and supramodal representations, the latter supported by the gradual convergence of information throughout large regions of temporal and inferior parietal association cortex. These supramodal convergences support a variety of conceptual functions including object recognition, social cognition, language and the uniquely human capacity to construct mental simulations of the past and future.
A major issue in the study of semantic memory concerns the nature of concept representations. Efforts in the last century to develop artificial intelligence focused on knowledge representation in the form of abstract symbols . This approach led to powerful new techniques for information representation and manipulation (e.g., semantic nets, feature lists, ontologies, schemata). Recent advances in this area used machine learning techniques together with massive verbal inputs to create a highly flexible, probabilistic symbolic network that can respond to general questions in a natural language format . Scientists interested in human brains, on the other hand, have long assumed that the brain represents concepts at least partly in the form of sensory and motor experiences. Nineteenth-century neurologists, for example, pictured a widely distributed `concept field' in the brain where visual, auditory, tactile, and motor `images' associated with a concept were activated in the process of word comprehension [12, 13]. A major advantage of such a theory over a purely symbolic representation is that it provides a parsimonious and biologically plausible mechanism for conceptual learning. Over the course of many similar experiences with entities from the same category, an idealized sensory or motor representation of the entity develops by generalization across unique exemplars, and reactivation or `simulation' of these modality-specific representations forms the basis of concept retrieval .
In addition to these issues concerning representation of information, questions arise about the mechanisms that control semantic information retrieval. Clearly not all knowledge associated with a concept is relevant in all contexts, thus mechanisms are needed for selecting or attending to task-relevant information [15, 16]. Some conceptual tasks also place strong demands on creativity, a term we use here to refer to flexible problem solving in the absence of strong constraining cues. Creative invention through technological innovation, art, and `brainstorming' are uniquely human endeavors that require fluent conceptual retrieval and flexible association of ideas. Even everyday conversation requires a logical but relatively unconstrained flow of ideas, in which one topic leads to another through a series of associated concepts. This type of flexible association and combining of concepts, though ubiquitous in everyday life, has largely been overlooked in functional imaging studies, which tend to focus on highly constrained retrieval tasks involving recognition of words and objects.
The idea that sensory and motor experiences form the basis of conceptual knowledge has a long history in philosophy, psychology, and neuroscience [1, 3, 12, 13]. In recent years, this proposal has gained new steam under the rubric of `embodied' or `situated' cognition, supported by numerous neuroimaging and behavioral studies. Some of the imaging studies showing modality-specific activations during language processing are summarized in Figure 1. A number of these address action concepts and show that processing action-related language activates brain regions involved in executing and planning actions. Motion, sound, color, olfaction, and gustatory concept processing have also been addressed, and also tend to show activation in or near regions that process these perceptual modalities (see legend, Figure 1).
Challenges to the embodiment view have also arisen. One objection is that activations observed in imaging experiments could be epiphenomenal and not causally related to comprehension . This hypothesis has been tested in patients with various forms of motor system damage. Initial results indicate a selective impairment for comprehending action verbs in patients with Parkinson's disease , progressive supranuclear palsy , apraxia , and motor neuron disease [21, 22]. Several studies employing transcranial magnetic stimulation to induce transient lesions in the primary motor cortex or inferior parietal lobe provide converging results [23–28]. Thus, involvement of the motor system during action word processing contributes to comprehension and is not a mere by-product. A related argument is that the activations represent post-comprehension imagery. In studies using imaging methods with high temporal resolution, however, the activation of motor regions during action word processing appear to be rapid, approximately 150–200 ms from word onset [29–32], suggesting that it is part of early semantic access rather than a result of post-comprehension processes. These converging results provide compelling evidence that sensory-motor cortices play an essential role in conceptual representation.
Although it is often overlooked in reviews of embodied cognition, emotion is as much a modality of experience as sensory and motor processing . Words and concepts vary in the magnitude and specific type of emotional response they evoke, and these emotional responses are a large part of the meaning of many concepts. Purple dots in Figure 1 represent activation peaks from 14 imaging studies that examined activation as a function of the emotional content of words or phrases. There is a clear preponderance of activations in the temporal pole (13 studies) and ventromedial prefrontal cortex (10 studies), both of which play a central role in emotion [34, 35]. Involvement of the temporal pole in high-level representation of emotion may also explain activation in this region associated with social concepts [36, 37], which tend to have strong emotional valence.
In addition to modality-specific simulations, we propose that the brain uses abstract, supramodal representations during conceptual tasks. One compelling argument for this view is that the human brain possesses large areas of cortex that are situated between modal sensory-motor systems and thus appear to function as information `convergence zones' . These heteromodal areas include the inferior parietal cortex (angular and supramarginal gyri), large parts of the middle and inferior temporal gyri, and anterior portions of the fusiform gyrus . These areas have expanded disproportionately in the human brain relative to the monkey brain, `taking over' much of the temporal lobe from the visual system . Advocates of a strictly embodied theory of conceptual processing have largely ignored these brain regions, yet they occupy a substantial proportion of the posterior cortex in humans.
A second body of evidence comes from patients with damage in the inferior and lateral temporal lobe, particularly patients with semantic dementia, a syndrome characterized by progressive temporal lobe atrophy and multimodal loss of semantic memory [40, 41]. These patients are unable to retrieve names of objects, categorize objects or judge their relative similarity, identify the correct color or sound of objects, or retrieve knowledge about actions associated with objects [42–45]. Critically, the deficits do not appear to be category-specific  – further evidence that the semantic impairment does not involve strongly modal representations.
A third large body of evidence comes from functional imaging studies that target general semantic rather than modality-specific semantic processes. For example, many imaging experiments have contrasted words against pseudowords, related against unrelated word pairs, meaningful against nonsensical sentences, and sentences against random word strings. In another type of general semantic contrast, a semantic task (e.g., a semantic decision) is contrasted with a phonological control task (e.g., rhyme decision). What is important to understand about all of these `general' contrasts is that although they elicit differences in the degree of access to semantic information, they include no manipulation of modality-specific information. In the absence of systematic biases affecting stimulus selection, the activations resulting from these contrasts are unlikely to reflect modality-specific representations.
A quantitative meta-analysis of 120 of these studies was recently performed . Studies were included only if they satisfied strict criteria for a semantic contrast. Studies were excluded if the stimuli in the contrasting conditions were not matched on orthographic or phonological properties, or if the activations could be explained by differences in attention or working memory demands. Reliability of the activation sites reported in the studies was analyzed using a volume-based technique called activation likelihood estimation .
The results showed remarkable consistency across studies, with reliable activation throughout the left temporal and parietal heteromodal cortex (Figure 2). These locations are consistent with the location of pathological changes in semantic dementia, as well as with temporal and parietal vascular lesions causing semantic impairments [49–53]. Other consistent sites of activation included the dorsomedial prefrontal cortex (superior frontal gyrus), ventromedial prefrontal cortex, inferior frontal gyrus, and the posterior cingulate gyrus and precuneus. The results offer compelling evidence for high-level convergence zones in the inferior parietal, lateral temporal, and ventral temporal cortex. These regions are far from primary sensory and motor cortices and appear to be involved in processing general rather than modality-specific semantic information.
Figure 3 illustrates several prominent theories that differ in the proposed level of separation between conceptual and perceptual representations. Models based on disembodied, symbolic conceptual representations [9, 10] are often criticized on the grounds that such symbols are ultimately devoid of content . From an empirical standpoint, the extensive evidence for involvement of modality-specific sensory, action, and emotion systems during language comprehension is also inconsistent with such a model.
At the other end of the spectrum are `strong embodiment' models in which perceptual and conceptual processes are carried out by the same (perceptual) system [55, 56]. These models are inconsistent with the evidence for modality-independent semantic networks reviewed above. Furthermore, conceptual deficits in patients with sensory-motor impairments, when present, tend to be subtle rather than catastrophic. In a recent study of aphasic patients , lesions in both sensory-motor and temporal regions were correlated with impairment in a picture-word matching task involving action words. This evidence is incompatible with a strong version of the embodiment account, in which sensory-motor regions are necessary and sufficient for conceptual representation.
Other theories propose that amodal representations derive their content from close interactions with modal perceptual systems [7, 14]. The purpose of amodal representations in these latter models is to bind and efficiently access information across modalities rather than to represent the information itself . The need for distinct amodal representations in such a model has been sharply questioned, however, as multimodal perceptual representations could fulfill the same role [55, 56].
We suggest that the current evidence is most compatible with a view we term `embodied abstraction,' briefly sketched here (see [59, 60] for similar proposals). In this view, conceptual representation consists of multiple levels of abstraction from sensory, motor, and affective input. All levels are not automatically accessed or activated under all conditions. Rather, this access is subject to factors such as context, frequency, familiarity, and task demands. The top level contains schematic representations that are highly abstracted from detailed representations in the primary perceptual-motor systems. These representations are `fleshed out' to varying degrees by sensory-motor-affective contributions in accordance with task demands. In highly familiar contexts, the schematic representations are sufficient for adequate and rapid processing. In novel contexts or when the task requires deeper processing, sensory-motor-affective systems make a greater contribution in fleshing out the representations (Box 1).
Figure 4 outlines a neuroanatomical model of semantic memory consistent with a broad range of available data. Modality-specific representations (yellow areas in Figure 4), located near corresponding sensory, motor, and emotion networks, develop as a result of experience with entities and events in the external and internal environment. These representations code recurring spatial and temporal configurations of lower-level modal representations. Although depicted as somewhat modular, we view these systems as an interactive continuum of hierarchically ordered neural ensembles, supporting progressively more combinatorial and idealized representations. These systems correspond to Damasio's local convergence zones  and to Barsalou's unimodal perceptual symbol systems . In addition to bottom-up input in their associated modality, they receive a range of top-down input from other modal systems and from attention. They are modal in the sense that the information they represent is an analog of (i.e., isomorphic with) their bottom-up input .
These modal convergence zones then converge with each other in higher-level cortices located in the inferior parietal lobe and much of the ventral and lateral temporal lobe (red areas in Figure 4). One function of these high-level convergences is to bind representations from two or more modalities, such as the sound and visual appearance of an animal, or the visual representation and action knowledge associated with a hand tool [7, 12, 14, 55]. Such supramodal representations capture similarity structures that define categories, such as the collection of attributes that place `pear' and `light bulb' in different categories despite a superficial similarity of appearance, and `pear' and `pineapple' in the same category despite very different appearances . More generally, supramodal representations allow the efficient manipulation of abstract, schematic conceptual knowledge that characterizes natural language, social cognition, and other forms of highly creative thinking [59, 60].
These modal and supramodal convergence zones store the actual content of semantic knowledge, whereas the prefrontal regions colored blue in Figure 4 control top-down activation and selection of the content in posterior stores (Box 2). The posterior cingulate gyrus and adjacent precuneus (green area in Figure 4) consistently show semantic effects in imaging experiments and have also been implicated in a wide variety of other processes, as discussed below. Given the strong reciprocal connections this region has with the hippocampal formation, it likely plays a role in encoding semantically and emotionally meaningful events in episodic memory , though its precise function remains a topic for future research.
Our view of semantic processing in posterior cortical regions is similar to the `hub and spoke' model of Patterson, Rogers, Lambon Ralph, and colleagues [7, 46, 58] and to the convergence zone model of Damasio , but differs in two important respects. First, we do not believe the data support a central role for the temporal pole as the highest level in the convergence zone hierarchy (Box 3). As shown in Figures 2 and and4,4, multimodal convergence of information processing streams occurs throughout much of the lateral and ventral temporal cortex, as well as in the inferior parietal lobe, whereas the temporal pole receives strong affective input from the ventral frontal lobe and amygdala and is better characterized as a modal region for processing emotion and social concepts [34, 36, 37]. Second, proponents of the hub and spoke model explicitly deny a role for the inferior parietal lobe in representation of semantic information . We believe that the anatomical and functional imaging evidence for semantic memory storage in the inferior parietal lobe is difficult to deny, even though the nature of the information represented in this region is still unclear (Box 4).
The network of brain regions we associate here with semantic processing has also been linked with more specific functions. Nearly all parts of the network have been implicated in aspects of social cognition, including theory-of-mind (processing of knowledge pertaining to mental states of other people), emotion processing, and knowledge of social concepts [36, 37, 63–67]. Much of the network has been implicated in retrieval of episodic and particularly autobiographical memories [68, 69], leading to the hypothesis that these regions function to retrieve event memories through a process of `scene construction' . The same scene construction processes have been proposed as the basis for `prospection,' i.e., imagining future scenarios for the purpose of planning and goal attainment [71, 72]. Finally, the association of these regions with autobiographical, `self-projection,' and self-referential processes has led to suggestions that they are specifically involved in processing self knowledge [73, 74]. Several recent reviews and meta-analyses attest to the high degree of neuroanatomical overlap between the networks supporting these purportedly distinct processes [67, 75–77].
Given this overlap, it is logical to ask whether there is a process common to all of these cognitive functions. A model based on self-referential processing cannot easily explain activation of the same regions by theory-of-mind tasks, which by definition emphasize knowledge pertaining to others. The general process of mental scene construction is common to episodic memory retrieval, prospection, and many theory-of-mind tasks, but this model cannot explain the consistent activation of these regions by single word comprehension tasks, as shown above in Figure 2. Indeed, the contrasts analyzed in Figure 2 focused on general semantic knowledge (especially knowledge about object concepts) and did not emphasize episodic, autobiographical, social, emotional, self, or any other specific knowledge domain.
One process shared across semantic, social cognition, episodic memory, scene construction, and self-knowledge tasks is the retrieval of conceptual knowledge. The scene construction posited to underlie episodic memory retrieval and prospection refers to a partial, internal simulation of prior experience. But the construction of a scene requires content. The content of such a simulation is conceptual knowledge about particular entities, events, and relationships. The variety of this content is impressive, encompassing object, action, social, self, spatial, and other domains, yet these types of content all share a common basis in sensory-motor experience, learning through generalization across individual exemplars, and progressive abstraction from perceptual detail. We propose that the essential function of the high-level convergence zone network is to store and retrieve this conceptual content, which is employed over a variety of domain-specific tasks.
This network of high-level convergence zones also overlaps extensively with the `default mode network' of regions that show higher levels of activity during passive and `resting' states than during attentional tasks [47, 74–76, 78]. The similarity between all of these networks lends strong support to proposals that `resting' is a cognitively complex condition characterized by episodic and autobiographical memory retrieval, manipulation of semantic and social knowledge, creativity, problem solving, prospection, and planning [75, 78–81]. Several authors have emphasized the profound adaptive value of these processes, which not only enable the attainment of personal goals but are also responsible for all of human cultural and technological development [78, 80, 81].
This review proposes a large-scale brain model of semantic memory organization in the human brain based on synthesis of a large body of empirical imaging data with a modified embodiment theory of knowledge representation. In contrast to strong versions of embodiment theory, the data show that large areas of heteromodal cortex participate in semantic memory processes. The multimodal convergence of information toward these brain areas enables progressive abstraction of conceptual knowledge from perceptual experience, enabling rapid and flexible manipulation of this knowledge for language and other highly creative tasks. In contrast to models that identify the temporal pole as the principal site of this information convergence, the evidence suggests involvement of heteromodal regions throughout the temporal and the inferior parietal lobes. We hope this anatomical-functional model provides a useful framework for several future lines of research (Box 5).
Modality-specific simulation provides a plausible mechanism for retrieval of concrete object concepts, but difficulties arise in considering abstract concepts. What sensory, motor, or emotional experience is reactivated in comprehending words such as `abstract', `concept', `modality', and `theory'? Another potential difficulty arises from the speed of spoken language, which is easily understood at rates of 3–4 words per second . It is far from clear that an extended sensory, motor, or emotional simulation of each word is possible at such speeds, or even necessary. Thus there is a strong rationale for considering theories that allow both sensory-motor-emotional simulation and manipulation of more abstract representations as a basis for conceptual processing, depending on the exigencies of a given task [59, 60]. At one end of this continuum are tasks that encourage simulation by explicitly requiring mental imagery of an object or event. At the other end are tasks requiring comprehension of rapidly presented, abstract verbal materials that evoke little or no mental simulation. Imagine, for example, that you are deciding which of two cars to buy. This task is likely to engage extended mental simulation of the sensory and motor experiences of examining and test-driving each car. In contrast, imagine hearing someone say, “I don't really have any need or money for a car right now, so it's low on my priority list”. This statement is perfectly understandable and full of meaning, but how extensively must the sensory attributes of `car' be simulated for full comprehension to occur, or simulation of words such as `need', `now', `low', and `priority'? Another factor that likely modulates `depth' of simulation during language comprehension is the familiarity of an expression. Imagine that instead of the statement about a car, you hear, “I don't really have any need or money for a llama right now, so it's low on my priority list”. Comprehending the word `llama' is likely to require an extended visual simulation and the unfamiliarity of the statement itself is likely to elicit a range of simulations involving possible uses for a llama. In general, the involvement of sensory-motor systems in language comprehension seems to change through a gradual abstraction process whereby relatively detailed simulations are needed for unfamiliar or infrequent concepts and these simulations become less detailed as familiarity and contextual support increases .
Studies of patients with semantic dementia have drawn attention to the temporal pole (TP) and the proposal that it functions as a central `hub' housing amodal semantic representations . Emphasis on the TP is also consistent with a longstanding view of this region as the zenith of a caudal-to-rostral convergence of information in the temporal lobe . There are, however, several reasons to question claims that the TP is the sole or principal focus of high-level information convergence. The concept of the TP as an anatomical convergence zone is based mainly on two sources of information: the caudal-rostral progression of information processing in the primate ventral visual system  and the convergence of massive multimodal inputs on anterior medial temporal regions, particularly perirhinal cortex [85, 86]. Although a caudal-rostral hierarchy of information complexity in the primate visual system is undeniable, the proportion of the temporal lobe devoted to unimodal visual processing is considerably less in the human than in the monkey brain . In contrast to the monkey visual system, which occupies ventral and lateral temporal cortex all the way to the temporal pole, the human visual system is largely confined to occipital cortex and posterior ventral temporal lobe. Apart from modal auditory cortex in the superior temporal gyrus, remaining regions in the human temporal lobe are not clearly modality-specific, therefore multimodal convergences are likely to occur along the entire length of the temporal lobe. The convergence of inputs on the anterior medial temporal lobe, though sometimes construed as serving a conceptual function , are more likely to represent input to the hippocampal system for the purpose of episodic memory encoding .
Pathological evidence regarding the TP is also somewhat ambiguous. Although atrophy in semantic dementia is typically most severe in the TP, the total area involved is usually much larger, including most of the ventral temporal cortex [88–90]. Regions showing the strongest correlation between atrophy and semantic deficits are actually closer to the mid-temporal lobe than the TP [90, 91]. Finally, the TP, ventromedial prefrontal cortex and lateral orbitofrontal cortex constitute a tightly interconnected network [34, 92] implicated in processing modality-specific emotional aspects of word meaning (see Figure 1). Considered together, these data suggest that the most anterior parts of the temporal lobe, including the TP and anteromedial temporal regions, are unlikely to be a critical hub for retrieval of multimodal semantic knowledge.
Imaging studies identify reliable semantic processing effects in the left inferior frontal gyrus (IFG) and in a larger dorsomedial prefrontal region extending from the posterior middle frontal gyrus laterally to the superior frontal gyrus (SFG) medially (see blue areas in Figure 4). Numerous experiments indicate that the IFG is engaged when tasks require effortful selection of semantic information, as when many alternative responses are possible or lexical ambiguity gives rise to competing semantic representations [15, 16, 93, 94]. Consistent with prior reviews [95, 96], the meta-analytical data presented here show more reliable activation of anterior and ventral aspects of the IFG (pars orbitalis and triangularis) in semantic studies compared to posterior IFG.
The role of dorsomedial prefrontal cortex in semantic processing has been much less studied, although this region has been a focus of attention in research on emotion processes, social cognition, self-referential processing and the default mode [67, 73, 74, 97–99]. Ischemic lesions to the left SFG cause transcortical motor aphasia, a syndrome characterized by sparse speech output [100, 101]. There is typically a striking disparity between cued and uncued speech production, in that patients can repeat words and name objects relatively normally, but are unable to generate lists of words within a category or invent non-formulaic responses in conversation. That is, patients perform well when a simple response is fully specified (a word to be repeated or object to be named) but poorly when a plan must be created for generating a response . This pattern suggests a specific deficit of self-initiated, self-guided retrieval of semantic information. The SFG lies between ventromedial prefrontal areas (rostral cingulate gyrus and medial orbitofrontal cortex) involved in emotion and reward and lateral prefrontal networks involved in cognitive control, and may act as an intermediary link between these processing systems. We propose that a key role of this region in semantic processing is to translate affective drive states into a coordinated plan for knowledge retrieval, that is, a plan for top-down activation of semantic fields relevant to the problem at hand. Damage to this region causes no loss of stored knowledge per se, but impairs the ability to access this knowledge for creative problem solving. As noted earlier, generating creative solutions in open-ended situations – including interpersonal conflicts, mechanical problems, future plans, even trivial conversational exchanges – is a relatively common task in daily life and also appears to be a large component of the conscious `resting' state.
The inferior parietal cortex lies at a confluence of visual, spatial, somatosensory and auditory processing streams. Human functional imaging studies implicate this region specifically in representational aspects of semantic memory. For example, the AG responds more strongly to words than to matched pseudowords , more to high-frequency than low-frequency words , more to concrete than abstract words  and more to meaningful than meaningless sentences . Thus the level of AG activation seems to reflect the amount of semantic information that can be successfully retrieved from a given input.
Whether the parietal and temporal convergence zones play distinct roles in representing meaning remains a topic for future research, although available evidence offers some intriguing clues. One clue comes from differences in the location and anatomical connectivity of these regions, which to some degree parallel well-established differences between the ventral and dorsal visual networks. The temporal lobe convergence zone receives heavy input from the ventral visual object identification network and from the auditory `what' pathway , suggesting that its main role in semantic memory concerns conceptual representation of concrete objects. In contrast, the AG is bounded by dorsal attention networks that play a central role in spatial cognition, anterior parietal regions concerned with representation of action and posterior temporal regions supporting movement perception . This suggests that the AG may play a unique role in representation of event concepts. Semantic memory research has focused overwhelmingly on knowledge about static concrete entities (i.e., objects, object features, categories), yet much of human knowledge concerns events in which entities interact in space and time. For example, the concept `birthday party' does not refer to a static entity but is instead defined by a configuration of entities (people, cake, candles, presents) and a series of events unfolding in time and space (lighting candles, singing, eating, opening presents). These spatial and temporal configurations define `birthday party' and distinguish it from similar concepts like `picnic' or `office party' in the same way that sets of sensory and motor features distinguish one object from another.
This hypothesis is consistent with recent evidence showing involvement of the AG in retrieval of episodic memories and in understanding theory-of-mind stories. The content of both episodic memory and socially complex stories consists largely of events. Several other studies show specific involvement of the AG in processing temporal and spatial information in stories [108, 109].
Supported by NIH grants R01 NS33576 and R01 DC010783. Thanks to Lisa Conant, Will Graves, Colin Humphries, Tim Rogers, and Mark Seidenberg for helpful discussions.
Publisher's Disclaimer: This is a PDF file of an unedited manuscript that has been accepted for publication. As a service to our customers we are providing this early version of the manuscript. The manuscript will undergo copyediting, typesetting, and review of the resulting proof before it is published in its final citable form. Please note that during the production process errors may be discovered which could affect the content, and all legal disclaimers that apply to the journal pertain.
Appendix A. Supplementary data Supplementary data associated with this article can be found, in the online version, at