Abstract
The human parahippocampal cortex has been ascribed central roles in both visuospatial and mnemonic processes. More specifically, evidence suggests that the parahippocampal cortex subserves both the perceptual analysis of scene layouts as well as the retrieval of associative contextual memories. It remains unclear, however, whether these two functional roles can be dissociated within the parahippocampal cortex anatomically. Here, we provide evidence for a dissociation between neural activation patterns associated with visuospatial analysis of scenes and contextual mnemonic processing along the parahippocampal longitudinal axis. We used fMRI to measure parahippocampal responses while participants engaged in a task that required them to judge the contextual relatedness of scene and object pairs, which were presented either as words or pictures. Results from combined factorial and conjunction analyses indicated that the posterior section of parahippocampal cortex is driven predominantly by judgments associated with pictorial scene analysis, whereas its anterior section is more active during contextual judgments regardless of stimulus category (scenes vs objects) or modality (word vs picture). Activation maxima associated with visuospatial and mnemonic processes were spatially segregated, providing support for the existence of functionally distinct subregions along the parahippocampal longitudinal axis and suggesting that, in humans, the parahippocampal cortex serves as a functional interface between perception and memory systems.
SIGNIFICANCE STATEMENT The functional neuroanatomy of the parahippocampal cortex is still subject to considerable debate. Specifically, its relative contributions to visuospatial and mnemonic functions remain unclear. This study constitutes the first evidence for the existence of distinct information-processing properties along the parahippocampal longitudinal axis. Our findings implicate the posterior section of the parahippocampus in visuospatial perception and the anterior section in contextual mnemonic processes. Our study provides novel neuroanatomical information critical for understanding the diversity of the purported functions of the human parahippocampal cortex.
Introduction
Various lines of evidence indicate that the human parahippocampal cortex plays a key role in both visuospatial (Epstein et al., 1999, 2003; Park and Chun, 2009; Rajimehr et al., 2011) and mnemonic (Bar and Aminoff, 2003; Düzel et al., 2003; Janzen and van Turennout, 2004; Baumann et al., 2010; Wegman and Janzen, 2011) processes. Evidence for a role of the parahippocampal gyrus in visuospatial analysis derives mainly from human neuroimaging experiments, which have shown that this region responds strongly to visual scenes such as landscapes, cityscapes, and buildings (giving rise to the name parahippocampal place area, PPA; Epstein and Kanwisher, 1998; Epstein, 2005; Epstein and Ward, 2010). Increased activity within the posterior aspect of the parahippocampal cortex during viewing of scenes relative to objects has been suggested to reflect regional specialization for visuospatial processes related to the extraction of scene layout information (Epstein, 2008). The idea that the posterior parahippocampal cortex serves a dedicated role in visuospatial processing is further supported by studies showing that this brain region is sensitive to changes in basic visuospatial properties such as the shape and spatial frequency of abstract geometric objects and layouts (Rajimehr et al., 2011; Zeidman et al., 2012).
In addition to its role in visuospatial analysis, the parahippocampal cortex has also been implicated in the integration and retrieval of contextual associations in memory (Aminoff et al., 2013; Bar and Aminoff, 2003; Bar et al., 2008a, 2008b). Contextual associations are defined as relations between objects and conditions that describe, represent, and bring meaning to an environment (Aminoff et al., 2013). Examples of such associations are objects that commonly co-occur, such as a refrigerator and an oven (typically encountered in a kitchen) or familiar spatial configurations of objects, such as rows of tables and chairs in a school classroom. According to this view, activity in the parahippocampal cortex reflects modality-independent retrieval of contextual information from memory, such as which objects typically occur in a given scene and where they are likely to be located relative to one another.
Although there is evidence that the parahippocampal cortex is involved in the visuospatial analysis of scenes as well as contextual processing, it remains unknown whether these two functions can be dissociated anatomically. A recent fMRI study (Baldassano et al., 2013) reported that anterior voxels in the parahippocampal cortex are more strongly functionally connected to parietal and medial temporal networks, whereas posterior voxels are more strongly functionally connected to occipital regions. This anatomical segregation of function suggests a predominantly perceptual role for the posterior parahippocampal cortex and a mnemonic role for the anterior parahippocampal cortex. (Note that the anterior parahippocampal cortex should not be confused with the anterior portion of the parahippocampal gyrus, which includes the perirhinal cortex; Ranganath and Ritchey, 2012.) Here, we tested this hypothesis directly by determining whether neural activity associated with visuospatial analysis of scenes can be functionally dissociated from activity associated with the processing of mnemonic contextual information.
We used fMRI to measure parahippocampal responses while participants engaged in a task that required them to judge the contextual relatedness of scene and object pairs (e.g., “bike” and “bell” are a contextually congruent object–object pair, whereas “sofa” and “microscope” are incongruent; “boat” and “lake” are a contextually congruent object–scene pair, whereas “tractor” and “hotel” are incongruent). Each pair was presented either in printed-word format or in a combined word–picture format, allowing us to present pictorial objects and scenes in isolation (Fig. 1). To identify visuospatial regions of the parahippocampal cortex, we tested for voxels with greater activity for picture–scene trials than for picture–object trials over and above activity differences elicited by the comparison of word–scene trials and word–object trials. In addition, to locate context-related regions of the parahippocampal cortex, we tested for voxels that were selectively active during the contextual judgment task regardless of the type of stimulus (scenes vs objects) or presentation modality (word vs picture).
Materials and Methods
Participants.
Twenty-four healthy, adult volunteers gave their informed consent to participate in the study, which was approved by the Human Research Ethics Committee of The University of Queensland. The participants (12 females) ranged in age from 18 to 33 years (mean, 21 years) and all were right-handed. All participants were fluent English readers.
Stimuli and procedure.
Each trial consisted of a 2 s presentation of an object–object pair or object–scene pair presented either in printed-word format or in a combined word–picture format (Fig. 1). This design allowed us to present pictorial objects and scenes in isolation and to contrast their visually evoked activity patterns. The participants' task was to judge whether a stimulus pair was contextually congruent or incongruent (e.g., “boat” and “lake” are contextually congruent, whereas “tractor” and “hotel” are not) and to indicate their decision via button press within 4 s of stimulus onset. The response period was followed by a variable blank interval of 0–2 s. In addition to the verbal and pictorial task conditions, we used an active baseline condition that required participants to compare two random-letter strings (one uppercase and one lowercase) and determine whether they were the same or different (Fig. 1). This condition was comparable in visual and response demands to the verbal condition, but did not involve contextual retrieval processes (i.e., access to semantic information), thus providing a baseline to test for a conjoint effect of contextual processing across the other eight task conditions.
Each object–object and object–scene pair was presented only once to each participant, either as printed words or pictorially, to avoid learning effects. To control for the effects of individual stimuli, object–object and object–scene pairs that were presented to half of the participants as words were presented to the other participants as pictures and vice versa. Thirty trials were presented for each of the ten task conditions. The resulting 300 trials were divided into 4 imaging runs lasting ∼6.5 min each.
Stimuli were color photographs of everyday objects and scenes (425 pixels in their longest dimension) and letter strings (number of characters, 3–17; font, Arial; font size, 45). Stimuli were presented either in printed-word format or in a combined word–picture format (with pictures always on the right side) using Presentation software (Neurobehavioral Systems; http://www.neurobs.com) running on a PC connected to a liquid crystal display projector (1024 × 768 resolution) that back projected stimuli onto a screen positioned at the head end of the scanner bed. Participants lay on their back within the bore of the magnet and viewed the stimuli via a mirror that reflected the images displayed on the screen. The distance to the screen was 90 cm (12 cm from eyes to mirror) and the visible part of the screen encompassed ∼22.0° × 16.4° of visual angle (35.5 × 26 cm).
MRI acquisition.
Brain images were acquired on a 3T MR scanner (Trio; Siemens) fitted with a 12-channel head coil. For the functional data, 29 axial slices (slice thickness, 3 mm) were acquired in a descending order using a gradient echo echoplanar T2*-sensitive sequence (repetition time, 1.8 s; echo time, 30 ms; flip angle, 80°; matrix, 64 × 64; field of view, 192 × 192 mm; in-plane resolution, 3×3 mm; phase encoding polarity, positive). In each of four runs, 219 volumes were acquired for each participant; the first four images were discarded to allow for T1 equilibration. The sequence was optimized to minimize signal dropouts in the medial temporal lobes (Weiskopf et al., 2006) and geometric distortions in the EPI images caused by magnetic field inhomogeneities were corrected using a point-spread mapping approach (Zeng and Constable, 2002; Zaitsev et al., 2003). We also acquired a T1-weighted structural MPRAGE scan. To minimize head movement, all participants were stabilized with tightly packed foam padding surrounding the head.
Data analysis.
Image processing and statistical analyses were performed using SPM12 (Wellcome Department of Imaging Neuroscience, University College London). Functional data volumes were slice-time corrected and realigned to the first volume. A T2*-weighted mean image of the unsmoothed images was coregistered with the corresponding anatomical T1-weighted image from the same individual. The individual T1 image was used to derive the transformation parameters for the stereotaxic space using the SPM12 template (Montreal Neurological Institute template), which was then applied to the individual coregistered EPI images. Given our hypothesis-driven approach and to maximize the sensitivity of our fMRI analyses, we used region of interest (ROI) analyses focused on the left and right parahippocampal gyrus. ROIs were defined using the AAL atlas (Tzourio-Mazoyer et al., 2002) and WFU pickatlas tool (Maldjian et al., 2003).
In the first-level analysis, we generated a model incorporating 10 experimental task regressors (baseline-same, baseline-different, word-object-congruent, word-object-incongruent, word-scene-congruent, word-scene-incongruent, picture-object-congruent, picture-object-incongruent, picture-scene-congruent, and picture-scene-incongruent; Fig. 1a). In the second-level analysis, an ROI-based SPM full factorial ANOVA with factors of stimulus category (object–object, object–scene) and modality (word, picture) was conducted. To identify parahippocampal regions that were predominantly engaged during visuospatial scene analysis (i.e., regions that showed greater activity for picture–scene trials than picture-object trials over and above activity differences elicited by the comparison of word-scene trials and word-object trials), we tested whether any voxels within the parahippocampal gyrus showed a significant interaction of the factors modality and stimulus category (threshold, p ≤ 0.05, FWE corrected at the voxel level, extend threshold, 10), followed by post hoc extraction of parameter estimates of regional peak activity. Finally, to identify regions of the parahippocampal gyrus that were significantly engaged across all contextual judgment tasks (i.e., activity common to all tasks), we performed a conjunction analysis (using a conjunction-null hypothesis) across these task conditions relative to the random-letter baseline conditions (Nichols et al., 2005; threshold, p ≤ 0.05 FWE corrected at the voxel-level, extent threshold, 10). The role of a baseline condition in a conjunction analysis is to remove common activity of no interest, which in this case is activity related to general attentional and response-related processes.
Results
Behavioral performance
For the categorization of object–object and object–scene pairs as either contextually congruent or incongruent, there was a remarkable degree of consistency across the participants' judgments. In the eight contextual judgment task conditions, participants made, on average, only 2.27% incorrect judgments. Errors of omission were even rarer (0.17%). It took participants, on average, 1.41 s (SE = 0.04 s) to respond in the eight contextual judgment task conditions. We compared response times across all conditions (paired t test, threshold, p ≤ 0.05) and found no significant difference between them, with the exception of the picture–object conditions, in which participants responded, on average, 0.09 s faster than in the other conditions.
fMRI results
Parahippocampal gyrus
To identify subregions of the parahippocampal cortex associated with picture–scene processing, we conducted a two-step statistical analysis. First, we used a factorial analysis to identify voxels that displayed a significant interaction between the factors of stimulus category (objects or scenes) and modality (pictures or words). This analysis identified two activity clusters located in the posterior section of the left (peak voxel = −28, −44, −6, cluster size = 97 voxels, F = 94.03) and right parahippocampal cortex (peak voxel = 28, −42, −8, cluster size = 189 voxels, F = 183.58; Fig. 2a). In a second step, we extracted the parameter estimates from the two activation maxima yielded by the factorial analysis, separately for congruent and incongruent trials, to identify which condition was driving the interaction. As can be seen in Figure 2b, the interaction was clearly driven by stronger activity in the picture–scene conditions relative to the other task conditions (Fig. 2b). The parahippocampal response was significantly stronger bilaterally for object–scene than object–object pairs in the pictorial conditions, but not in the word conditions (two-tailed dependent t test, threshold, p ≤ 0.05). There was no significant effect of congruency (congruent vs incongruent trials; repeated-measures ANOVA, threshold, p ≤ 0.05; Fig. 2b). This suggests that the posterior section of the parahippocampal cortex underlies perception-based scene analysis, but not more general contextual mnemonic processing. In addition, voxelwise analysis of the parahippocampal ROIs showed that activity in the word conditions was not significantly different from activity in the baseline conditions. Together, these findings suggest that a posterior section of the parahippocampal cortex underlies perception-based scene analysis, but not more general contextual mnemonic processing.
To identify any subregions of the parahippocampal cortex engaged by contextual mnemonic processes, regardless of stimulus category (objects or scenes) or modality (pictures or words), we conducted a conjunction analysis, testing for common activity across all contextual judgment task conditions relative to baseline. This analysis revealed a single left-hemisphere cluster located in the anterior parahippocampal cortex (peak voxel = −28, −32, −18, cluster size = 58 voxels, t = 8.01; Fig. 2a,c). This result suggests that a more anterior section of the left parahippocampal cortex underlies contextual mnemonic processing regardless of the visual properties (picture vs word) and stimulus category (objects or scenes) of the stimulus material. As shown in Figure 2a, the activation maxima associated with visuospatial and mnemonic processes were spatially well separated and there was virtually no overlap between the activation clusters, providing further support for the existence of functionally distinct subregions in the posterior and anterior parahippocampal cortex. To further statistically investigate the functional dissociation between these two putative subregions, we extracted and compared the peak voxel parameter estimates of the interaction and conjunction analyses separately for the anterior (−28, −32, −18) and posterior (−28, −44, −6) parahippocampal ROIs. The interaction effect was significantly stronger in the posterior ROI (average parameter estimate = 24.36, SEM = 2.51) than in the anterior ROI (average parameter estimate = 4.76, SEM = 1.66; independent t test, t = 6.51, p=<0.0001). In contrast, the conjunction effect was significantly stronger in the anterior ROI (average parameter estimate = 12.11, SEM = 1.51) than in the posterior ROI (average parameter estimate = 2.46, SEM = 2.29; independent t test, t = 3.52, p < 0.001). It is important to note that we have shown peak-voxel parameter estimates here to highlight the maximal differences in response characteristics of the two subregions. We note that these differences are less pronounced for voxels at the borders between the anterior and posterior clusters and that the precise location and extent of anterior and posterior subregions is likely to vary between individual participants. Finally, it should be noted that, as is the case with all voxelwise fMRI investigations, the fact that voxels within the anterior and posterior parahippocampal cortex are significantly activated by different fMRI contrasts does not logically demonstrate a functional double dissociation between these two regions.
To assess the relative location of the two identified parahippocampal subregions relative to the location of the PPA (Epstein and Kanwisher, 1998), we used the contrast picture-scene > picture-object to approximate its location (see white outline in Fig. 2a). It is evident that the posterior scene-sensitive subregions of the parahippocampal cortex fall entirely within the PPA, whereas the anterior context-related parahippocampal cortex falls only partially within the PPA. It should be noted, however, that the picture-scene > picture-object contrast can only serve as an approximation for the PPA because it was not independently derived. In addition, given that our experimental task involved contextual judgments, the contrast of picture-scene > picture-object is likely to have encapsulated not only visuospatial influences, but also contextual differences between scenes and objects.
Whole brain
To put our results into the context of more widespread visual and memory networks, we also performed exploratory whole-brain analyses. Using the factorial analysis described in the previous section, we identified several additional clusters that displayed a significant interaction between the factors of stimulus category (objects or scenes) and modality (pictures or words). The highest level of activity was observed in a bilateral cluster encompassing the lateral retrosplenial cortices (Brodmann areas 23, 29, and 30; Fig. 3a, Table 1). In addition, we identified seven clusters in occipital cortex corresponding to Brodmann areas 17, 18, and 19; Table 1).
Using the conjunction analysis described in the previous section, we identified several additional clusters that were engaged by contextual mnemonic processes regardless of stimulus category or modality. The highest level of activity was observed in a large inferior frontal cluster, corresponding approximately to Broca's area (Table 2). In addition, we observed activity in the medial retrosplenial cortex (Brodmann area 30; Fig. 3b, Table 2), as well as the medial frontal and middle temporal cortex (Table 2).
The retrosplenial cortex has been implicated in both the perceptual analysis of scene layouts (Epstein, 2008) and the retrieval of associative contextual memories (Bar and Aminoff, 2003). Our results show that the lateral retrosplenial cortex is engaged during visual scene processing, whereas its medial portion is involved in contextual retrieval, suggesting the existence of functionally segregated subregions for contextual and visuospatial processing.
Discussion
Previous research has shown that the parahippocampal cortex is involved in visuospatial analysis of scenes (Epstein, 2008), as well as contextual mnemonic processing (Aminoff et al., 2013). Until now, however, it has remained unclear whether these two processes are functionally and anatomically segregated within the parahippocampal cortex. We observed that, whereas the posterior parahippocampal cortex was strongly modulated bilaterally by the visual properties of the stimulus material, as evidenced by strong responses in the picture-scene conditions and no significant response in the word-object conditions, the left anterior parahippocampal cortex was active in all contextual judgment task conditions. This suggests a functional segregation, with posterior sections of the parahippocampal cortex supporting visuospatial processing of scenes and anterior sections of the left parahippocampal cortex involved in contextual associative processing. It is important to note that, whereas our study only examined retrieval-related activity for object–object and object–scene relationships, findings from prior studies suggest that the functions of the anterior parahippocampal cortex are not limited to these types of material, but should also be involved in the retrieval of associations involving other stimulus categories such as faces (Düzel et al., 2003; Kirwan and Stark, 2004).
A recent study by Aminoff and Tarr (2015) found activity within the posterior parahippocampal cortex when participants were shown previously viewed spatial configurations of meaningless shapes (regardless of their individual identities), whereas the anterior parahippocampal cortex was more active during the display of previously viewed meaningless shapes (regardless of their spatial configuration). Consistent with our results, these findings can also be interpreted as suggestive of anterior parahippocampal contextual processes and posterior parahippocampal visuospatial processes. Similarly, using an fMRI searchlight analysis, Linsley and McEvoy (2015) found that activity patterns in the anterior parahippocampal cortex were strongly modulated by the presence of objects in scenes, whereas those in the posterior parahippocampal gyrus were not, again suggesting that the anterior parahippocampal cortex is more concerned with contextual than visuospatial analysis of scene images.
Our proposal of a functional segregation along the parahippocampal longitudinal axis is also consistent with findings from a recent fMRI study (Baldassano et al., 2013), which found that anterior voxels in the parahippocampal cortex were more strongly functionally connected to parieto-medial temporal networks (i.e., regions supporting episodic, semantic, and spatial memories), whereas posterior voxels were more strongly functionally connected to occipital networks (i.e., regions concerned with visual processing). Together with the results of the present study, the findings of Baldassano et al. (2013) suggest that the functional topographic segregation of the parahippocampal cortex is underpinned by corresponding segregated connectivity patterns. More broadly, the findings are consistent with a model of parahippocampal function in which visual information is projected from occipital and temporal areas to the posterior parahippocampal cortex (Libby et al., 2012), where encoding of visuospatial information takes place, a process that is computationally more demanding for complex stimuli such as scenes than for simple ones such as objects (Chai et al., 2010). Conversely, we suggest that the anterior parahippocampal cortex encodes incoming information in an amodal format and acts as an interface with other memory regions, including the hippocampal, retrosplenial, and perirhinal memory systems (for an overview, see Ranganath and Ritchey, 2012).
Damage to the parahippocampal cortex can cause impairments in the identification of scenes, as well as deficits in forming new associative memories (Landis et al., 1986; Malkova and Mishkin, 2003). To date, however, human neuropsychological studies have not clearly differentiated between cases in which damage affected anterior or more posterior regions of the parahippocampal cortex (Bohbot et al., 1998, 2000). Based on our results, we would predict that focal lesions of the posterior parahippocampal cortex should impair performance selectively for conditions in which patients must encode or retrieve pictorial scenes because object- and word-processing pathways would remain intact and thus continue to mediate encoding processes for this type of stimulus material. In contrast, a focal lesion of the anterior parahippocampal cortex should lead to wide-ranging impairments in retrieving semantic associations because it would directly disrupt contextual associative memory processes for a wider range of stimulus categories. Future patient studies using high-resolution lesion symptom mapping should be able to test this prediction (Rorden et al., 2007).
Although our data suggest that the posterior parahippocampal cortex is involved in processing spatial features of visual scenes, it is still unclear how visuospatial properties are encoded at the neural level. Previous neuroimaging studies in humans and monkeys have shown that the parahippocampal cortex is robustly engaged in response to depictions of empty 3D spaces devoid of objects, scene layouts, or contextual associations, provided the stimulus contains sufficient high spatial frequency information (Rajimehr et al., 2011; Zeidmann et al., 2012). Further, a recent study by Nasr and colleagues (2014) showed that areas corresponding to the posterior PPA responded more strongly to simple rectilinear visual features, whereas those corresponding to the anterior PPA responded more strongly to complex visuospatial stimuli. However, it has not yet been established whether the parahippocampal cortex responds selectively to different visuospatial features such as textures, colors, and boundaries (Epstein, 2008). It also remains unclear whether the manner in which the anterior parahippocampal cortex encodes and retrieves associative memories differs from that of other regions known to be involved in these processes, such as the hippocampal and retrosplenial regions (Mayes et al., 2007; Vann et al., 2009). Moreover, it is conceivable that different categories of contextual associations (e.g., spatial, temporal, affective, etc.) cluster in different subregions within the anterior parahippocampal cortex. High-resolution fMRI and analysis techniques such as adaptation fMRI (Krekelberg et al., 2006) or multivoxel pattern classification (Norman et al., 2006) could be used to determine any such fine-grained functional topography within the parahippocampal cortex.
Footnotes
This work was supported by the Australian Research Council (ARC) (Discovery Early Career Researcher Award DE120100535 to O.B; ARC Australian Laureate Fellowship FL110100103 to J.B.M.), the ARC-SRI Science of Learning Research Centre (Grant SR120300015), and the ARC Centre of Excellence for Integrative Brain Function (ARC Centre Grant CE140100007).
The authors declare no competing financial interests.
- Correspondence should be addressed to Oliver Baumann, Queensland Brain Institute, the University of Queensland, St Lucia, Queensland, 4072, Australia. o.baumann{at}uq.edu.au