Abstract
The somatotopical organization of the postcentral gyrus is well known, but less is known about the somatotopical organization of area 2, the somatosensory association areas in the postparietal cortex, and the parietal operculum. The extent to which these areas are modulated by attention is also poorly understood. For these reasons, we measured the BOLD signal when rectangular parallelepipeds of varying shape were presented to the immobile right hand or right foot of 10 subjects either discriminating these or just being stimulated. Activation areas in each subject were mapped against cytoarchitectural probability maps of area 2, IP1, and IP2 along the intraparietal sulcus and the parietal opercular areas OP1-OP4.
In area 2, the somatotopical representation of the hand and foot were distinctly separate, whereas there was considerable overlap in IP1 and no clear evidence of separate representations in OP1, OP4, and IP2. The overlap of hand and foot representations increased in the following order: area 3a, 3b, 1, 2, IP1, OP4, IP2, and OP1. There were significant foot representations but no hand representations in right (ipsilateral) areas 3a, 3b, and 1. Shape discrimination using the foot as opposed to stimulation enhanced the signal in OP4 bilaterally, whereas discrimination with the hand enhanced the signal bilaterally in area 2, IP1, and IP2. These results indicate that somatosensory areas in humans are arranged from strong somatotopy into no somatotopy in the following order: 3a, 3b, 1, 2, IP1, OP4, IP2, and OP1. Higher order areas such as IP1, IP2, and OP4 showed task-related attentional enhancement.
- somatotopy
- attention
- somatosensory
- parietal
- operculum
- cortex
Introduction
The human postcentral gyrus is associated with somatic sensation, and an orderly representation of the skin surface (somatotopy) has been demonstrated with various methods (Head and Holmes, 1911; Foerster, 1936; Penfield and Boldrey, 1937; Corkin et al., 1970). Using an observer-independent method in a small population, a subdivision of the human postcentral gyrus has recently revealed probability maps for areas 3a, 3b, 1, and 2 (Geyer et al., 1999; Grefkes et al., 2001). It is not known whether each of these areas contain a full representation of the skin surface.
There are additional somatosensory areas in the parietal operculum and perhaps the retroinsular cortex, but their number and location are still unresolved (Penfield and Rasmussen, 1950; Hari et al., 1983; Roland and Mortensen, 1987; Seitz and Roland, 1992; Burton et al., 1993, 1997; Lin et al., 1996; Hodge et al., 1998; Maldjian et al., 1999; Polonara et al., 1999; Disbrow et al., 2000; Francis et al., 2000; Corbetta et al., 2002; Hagen and Pardo, 2002). Similarly, areas with predominantly somatosensory functions have been described lining the anterior part of the intraparietal sulcus and in the superior parietal lobule (Penfield and Rasmussen, 1952; Roland, 1984; Seitz et al., 1991; O'Sullivan et al., 1994; Burton et al., 1996; Matsumura et al., 1996; Roland et al., 1998). For these areas to qualify as somatosensory areas in humans, a prerequisite would be a full representation of the skin surface (Kaas et al., 1979). In addition, these tentative somatosensory areas may be arranged on the basis of progressive larger receptive fields. A consequence of larger receptive fields could be a decreasing somatotopical organization. One may even expect a transition from somatotopical organization to a nonsomatotopical organization in analogy with higher order visual areas in the inferotemporal cortex of the monkey (Wang et al., 1996). An increase in focal attention can also enhance activity in somatosensory areas in the postcentral gyrus (Roland, 1981), and it is claimed that the effects of attention in behavioral tasks are stronger for higher order sensory areas.
Choi et al. (2002) and Eickhoff et al. (2002) have recently published a cytoarchitectural division of areas lining the anterior and middle part of the human intraparietal sulcus, area IP1 and IP2, the parietal operculum, areas OP1, OP2, OP3, and OP4. Based on this, we examined the hypotheses that (1) each of the presently known and newly delimited cytoarchitectural areas in the postcentral gyrus, anterior parts of the parietal lobules, and parietal operculum contain somatosensory representations of the hand and foot; (2) some of these areas (2, IP1, IP2, OP1-OP4) will have less or no somatotopical organization, and hence be classified as higher order somatosensory areas; and (3) areas, according to hypothesis 2, classified as higher will show attention-dependent signal enhancement in task-relevant conditions. To examine these hypotheses, we compared passive stimulation with rectangular parallelepipeda to either the right hand or foot with a condition of active discrimination. These procedures have been shown previously to activate potential somatosensory areas in the parietal cortex (Roland and Larsen, 1976; Bodegard et al., 2001).
Materials and Methods
The Karolinska Hospital Ethics Committee approved this study. Ten healthy subjects, six males and four females, 21-32 years of age (mean age, 26.8 years) gave informed written consent in accordance with the Helsinki Declaration. All subjects were strongly right handed according to the Edinburgh Inventory (Oldfield, 1971), which was given in English or a version translated into Swedish, depending on the subjects' native language. None of the subjects had a previous history of neurological disease and had normal magnetic resonance (MR) brain images. Psychophysical measurements of the behavioral tasks were performed on all subjects before MR scanning.
Stimuli. Rectangular parallelepipeds were used as stimuli. Fifteen rectangular parallelepipeds varied in shape from cubic (22.57 × 22.54 mm) to more oblong (44.53 × 16.06 mm) (Roland, 1975; Roland and Mortensen, 1987). The shapes were custom made from polyvinyl chloride material compatible with the MR environment. All parallelepipeds were of the same weight, microgeometric surface properties, and thermal properties. Therefore, the objects differed only in shape. Subjects were only shown the objects on completion of the experiment.
Before functional MR scanning, psychophysics was performed on all subjects to determine the discrimination thresholds (Roland and Mortensen, 1987) of the rectangular parallelepipeds. From these psychophysical results, six parallelepiped objects were selected for comparison with the most cubic object (Table 1) such that subjects could determine the cubic object with a probability of detection between 0.7 and 0.8. A set of six objects was used on the hand, and another set of six objects was used for the foot (Fig. 1).
Stimulus location. Stimulation was performed on the right hand or the right foot. To determine the area of stimulation, a parallepiped was dipped in stamp ink and the shape was passively rotated on the right hand and foot. On the right hand, the area of stimulation included the distal palmar surface adjacent to and including the proximal phalanges of digits 2, 3, and 4. Depending on the size of the object, the area of stimulation of the palmar surface ranged from 2036 to 2797 mm2 for all subjects. On the right foot, the area of stimulation included the proximal and distal phalanges of digits 1, 2, and 3 and adjacent parts of the planta. For all subjects, the area of stimulation of the planta ranged from 2036 to 3584 mm2 for the most cubic to the most oblong object, respectively.
Presentation of parallelepiped objects. To minimize any bias in the object stimulations, the following steps were taken. All stimulations were performed by the same author, and identical stimulations were applied on two different occasions. For psychophysical testing and functional MR imaging (fMRI) scanning so that both the author and subject were familiar with all tasks, all subjects had to complete psychophysical testing with a probability of being able to discriminate correctly the cube with an accuracy of at least 0.7 on both the hand and foot, and psychophysical performance was measured during the scanning sessions. Timing was kept by the MRI clock displaying the remaining scan time.
All object presentations were standardized and performed on the subject's right hand or foot. Each parallelepiped was passively rotated over four surfaces in a systematic order. The object was always in contact with the skin while being rotated. The rotation consisted of five systematic moves. For the hand, the shortest side was first placed on the distal palmar surface and sequentially rotated onto the long side on the proximal phalanges 2 and 3. Then, the object was rotated onto the next long side on phalanges 3 and 4 followed by a rotation back to the same short side on the distal palmar surface adjacent to digits 3 and 4 and then rotated onto the long side on the distal palmar surface adjacent to digits 2 and 3. The object was then removed. The stimulus duration was 3.5 sec. An identical procedure was performed on the right foot area of stimulation.
A two-alternative forced choice (2-AFC) paradigm was used to determine the discrimination limits of distinguishing rectangular parallaepipeds from the cube. In each trial, a parallelepiped was compared with the cube. Each trial lasted 10 sec. In the first 3.5 sec, either the parallelepiped or the cube was presented; in the next 3.5 sec, the object for comparison was presented, which was the cube in the case where the parallelepiped was presented in the first interval, or the parallepiped in case where the cube was presented in the first interval. Thus, the cube was the standard appearing in each 2-AFC trial. Immediately after the second object presentation, a verbal response of “one” or “two” was required, depending on which object, the first or the second, respectively, the subject thought was cubic.
Conditions. There were five experimental conditions: (1) discriminating objects on the hand (HD) and (2) on the foot (FD), and (3) object stimulation on the hand (HS) and (4) on the foot (FS), and (5) a rest condition. During all conditions, subjects closed their eyes and wore a blindfold and headphones, through which they were informed at all times about the condition they were about to perform. The order of which the tasks were presented to the subjects was randomized for each scanning session.
Shape discrimination on the hand. For this condition, subjects were instructed that two objects would be passively presented on their right hand, one after the other. They could decide which of the two objects, the first or the second, was the cube. When the second object presentation was complete, they could verbally respond one or two, depending on which of the two objects, the first or the second, respectively, was more cubic.
Shape discrimination on the foot. This condition is the same as the HD condition, except that the instructions and stimulation with the object were adopted for the right foot.
Same shape on the hand. The subjects were instructed that an object would be passively presented twice to the right hand. Because the object was identical both times, no discrimination was needed. On completion of the second presentation, subjects verbally responded with two.
Same shape on the foot. This condition is similar to the HS condition; however, the difference was that the right foot was used for stimulation. The subjects were instructed that an object would be passively presented twice to the right foot. Because the object was identical both times, no discrimination was needed. On completion of the second presentation, the subjects verbally responded with two.
Rest. During the rest condition, subjects were instructed to close their eyes, relax, and not to think of anything in particular. Headphones and a blindfold were worn over the ears and eyes, respectively. On completion of a scan, subjects were questioned as to whether they heard or felt anything during the scan that disturbed them.
Scanning procedure. Subjects lay comfortably in the supine position in the MR camera. A thermoplastic bite piece was used to stabilize the subject's head for the duration of the experiment. The subject's feet and hands were exposed and rested in a comfortable position so that the palmar surface of the hand faced upwards and the sole of the foot was vertical. Subjects were blindfolded and wore noise-attenuating headphones. Instructions were given to the subject via the headphones before each condition.
MR data were acquired on a clinical GE Signa 1.5 T scanner (Signa Horizon Echospeed; General Electric, Milwaukee, WI) fitted with a head coil. An anatomical, high-resolution, three-dimensional, gradient-echo, T1-weighted volume of the whole brain was acquired with a spoiled gradient-recalled acquisition in a steady state sequence [echo time (TE), 4.2 msec; repetition time (TR), 13 msec; field of view (FOV), 240 mm; flip angle, 40; voxel size, 0.96 × 0.96 × 2.0 mm3]. Functional data were acquired with T2-weighted blood-oxygenated level-dependent (BOLD) signals measured with an echo-planar image sequence (TE, 60 msec; TR, 5000 msec; flip angle, 90; 64 × 64 matrix; FOV, 240 mm; voxel size, 3.4 × 3.4 × 5mm3). A functional volume comprised 21 contiguous axial slices. The first 20 sec of data were discarded. A functional scanning session lasted 270 sec (54 volumes). A block design paradigm was used so that one scanning session consisted of 9 × 30 sec blocks. Each 10 sec trial was repeated three times within a block. Odd blocks were rest conditions. Each of the four even blocks was one of the HD, FD, HS, or FS conditions randomized across scanning sessions. Scan sessions were repeated 16 times for each subject. Of the 864 volumes acquired in total, 480 volumes were of the rest state and 96 volumes were acquired for each of the HD, FD, HS, and FS conditions.
Image analysis. Analysis of the reconstructed functional imaging data was performed with statistical parametric mapping 99 (SPM99) (Well-come Department of Cognitive Neurology, UK; http://www.fil.ion.ucl.ac.uk/spm). Data from each subject were treated separately. First, functional data for each subject was corrected for movement with a least squares fit and six-parameter rigid body spatial transformation. Functional and high-resolution anatomical images were coregistered. Anatomical images and subsequently the functional images were aligned to the space of Talairach and Tournoux (1988) with the standard brain (Roland et al., 2001) as a template. Alignment involved the estimation of the optimum 12-parameter affine registration followed by the estimation of nonlinear deformations and three-dimensional discrete cosine transform basis functions. The data were smoothed with an isotropic Gaussian filter with a full width at half maximum (FWHM) of 6 mm. Time series data were temporally smoothed with a Gaussian filter with FWHM of 4 sec.
Individual analysis. Condition-specific effects for each subject were estimated with a fixed-effect general linear model convolved with a 4 sec delayed boxcar waveform as implemented in SPM99. Realignment parameters were included as covariates in the linear model.
Statistical inferences were made at the cluster level such that the omnibus probability of detecting a cluster >100 voxels is <0.05. From the cytoarchitectural probability map, a volume of interest covering area 6, 4a, 4p, 3a, 3b, 1, 2, OP areas, IP areas, and the parietal lobe was used to correct the p values. Boolean intersections were made between brain activation results from multiple contrasts.
Group analysis. A random-effects analysis (Friston et al., 1999) was performed to reveal the group activations for somatotopy and the effect of task-related attention. To examine somatotopy of the group, a random-effect analysis of the contrasts [(HD + HS) - 2 R] and [(FD + FS) - 2 R] was made (Table 2). To examine the effect of task-related attention from the group, a random-effect analysis of the contrasts [(HD - HS) and (FD - FS)] was used.
Cytoarchitectonic mapping. Cytoarchitecturally delineated areas 6, 4a, 4p, 3a, 3b, 1, 2, IP, and OP areas were used for localizing the activations. Delineation of all cytoarchitectonic areas used in this study, including areas 4a and 4p (Geyer et al., 1996), 3a, 3b, 1, (Geyer et al., 1999), and 2 (Grefkes et al., 2001) that have been published previously, all use the same procedures described below. Briefly, histological serial sections (20 μm thick) of 10 postmortem adult human brains were silver stained for cell bodies. The histological sections were matched with a previously acquired MR image of the same brain before fixation and sectioning to account for histological procedural artifacts. A quantitative observer-independent method based on the gray level index was used to measure cortical laminar densities of neurons and to determine borders between cytoarchitectonic areas for each brain (Schleicher et al., 1999). Each delineated cytoarchitectural area was reconstructed as a three-dimensional digital image aligned to the standard brain (Roland et al., 2001) and filtered with an isotropic Gaussian filter with an FWHM of 6 mm. A population map (Roland and Zilles, 1998) was generated by superimposing corresponding cytoarchitectural areas of each postmortem brain in standard anatomical three-dimensional space of the human brain atlas (Roland et al., 2001). From the population map, each cytoarchitectural area was defined by voxels having a higher probability of belonging to this area than to any other area. This produced the 6 mm filtered cytoarchitectural probability map (Bodegard et al., 2001).
The Boolean intersection of all foot versus rest and all hand versus rest was made: [(FD + FS) - 2 R] ∩ [(HD + HS) - 2 R] and [(FD + FS) - 2 R] ∩ [(HD + HS) - 2 R)]. To localize the activations, the intersected images were superimposed on the 6 mm filtered cytoarchitectural probability map. Overlap of all hand intersections with all foot (H∩F) and all hand unions with all foot (H∪F) of the activation images with each area in the cytoarchitectural probability map was calculated (see Fig. 6). The H∩F/H∪F ratio represents the commonality between hand and foot activations for each subject.
Results
Psychophysics
During the scanning, the subjects were on average able to discriminate correctly the cube on the right foot (FD) with 0.80 probability and on the right-hand (HD) with 0.81 probability. A psychophysical curve averaged over 10 subjects is shown in Figure 2.
Imaging representations of hand and foot in individual subjects
A probability map of cytoarchitectural areas 3a, 3b, 1, 2, IP1, IP2, and OP1-OP4 of the cerebral cortex is depicted in Figure 3. The probability map gives the best estimate of localizing a cytoarchitectural area. We will focus on hand and foot representations within the somatosensory and putative somatosensory areas: 3a, 3b, 1, 2, IP1, IP2, and OP1-OP4. The location of areas 3a, 3b, and 1 (Geyer et al., 1999) and area 2 (Grefkes et al., 2001) has been reported previously. More caudally and abutting area 2 is area IP1, lining the rostral part of the intraparietal sulcus. Caudally and somewhat medially to IP1 and abutting IP1 is area IP2 (Choi et al., 2002) (Fig. 3a,c). The four opercular areas are situated in the parietal operculum in the cortex lining the foot of the lateral sulcus (Eickhoff et al., 2002). OP2 and OP3 are localized most medial in the parietal operculum, with OP3 rostrally and OP2 caudally. OP4 is localized laterally in the operculum, lateral to OP3 and extending somewhat into the lateral surface of the supramarginal gyrus. OP1 is located caudally to OP4 also laterally in the parietal operculum and extending into the supramarginal gyrus (Fig. 3b).
Individual analysis
Using these cytoarchitectural areas as the best estimate of localization of somatosensory areas, we tested the hypothesis that each of these areas contains a representation of the hand and foot. For this purpose, the contrast of all foot versus rest [(FD + FS) - 2R] was made to reveal foot representations in each subject. Similarly, the contrast of all hand versus rest [(HD + HS) - 2R] was made to reveal the hand representations in each subject. A subject was considered to have a hand or a foot representation if the respective contrast revealed a statistically significant cluster size (at omnibus p < 0.05). A summary of the 10 subjects is given below, and results from two complete individual cases are presented to illustrate significant clusters (Fig. 4). Cluster locations, sizes, and significance are also reported from a random-effects analysis of all subjects (Tables 3).
Foot
As the right planta and toes were stimulated, one would expect to find representations of the foot in the left hemisphere somatosensory areas. Indeed, there were foot representations in 9 of 10 subjects in the cortex of areas 3b and 1 on the mesial surface of the left hemisphere. (Fig. 4, a and b, illustrates these activations for subjects 1 and 2, respectively). For area 2, four subjects had a significant activity representing the foot located in the top part of the left area 2, above z = 50 (Fig. 4a). Because area 2 does not extend into the mesial surface, these foot activations thus were located on the lateral surface of the hemisphere. Eight subjects had foot representations in the left IP1 and nine subjects had representations in the left IP2 (Fig. 4a,b). Eight of the 10 subjects had a representation of the foot in the left OP1 and OP4 (Fig. 4a,b). There were sparse representations of the foot in the parietal opercular areas OP2 and OP3 on the contralateral side. The right (ipsilateral) area OP3 primarily contained a representation of the foot.
As expected, the hand was represented in area 3b and 1 clustered between Talairach coordinates z = 36 and 60. In addition, the hand representation of area 2 was in approximately the same position mediolaterally (Fig. 4a,b). The majority of the 10 subjects had statistically significant representations of the hand in areas IP2, OP1, and OP4 in the left hemisphere (Fig. 4a,b). There were many ipsilateral activations in these subjects. For the foot stimulation, these appeared even in ipsilateral areas 3b, 1, and 2 (Fig. 4d,e,g). As one proceeds further caudally, almost all subjects had significant clusters in ipsilateral areas IP1, IP2, OP1, and OP4 (Fig. 4). In these areas, there was no obvious segregation in the location of the clusters for the foot and the hand, either contralaterally or ipsilaterally. This is illustrated by the two individual subject results (Fig. 4a,b).
Somatotopy
From the previous analysis of all foot versus rest and all hand versus rest, one can see that the hand and foot representations are somewhat segregated near the superolateral border of the left hemisphere, whereas they seem to engage similar space in the left parietal operculum. If an area is somatotopically organized and the BOLD signal from the hand stimulation and the foot stimulation is of comparable amplitudes, one should find a foot representation by examining the contrast of all foot versus all hand [(FD + FS) - (HD + HS)] in each subject. Conversely, one should see a representation of the hand by examining the contrast of all hand versus all foot [(HD + HS) - (FD + FS)]. The result was that areas 3b and 1 were clearly somatotopically organized, whereas it was not obvious whether the remaining areas were as strongly somatotopical.
Another and more descriptive way to examine somatotopy is to evaluate the extent to which the cluster associated with the foot stimulation would overlap the cluster associated with hand stimulation. We re-examined the contrasts of all foot versus rest [(FD + FS) - 2R] and all hand versus rest [(HD + HS) - 2R] in each subject. For each subject and each cytoarchitectural area, a ratio was calculated to show the extent of hand and foot cluster overlap. A high ratio indicates a large overlap of clusters and, therefore, little somatotopical organization. The ratio was calculated by the intersection of foot and hand activations divided by the union of these two activations (i.e., [(FD + FS) - 2R] ∩ [(HD + HS) - 2R]/[(FD + FS) - 2R] ∪ [(HD + HS) - 2R]). Only if both clusters within the respective cytoarchitectural area exceeded 800 mm3 was this entity calculated (see Materials and Methods). The results are depicted in Figure 4. Areas having a low ratio and, hence, a small overlap of hand and foot clusters indicating a high degree of separation of these regions were areas 3a, 3b, 1, and 2 in decreasing order of somatotopy. The clusters encroaching into areas OP2 and OP3 were, in all subjects, so small for both hand and foot stimulations that they failed to reach the criterion. Therefore, it is not possible to make any statements about the somatotopy of these areas. For the hemisphere contralateral to the stimulated extremities, the ratio increased in the following order: IP1, OP4, IP2, OP1, indicating less and less somatotopy.
The use of a search space restricted to the somatosensory and motor areas of the brain does not permit one to examine whether there were representations of hand and foot outside these areas. When the search space was increased to cover the whole space of the brain, and the significance limit was set to p < 0.05 omnibus, there were representations of the hand and foot caudally to OP2, perhaps located to the retroinsular cortex (Roland and Mortensen, 1987).
Group analysis: the effect of task-related attention
Because no discrimination was required during the HS and FS conditions as identical objects were presented to either the right hand or foot, and because discrimination was required during the HD and FD conditions to discriminate between objects, one could isolate the discrimination component for the hand and foot tasks by contrasting these conditions. This analysis was performed as a random effects analysis of all subjects. The contrasts (HD - HS) and (FD - FS) were made. Significant clusters are listed in Table 3. For the (HD - HS) contrast, significant clusters were found in area 2, IP1, and IP2 (Fig. 5a). A contrast of (FD - FS) revealed peaks in OP4 bilaterally (Fig. 5b).
Discussion
In this study, we examined the known somatosensory areas in the postcentral gyrus, 3a, 3b, 1 (Geyer et al., 1999), and 2 (Grefkes et al., 2001), parcellated by an observer-independent method (Schleicher et al., 1999). Somatosensory stimulation was performed on the immobile palm and planta. In addition, new and putative somatosensory areas, IP1, IP2, and OP1-OP4, parcellated cytoarchitecturally were examined for activations caused by stimulation of the hand and foot. All areas parcellated in the hemisphere contralateral to the stimulation did have representations of the hand and foot in the majority of the subjects, with the exception of areas 3a, OP2, and OP3. The right OP3 primarily contained a representation of the foot. In areas 3b, 1, and 2, representations of the hand and foot were relatively distinct. Although the area 2 foot representation was on the superolateral side of the brain. Areas IP1, IP2, OP1, and OP4 had little or no somatotopical organization (Fig. 4a,b). Discrimination by the use of the foot as opposed to shape stimulation with identical objects enhanced the signal in OP4 bilaterally, whereas discrimination by the use of the hand enhanced the signal bilaterally in area 2, IP1, and IP2 (Fig. 5, Table 3).
A voxel in the probability map of the cerebral cortex only indicates the cytoarchitectural area found in the postmortem brains as being more frequently represented than any other area. For example, the space allocated to area IP1 in the probability map will also contain, with lesser but not negligible probability, representations of adjacent areas 2 and IP2. Furthermore, the individual brain images in this study were transformed into a standard anatomical space the same way, but not with the same software, as the postmortem brains. This might have caused a slight discrepancy in the coregistration of the cytoarchitectural areas and the functional images. However, coregistration errors were small, for example, the positioning of the bottom of the central sulcus where the border between area 3a and 4p is located (Bodegard et al., 2001; Crivello et al., 2002). Smoothing the functional maps and cytoarchitectural maps compensated for any coregistration effects but at the expense of contaminating the borders of BOLD signals and cytoarchitectural areas in adjacent locations. Age-related brain modifications may also exist between the population of subjects in this experiment and the postmortem brain donors. Therefore, the cytoarchitectural areas of the probability map are no more than the best estimate of the location of these areas currently available.
Areas 3a, 3b, 1, and 2 in humans are regarded as somatosensory areas. This is because of their consistent responses to mechanoreceptive stimulation and previous evidence that these areas contain a full representation of the skin surface (Roland and Larsen, 1976; Greenberg et al., 1979; Reivich et al., 1983; Fox et al., 1987; Ginsberg et al., 1987). The present demonstration that each of the areas 3b, 1, and 2 contain a distinct representation of both the hand and the foot is new. This indicates that these areas in humans may be organized after the same scheme of multiple representations of the skin surface as in other primates (Kaas, 1983). However, we cannot conclude that areas 3a, 3b, 1, and 2 and areas IP1, IP2, OP1, and OP4 have a full representation of the skin surface, somatotopical or not, although they all have hand and foot representations and thus qualify as candidates for somatosensory areas. Areas IP1 and IP2 have been activated previously by tasks in which somatosensory stimulation has been predominant, but in these cases, the stimulation was presented to the hand (Roland et al., 1998; Bodegard et al., 2001). Here, the new finding is that these areas are also activated by somatosensory stimulation of the foot. The terminology has also been somewhat confusing as different labels, such as intraparietal area and anterior intraparietal area, and descriptive terms of localization within the supramarginal gyrus have been used for the anterior part of the IP region (Roland et al., 1998; Bodegard et al., 2000, 2001; Fink et al., 2000; Grefkes et al., 2002). Here, we propose that the terminology should be IP1 for the rostral area and IP2 for the area immediately caudal to IP1.
For area 3a, less than half of the subjects had a representation of either the hand or the foot. This could be because our stimuli did not efficiently activate this area, which might be more readily activated by proprioceptive afferents (Naito et al., 1999, 2000, 2002). Thus, we cannot exclude that the failure to activate OP2 could have a similar cause, because temperature or roughness stimuli may be more appropriate for this region (O'Sullivan et al., 1994; Ledberg et al., 1995; Roland et al., 1998). In any case, our failure to activate these areas consistently does not imply that they are not somatosensory.
There have been no studies examining the somatotopy of the region just posterior to area 2. In the newly delimited areas IP1 and IP2, the hand and the foot representations overlapped, and the peak activities for hand and foot showed no consistent segregation. Each of these areas contains subdivisions that may be somatotopically organized but beyond the resolution of the fMRI method used, or that IP1 and IP2 are indeed single areas with large receptive fields and a coarse somatotopy or no somatotopy.
There are additional somatosensory areas in the parietal operculum and perhaps the retroinsular cortex, but their number and location in humans are still unresolved (Penfield and Rasmussen, 1950; Hari et al., 1983; Roland and Mortensen, 1987; Seitz and Roland, 1992; Burton et al., 1993, 1997; Lin et al., 1996; Hodge et al., 1998; Maldjian et al., 1999; Polonara et al., 1999; Disbrow et al., 2000; Francis et al., 2000). Investigation of secondary and associational somatosensory areas has been hindered by the selective nature of these areas and their tendency to respond only to certain stimuli. Activation of the parietal operculum (PO) has been reported by somatosensory discrimination of hardness (Servos et al., 2001), texture (Roland et al., 1998), electrical stimulation of finger and toes (Ruben et al., 2001; Del Gratta et al., 2002; Hagen and Pardo, 2002), and stimulation with moving sponges (Disbrow et al., 2001). At present, it is impossible to ascertain which of the areas OP1-OP4 would represent SII, the second somatosensory area, which at least in monkeys is somatotopically organized (Woolsey and Fairman, 1946; Whitsel et al., 1969; Friedman et al., 1980; Robinson and Burton, 1980; Burton et al., 1995; Krubitzer et al., 1995). A somatotopic organization of SII has been reported in humans (Disbrow et al., 2000; Ruben et al., 2001; but see Gelnar et al., 1998; Hagen and Pardo, 2002). However, cytoarchitectural delineation of human PO reveals four regions within the PO, whereas these studies reported only a single somatotopical field.
From Figure 4, it is evident that motor areas PM, SMA, 4a, and 4p and area 44 were also activated. These motor activations may be attributable to the verbal response required from the subjects on completion of each condition; however, activation of motor areas during passive somatosensory stimulation has been reported previously (Naito et al., 1999; Bodegard et al., 2000; Oishi et al., 2003).
The present data do not support a clear somatotopical organization of the OP1 and OP4 for which we found strong responses from the hand and foot stimulation. Neither did the distribution of hand and foot peak activities in the parietal operculum give reasons to suggest another parcellation that could accommodate somatotopy. OP3 was mostly activated by stimulation of the foot ipsilaterally. Ledberg et al. (1995) found three separate representations of the hand in response to roughness discrimination and brief touch. The present data could be in agreement with this previous finding that was unable to demonstrate any somatotopy. We cannot be certain that this area simply did not respond to our stimulation of the hand and foot and is, in fact, a candidate for SII.
In summary, the present data suggest varying degrees of somatotopy. The decreasing segregation of the hand and foot peak activations (Fig. 6), increasing overlap of hand- and foot-associated activation clusters (Fig. 6), and increasing merging of the hand and foot activations (Fig. 4a,b; Table 2) all suggest the same ranking of somatotopy: area 3b, area1, area 2, area IP1, area OP4, area IP2, and area OP1. If the somatosensory cortex is organized according to the same principle as the visual cortex, areas that are low in the hierarchy will have strong topological organization, whereas areas high in the hierarchy will have little or no topological organization (Boussaoud et al., 1991). According to this, the above ranking of areas should reflect a somatotopical hierarchy of somatosensory areas. The fact that areas high in the hierarchy are more susceptible to attentional modulation than areas lower in the hierarchy is known from the visual cortex (Maunsell and Cook, 2002) and has also been observed in the somatosensory system (Fujiwara et al., 2002). Our finding of task-related attentional modulation of areas 2, IP1, and IP2 and of area OP4 is in accordance with a proposed somatotopical hierarchy.
Footnotes
- Received June 26, 2003.
- Revision received April 29, 2004.
- Accepted April 29, 2004.
This project was supported by European Union Grant DCG-CT-1999-00677 Neurogenerator and by a grant from the Swedish Research Council. The cytoarchitectonic part of this work was supported by Deutsche Forschungsgemeinschaft grants (K.Z.), and Human Brain Project/Neuroinformatics was funded by the National Institute of Biomedical Imaging and Bioengineering, the National Institute of Neurological Disorders and Stroke, and the National Institute of Mental Health (K.Z.).
Correspondence should be addressed to Jeremy Young, Division of Human Brain Research, Department of Neuroscience, Retzius väg 8, A3:3, Karolinska Institute, Stockholm, Sweden S-171 77. E-mail: Jeremy.Young{at}neuro.ki.SE.
Copyright © 2004 Society for Neuroscience 0270-6474/04/245391-09$15.00/0