Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Articles, Behavioral/Systems/Cognitive

Spatiotemporal Convergence of Semantic Processing in Reading and Speech Perception

Johanna Vartiainen, Tiina Parviainen and Riitta Salmelin
Journal of Neuroscience 22 July 2009, 29 (29) 9271-9280; https://doi.org/10.1523/JNEUROSCI.5860-08.2009
Johanna Vartiainen
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Tiina Parviainen
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Riitta Salmelin
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Retrieval of word meaning from the semantic system and its integration with context are often assumed to be shared by spoken and written words. How is modality-independent semantic processing manifested in the brain, spatially and temporally? Time-sensitive neuroimaging allows tracking of neural activation sequences. Use of semantically related versus unrelated word pairs or sentences ending with a semantically highly or less plausible word, in separate studies of the auditory and visual modality, has associated lexical-semantic analysis with sustained activation at ∼200–800 ms. Magnetoencephalography (MEG) studies have further identified the superior temporal cortex as a main locus of the semantic effect. Nevertheless, a direct comparison of the spatiotemporal neural correlates of visual and auditory word comprehension in the same brain is lacking. We used MEG to compare lexical-semantic analysis in the visual and auditory domain in the same individuals, and contrasted it with phonological analysis that, according to models of language perception, should occur at a different time with respect to semantic analysis in reading and speech perception. The stimuli were lists of four words that were either semantically or phonologically related, or with the final word unrelated to the preceding context. Superior temporal activation reflecting semantic processing occurred similarly in the two modalities, left-lateralized at 300–450 ms and thereafter bilaterally, generated in close-by areas. Effect of phonology preceded the semantic effect in speech perception but not in reading. The present data indicate involvement of the middle superior temporal cortex in semantic processing from ∼300 ms onwards, regardless of input modality.

Introduction

The processing of written and spoken words starts with modality-specific analysis of visual features or acoustic signals. These stages are assumed to be followed by further modality-specific processing, i.e., analysis of graphemes of written words and phonetic features of spoken words. When entering core linguistic processes, such as semantic analysis, the question arises at what point, if ever, neural processing becomes independent of the input modality. In models of language processing, and based on the secondary status of reading in language development, retrieval of word meaning from a semantic system and its integration with the context are often assumed to be common for spoken and written words (Patterson and Shewell, 1987).

Magnetoencephalography (MEG) studies that provide information of both timing and location of cortical activation have associated lexical-semantic processing most consistently with sustained activation of the middle superior temporal cortex at ∼200–800 ms, both in the case of written (Simos et al., 1997; Helenius et al., 1998; Halgren et al., 2002; Pylkkänen et al., 2006) and spoken words (Helenius et al., 2002; Uusvuori et al., 2008). However, these studies have focused on either the visual or auditory input modality.

In the present study, we directly compared neural correlates of lexical-semantic processing for written and spoken words in the same individuals using MEG. The specific functional manipulation was realized with semantic priming that has been widely used in behavioral and neuroimaging experiments. In typical semantic priming experiments, the target word is preceded by a prime word that has related meaning (Nobre and McCarthy, 1994). Electroencephalography (EEG) and MEG studies have also used sentences that end with the expected word or a semantically inappropriate one (Kutas and Hillyard, 1980; Connolly et al., 1995; Helenius et al., 1998). To build a strong semantic expectation of the target word, while avoiding contamination by expectation of sound form present in sentence tasks, we used lists of four words. The first three words of the list were semantically related, and the final word either agreed with that context or was semantically unrelated. The areas and time-windows that show suppression of activation over the course of the first three words and increase of activation to a semantically unrelated list-final word may be interpreted as reflecting sensitivity to semantic information. The semantic content of the word must be available in the time window in which the suppression/increase occurs.

Semantic processing was contrasted with phonological analysis that is assumed to precede semantic analysis in speech perception (Patterson and Shewell, 1987) and co-occur with it in reading (Coltheart et al., 1993; Harm and Seidenberg, 2004). In typical phonological priming experiments, the prime and the target share phonemes (Rugg and Barrett, 1987). Our study included phonological lists in which the words had the same initial phonemes but were not related semantically. If suppression/increase is specific to semantic processing, it should not appear in the phonological lists.

Materials and Methods

Subjects.

Ten right-handed Finnish-speaking subjects (5 females; 21–29 years, mean 25 years) participated in the experiment. The subjects had normal or corrected-to-normal vision and normal hearing, and none of them reported a history of neurological abnormalities. Informed consent was obtained from all subjects, in agreement with the prior approval of the Helsinki and Uusimaa Ethics Committee.

Stimuli and experimental design.

Stimuli were lists of four words. The first three words had either a related meaning or began with the same two letters/phonemes (the Finnish language has a one-to-one correspondence between letters and phonemes). The final word of the list either matched the framework set by the first three words or differed from it semantically or phonologically. The lists thus fell into four categories, according to the type of the final word: (1) semantically related, (2) semantically unrelated, (3) phonologically related, and (4) phonologically unrelated. Figure 1 gives an example of each list type (actual stimuli in Finnish). There were 87 word lists per category and 348 lists in total. The subject's task was to press a button when s/he detected a word list in which one word appeared twice (detection rate on average 85%). The probability of these target lists was 6% (20 lists). The target lists were not included in the analysis.

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Examples of the word-list stimuli. In addition to the four list types, there were target lists in which one word appeared twice (probability 6%; data not shown). The actual stimuli were in Finnish. In the earlier auditory version of this experiment (Uusvuori et al., 2008), the same word lists were presented aurally.

The words were bisyllabic 4- to 5-letter common Finnish nouns beginning with a consonant, chosen from a Finnish newspaper corpus (WordMill Lexical Search Program) (Laine and Virtanen, 1999; Uusvuori et al., 2008). Each word (1392 in total) appeared in the experiment once. Semantic lists were created by selecting from the corpus word groups that had a clearly related meaning. In semantic word lists the first phoneme of every word differed from that of every other word more than by voicing only. Most semantic lists included both concrete and abstract words, but in case all the prime words were concrete/abstract the list-final word was also concrete/abstract. There was a 1000 ms interval between the onsets of the successive words in a list. Interval between the onset of the list-final word and the initial word of the next list was 2100 ms.

The auditory data set had been recorded previously (Uusvuori et al., 2008). In the present visual experiment, the words were displayed on a back-projection screen placed at a distance of 1 m from the subject's eyes (visual angle <4°). The MEG response timing was corrected for the 34 ms delay from the stimulus trigger (Presentation software, Neurobehavioral Systems) to picture a display by the projector (VistaPro, Electrohome Limited). The words were written in black Helvetica font on gray background. Each word was shown for 300 ms. The word-list stimuli were presented in a pseudo-randomized order, i.e., neither phonological nor semantic lists appeared more than three times in a row. The subjects were informed about the four types of word lists and instructed to read the lists carefully and detect the targets. The total measurement time was ∼35 min.

MEG recording and data analysis.

MEG recording was conducted in a magnetically shielded room with a vector-view whole-head MEG device (Elekta Neuromag). The system contains 102 triple sensor elements composed of two orthogonal planar gradiometers and one magnetometer. The gradiometers detect the maximum signal directly above an active cortical area. The signals were bandpass filtered at 0.03–200 Hz and digitized at 600 Hz. Horizontal and vertical eye movements were monitored [electro-oculogram (EOG)].

The MEG data were averaged off-line across trials from −0.2 s to 4.2 s relative to the onset of the list-initial word. The averaged MEG responses were baseline corrected to the 200 ms interval immediately preceding the list onset and low-pass filtered at 40 Hz. Trials with MEG amplitude exceeding 3000 fT/cm during the presentation of the list-final word were discarded automatically; thereafter, trials during which major disturbances remained (∼10%) were removed from the data manually. In addition, artifactual, slow ∼0.1 Hz shifts were detected in the data. The disturbances were removed by applying a high-pass filter to the nonaveraged data (center 0.15 Hz, width 0.1 Hz); the high-pass filtering did not affect the experimental effects but removed noise from sensors at the edge of the sensor helmet. Artifacts resulting from eye movements were removed from the data by rejecting trials during which the subject blinked (EOG >150 μV). Additionally, principal component analysis (PCA) (Uusitalo and Ilmoniemi, 1997) was performed on MEG signals averaged off-line with respect to eye movements, and the PCA component that contained artifacts was removed from the responses to the word stimuli.

The experiment yielded on average 65 (at least 48) artifact-free trials for each list type in all subjects. For source modeling, the signal-to-noise ratio for the response to the first word was enhanced by additionally averaging the data across the four word lists. This procedure increased the number of trials to 261 on average (at least 204 trials). Furthermore, the responses to the second and third words were averaged for the two semantic lists and for the two phonological lists, amounting to on average 131 (at least 99) trials.

Sensor-level analysis.

To obtain an initial overview of the data, areal mean signals were calculated over seven areas of interest: left and right frontal, temporal, and occipital areas, and the parietal area. First, vector sums of each gradiometer pair were computed by squaring the two MEG signals, summing them together, and calculating the square root of this sum. The areal mean signals were obtained by averaging these vector sums for each area of interest, individually for each subject. Finally, the areal mean signals were averaged across subjects. Because of the way the sensor-level areal mean signals are calculated (square root of sum of squared signals) they always have a positive value (>0). The areal mean signals were characterized by measuring, for each of the seven areas, the mean amplitude in three time windows, 50–250 ms, 250–400 ms, and 400–650 ms after the onset of each word in the list.

Equivalent current dipole analysis.

An equivalent current dipole (ECD) represents the center of an active cortical patch, and the mean orientation and strength of electric current in that area (Hämäläinen et al., 1993). The ECD analysis was done with the Elekta Neuromag software package, and proceeded as described previously (Salmelin et al., 1996; Helenius et al., 1998; Wydell et al., 2003). ECDs were localized individually in each subject. The whole-head magnetic field patterns were visually inspected for dipolar field patterns, and the ECDs were identified one by one at time points at which the specific field patterns were clearest. ECDs were determined from a subset of planar gradiometers that covered the spatially and/or temporally distinct magnetic field patterns. The 95% confidence volume of source location, a measure reflecting the signal-to-noise-ratio, was automatically obtained for each source. Thereafter, time courses of activation in those brain areas (source waveforms) were estimated by including the ECDs simultaneously in a multidipole model: The locations and orientations of the ECDs were fixed, whereas their amplitudes were allowed to vary to best account for the measured data in each condition with the selected model. The final multidipole models typically accounted for >85% (at least 75%) of the total magnetic field variance at the activation peak in each condition. In each subject, the cortical activation evoked by all words in the four types of lists was well represented by the same set of ECDs. The final models were composed of 10–13 ECDs (mean 12).

The location of the ECDs was defined in the head coordinate system that was set by the nasion and two reference points anterior to the ear canals. Before the MEG recording, four head position indicator (HPI) coils were attached to the subject's head and their locations were measured with a three-dimensional digitizer (Polhemus). At the beginning of the recording, the HPI coils were briefly energized to determine their location with respect to the MEG helmet. For visualization and comparison of the sources between subjects, the ECDs were transformed to a standard brain (Roland and Zilles, 1996) using elastic transformation (Schormann et al., 1996; Woods et al., 1998).

Neural activation within the first 300 ms was typically transient and was adequately characterized by the maximum level of activation and the time at which it was reached (peak amplitude and peak latency of the source waveforms); when more sustained activation in the posterior brain areas was detected after 300 ms it was described by measuring the mean amplitude between 300 and 600 ms post stimulus. Sustained activation in the temporal areas, of particular interest for the research question in the present study, was characterized with a number of measures. Its time course was described by the peak latency and the time points at which the waveform reached 50% of the peak amplitude in the ascending and descending slopes (onset and offset latencies). As for the activation strength, the peak amplitude was measured and the mean amplitude was calculated separately for the ascending and descending slopes, over 200 ms intervals before and after the group mean peak latency. The signal-to-noise ratio of the temporal activation in the right hemisphere was too low to allow definition of peak latency; the left-hemisphere time windows were used for the right hemisphere as well. All measures were collected individually for each subject and separately for each condition.

Minimum current estimates.

Source locations were additionally determined using minimum current estimates (MCEs) to verify the spatial distribution of activity obtained with ECD modeling and to enable comparison with MEG experiments that have used distributed modeling approaches (Halgren et al., 2002; Marinkovic et al., 2003). The minimum current estimate (Uutela et al., 1999) is an implementation of the minimum L1-norm estimate (Matsuura and Okabe, 1995). The measured signals are accounted for by a distribution of electric current that has the minimum total amplitude. MCEs were computed separately for each subject's data and, thereafter, averaged across subjects.

Statistical tests.

The following statistical tests were conducted on the sensor level data and on the ECD source waveforms. The effects on the strength and timing of activation in the different areas were tested using two separate tests, one for the first three words of the list and the other for the list-final word: The buildup of semantic or phonological context was investigated during the first three words of the list using a repeated-measures ANOVA with within-subjects factors Condition (semantic list, phonological list) and Position (first word, second word, third word). When a significant main effect of Position was detected, contrasts (linear and quadratic trend) were evaluated. When an interaction Condition × Position was detected, contrasts (linear and quadratic trend) were evaluated separately for semantic and phonological lists to detect possible effects specific to semantic or phonological manipulation. For the list-final word, the effects were tested using a repeated-measures ANOVA with within-subjects factors Condition (semantic, phonological) and Congruence (related, unrelated). When the sphericity assumption was violated (Mauchly's test of sphericity), Greenhouse–Geisser correction was used.

When the visual data set showed activation in approximately the same brain area and time window as the previously measured auditory data set (Uusvuori et al., 2008), activation strength and timing of the ECDs were evaluated using a repeated-measures ANOVA with Modality (auditory, visual) and Hemisphere (left, right) as within-subjects factors. When data from only one hemisphere was available, the repeated-measures ANOVA was conducted with the within-subjects factor Modality (auditory, visual). The ECD source locations were compared between the auditory and visual experiment in the anterior–posterior, superior–inferior, and left–right directions using t test. To check whether neural priming occurred in the same individuals in both the visual and auditory modality, the significance of the semantic and phonological priming effects was evaluated separately in each subject. A permutation test (Holmes et al., 1996; Nichols and Holmes, 2002; Laaksonen et al., 2008) was applied to the mean activation strength in the time window extending from 200 ms before to 200 ms after the group mean peak latency. Separately for the left and right temporal source areas, the activation strengths in the single trials corresponding to semantically related (SemRel) and unrelated (SemUnrel) list-final words were randomly shuffled 3000 times, and after each shuffle, the trials were divided into two samples that were compared using a t test. The t value of the original comparison of SemRel versus SemUnrel samples was compared against the distribution of t values derived from the permutations, and the SemRel and SemUnrel samples were considered significantly different if the original t value was not contained within 95% of the shuffled t values (p < 0.05). A similar test was conducted for the phonologically related versus unrelated trials.

Results

Sequence of cortical activation in reading

Figure 2 displays the areal mean signals for the different word-list types, averaged across all subjects. The strongest signals were detected over the occipital and left temporal areas. Each word of the list elicited a response at ∼100 ms after the stimulus onset over the occipital area followed by a more sustained response over the left temporal area that reached the maximum at ∼400 ms after the stimulus onset. Salient stimulus effects occurred between 300 and 600 ms over the left temporal area.

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Group averaged areal mean signals in reading. Group averaged areal mean signals calculated over the left and right occipital, temporal, and frontal cortex, and the parietal cortex. Note the marked effect of stimulus type over the left temporal cortex.

Figure 3 illustrates the active areas that were found consistently (in at least six subjects) and their mean time course of activation, as modeled by equivalent current dipoles (ECDs). Occipital activation at ∼100 ms poststimulus was followed by left occipitotemporal activation at ∼150 ms. Right occipital cortex showed a transient response at ∼150–200 ms and a more sustained signal at ∼400–800 ms after each word. Sources around the parieto–occipital midline displayed a similar sequence of transient and sustained activation. The only sources that systematically demonstrated meaningful stimulus effects, i.e., priming-induced suppression over the first three words and stronger activation to unrelated than related final words, were located in the superior temporal cortex, with the sustained activity peaking at ∼400–500 ms. The time course of the strong activation in the left temporal cortex was practically identical to the areal mean signal recorded over the left temporal lobe (Fig. 2). The right-hemisphere activation showed similar stimulus effects but was relatively weak. The location, orientation, and time course of the superior temporal ECDs suggest that they correspond to the N400m sources described in earlier MEG studies of reading (Simos et al., 1997; Helenius et al., 1998). In the following, we refer to the sustained activity peaking at ∼400–500 ms as the N400m response.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Source clusters in reading. Clusters were formed on the basis of location and timing and include data from at least six subjects. White dots and black tails indicate the individual source locations and directions of current flow, respectively. The curves show the mean time course of activation for each source cluster. For the list-initial word, the response was averaged over all word lists, and for the second and third word, over semantic or phonological lists (see Materials and Methods). For the list-final word, all four categories are plotted separately. In the auditory experiment (data not shown), clusters of active areas were detected in the temporal lobes only. subj, Subjects.

The sequence of activation observed using MCEs agreed with that detected using ECDs and confirmed the dominant role of the superior temporal activation (Fig. 4A; supplemental Figs. S1, S2, available at www.jneurosci.org as supplemental material). The group-level MCE additionally suggested activation in the left and right temporal pole or anterior medial temporal cortex, directly behind the eyes. However, inspection at the individual level revealed that activation in this area was typically either very short-lived (<50 ms) or it was localized outside of the brain, in or near the extraocular muscles.

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

Superior temporal activation in the visual versus auditory modality and effect of semantic and phonological manipulation. A, Minimum current estimate of the response to the list-initial word in the visual (top) and auditory (bottom) experiment. The top row in each set shows the MCE for a single subject and the bottom row MCE averaged across 10 subjects. Distribution of electric current was projected on a triangle mesh representing the brain surface and integrated over the time interval of interest (400–500 ms after the stimulus onset). The encircled colored areas denote the surface projection of the regions of interest with the maximum activation in the left and right hemisphere. In the visual modality, large interindividual variability in source location led to a weak and blurred group-level estimate, as opposed to the concise source cluster and strong group-level MCE in the auditory modality. B, N400m source locations in the visual and auditory experiment, as revealed by ECD analysis (compare Fig. 3). C, The relative locations of the left-hemisphere N400m sources in the visual (triangles) and auditory (circles) experiment, with a line connecting the two sources of each subject. The inset shows the relative locations when the auditory N400m sources were placed at the origin. D, N400m source waveforms for the four types of list-final words in the two modalities. The gray horizontal bar indicates the noise level (±2 SD of the signal level during the 200 ms period immediately preceding the list onset).

Effects of semantic and phonological priming in reading

Sensor-level effects

Effects of semantic and phonological manipulation were most marked over the left temporal region (Fig. 2; Tables 1, 3). In the list-final word, the response was increased to a semantically incongruent word at 250–650 ms (250–400 ms: Condition × Congruence F(1,9) = 9.0, p < 0.05, semantic lists: Congruence F(1,9) = 38.1, p < 0.001; 400–650 ms: Condition × Congruence F(1,9) = 6.4, p < 0.05, semantic lists: Congruence F(1,9) = 16.7, p < 0.005). During the first three words, the responses were weaker to the semantic than phonological lists (400–650 ms: Condition F(1,9) = 7.7, p < 0.05). For phonological list-final words, an increase to a phonologically incongruent word approached significance (250–400 ms, phonological lists: Congruence F(1,9) = 4.2, p = 0.070; 400–650 ms, phonological lists: Congruence F(1,9) = 3.4, p = 0.096). Over the left frontal region, the response was increased to incongruent list-final words, regardless of the list type (semantic or phonological) (250–400 ms: Congruence F(1,9) = 12.6, p < 0.01; 400–650 ms: Congruence F(1,9) = 7.9, p < 0.05).

View this table:
  • View inline
  • View popup
Table 1.

Statistically significant effects in the visual experiment: sensor-level analysis

View this table:
  • View inline
  • View popup
Table 2.

Statistically significant effects in the visual experiment: source-level analysis

View this table:
  • View inline
  • View popup
Table 3.

Statistically significant effects in the visual experiment: sensor-level analysis

View this table:
  • View inline
  • View popup
Table 4.

Statistically significant effects in the visual experiment: source-level analysis

In addition to the semantic and phonological priming effects relevant for the present research question, general effects of word position in the list were detected as well. The response was increased when advancing from the first to the second word of the list and thereafter diminished when advancing from the second to the third word, regardless of the list type (semantic or phonological). These effects (main effect of Position and quadratic trend) were significant in several time windows over all the left-hemisphere regions, and right frontal region (Table 1).

Source-level effects

Stimulus effects were observed in the superior temporal cortex bilaterally (Fig. 3; Tables 2, 4). Suppression of activity by semantic or phonological priming over the course of the first three words and enhanced activation to an unrelated list-final word were detected as follows:

Effects of semantic manipulation emerged in the left hemisphere. The N400m activation was attenuated over the course of the first three words for the semantic lists but not for the phonological lists [peak amplitude: Condition F(1,9) = 10.1, p < 0.05, Condition × Position F(2,18) = 5.6, p < 0.05, semantic lists: linear trend F(1,9) = 7.2, p < 0.05, phonological lists: all contrasts not significant (n.s.); mean amplitude in the descending slope: Condition F(1,9) = 7.7, p < 0.05, Condition × Position F(2,18) = 5.2, p < 0.05, semantic lists: linear trend F(1,9) = 8.6, p < 0.05, phonological lists: all contrasts n.s.]. In the list-final word, the left-hemisphere N400m activation was increased to a semantically incongruent word, both during the ascending and descending slope (mean amplitude in the ascending slope: Congruence F(1,9) = 5.9, p < 0.05; peak amplitude: Condition × Congruence F(1,9) = 15.1, p < 0.005, semantic lists: Congruence F(1,9) = 46.2, p < 0.001; mean amplitude in the descending slope: Condition × Congruence F(1,9) = 10.1, p < 0.05, semantic lists: Congruence F(1,9) = 33.7, p < 0.001). Semantic effects were detected also in the duration of the response: The shortest duration was observed for the semantically related list-final word (offset earlier than for the other word types) (Condition × Congruence F(1,9) = 12.7, p < 0.01, semantic lists: Congruence F(1,9) = 15.7, p < 0.005). Phonological effect was observed in the ascending slope of the left-hemisphere N400m activation as increased mean amplitude to an incongruent list-final word (Congruence F(1,9) = 5.9, p < 0.05). The N400m response also reached the maximum earlier for the phonologically unrelated than related list-final words (Condition × Congruence F(1,9) = 11.5, p < 0.01, phonological lists: Congruence F(1,9) = 8.3, p < 0.05).

In the right hemisphere, a main effect of Condition (F(1,6) = 7.1, p < 0.05) and an interaction Condition × Position (F(2,12) = 5.6, p < 0.05) were detected in the descending slope but contrasts did not reach significance for either the semantic or phonological lists. In the list-final word, both semantic and phonological priming effects were significant; the peak amplitude and the descending slope of the N400m activation were weaker to the related than unrelated final words regardless of the list type (peak amplitude: Congruence F(1,6) = 9.3, p < 0.05, mean amplitude in the descending slope: Congruence F(1,6) = 9.7, p < 0.05).

In addition to the superior temporal priming effects, an effect of word position in the list was detected in the timing of the occipital activation peaking at ∼100 ms. This response reached the maximum earlier when advancing in the word list, regardless of the list type (semantic or phonological) (Position F(1.2,10.9, Greenhouse–Geisser ε 0.6) = 9.3, p < 0.01, linear trend F(1,9) = 10.6, p < 0.05).

Written versus spoken words

Auditory version of the experiment

MEG data of auditory word-list processing was collected with the same device and the data were analyzed similarly as in the present visual experiment (Uusvuori et al., 2008). In the auditory data, activation was concentrated to the left and right superior temporal cortex. These areas were involved in both analysis of sound form and meaning but the role of each hemisphere varied over time. Sensitivity to sound form was observed at ∼100 ms after word onset, followed by sensitivity to semantic aspects from ∼250 ms onwards, in the left hemisphere. From ∼450 ms onwards, the picture was changed, with semantic effects present bilaterally, accompanied by a late effect of sound form in the right hemisphere. These results obtained with ECD analysis (Uusvuori et al., 2008) (supplemental Text S1, available at www.jneurosci.org as supplemental material) were corroborated by minimum current estimates (supplemental Figs. S3, S4, available at www.jneurosci.org as supplemental material).

Comparison of the visual and auditory data

The left superior temporal cortex showed activation in both the visual and auditory tasks in all 10 subjects (Fig. 4). The right superior temporal cortex was active in all subjects during speech perception; in reading, corresponding activation in the right hemisphere was detected in seven subjects (auditory–visual difference: p = 0.25, n.s., nonparametric McNemar test).

At the group level, the N400m source location (Fig. 4C) was not statistically significantly different in the visual and auditory modalities (left hemisphere, anterior–posterior t(9) = 0.9, p = 0.4, n.s, inferior–superior t(9) = 1.8, p = 0.1, n.s, lateral–medial t(9) = 0.1, p = 0.9, n.s.). However, the N400m sources were tightly clustered in the auditory experiment (SD, 10, 6, and 5 mm in the anterior–posterior, inferior–superior, and lateral–medial direction) but somewhat more scattered in the visual experiment (SD, 12, 9, and 6 mm). The distance between the left auditory and visual N400m sources of the same subject was 6 ± 1 mm on average, and in 8 out of 10 subjects the 95% confidence volumes of the two ECDs did not overlap. The larger interindividual variability of source location (reflecting the center of an active cortical patch) in the visual than auditory task was also evident in the MCE (Fig. 4A). At the level of a single subject, there was clear focal activation of the left temporal cortex both in the visual and auditory experiment. At the group level, however, the picture remained salient in the auditory task but became blurred in the visual task.

The ECDs, representing the centers of the active cortical patches in the individual subjects, typically fell on the STG (left hemisphere: auditory six subjects, visual five subjects; right hemisphere: auditory four subjects) or in the neighboring sulci (left hemisphere: auditory four subjects, visual two subjects; right hemisphere: auditory four subjects). In the visual experiment, a right-hemisphere N400m source was detected in seven individuals, but its location was quite variable because of the relatively low signal-to-noise ratio. MEG detects the center of an active cortical patch but does not provide direct information of its extent or shape, regardless of the analysis method. The localization is most accurate perpendicular to the direction of current flow (here anterior–posterior) and least accurate along it (here inferior–superior). Furthermore, because of the orientation of the apical dendrites of the pyramidal cells that mainly generate the MEG signal and the high sensitivity of MEG to current flow oriented tangential to the brain surface, MEG is most sensitive to activation in the sulci (Hämäläinen et al., 1993). Within the spatial accuracy of the MEG method, the present data thus point to activation of the middle section of the STG and the flanking sulci (superior temporal sulcus and Sylvian fissure) as the most consistent neural source of the N400m response.

Figure 4d illustrates the mean time course of activation to the list-final words in the left and right hemisphere, for the visual and auditory modality. The left-hemisphere N400m semantic priming effect on the list-final word was an extremely robust finding as it was evident also in the responses of each individual subject. A permutation test for stronger activation to unrelated than related list-final words (see Materials and Methods) confirmed a robust left-hemisphere semantic effect in the visual (auditory) experiment that was significant (p < 0.05) in seven (six) subjects and approached significance (p < 0.09) in two (three) subjects. In one subject, the semantic effect did not reach significance in either modality. The effects of semantics in the right hemisphere and of phonology in either hemisphere were each significant in up to five individuals.

The superior temporal activation was overall significantly stronger to the auditory than visual words, regardless of the hemisphere (N400m peak amplitude: Modality F(1,6) = 13.7, p < 0.05). In the visual experiment, the N400m activation seemed weaker in the right than left hemisphere, with poorer signal-to-noise ratio (Fig. 4D). Such differences would tend to suggest stronger lateralization in the visual task than in the auditory task in which the activation appeared more bilateral. The hemisphere difference, however, did not reach significance (N400m peak amplitude: Modality × Hemisphere F(1,6) = 0.2, p = 0.7, Hemisphere F(1,6) = 1.8, p = 0.2).

The duration of the left-hemisphere N400m response was longer in the auditory experiment (∼450 ms, full width at half maximum) than in the visual experiment (∼150 ms). This difference resulted from significantly earlier onset latencies (Modality F(1,9) = 6.1, p < 0.05) and delayed offset latencies (Modality F(1,9) = 242.7, p < 0.001) in the auditory than visual experiment (average onset and offset latencies across subjects, list types and words in the list: auditory 285 ± 3 ms and 720 ± 7 ms, visual 326 ± 5 ms and 471 ± 7 ms). The left N400m response reached the maximum ∼80 ms earlier for visual (393 ± 5 ms) than auditory presentation (471 ± 5 ms) (F(1,9) = 39.3, p < 0.001). In the right hemisphere, the N400m response to auditorily presented words peaked at 448 ± 9 ms (left vs right hemisphere n.s.); in the visual experiment no latency values were extracted in the right hemisphere because of the poor signal-to-noise ratio.

Discussion

The pattern of activation evoked by visually presented words indicated that, following initial modality-specific analysis, activation reflecting semantic processing converged on that evoked by spoken words, left-lateralized at 300–450 ms and thereafter bilaterally, generated in close-by areas in the superior temporal cortex (Fig. 5). The left-hemisphere semantic effect was an extremely robust finding both at the group and individual level, and it emerged at the sensor and source level. Analysis of sound form, although manifested in activation of the same general area, displayed modality-specific differences in timing. For written words, phonological priming influenced the left-hemisphere responses at ∼300–450 ms, whereas for spoken words, phonological priming effect was detected at ∼100 ms. From ∼450 ms onwards, phonological priming effect was observed in the right hemisphere in both modalities. Effects specific to semantic and phonological manipulation and auditory and visual modality were thus concentrated primarily to the left hemisphere, whereas general effects of priming were seen in the right hemisphere from ∼450 ms onwards.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Summary of the main results. Schematic representation of the time windows and hemispheric interplay of phonological and semantic effects, overlaid on the N100m/N400m source waveforms. The dark and light gray bars indicate the time windows in which phonological and semantic priming, respectively, influenced the response. The areas and time windows that showed increased activation to semantically (phonologically) unrelated list-final words were interpreted to reflect sensitivity to semantic (phonological) information. These effects were corroborated by suppression of activation over the course of the first three words of the semantically (phonologically) related lists.

The overall spatiotemporal sequence of cortical activation for written words is in line with previous MEG reports (for review, see Salmelin, 2007). Activation in the occipital cortex at ∼100 ms was observed in all subjects, presumably reflecting visual feature analysis (Tarkiainen et al., 1999). This response was followed by activation in the left inferior occipitotemporal cortex at ∼150 ms in 6/10 subjects, closely resembling the letter-string specific response described earlier (Tarkiainen et al., 1999). We also detected sources in the right inferior occipitotemporal cortex (7/10 subjects) at ∼150–200 ms; activation in this area and time window has been previously reported in reading tasks, equally strong to letter and symbol strings but sensitive to vertical shifts in the letter positions (Salmelin et al., 1996; Tarkiainen et al., 1999; Cornelissen et al., 2003). From 200–300 ms onwards, processing of written and spoken words converged to superior temporal cortex bilaterally (N400m response), but with more individuals showing activation in the left than right hemisphere in the visual modality.

Semantic priming effects for written and spoken words, when measured in the same individuals in the two modalities, were similar in function, timing, general location, and lateralization. This result is consistent with information obtained from EEG and MEG experiments that have separately investigated lexical-semantic processing of written or spoken words and detected semantic effects in the N400 time window (∼200–800 ms) in the visual (Kutas and Hillyard, 1980; Helenius et al., 1998) and auditory (Connolly and Phillips, 1994; Helenius et al., 2002) modality. The data thus point to shared neural representation of word meaning between the two modalities that has been suggested by word-processing models (Patterson and Shewell, 1987).

The relative locations of the source areas in the visual and auditory modality were evaluated in the left hemisphere where all participants had salient N400m responses in both modalities. The locations were highly similar although not identical in individual subjects. Neurophysiological imaging methods (MEG, EEG) detect the center of an active brain area but do not provide direct information about its extent or shape, regardless of the analysis method. The observed small differences in source location between the two modalities thus suggest a relative shift in the centers of the active areas, which may, overall, be largely overlapping. There was no systematic direction of the shift across individuals. Based on the present MEG data, therefore, it is not possible to define a neuroanatomical distinction between the two modalities. The data thus support a view of a largely shared semantic system in auditory and visual language perception, in line with several hemodynamic studies on language comprehension that have found overlapping activation between modalities (Booth et al., 2002; Spitsyna et al., 2006) [meta-analysis of Vigneau et al. (2006)].

In the present experiment, activation associated with semantic processing was found in the superior temporal cortex, with the center of activation most consistently falling on the middle section of the superior temporal gyrus and in the neighboring sulci, in line with previous MEG findings (Simos et al., 1997; Helenius et al., 1998; Pylkkänen et al., 2006) and lesion work (Hagoort et al., 1996; Friederici et al., 1998) (for review, see Van Petten and Luka, 2006). The present findings are in general agreement also with hemodynamic studies that have typically indicated activation in the anterior, posterior, and inferior parts of the temporal cortex (Chee et al., 1999; Booth et al., 2002; Jobard et al., 2003; Spitsyna et al., 2006; Vandenbulcke et al., 2007; Lau et al., 2008).

We found no effects of semantic manipulation in the frontal cortex, in contrast with hemodynamic data that typically show overlap between modalities in semantic tasks in this region, particularly in the left inferior frontal gyrus (Chee et al., 1999; Booth et al., 2002). The lack of systematic semantic activation in the frontal lobe in the present study is, however, in agreement with hemodynamic studies that have investigated implicit semantic processing (Crinion et al., 2003; Spitsyna et al., 2006). Based on hemodynamic measurements and lesion data that show severe semantic impairments mainly resulting from temporal lesions (for review, see Thompson-Schill et al., 1998), it has been suggested that the left inferior frontal cortex is involved in semantic processing when selection or manipulation of such information is required, whereas temporal areas, in particular the middle and inferior temporal cortex, would be the primary store of semantic representations (Fiez, 1997; Poldrack et al., 1999; Rissman et al., 2003; Heim et al., 2005). Such a division of labor could explain the lack of frontal activation in the current experiment in which the semantic manipulation was built in the stimulus set but the task itself was nonsemantic (detection of words that were presented twice within a word list).

In contrast to semantic priming effects, the effect of phonological priming differed between the modalities. In line with previous EEG/MEG data gathered separately in the auditory (Näätänen et al., 1997; Parviainen et al., 2005) and visual modality (Rugg, 1984; Wydell et al., 2003), the analysis of sound form began clearly later for written words (at ∼300 ms) than for spoken words (at ∼100 ms after the word onset). However, it should be noted that whereas sensitivity to auditory phonological priming in our experimental design indicates processing at acoustic, phonetic, and/or phonological level, in the visual modality phonological priming effect should reflect visual, orthographic, or phonological analysis. As analysis of visual features and graphemes is reflected in cortical responses at ∼100–200 ms (Tarkiainen et al., 1999), the effects of phonological priming from ∼300 ms onwards imply analysis of phonological information.

Because of its long duration (from ∼200 to ∼800 ms), the N400m activation is unlikely to reflect a single uniform processing stage of written and spoken words. Indeed, the present data revealed two functionally distinct phases. This result is in line with previous auditory data indicating unequal sensitivity to experimental manipulations in the ascending and descending slopes of the N400m response (Bonte et al., 2006) or two functionally distinct EEG responses within this time window (Connolly and Phillips, 1994; Hagoort and Brown, 2000). Based on those results, it has been suggested that the memory representations of words are activated and selected within ∼300 ms from the word onset, and that the activation at ∼400 ms reflects postlexical integration of the retrieved semantic information to the context (Connolly and Phillips, 1994). This issue is not resolved, however, as other authors have suggested that the N400 response mainly reflects lexical access (Lau et al., 2008), or that both access and integration occur within ∼200 ms from the word recognition point (Hauk et al., 2006; Pulvermüller et al., 2006). In the visual modality, it has been proposed that the monophasic N400 response measured using EEG and the superior temporal activation detected with MEG reflect both lexical access and postlexical integration (Anderson and Holcomb, 1995; Hagoort and Brown, 2000; Kiefer, 2002; Pylkkänen and Marantz, 2003). The present data support the interpretation that at least two different processing stages are reflected in the N400m activation as evoked by both auditory and visual words. The ascending slope of the N400m response (∼200–450 ms) lies in the time window when lexical-semantic information of written and spoken words is likely to be accessed (Lau et al., 2008). However, as behavioral data indicate that both written and spoken words can be recognized within ∼500 ms or less (for review, see Tyler et al., 2002; Balota et al., 2004), the effects of both meaning and sound form from 450 ms onwards are likely to reflect postlexical integration of all available information in the context created by the preceding words that may well occur similarly for written and spoken words.

The present MEG data indicate involvement of the middle part of the superior temporal cortex in semantic processing from ∼300 ms onwards, regardless of input modality. The superior temporal cortex was involved in the analysis of phonological structure as well but with clearly different timing between modalities, in concordance with theoretical models of language perception.

Footnotes

  • This study was financially supported by the Finnish Cultural Foundation, the Academy of Finland (National Centres of Excellence Programme 2006–2011, Grant 129160), the Sigrid Juselius Foundation, and the Finnish Ministry of Education (Finnish Graduate School of Neuroscience). We thank Kimmo Uutela and Lauri Parkkonen for valuable advice on data analysis, and Hannu Laaksonen for supplying us with the permutation test.

  • Correspondence should be addressed to Johanna Vartiainen, Brain Research Unit, Low Temperature Laboratory, Helsinki University of Technology, PO Box 5100, 02015 TKK, Espoo, Finland. juusvuor{at}neuro.hut.fi

References

  1. ↵
    1. Anderson JE,
    2. Holcomb PJ
    (1995) Auditory and visual semantic priming using different stimulus onset asynchronies: an event-related brain potential study. Psychophysiology 32:177–190.
    OpenUrlCrossRefPubMed
  2. ↵
    1. Balota DA,
    2. Cortese MJ,
    3. Sergent-Marshall SD,
    4. Spieler DH,
    5. Yap M
    (2004) Visual word recognition of single-syllable words. J Exp Psychol Gen 133:283–316.
    OpenUrlCrossRefPubMed
  3. ↵
    1. Bonte M,
    2. Parviainen T,
    3. Hytönen K,
    4. Salmelin R
    (2006) Time course of top-down and bottom-up influences on syllable processing in the auditory cortex. Cereb Cortex 16:115–123.
    OpenUrlAbstract/FREE Full Text
  4. ↵
    1. Booth JR,
    2. Burman DD,
    3. Meyer JR,
    4. Gitelman DR,
    5. Parrish TB,
    6. Mesulam MM
    (2002) Modality independence of word comprehension. Hum Brain Mapp 16:251–261.
    OpenUrlCrossRefPubMed
  5. ↵
    1. Chee MW,
    2. O'Craven KM,
    3. Bergida R,
    4. Rosen BR,
    5. Savoy RL
    (1999) Auditory and visual word processing studied with fMRI. Hum Brain Mapp 7:15–28.
    OpenUrlCrossRefPubMed
  6. ↵
    1. Coltheart M,
    2. Curtis B,
    3. Atkins P,
    4. Haller M
    (1993) Models of reading aloud: dual-route and parallel-distributed-processing approaches. Psychol Rev 100:589–608.
    OpenUrlCrossRef
  7. ↵
    1. Connolly J,
    2. Phillips N
    (1994) Event-related potential components reflect phonological and semantic processing of the terminal word of spoken sentences. J Cogn Neurosci 6:256–266.
    OpenUrlCrossRefPubMed
  8. ↵
    1. Connolly JF,
    2. Phillips NA,
    3. Forbes KA
    (1995) The effects of phonological and semantic features of sentence-ending words on visual event-related brain potentials. Electroencephalogr Clin Neurophysiol 94:276–287.
    OpenUrlCrossRefPubMed
  9. ↵
    1. Cornelissen P,
    2. Tarkiainen A,
    3. Helenius P,
    4. Salmelin R
    (2003) Cortical effects of shifting letter position in letter strings of varying length. J Cogn Neurosci 15:731–746.
    OpenUrlCrossRefPubMed
  10. ↵
    1. Crinion JT,
    2. Lambon-Ralph MA,
    3. Warburton EA,
    4. Howard D,
    5. Wise RJ
    (2003) Temporal lobe regions engaged during normal speech comprehension. Brain 126:1193–1201.
    OpenUrlAbstract/FREE Full Text
  11. ↵
    1. Fiez JA
    (1997) Phonology, semantics, and the role of the left inferior prefrontal cortex. Hum Brain Mapp 5:79–83.
    OpenUrlCrossRefPubMed
  12. ↵
    1. Friederici AD,
    2. Hahne A,
    3. von Cramon DY
    (1998) First-pass versus second-pass parsing processes in a Wernicke's and a Broca's aphasic: electrophysiological evidence for a double dissociation. Brain Lang 62:311–341.
    OpenUrlCrossRefPubMed
  13. ↵
    1. Hagoort P,
    2. Brown CM
    (2000) ERP effects of listening to speech: semantic ERP effects. Neuropsychologia 38:1518–1530.
    OpenUrlCrossRefPubMed
  14. ↵
    1. Hagoort P,
    2. Brown CM,
    3. Swaab TY
    (1996) Lexical-semantic event-related potential effects in patients with left hemisphere lesions and aphasia, and patients with right hemisphere lesions without aphasia. Brain 119:627–649.
    OpenUrlAbstract/FREE Full Text
  15. ↵
    1. Halgren E,
    2. Dhond RP,
    3. Christensen N,
    4. Van Petten C,
    5. Marinkovic K,
    6. Lewine JD,
    7. Dale AM
    (2002) N400-like magnetoencephalography responses modulated by semantic context, word frequency, and lexical class in sentences. Neuroimage 17:1101–1116.
    OpenUrlCrossRefPubMed
  16. ↵
    1. Hämäläinen M,
    2. Hari R,
    3. Ilmoniemi R,
    4. Knuutila J,
    5. Lounasmaa O
    (1993) Magnetoencephalography - theory, instrumentation, and applications to noninvasive studies of the working human brain. Rev Mod Phys 65:413–497.
    OpenUrlCrossRef
  17. ↵
    1. Harm MW,
    2. Seidenberg MS
    (2004) Computing the meanings of words in reading: cooperative division of labor between visual and phonological processes. Psychol Rev 111:662–720.
    OpenUrlCrossRefPubMed
  18. ↵
    1. Hauk O,
    2. Davis MH,
    3. Ford M,
    4. Pulvermüller F,
    5. Marslen-Wilson WD
    (2006) The time course of visual word recognition as revealed by linear regression analysis of ERP data. Neuroimage 30:1383–1400.
    OpenUrlCrossRefPubMed
  19. ↵
    1. Heim S,
    2. Alter K,
    3. Ischebeck AK,
    4. Amunts K,
    5. Eickhoff SB,
    6. Mohlberg H,
    7. Zilles K,
    8. von Cramon DY,
    9. Friederici AD
    (2005) The role of the left Brodmann's areas 44 and 45 in reading words and pseudowords. Brain Res Cogn Brain Res 25:982–993.
    OpenUrlCrossRefPubMed
  20. ↵
    1. Helenius P,
    2. Salmelin R,
    3. Service E,
    4. Connolly JF
    (1998) Distinct time courses of word and context comprehension in the left temporal cortex. Brain 121:1133–1142.
    OpenUrlAbstract/FREE Full Text
  21. ↵
    1. Helenius P,
    2. Salmelin R,
    3. Service E,
    4. Connolly JF,
    5. Leinonen S,
    6. Lyytinen H
    (2002) Cortical activation during spoken-word segmentation in nonreading-impaired and dyslexic adults. J Neurosci 22:2936–2944.
    OpenUrlAbstract/FREE Full Text
  22. ↵
    1. Holmes AP,
    2. Blair RC,
    3. Watson JD,
    4. Ford I
    (1996) Nonparametric analysis of statistic images from functional mapping experiments. J Cereb Blood Flow Metab 16:7–22.
    OpenUrlCrossRefPubMed
  23. ↵
    1. Jobard G,
    2. Crivello F,
    3. Tzourio-Mazoyer N
    (2003) Evaluation of the dual route theory of reading: a metanalysis of 35 neuroimaging studies. Neuroimage 20:693–712.
    OpenUrlCrossRefPubMed
  24. ↵
    1. Kiefer M
    (2002) The N400 is modulated by unconsciously perceived masked words: further evidence for an automatic spreading activation account of N400 priming effects. Brain Res Cogn Brain Res 13:27–39.
    OpenUrlCrossRefPubMed
  25. ↵
    1. Kutas M,
    2. Hillyard SA
    (1980) Reading senseless sentences: brain potentials reflect semantic incongruity. Science 207:203–205.
    OpenUrlAbstract/FREE Full Text
  26. ↵
    1. Laaksonen H,
    2. Kujala J,
    3. Salmelin R
    (2008) A method for spatiotemporal mapping of event-related modulation of cortical rhythmic activity. Neuroimage 42:207–217.
    OpenUrlCrossRefPubMed
  27. ↵
    1. Laine M,
    2. Virtanen P
    (1999) WordMill Lexical Search Program Ed. 2.2.
  28. ↵
    1. Lau EF,
    2. Phillips C,
    3. Poeppel D
    (2008) A cortical network for semantics: (de)constructing the N400. Nat Rev Neurosci 9:920–933.
    OpenUrlCrossRefPubMed
  29. ↵
    1. Marinkovic K,
    2. Dhond RP,
    3. Dale AM,
    4. Glessner M,
    5. Carr V,
    6. Halgren E
    (2003) Spatiotemporal dynamics of modality-specific and supramodal word processing. Neuron 38:487–497.
    OpenUrlCrossRefPubMed
  30. ↵
    1. Matsuura K,
    2. Okabe Y
    (1995) Selective minimum-norm solution of the biomagnetic inverse problem. IEEE Trans Biomed Eng 42:608–615.
    OpenUrlCrossRefPubMed
  31. ↵
    1. Näätänen R,
    2. Lehtokoski A,
    3. Lennes M,
    4. Cheour M,
    5. Huotilainen M,
    6. Iivonen A,
    7. Vainio M,
    8. Alku P,
    9. Ilmoniemi RJ,
    10. Luuk A,
    11. Allik J,
    12. Sinkkonen J,
    13. Alho K
    (1997) Language-specific phoneme representations revealed by electric and magnetic brain responses. Nature 385:432–434.
    OpenUrlCrossRefPubMed
  32. ↵
    1. Nichols TE,
    2. Holmes AP
    (2002) Nonparametric permutation tests for functional neuroimaging: a primer with examples. Hum Brain Mapp 15:1–25.
    OpenUrlCrossRefPubMed
  33. ↵
    1. Nobre AC,
    2. McCarthy G
    (1994) Language-related ERPs: scalp distributions and modulation by word type and semantic priming. J Cogn Neurosci 6:233–255.
    OpenUrlCrossRef
  34. ↵
    1. Parviainen T,
    2. Helenius P,
    3. Salmelin R
    (2005) Cortical differentiation of speech and nonspeech sounds at 100 ms: implications for dyslexia. Cereb Cortex 15:1054–1063.
    OpenUrlAbstract/FREE Full Text
  35. ↵
    1. Coltheart M,
    2. Sartori G,
    3. Job R
    1. Patterson K,
    2. Shewell C
    (1987) in The cognitive neuropsychology of language, Speak and spell: dissociations and word class effects, eds Coltheart M, Sartori G, Job R (Erlbaum, London), pp 273–294.
  36. ↵
    1. Poldrack RA,
    2. Wagner AD,
    3. Prull MW,
    4. Desmond JE,
    5. Glover GH,
    6. Gabrieli JD
    (1999) Functional specialization for semantic and phonological processing in the left inferior prefrontal cortex. Neuroimage 10:15–35.
    OpenUrlCrossRefPubMed
  37. ↵
    1. Pulvermüller F,
    2. Shtyrov Y,
    3. Ilmoniemi RJ,
    4. Marslen-Wilson WD
    (2006) Tracking speech comprehension in space and time. Neuroimage 31:1297–1305.
    OpenUrlCrossRefPubMed
  38. ↵
    1. Pylkkänen L,
    2. Marantz A
    (2003) Tracking the time course of word recognition with MEG. Trends Cogn Sci 7:187–189.
    OpenUrlCrossRefPubMed
  39. ↵
    1. Pylkkänen L,
    2. Llinás R,
    3. Murphy GL
    (2006) The representation of polysemy: MEG evidence. J Cogn Neurosci 18:97–109.
    OpenUrlCrossRefPubMed
  40. ↵
    1. Rissman J,
    2. Eliassen JC,
    3. Blumstein SE
    (2003) An event-related fMRI investigation of implicit semantic priming. J Cogn Neurosci 15:1160–1175.
    OpenUrlCrossRefPubMed
  41. ↵
    1. Roland PE,
    2. Zilles K
    (1996) The developing European computerized human brain database for all imaging modalities. Neuroimage 4:39–47.
    OpenUrlCrossRef
  42. ↵
    1. Rugg MD
    (1984) Event-related potentials in phonological matching tasks. Brain Lang 23:225–240.
    OpenUrlCrossRefPubMed
  43. ↵
    1. Rugg MD,
    2. Barrett SE
    (1987) Event-related potentials and the interaction between orthographic and phonological information in a rhyme-judgment task. Brain Lang 32:336–361.
    OpenUrlCrossRefPubMed
  44. ↵
    1. Salmelin R
    (2007) Clinical neurophysiology of language: The MEG approach. Clin Neurophysiol 118:237–254.
    OpenUrlCrossRefPubMed
  45. ↵
    1. Salmelin R,
    2. Service E,
    3. Kiesilä P,
    4. Uutela K,
    5. Salonen O
    (1996) Impaired visual word processing in dyslexia revealed with magnetoencephalography. Ann Neurol 40:157–162.
    OpenUrlCrossRefPubMed
  46. ↵
    1. Schormann T,
    2. Henn S,
    3. Zilles K
    (1996) A new approach to fast elastic alignment with applications to human brains. Lect Notes Comput Sci 1131:337–342.
    OpenUrlCrossRef
  47. ↵
    1. Simos PG,
    2. Basile LF,
    3. Papanicolaou AC
    (1997) Source localization of the N400 response in a sentence-reading paradigm using evoked magnetic fields and magnetic resonance imaging. Brain Res 762:29–39.
    OpenUrlCrossRefPubMed
  48. ↵
    1. Spitsyna G,
    2. Warren JE,
    3. Scott SK,
    4. Turkheimer FE,
    5. Wise RJ
    (2006) Converging language streams in the human temporal lobe. J Neurosci 26:7328–7336.
    OpenUrlAbstract/FREE Full Text
  49. ↵
    1. Tarkiainen A,
    2. Helenius P,
    3. Hansen PC,
    4. Cornelissen PL,
    5. Salmelin R
    (1999) Dynamics of letter string perception in the human occipitotemporal cortex. Brain 122:2119–2132.
    OpenUrlAbstract/FREE Full Text
  50. ↵
    1. Thompson-Schill SL,
    2. Swick D,
    3. Farah MJ,
    4. D'Esposito M,
    5. Kan IP,
    6. Knight RT
    (1998) Verb generation in patients with focal frontal lesions: a neuropsychological test of neuroimaging findings. Proc Natl Acad Sci U S A 95:15855–15860.
    OpenUrlAbstract/FREE Full Text
  51. ↵
    1. Tyler LK,
    2. Moss HE,
    3. Galpin A,
    4. Voice JK
    (2002) Activating meaning in time: the role of imageability and form-class. Lang Cogn Proc 17:471–502.
    OpenUrlCrossRef
  52. ↵
    1. Uusitalo MA,
    2. Ilmoniemi RJ
    (1997) Signal-space projection method for separating MEG or EEG into components. Med Biol Eng Comput 35:135–140.
    OpenUrlCrossRefPubMed
  53. ↵
    1. Uusvuori J,
    2. Parviainen T,
    3. Inkinen M,
    4. Salmelin R
    (2008) Spatiotemporal interaction between sound form and meaning during spoken word perception. Cereb Cortex 18:456–466.
    OpenUrlAbstract/FREE Full Text
  54. ↵
    1. Uutela K,
    2. Hämäläinen M,
    3. Somersalo E
    (1999) Visualization of magnetoencephalographic data using minimum current estimates. Neuroimage 10:173–180.
    OpenUrlCrossRefPubMed
  55. ↵
    1. Vandenbulcke M,
    2. Peeters R,
    3. Dupont P,
    4. Van Hecke P,
    5. Vandenberghe R
    (2007) Word reading and posterior temporal dysfunction in amnestic mild cognitive impairment. Cereb Cortex 17:542–551.
    OpenUrlAbstract/FREE Full Text
  56. ↵
    1. Van Petten C,
    2. Luka BJ
    (2006) Neural localization of semantic context effects in electromagnetic and hemodynamic studies. Brain Lang 97:279–293.
    OpenUrlCrossRefPubMed
  57. ↵
    1. Vigneau M,
    2. Beaucousin V,
    3. Hervé PY,
    4. Duffau H,
    5. Crivello F,
    6. Houdé O,
    7. Mazoyer B,
    8. Tzourio-Mazoyer N
    (2006) Meta-analyzing left hemisphere language areas: phonology, semantics, and sentence processing. Neuroimage 30:1414–1432.
    OpenUrlCrossRefPubMed
  58. ↵
    1. Woods RP,
    2. Grafton ST,
    3. Watson JD,
    4. Sicotte NL,
    5. Mazziotta JC
    (1998) Automated image registration: II. Intersubject validation of linear and nonlinear models. J Comput Assist Tomogr 22:153–165.
    OpenUrlCrossRefPubMed
  59. ↵
    1. Wydell TN,
    2. Vuorinen T,
    3. Helenius P,
    4. Salmelin R
    (2003) Neural correlates of letter-string length and lexicality during reading in a regular orthography. J Cogn Neurosci 15:1052–1062.
    OpenUrlCrossRefPubMed
Back to top

In this issue

The Journal of Neuroscience: 29 (29)
Journal of Neuroscience
Vol. 29, Issue 29
22 Jul 2009
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Spatiotemporal Convergence of Semantic Processing in Reading and Speech Perception
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Spatiotemporal Convergence of Semantic Processing in Reading and Speech Perception
Johanna Vartiainen, Tiina Parviainen, Riitta Salmelin
Journal of Neuroscience 22 July 2009, 29 (29) 9271-9280; DOI: 10.1523/JNEUROSCI.5860-08.2009

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Spatiotemporal Convergence of Semantic Processing in Reading and Speech Perception
Johanna Vartiainen, Tiina Parviainen, Riitta Salmelin
Journal of Neuroscience 22 July 2009, 29 (29) 9271-9280; DOI: 10.1523/JNEUROSCI.5860-08.2009
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Articles

  • Memory Retrieval Has a Dynamic Influence on the Maintenance Mechanisms That Are Sensitive to ζ-Inhibitory Peptide (ZIP)
  • Neurophysiological Evidence for a Cortical Contribution to the Wakefulness-Related Drive to Breathe Explaining Hypocapnia-Resistant Ventilation in Humans
  • Monomeric Alpha-Synuclein Exerts a Physiological Role on Brain ATP Synthase
Show more Articles

Behavioral/Systems/Cognitive

  • Identification and Characterization of a Sleep-Active Cell Group in the Rostral Medullary Brainstem
  • Gravin Orchestrates Protein Kinase A and β2-Adrenergic Receptor Signaling Critical for Synaptic Plasticity and Memory
  • Generation of Intensity Selectivity by Differential Synaptic Tuning: Fast-Saturating Excitation But Slow-Saturating Inhibition
Show more Behavioral/Systems/Cognitive
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.