Skip to main content

Umbrella menu

  • SfN.org
  • eNeuro
  • The Journal of Neuroscience
  • Neuronline
  • BrainFacts.org

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Current Issue
    • Issue Archive
    • Video Archive
    • Collections
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • Subscriptions
    • Advertise
    • For the Media
    • Permissions
    • Contact Us
    • Feedback
  • SfN.org
  • eNeuro
  • The Journal of Neuroscience
  • Neuronline
  • BrainFacts.org

User menu

  • Log in
  • Subscribe
  • My alerts

Search

  • Advanced search
Journal of Neuroscience
  • Log in
  • Subscribe
  • My alerts
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Current Issue
    • Issue Archive
    • Video Archive
    • Collections
  • ALERTS
  • FOR AUTHORS
  • ABOUT
    • Overview
    • Editorial Board
    • Subscriptions
    • Advertise
    • For the Media
    • Permissions
    • Contact Us
    • Feedback
PreviousNext
Articles, Behavioral/Systems/Cognitive

Action Representation of Sound: Audiomotor Recognition Network While Listening to Newly Acquired Actions

Amir Lahav, Elliot Saltzman and Gottfried Schlaug
Journal of Neuroscience 10 January 2007, 27 (2) 308-314; DOI: https://doi.org/10.1523/JNEUROSCI.4822-06.2007
Amir Lahav
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Elliot Saltzman
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Gottfried Schlaug
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

The discovery of audiovisual mirror neurons in monkeys gave rise to the hypothesis that premotor areas are inherently involved not only when observing actions but also when listening to action-related sound. However, the whole-brain functional formation underlying such “action–listening” is not fully understood. In addition, previous studies in humans have focused mostly on relatively simple and overexperienced everyday actions, such as hand clapping or door knocking. Here we used functional magnetic resonance imaging to ask whether the human action-recognition system responds to sounds found in a more complex sequence of newly acquired actions. To address this, we chose a piece of music as a model set of acoustically presentable actions and trained non-musicians to play it by ear. We then monitored brain activity in subjects while they listened to the newly acquired piece. Although subjects listened to the music without performing any movements, activation was found bilaterally in the frontoparietal motor-related network (including Broca's area, the premotor region, the intraparietal sulcus, and the inferior parietal region), consistent with neural circuits that have been associated with action observations, and may constitute the human mirror neuron system. Presentation of the practiced notes in a different order activated the network to a much lesser degree, whereas listening to an equally familiar but motorically unknown music did not activate this network. These findings support the hypothesis of a “hearing–doing” system that is highly dependent on the individual's motor repertoire, gets established rapidly, and consists of Broca's area as its hub.

  • Broca's area
  • mirror neuron system
  • auditory
  • fMRI
  • premotor
  • sensorimotor

Introduction

When we observe someone performing an action, our brain may produce neural activity similar to that seen when we perform it ourselves. This neural simulation happens via a multimodal mirror neuron system, originally found in the monkey ventral premotor cortex (Gallese et al., 1996; Rizzolatti et al., 1996) and recently extended to humans in a variety of action-observation tasks (Decety and Grezes, 1999; Nishitani and Hari, 2000; Buccino et al., 2001; Grezes et al., 2003; Kilner et al., 2003; Gangitano et al., 2004; Hamilton et al., 2004; Calvo-Merino et al., 2005; Haslinger et al., 2005; Iacoboni, 2005; Nelissen et al., 2005; Thornton and Knoblich, 2006). Mirror neurons, however, may not only be triggered by visual stimuli. A subgroup of premotor neurons in monkeys also responded to the sound of actions (e.g., peanut breaking), and it has been recently suggested that there might also be a cross-modal neural system that formally orchestrates these neurons in humans (Kohler et al., 2002; Keysers et al., 2003). However, the whole-brain functional formation and mechanism underlying such “action–listening” is not fully understood.

Action listening takes part in many of our daily activities. Consider, for example, listening to door knocking or finger snapping. You would probably recognize the sound, although at the same time, your brain may also simulate the action (Aziz-Zadeh et al., 2004). Previous studies, however, have mostly focused on actions that we all learn from infancy and that are typically overexperienced, for example, hand clapping (Pizzamiglio et al., 2005), tongue clicking (Hauk et al., 2006), and speech (Fadiga et al., 2002; Wilson et al., 2004; Buccino et al., 2005). One other concern, at least with regard to the sound of speech, is that speech is not representative of all sounds; it carries meaning and is limited to communicative mouth actions, which all together, may activate different types of neural circuits than nonspeech sounds (Pulvermuller, 2001; Zatorre et al., 2002; Thierry et al., 2003; Schon et al., 2005; Ozdemir et al., 2006). Thus, in the present study, we ask whether and how the mirror neuron system will respond to actions and sounds that do not have verbal meaning and, most importantly, are well controlled and newly acquired.

To address this, we chose piano playing as a model task and a piece of music as a set of acoustically presentable novel actions. The use of music making as a sensory–motor framework for studying the acquisition of actions has been demonstrated in several previous studies (Stewart et al., 2003; Bangert et al., 2006b). Typically, when playing the piano, auditory feedback is naturally involved in each of the player's movements, leading to an intimate coupling between perception and action (Bangert and Altenmuller, 2003; Janata and Grafton, 2003). We therefore hypothesized that music one knows how to play (even if only recently learned) may be strongly associated with the corresponding elements of the individual's motor repertoire and might activate an audiomotor network in the human brain (Fig. 1).

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Action–listening illustration. A, Music performance can be viewed as a complex sequence of both actions and sounds, in which sounds are made by actions. B, The sound of music one knows how to play can be reflected, as if in a mirror, in the corresponding motor representations.

Materials and Methods

Experimental design

The present study had two stages. First, we trained musically naive subjects to play a novel piano piece (“trained-music”) and measured their learning progress over a period of 5 d. Next, on day 5, we performed functional magnetic resonance imaging (fMRI) scans on subjects while they passively listened to short passages taken from the newly acquired piece (this was considered the action–listening condition). Similar passages taken from two other original musical pieces that subjects had only listened to (to control for familiarity) but had never learned to play were used for the (two) control conditions (Fig. 2B).

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Experimental setup. A, Learning times of the trained-music are shown for individual subjects (n = 9) over five daily piano-training sessions. Learning time was influenced by the number of errors made during a session, until reaching error-free performance with the musical piece. A plateau level of performance is notable in sessions 4–5, obtaining the minimum possible learning time (12 min) set by the software. B, Examples of three fMRI experimental conditions. Sample short passages extracted from the trained-music that subjects learned how to play (blue) and two untrained control musical pieces are shown. These controls include an untrained-same-notes-music, in which the same notes were used in a different order to create a new melody (cyan) and an untrained-different-notes-music, composed of a completely different set of notes (red). C, A sample fMRI time series is shown between two imaging time points (TR, 18 s; TA, 1.75 s; red) within a single run. Images were acquired immediately after listening to short passages (5–8 s each; yellow) extracted from all three musical pieces. A behavioral control task (green) was used to control for listening attention; subjects heard three piano notes and had to press a button with their left hand if these notes had appeared as a subsequence in the preceding musical passage (yellow) they had heard. Intentionally, no images reflecting this task were acquired. We varied pause times between scans (gray) so that the TR remained fixed. D, Results of the behavioral control task are shown for all listening conditions. The graph shows the percentage of correct button presses from the group (n = 9) for each listening condition (24 trials per subject). No significant differences were found for mean correct responses across musical pieces (p = 0.489).

Subjects

Nine non-musicians (six females and three males; mean age, 22.4 ± 2.2 years old) participated in the study. All subjects were right-handed [assessed by a questionnaire (Oldfield, 1971)], had no previous musical training (including voice), and had no history of neurological, psychiatric, or auditory problems. This study was approved by the Institutional Review Boards of Beth Israel Deaconess Medical Center and Boston University, and all subjects gave written informed consent.

Experimental training

Drawing on previous MIDI-based interactive software tools (Bangert et al., 2001), we have developed our own MIDI-based interactive software for learning to play by ear (Lahav et al., 2005) and have trained subjects to play the piano part of a novel musical piece (trained-music) with no sight-reading required. This was done to abolish the requirement for visuomotor translation of musical notation into key presses and to enhance auditory–motor learning. During piano training, subjects learned to play the piano part (along with a prerecorded accompaniment), using their right hand and a set of five keys in a fixed fingering position (F, G, A, Bb, C) (Fig. 1A, gray keys) (the same finger always hit the same key). To complete a piano session, subjects had to gradually go through a series of trials until reaching error-free performance with the musical piece, while a computer notified them of a note (wrong key press) or timing (> 116 of a beat) error after each playing trial.

Two behavioral tests to assess auditory–motor learning

Measuring learning time.

The learning time of the musical piece was a function of the number of errors made during a given session. We measured learning progress in subjects over 5 consecutive days (one session per day), specifically looking at the duration (minutes) it took for each subject to achieve error-free performance and mastery of the piece.

Pitch-recognition–production test.

Subjects heard 30 single piano notes [from the set of five notes used in the trained-music: F, G, A, Bb, C] and had to press the corresponding piano key with the matching right-hand finger for each note at a time. The notes were played in random order, out of musical context. To rule out possible learning effects, subjects did not receive knowledge of the results (auditory feedback) when pressing the piano keys. This test was done to assess pitch–key mapping ability and was performed before and after the 5 d piano-training period.

Auditory stimuli

Three musical pieces were involved in this study. Subjects learned how to play only one musical piece (trained-music), whereas in addition, they listened to (to control for familiarity) but did not physically train with two other control musical pieces: (1) a musical piece composed of a completely different set of notes (F#, G#, B, C#, D#) than the one used in the trained-music (i.e., “untrained-different-notes-music”); and (2) a musical piece in which the same notes (F, G, A, Bb, C) were used in a different order to compose a new melody (i.e., untrained-same-notes-music). The auditory exposure time for all three musical pieces was equivalent. All musical pieces were novel and composed specifically for this study based on principles of western music; they were all of the same length (24 s, eight measures, 15 notes in total) and same tempo (80 beats per minute) and were played by piano accompanied by guitar, base, and drums. Short samples of the three musical pieces are shown in Figure 2B. The first three bars of the trained-music are shown in Figure 1A; for a fully orchestrated score and detailed description of software setup, see Lahav et al. (2005).

Motion-tracking system

To ensure motionless listening, we trained subjects to listen passively to music in a simulated scanning position (supine, palms facing up, fixated wrist) and digitally monitored their finger movements. We used a specially designed motion-tracking system plus a passive-marker glove, implemented in the EyesWeb development environment (www.eyesweb.org). Subjects wore a cotton glove with red markers on the fingertips and listened to music in a simulated scanning position just before the fMRI procedure. A camera (Quickcam Pro 4000; Logitech, Fremont, CA), faced downward, detected possible changes in finger position (<1 mm), based on real-time computation of pixel color change, collecting 352 × 288 pixel frames at a rate up to 30 Hz in the RGB (red/green/blue) color space.

fMRI acquisition and data analysis

A 3T GE whole-body system (GE Medical Systems, Milwaukee, WI) was used to acquire a set of high-resolution T1-weighted anatomical images (voxel size, 0.93 × 0.93 × 1.5 mm) and functional magnetic resonance images using a gradient echo-planar T2* sequence sensitive to the blood oxygenation level-dependent contrast (voxel size, 2 × 2 × 4 mm). To reduce scanner noise artifacts and interference with the music, we used a sparse temporal sampling technique [repetition time (TR), 18 s], acquiring 28 axial slices in a cluster [acquisition time (TA), 1.75 s]. MR images were acquired after listening to 5, 6, 7, and 8 s musical passages. A total of 32 fully orchestrated short passages extracted from three musical pieces were presented in a counterbalanced block design, with a total of 108 sets of axial images acquired during nine functional runs (see Fig. 2B,C for details). During each run, we acquired 12 sets of 28 axial slices (eight listening scans and four rest scans; total run time, 234 s). To reduce the chance of movement artifacts, subjects lay supine with their eyes closed and their palms facing up (an “uninviting” playing position) and followed instructions to stay as still as possible. fMRI data were analyzed using the SPM99 software package (Institute of Neurology, London, UK). Each set of axial images for each subject was realigned to the first image, coregistered with the corresponding T1-weighted data set, spatially normalized to the T1 template, and smoothed with an isotropic Gaussian kernel (8 mm full-width at half-maximum). Subject and condition effects were estimated using a general linear model. Global differences in scan intensity were removed by scaling each scan in proportion to its global intensity, and low-frequency drifts were removed using the default temporal high-pass filter.

A behavioral control task during the fMRI procedure

To ensure attentive listening to music, we included a behavioral control task during the fMRI procedure. After listening to each musical passage, subjects heard a three-tone sequence and had to press a button with their left hand if these notes had appeared as a subsequence in the preceding musical passage they had heard. Approximately 50% of the time, the three-tone sequence was part of the musical passage heard before. We intentionally arranged the acquisition of images so that no acquisition would reflect the neural activity induced by this control task (for details of the task design, see Fig. 2C).

Results

Action–sound training: learning the musical piece

In the first piano session, the time required to reach error-free performance with the musical piece was highly variable across subjects (mean, 29.11 min; SD, 5.53). The following sessions (2–5) were used to ensure mastery of the musical piece and to reduce performance variability between subjects during their fMRI scanning session (Fig. 2A). A plateau level of performance is notable in sessions 4–5, obtaining (and staying within) the minimum possible learning time (12 min) set by the software.

Pitch-recognition–production test

Results of the pitch-recognition–production (PRP) test indicate that as a byproduct of learning to play by ear, subjects also developed (albeit not to a perfection) a pitch–key mapping ability for the five pitches–keys used during piano training (i.e., the ability to recognize pitches and identify them in real time on the piano keyboard). Subjects significantly improved their PRP scores from 24% before the piano training period (which is around the level of chance) to 77% after training was completed.

Behavioral control during fMRI

The three-tone recognition task revealed similar performances across all three musical pieces. The motivation for implementing such a task was to make sure subjects are indeed listening and attending to the music (and not just hearing it in the background). The proportion of errors during this task did not differ across fMRI listening conditions, indicating that the differences in neural activation across music conditions were not caused by possible variations in the level of attention or auditory engagement with a given musical piece (Fig. 2D) [repeated-measure ANOVA with conditions (trained-music, untrained-same-notes music, untrained-different-notes-music) as within-subjects factors; F(2,16) = 1.536; p = 0.489].

Contrasting trained-music versus untrained-different-notes-music

All listening conditions (compared with rest) showed a similar activation pattern in primary and secondary auditory cortices bilaterally (Fig. 3A,B). However, when subjects listened to the trained-music (but not to the untrained-different-notes-music), activation was found in additional frontoparietal motor-related regions including, prominently, the posterior inferior frontal gyrus (IFG; Broca's region and Broca's homolog on the right; BA44, BA45) as well as the posterior middle premotor region and, to a lesser degree, the inferior parietal lobule (supramarginal gyrus and angular gyrus) bilaterally and cerebellar areas on the left (Fig. 3A,C) [trained-music > untrained-different-notes-music; p < 0.05, false discovery rate (FDR) corrected]. The complete lack of primary motor cortex (M1) activation in the hand/finger region during all conditions may be taken as evidence that subjects indeed did not physically move, as instructed (this was additionally verified by an observer in the scanning room).

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Action–listening activation. A, B, Extensive bilateral activation in the frontoparietal motor-related brain regions was observed when subjects listened to the trained-music they knew how to play (A), but not when they listened to the never-learned untrained-different-notes-music (B) (contrasted against rest baseline; p < 0.05, FDR corrected). C, Activation maps are shown in areas that were significantly more active during listening to the trained-music versus the untrained-different-notes-music. Surface projection of group mean activation (n = 9) are rendered onto an individual standardized brain (p < 0.05, FDR corrected). L, Left; R, right.

Contrasting trained-music versus untrained-same-notes-music

To further investigate the tuning of the action-recognition system, we compared brain activity in subjects during listening to the trained-music versus the untrained-same-notes-music (in which the notes of the trained-music were arranged in a different order). Interestingly, we found significant activation in the left posterior premotor cortex, as well as in the posterior part of the IFG containing Broca's area and in its right-hemispheric homolog (Fig. 4B) (p < 0.05, FDR corrected). Yet, despite those differences, several premotor and parietal regions were still active bilaterally (although to a much lesser degree) even when subjects listened to the untrained-same-notes-music (Fig. 4A). In addition, a region-of-interest analysis of the pars opercularis of the IFG indicates significant pick activations on the left only when subjects listen to the trained-music, whereas the right IFG remained fairly active across listening conditions (Fig. 4C) [repeated-measure ANOVA shows a significant condition effect for the left IFG (F(2,16) = 10.324; p = 0.001) and no effect for the right IFG (F(2,16) = 0.026; p = 0.973)].

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

A,Areas activated during listening to the untrained-same-notes-music contrasted against rest (p < 0.05, FDR corrected). B, Contrasted image of group mean activation is presented in areas that were significantly more active during listening to trained-music compared with untrained-same-notes-music. This included the left premotor region as well as Broca's area and its right hemispheric homolog (green arrows), shown also in the corresponding coronal view (middle) (p < 0.05, FDR corrected). C, Parameter estimates (β values) of the left (−50, 18, 16; magenta) and right (52, 18, 16; cyan) IFG across listening conditions. Results indicate significant pick activations on the left IFG only when subjects listen to the trained-music they knew how to play (p = 0.001), whereas the right IFG remained fairly active across listening conditions (p = 0.973).

Discussion

A unique aspect of the present study is the use of completely novel actions for studying auditory operations of the mirror neuron system. Previous EEG (Pizzamiglio et al., 2005; Hauk et al., 2006) and transcranial magnetic stimulation (Aziz-Zadeh et al., 2004) studies have mainly focused on the sound of over-experienced actions laying the groundwork for this field. The advantage, however, of training subjects with novel actions is twofold. First, it allows following the neural formation of auditory–motor learning, while eliminating possible influences from previous learning experiences. Second, it allows dealing with actions and sounds that are exclusively associated with a specific body part (here, the right hand). Such kind of control is almost impossible when investigating freestyle everyday actions that could potentially be bimanual (paper tearing) or mixed-handed (door knocking).

In addition, the use of music as an audible sequence of actions reveals important sequential aspects of the human action-recognition system. One should bear in mind that during the fMRI procedure, subjects did not listen to the musical piece in its entirety but to only short segments of it (5–8 s each) containing only part of the entire sequence of actions. The fact that listening to such short subsequences was still enough to activate an audiomotor recognition network suggests high levels of detection for action-related sounds (Fig. 3A,C). It is also striking that these audiomotor activation patterns are, in fact, within the core regions of the frontoparietal mirror neuron circuit previously found in humans in a variety of action-observation tasks (Grezes et al., 2003; Iacoboni et al., 2005; Lotze et al., 2006) including music observations, such as watching chord progression on the guitar (Buccino et al., 2004a) or finger-playing movements on the piano (Haslinger et al., 2005).

When subjects listened to the trained-music, a sound–action network became active but actions were not executed (Fig. 3A,C). However, when subjects listened to music they had never played before, they could not match it with existing action representations, and thus auditory activation was entirely dominant (Fig. 3B). These findings point to the heart of the action–listening mechanism and are in keeping with previous evidence for increased motor excitability (D'Ausilio et al., 2006) and premotor activity (Lotze et al., 2003; Bangert et al., 2006a) during listening to a rehearsed musical piece. Furthermore, our findings may be analogous to studies in the visuomotor domain, in which the mirror neuron system was involved only when the observed action was part of the observer's motor repertoire, such as in the case of dancers watching movements from their own dance style (but not from other styles) (Calvo-Merino et al., 2005) or in the case of humans watching biting actions (but not barking) (Buccino et al., 2004b).

It is particularly interesting that the posterior IFG, including Broca's area, was active only when subjects listened to the music they knew how to play. Even the presentation of music made of the same notes in an untrained order did not activate this area (Fig. 4B,C). These results reinforce the previous literature in two important ways. First, Broca's area is the human homolog of area F5 (ventral premotor cortex) in the monkey, where mirror neurons have been located (Rizzolatti and Arbib, 1998; Binkofski and Buccino, 2006). Our findings thus support the view that Broca's area is presumably a central region (“hub”) of the mirror neuron network (Iacoboni et al., 1999; Nishitani and Hari, 2000; Hamzei et al., 2003; Rizzolatti and Craighero, 2004; Nelissen et al., 2005), demonstrating here its multifunctional role in action listening. Second, mounting evidence suggest that in addition to its “classical” involvement in language, Broca's area functions also as a sensorimotor integrator (Binkofski and Buccino, 2004, 2006; Baumann et al., 2005), manual imitator (Krams et al., 1998; Heiser et al., 2003), supramodal sequence predictor (Maess et al., 2001; Kilner et al., 2004; Iacoboni et al., 2005), and internal simulator (Platel et al., 1997; Nishitani and Hari, 2000; Schubotz and von Cramon, 2004) of sequential actions. It is therefore possible that the activity in Broca's area during the action–listening condition reflects sequence-specific priming of action representations along with unconscious simulations and predictions as to the next action/sound to come. Such implicit motor predictions may be partially overlapped with functional operations of the mirror neuron system and thus could have hardly been made while listening to the never-learned musical pieces.

The significant activity in Broca's area during the action–listening condition strongly supports recent evidence for left-hemisphere dominance for action-related sounds (Pizzamiglio et al., 2005; Aziz-Zadeh et al., 2006). Region-of-interest analysis strongly confirms this laterality effect and also shows that the right IFG remained fairly active across all listening conditions (Fig. 4C). It is therefore likely that the observed activation in the right IFG may not fully reflect action representation per se but rather simply operations of music perception, consistent with the role of the right hemisphere in melodic/pitch processing (Zatorre et al., 1992). However, despite the left-hemispheric lateralization of the IFG, our data suggest that, at least with regard to the overall frontoparietal activation pattern, actions may be represented across the two hemispheres (Fig. 3A,C). The ipsilateral premotor activity, although seemingly illogical (because only the right hand was trained) is, in fact, in accordance with previous evidence for bilateral premotor representation of finger movements (Kim et al., 1993; Hanakawa et al., 2005), as well as with the general view that the mirror neuron system is bihemispheric in nature regardless of the laterality of the involved hand (Rizzolatti and Craighero, 2004; Molnar-Szakacs et al., 2005; Aziz-Zadeh et al., 2006). Support for such bilateral operation has mainly come from action-observation studies, whereas evidence from action listening is still quite limited. Having abstract representations of actions in both hemispheres may be essential for making actions more generalizable (Grafton et al., 2002; Vangheluwe et al., 2005) and potentially transferable from one limb to another (Rijntjes et al., 1999).

One may wonder what could be the reason for the premotor activity seen when subjects listened to the untrained-same-notes-music (Fig. 4A), especially because they had never played that piece. A possible explanation is that this premotor activity reflects the ability of subjects to link some of the notes they heard to the matching fingers and piano keys (77% matching accuracy in the PRP test; see Materials and Methods) (Fig. 5). Strikingly, this pitch–key linking mechanism seemed to be purely implicit because subjects were completely unaware (presumably because of their musical naivety) that the piece was composed of the same notes as the trained-music (as confirmed in post-fMRI interviews). As a result of this linking ability, the notes themselves appeared motorically familiar, which in turn was sufficient to activate a small action–sound circuit. This circuit supposedly demonstrates basic operations of audiomotor recognition at the level of one single action (finger press) and one sound (piano pitch). Nevertheless, this audiomotor recognition of individual notes, without the complete motor representation of the melodic sequence, was not enough to fully engage the hearing–doing mirror neuron system for action listening.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Scores of the PRP test. The percentage of correct responses is shown for individual subjects, before the initial (red) and after the last (blue) piano-training session. Mean group errors (solid line) and SEs (dashed line) are presented. An improvement trend is shown, from 24% before training (around the level of chance) to 77% after training (mean relative improvement of 209%; p < 10−4).

To what extent does action listening involve an implicit mental rehearsal component of the heard action? This may be somewhat difficult to answer based on the present study, because we have not included an explicit imagery condition. Yet, the lack of activation in the “classic” (although still controversial) motor imagery network [including the contralateral (left) M1/S1, the supplementary motor area, and the ipsilateral cerebellum] (Fig. 3A) argues against the possibility that subjects consciously imagined themselves playing the music (Grafton et al., 1996; Porro et al., 1996; Langheim et al., 2002). Furthermore, imagery was very unlikely, because our auditory control task during fMRI scanning sidetracked subjects from playing the music in their minds (as confirmed by post-fMRI interviews). Furthermore, evidence suggest that mental simulations and operations of the mirror neuron system might all be functionally related or possibly even a form of one another (Grezes and Decety, 2001; Patuzzo et al., 2003; Cisek and Kalaska, 2004). Additional studies are still needed to support this hypothesis.

Finally, what could be the functional advantage of having audiomotor recognition networks? It has been suggested that such audiomotor networks are essential for the acquisition of language, serving as a critical sensorimotor feedback loop during speech perception (Rizzolatti and Arbib, 1998; Theoret and Pascual-Leone, 2002). We suggest that, in addition, such networks may have been developed to protect the survival of all hearing organisms, allowing the understanding of actions even when they cannot be observed but can only be heard (for example, the sound of footsteps in the dark).

In conclusion, this study indicates that the human action-recognition system is highly sensitive to the individual's motor experience and has the tuning capabilities needed to discriminate between the sound of newly acquired actions and the sound of actions that are motorically unknown. Thus, acquiring actions that have an audible output quickly generates a functional neural link between the sound of those actions and the presumably corresponding motor representations. These findings serve as an important brain imaging addition to a growing body of research on the auditory properties of the mirror neuron system and may encourage additional investigations in the realm of action listening.

Footnotes

  • This work was supported in part by grants from the National Institutes of Health–National Institute of Neurological Disorders and Stroke (E.S. and G.S.) and the International Foundation for Music Research (G.S.) and by a Dudley Allen Sargent Research Grant (A.L.). We thank Adam Boulanger for computer music support, Marc Bangert for discussions at initial stages, Eugene Wong for help with graphic design, and Galit Lahav and Daniel Segre for insightful comments.

  • Correspondence should be addressed to Amir Lahav, Department of Neurology, Beth Israel Deaconess Medical Center and Harvard Medical School, 330 Brookline Avenue, Boston, MA 02215. alahav{at}bidmc.harvard.edu

References

  1. ↵
    1. Aziz-Zadeh L,
    2. Iacoboni M,
    3. Zaidel E,
    4. Wilson S,
    5. Mazziotta J
    (2004) Left hemisphere motor facilitation in response to manual action sounds. Eur J Neurosci 19:2609–2612.
    OpenUrlCrossRefPubMed
  2. ↵
    1. Aziz-Zadeh L,
    2. Koski L,
    3. Zaidel E,
    4. Mazziotta J,
    5. Iacoboni M
    (2006) Lateralization of the human mirror neuron system. J Neurosci 26:2964–2970.
    OpenUrlAbstract/FREE Full Text
  3. ↵
    1. Bangert M,
    2. Altenmuller EO
    (2003) Mapping perception to action in piano practice a longitudinal DC-EEG study. BMC Neurosci 4:26.
    OpenUrlCrossRefPubMed
  4. ↵
    1. Bangert M,
    2. Haeusler U,
    3. Altenmuller E
    (2001) On practice how the brain connects piano keys and piano sounds. Ann NY Acad Sci 930:425–428.
    OpenUrlPubMed
  5. ↵
    1. Bangert M,
    2. Peschel T,
    3. Schlaug G,
    4. Rotte M,
    5. Drescher D,
    6. Hinrichs H,
    7. Heinze HJ,
    8. Altenmuller E
    (2006a) Shared networks for auditory and motor processing in professional pianists: evidence from fMRI conjunction. NeuroImage 30:917–926.
    OpenUrlCrossRefPubMed
  6. ↵
    1. Bangert M,
    2. Jurgens U,
    3. Hausler U,
    4. Altenmuller E
    (2006b) Classical conditioned responses to absent tones. BMC Neurosci 7:60.
    OpenUrlCrossRefPubMed
  7. ↵
    1. Baumann S,
    2. Koeneke S,
    3. Meyer M,
    4. Lutz K,
    5. Jancke L
    (2005) A network for sensory-motor integration: what happens in the auditory cortex during piano playing without acoustic feedback? Ann NY Acad Sci 1060:186–188.
    OpenUrlCrossRefPubMed
  8. ↵
    1. Binkofski F,
    2. Buccino G
    (2004) Motor functions of the Broca's region. Brain Lang 89:362–369.
    OpenUrlCrossRefPubMed
  9. ↵
    1. Binkofski F,
    2. Buccino G
    (2006) The role of ventral premotor cortex in action execution and action understanding. J Physiol (Paris) 99:396–405.
    OpenUrlCrossRefPubMed
  10. ↵
    1. Buccino G,
    2. Binkofski F,
    3. Fink GR,
    4. Fadiga L,
    5. Fogassi L,
    6. Gallese V,
    7. Seitz RJ,
    8. Zilles K,
    9. Rizzolatti G,
    10. Freund HJ
    (2001) Action observation activates premotor and parietal areas in a somatotopic manner: an fMRI study. Eur J Neurosci 13:400–404.
    OpenUrlCrossRefPubMed
  11. ↵
    1. Buccino G,
    2. Vogt S,
    3. Ritzl A,
    4. Fink GR,
    5. Zilles K,
    6. Freund HJ,
    7. Rizzolatti G
    (2004a) Neural circuits underlying imitation learning of hand actions: an event-related fMRI study. Neuron 42:323–334.
    OpenUrlCrossRefPubMed
  12. ↵
    1. Buccino G,
    2. Lui F,
    3. Canessa N,
    4. Patteri I,
    5. Lagravinese G,
    6. Benuzzi F,
    7. Porro CA,
    8. Rizzolatti G
    (2004b) Neural circuits involved in the recognition of actions performed by nonconspecifics: an fMRI study. J Cogn Neurosci 16:114–126.
    OpenUrlCrossRefPubMed
  13. ↵
    1. Buccino G,
    2. Riggio L,
    3. Melli G,
    4. Binkofski F,
    5. Gallese V,
    6. Rizzolatti G
    (2005) Listening to action-related sentences modulates the activity of the motor system: a combined TMS and behavioral study. Brain Res Cogn Brain Res 24:355–363.
    OpenUrlCrossRefPubMed
  14. ↵
    1. Calvo-Merino B,
    2. Glaser DE,
    3. Grezes J,
    4. Passingham RE,
    5. Haggard P
    (2005) Action observation and acquired motor skills: an fMRI study with expert dancers. Cereb Cortex 15:1243–1249.
    OpenUrlAbstract/FREE Full Text
  15. ↵
    1. Cisek P,
    2. Kalaska JF
    (2004) Neural correlates of mental rehearsal in dorsal premotor cortex. Nature 431:993–996.
    OpenUrlCrossRefPubMed
  16. ↵
    1. D'Ausilio A,
    2. Altenmuller E,
    3. Olivetti Belardinelli M,
    4. Lotze M
    (2006) Cross-modal plasticity of the motor cortex while listening to a rehearsed musical piece. Eur J Neurosci 24:955–958.
    OpenUrlCrossRefPubMed
  17. ↵
    1. Decety J,
    2. Grezes J
    (1999) Neural mechanisms subserving the perception of human actions. Trends Cogn Sci 3:172–178.
    OpenUrlCrossRefPubMed
  18. ↵
    1. Fadiga L,
    2. Craighero L,
    3. Buccino G,
    4. Rizzolatti G
    (2002) Speech listening specifically modulates the excitability of tongue muscles: a TMS study. Eur J Neurosci 15:399–402.
    OpenUrlCrossRefPubMed
  19. ↵
    1. Gallese V,
    2. Fadiga L,
    3. Fogassi L,
    4. Rizzolatti G
    (1996) Action recognition in the premotor cortex. Brain 119:593–609.
    OpenUrlAbstract/FREE Full Text
  20. ↵
    1. Gangitano M,
    2. Mottaghy FM,
    3. Pascual-Leone A
    (2004) Modulation of premotor mirror neuron activity during observation of unpredictable grasping movements. Eur J Neurosci 20:2193–2202.
    OpenUrlCrossRefPubMed
  21. ↵
    1. Grafton ST,
    2. Arbib MA,
    3. Fadiga L,
    4. Rizzolatti G
    (1996) Localization of grasp representations in humans by positron emission tomography. 2. Observation compared with imagination. Exp Brain Res 112:103–111.
    OpenUrlPubMed
  22. ↵
    1. Grafton ST,
    2. Hazeltine E,
    3. Ivry RB
    (2002) Motor sequence learning with the nondominant left hand. A PET functional imaging study. Exp Brain Res 146:369–378.
    OpenUrlCrossRefPubMed
  23. ↵
    1. Grezes J,
    2. Decety J
    (2001) Functional anatomy of execution, mental simulation, observation, and verb generation of actions: a meta-analysis. Hum Brain Mapp 12:1–19.
    OpenUrlCrossRefPubMed
  24. ↵
    1. Grezes J,
    2. Armony JL,
    3. Rowe J,
    4. Passingham RE
    (2003) Activations related to “mirror” and “canonical” neurones in the human brain: an fMRI study. NeuroImage 18:928–937.
    OpenUrlCrossRefPubMed
  25. ↵
    1. Hamilton A,
    2. Wolpert D,
    3. Frith U
    (2004) Your own action influences how you perceive another person's action. Curr Biol 14:493–498.
    OpenUrlCrossRefPubMed
  26. ↵
    1. Hamzei F,
    2. Rijntjes M,
    3. Dettmers C,
    4. Glauche V,
    5. Weiller C,
    6. Buchel C
    (2003) The human action recognition system and its relationship to Broca's area: an fMRI study. NeuroImage 19:637–644.
    OpenUrlCrossRefPubMed
  27. ↵
    1. Hanakawa T,
    2. Parikh S,
    3. Bruno MK,
    4. Hallett M
    (2005) Finger and face representations in the ipsilateral precentral motor areas in humans. J Neurophysiol 93:2950–2958.
    OpenUrlAbstract/FREE Full Text
  28. ↵
    1. Haslinger B,
    2. Erhard P,
    3. Altenmuller E,
    4. Schroeder U,
    5. Boecker H,
    6. Ceballos-Baumann AO
    (2005) Transmodal sensorimotor networks during action observation in professional pianists. J Cogn Neurosci 17:282–293.
    OpenUrlCrossRefPubMed
  29. ↵
    1. Hauk O,
    2. Shtyrov Y,
    3. Pulvermuller F
    (2006) The sound of actions as reflected by mismatch negativity: rapid activation of cortical sensory-motor networks by sounds associated with finger and tongue movements. Eur J Neurosci 23:811–821.
    OpenUrlCrossRefPubMed
  30. ↵
    1. Heiser M,
    2. Iacoboni M,
    3. Maeda F,
    4. Marcus J,
    5. Mazziotta JC
    (2003) The essential role of Broca's area in imitation. Eur J Neurosci 17:1123–1128.
    OpenUrlCrossRefPubMed
  31. ↵
    1. Iacoboni M
    (2005) Neural mechanisms of imitation. Curr Opin Neurobiol 15:632–637.
    OpenUrlCrossRefPubMed
  32. ↵
    1. Iacoboni M,
    2. Woods RP,
    3. Brass M,
    4. Bekkering H,
    5. Mazziotta JC,
    6. Rizzolatti G
    (1999) Cortical mechanisms of human imitation. Science 286:2526–2528.
    OpenUrlAbstract/FREE Full Text
    1. Iacoboni M,
    2. Molnar-Szakacs I,
    3. Gallese V,
    4. Buccino G,
    5. Mazziotta JC,
    6. Rizzolatti G
    (2005) Grasping the intentions of others with one's own mirror neuron system. PLoS Biol 3:e79.
    OpenUrlCrossRefPubMed
  33. ↵
    1. Janata P,
    2. Grafton ST
    (2003) Swinging in the brain: shared neural substrates for behaviors related to sequencing and music. Nat Neurosci 6:682–687.
    OpenUrlCrossRefPubMed
  34. ↵
    1. Keysers C,
    2. Kohler E,
    3. Umilta MA,
    4. Nanetti L,
    5. Fogassi L,
    6. Gallese V
    (2003) Audiovisual mirror neurons and action recognition. Exp Brain Res 153:628–636.
    OpenUrlCrossRefPubMed
  35. ↵
    1. Kilner JM,
    2. Paulignan Y,
    3. Blakemore SJ
    (2003) An interference effect of observed biological movement on action. Curr Biol 13:522–525.
    OpenUrlCrossRefPubMed
  36. ↵
    1. Kilner JM,
    2. Vargas C,
    3. Duval S,
    4. Blakemore SJ,
    5. Sirigu A
    (2004) Motor activation prior to observation of a predicted movement. Nat Neurosci 7:1299–1301.
    OpenUrlCrossRefPubMed
  37. ↵
    1. Kim SG,
    2. Ashe J,
    3. Hendrich K,
    4. Ellermann JM,
    5. Merkle H,
    6. Ugurbil K,
    7. Georgopoulos AP
    (1993) Functional magnetic resonance imaging of motor cortex: hemispheric asymmetry and handedness. Science 261:615–617.
    OpenUrlAbstract/FREE Full Text
  38. ↵
    1. Kohler E,
    2. Keysers C,
    3. Umilta MA,
    4. Fogassi L,
    5. Gallese V,
    6. Rizzolatti G
    (2002) Hearing sounds, understanding actions: action representation in mirror neurons. Science 297:846–848.
    OpenUrlAbstract/FREE Full Text
  39. ↵
    1. Krams M,
    2. Rushworth MF,
    3. Deiber MP,
    4. Frackowiak RS,
    5. Passingham RE
    (1998) The preparation, execution and suppression of copied movements in the human brain. Exp Brain Res 120:386–398.
    OpenUrlCrossRefPubMed
  40. ↵
    1. Lahav A,
    2. Boulanger A,
    3. Schlaug G,
    4. Saltzman E
    (2005) The power of listening: auditory-motor interactions in musical training. Ann NY Acad Sci 1060:189–194.
    OpenUrlCrossRefPubMed
  41. ↵
    1. Langheim FJ,
    2. Callicott JH,
    3. Mattay VS,
    4. Duyn JH,
    5. Weinberger DR
    (2002) Cortical systems associated with covert music rehearsal. NeuroImage 16:901–908.
    OpenUrlCrossRefPubMed
  42. ↵
    1. Lotze M,
    2. Scheler G,
    3. Tan HR,
    4. Braun C,
    5. Birbaumer N
    (2003) The musician's brain: functional imaging of amateurs and professionals during performance and imagery. NeuroImage 20:1817–1829.
    OpenUrlCrossRefPubMed
  43. ↵
    1. Lotze M,
    2. Heymans U,
    3. Birbaumer N,
    4. Veit R,
    5. Erb M,
    6. Flor H,
    7. Halsband U
    (2006) Differential cerebral activation during observation of expressive gestures and motor acts. Neuropsychologia 44:1787–1795.
    OpenUrlCrossRefPubMed
  44. ↵
    1. Maess B,
    2. Koelsch S,
    3. Gunter TC,
    4. Friederici AD
    (2001) Musical syntax is processed in Broca's area: an MEG study. Nat Neurosci 4:540–545.
    OpenUrlCrossRefPubMed
  45. ↵
    1. Molnar-Szakacs I,
    2. Iacoboni M,
    3. Koski L,
    4. Mazziotta JC
    (2005) Functional segregation within pars opercularis of the inferior frontal gyrus: evidence from fMRI studies of imitation and action observation. Cereb Cortex 15:986–994.
    OpenUrlAbstract/FREE Full Text
  46. ↵
    1. Nelissen K,
    2. Luppino G,
    3. Vanduffel W,
    4. Rizzolatti G,
    5. Orban GA
    (2005) Observing others: multiple action representation in the frontal lobe. Science 310:332–336.
    OpenUrlAbstract/FREE Full Text
  47. ↵
    1. Nishitani N,
    2. Hari R
    (2000) Temporal dynamics of cortical representation for action. Proc Natl Acad Sci USA 97:913–918.
    OpenUrlAbstract/FREE Full Text
  48. ↵
    1. Oldfield RC
    (1971) The assessment and analysis of handedness: the Edinburgh inventory. Neuropsychologia 9:97–113.
    OpenUrlCrossRefPubMed
  49. ↵
    1. Ozdemir E,
    2. Norton A,
    3. Schlaug G
    (2006) Shared and distinct neural correlates of singing and speaking. NeuroImage 33:628–635.
    OpenUrlCrossRefPubMed
  50. ↵
    1. Patuzzo S,
    2. Fiaschi A,
    3. Manganotti P
    (2003) Modulation of motor cortex excitability in the left hemisphere during action observation: a single- and paired-pulse transcranial magnetic stimulation study of self- and non-self-action observation. Neuropsychologia 41:1272–1278.
    OpenUrlCrossRefPubMed
  51. ↵
    1. Pizzamiglio L,
    2. Aprile T,
    3. Spitoni G,
    4. Pitzalis S,
    5. Bates E,
    6. D'Amico S,
    7. Di Russo F
    (2005) Separate neural systems for processing action- or non-action-related sounds. NeuroImage 24:852–861.
    OpenUrlCrossRefPubMed
  52. ↵
    1. Platel H,
    2. Price C,
    3. Baron JC,
    4. Wise R,
    5. Lambert J,
    6. Frackowiak RS,
    7. Lechevalier B,
    8. Eustache F
    (1997) The structural components of music perception. A functional anatomical study. Brain 120:229–243.
    OpenUrlAbstract/FREE Full Text
  53. ↵
    1. Porro CA,
    2. Francescato MP,
    3. Cettolo V,
    4. Diamond ME,
    5. Baraldi P,
    6. Zuiani C,
    7. Bazzocchi M,
    8. di Prampero PE
    (1996) Primary motor and sensory cortex activation during motor performance and motor imagery: a functional magnetic resonance imaging study. J Neurosci 16:7688–7698.
    OpenUrlAbstract/FREE Full Text
  54. ↵
    1. Pulvermuller F
    (2001) Brain reflections of words and their meaning. Trends Cogn Sci 5:517–524.
    OpenUrlCrossRefPubMed
  55. ↵
    1. Rijntjes M,
    2. Dettmers C,
    3. Buchel C,
    4. Kiebel S,
    5. Frackowiak RS,
    6. Weiller C
    (1999) A blueprint for movement: functional and anatomical representations in the human motor system. J Neurosci 19:8043–8048.
    OpenUrlAbstract/FREE Full Text
  56. ↵
    1. Rizzolatti G,
    2. Arbib MA
    (1998) Language within our grasp. Trends Neurosci 21:188–194.
    OpenUrlCrossRefPubMed
  57. ↵
    1. Rizzolatti G,
    2. Craighero L
    (2004) The mirror-neuron system. Annu Rev Neurosci 27:169–192.
    OpenUrlCrossRefPubMed
  58. ↵
    1. Rizzolatti G,
    2. Fadiga L,
    3. Gallese V,
    4. Fogassi L
    (1996) Premotor cortex and the recognition of motor actions. Brain Res Cogn Brain Res 3:131–141.
    OpenUrlCrossRefPubMed
  59. ↵
    1. Schon D,
    2. Gordon RL,
    3. Besson M
    (2005) Musical and linguistic processing in song perception. Ann NY Acad Sci 1060:71–81.
    OpenUrlCrossRefPubMed
  60. ↵
    1. Schubotz RI,
    2. von Cramon DY
    (2004) Sequences of abstract nonbiological stimuli share ventral premotor cortex with action observation and imagery. J Neurosci 24:5467–5474.
    OpenUrlAbstract/FREE Full Text
  61. ↵
    1. Stewart L,
    2. Henson R,
    3. Kampe K,
    4. Walsh V,
    5. Turner R,
    6. Frith U
    (2003) Brain changes after learning to read and play music. NeuroImage 20:71–83.
    OpenUrlCrossRefPubMed
  62. ↵
    1. Theoret H,
    2. Pascual-Leone A
    (2002) Language acquisition: do as you hear. Curr Biol 12:R736–R737.
    OpenUrlCrossRefPubMed
  63. ↵
    1. Thierry G,
    2. Giraud AL,
    3. Price C
    (2003) Hemispheric dissociation in access to the human semantic system. Neuron 38:499–506.
    OpenUrlCrossRefPubMed
  64. ↵
    1. Thornton IM,
    2. Knoblich G
    (2006) Action perception: seeing the world through a moving body. Curr Biol 16:R27–R29.
    OpenUrlCrossRefPubMed
  65. ↵
    1. Vangheluwe S,
    2. Suy E,
    3. Wenderoth N,
    4. Swinnen SP
    (2005) Learning and transfer of bimanual multifrequency patterns: effector-independent and effector-specific levels of movement representation. Exp Brain Res, 1–12.
  66. ↵
    1. Wilson SM,
    2. Saygin AP,
    3. Sereno MI,
    4. Iacoboni M
    (2004) Listening to speech activates motor areas involved in speech production. Nat Neurosci 7:701–702.
    OpenUrlCrossRefPubMed
  67. ↵
    1. Zatorre RJ,
    2. Evans AC,
    3. Meyer E,
    4. Gjedde A
    (1992) Lateralization of phonetic and pitch discrimination in speech processing. Science 256:846–849.
    OpenUrlAbstract/FREE Full Text
  68. ↵
    1. Zatorre RJ,
    2. Belin P,
    3. Penhune VB
    (2002) Structure and function of auditory cortex: music and speech. Trends Cogn Sci 6:37–46.
    OpenUrlCrossRefPubMed
View Abstract
Back to top

In this issue

The Journal of Neuroscience: 27 (2)
Journal of Neuroscience
Vol. 27, Issue 2
10 Jan 2007
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Action Representation of Sound: Audiomotor Recognition Network While Listening to Newly Acquired Actions
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
Print
View Full Page PDF
Article Alerts
Sign In to Email Alerts with your Email Address
Citation Tools
Action Representation of Sound: Audiomotor Recognition Network While Listening to Newly Acquired Actions
Amir Lahav, Elliot Saltzman, Gottfried Schlaug
Journal of Neuroscience 10 January 2007, 27 (2) 308-314; DOI: 10.1523/JNEUROSCI.4822-06.2007

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article

Share
Action Representation of Sound: Audiomotor Recognition Network While Listening to Newly Acquired Actions
Amir Lahav, Elliot Saltzman, Gottfried Schlaug
Journal of Neuroscience 10 January 2007, 27 (2) 308-314; DOI: 10.1523/JNEUROSCI.4822-06.2007
del.icio.us logo Digg logo Reddit logo Twitter logo CiteULike logo Facebook logo Google logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Articles

  • Choice Behavior Guided by Learned, But Not Innate, Taste Aversion Recruits the Orbitofrontal Cortex
  • Maturation of Spontaneous Firing Properties after Hearing Onset in Rat Auditory Nerve Fibers: Spontaneous Rates, Refractoriness, and Interfiber Correlations
  • Insulin Treatment Prevents Neuroinflammation and Neuronal Injury with Restored Neurobehavioral Function in Models of HIV/AIDS Neurodegeneration
Show more 3

Behavioral/Systems/Cognitive

  • An Abrupt Shift in the Day/Night Cycle Causes Desynchrony in the Mammalian Circadian Center
  • Caspase Inhibitors Promote Vestibular Hair Cell Survival and Function after Aminoglycoside Treatment In Vivo
  • Differential Pattern of cAMP Response Element-Binding Protein Activation in the Rat Brain after Conditioned Aversion as a Function of the Associative Process Engaged: Taste versus Context Association
Show more 3
  • Home
  • Alerts
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Video Archive
  • Collections

For Authors

  • Information for Authors

About

  • Overview
  • Editorial Board
  • Subscriptions
  • For the Media
  • Permissions
  • Contact Us
  • Feedback
(JNeurosci logo)
(SfN logo)

Copyright © 2018 by the Society for Neuroscience.

JNeurosci   Print ISSN: 0270-6474   Online ISSN: 1529-2401