The hypothesis that two distinct neural pathways mediate visual processing has been very influential in the cognitive sciences of vision. The idea is that each pathway processes different aspects of visual information: the dorsal pathway, leading from the primary visual cortex to the posterior parietal cortex, processes visual information for action, whereas the ventral pathway, leading from the primary visual cortex to the inferior temporal cortex, processes visual information for perception (Ungerleider and Mishkin, 1982). This dissociation is based on both neurophysiological investigations in the monkey and behavioral studies in patients (Goodale et al., 2005). Despite its elegant simplicity, this hypothesis appears incompatible with evidence converging on functional links between perception and action, in particular between visual perception and oculomotor actions. For example, visuospatial attention is directed toward the target position of the eyes (Deubel and Schneider, 1996); the size of the motion aftereffect is correlated with the speed of smooth-pursuit eye movements (Braun et al., 2006); and judgments about an object's position in space are compressed toward the saccadic endpoint just before the saccade (Ross et al., 1997). These experiments suggest that the production of a body movement (and often, its mere planning), influences which parts of a visual scene we see best, how we see them, or what position in space objects seem to occupy. This line of experiments leads to the intriguing hypothesis that our perception relies not only on the physical attributes of the stimulus but also on the actions we perform on them. Thus, visual information used for perception and action would be similar. Although several studies have shown functional similarities between perceptual and oculomotor performance, similar performance can arise from different mechanisms.
In an article published recently in the Journal of Neuroscience, Eckstein et al. (2007) address this question in human subjects with a novel method that avoids some of the problems of equating similar performance with a common mechanism. Rather than comparing performance, the authors propose a classification image (CI) analysis. The basic idea of this analysis is that clues about what information is used to accomplish a task can be obtained from the stimuli and the participant's decisions on those stimuli. CI analysis depends on the use of noisy stimuli and on the analysis of errors, the idea being that when a distracter is mistaken for the target, its noise must have contained some information that fooled the decision mechanism into selecting it. By averaging the noise profiles of all the error trials, one can estimate the information on which the decision was based.
The task proposed by Eckstein et al. (2007) is a five-alternative forced choice (5-AFC) visual search task: participants must detect a brighter Gaussian-shaped target among four distracters. Spatially uncorrelated Gaussian noise is added to the stimuli. Two conditions are contrasted: in the eye-movement condition, participants make an oculomotor decision by saccading to the target as quickly as possible. The decision is the target closest to the endpoint of the first saccade. In the perception-only condition, participants select the target by a button press. Importantly, processing time is maintained equivalent across the two conditions by measuring the average saccade latency of the first saccade in the eye-movement condition and then presenting the visual stimulus for that time minus 80 ms in the perception-only condition. The 80 ms interval corresponds to the approximate neural delays for motor programming, once visual processing is complete. This duration ranged from ∼100 to 250 ms across participants.
In such a paradigm, CI analysis will reveal what spatial information is used to drive the decision, i.e., the shape of the receptive field (RF) of the estimated visual mechanisms driving the oculomotor and perceptual decisions. For each condition (eye-movement vs perception-only), CIs were computed by averaging the noise profiles of the distractor stimuli to which participants erroneously saccaded or chose as a target. Then, the CIs were again averaged over pixels of equal radius from the center. The apparent radial symmetry of the CIs [Eckstein et al. (2007), their Fig. 2 (http://www.jneurosci.org/cgi/content/full/27/6/1266/F2)] suggests that this is a good way to reduce the degrees of freedom in the data without losing important features. The radial profiles are shown in Eckstein et al. (2007), their Fig. 3 (http://www.jneurosci.org/cgi/content/full/27/6/1266/F3). Two important conclusions emerge and were confirmed statistically by fitting the profiles to difference of Gaussian functions. First, CIs for the eye-movement and perception-only conditions were indistinguishable. Indeed, both profiles showed lateral inhibition (they dipped down below zero at the edge of the stimulus location) and both profiles were slightly wider than the target profile, meaning that information at the center of the RF is particularly important for driving the response. These characteristics are reminiscent of neural RFs. Second, the eye-movement and perception-only profiles were more similar to each other (t(5) = 0.13) than they were to the target profile (t(5) = 1.08). This last result is probably the most interesting because, as the authors point out in the discussion, the simplicity of their target could have led to behavioral RFs that matched it; therefore, the similarity between the eye-movement RF and the perception-only RF could have been the result of their correlation with the target. This last analysis proves that this is not the case, because behavioral RFs differed systematically from the target but not from each other.
The RF of the visual mechanism selecting the saccade target and that of the visual mechanism selecting the perceptual target are indistinguishable. If eye movements and visual perception are controlled by different brain areas, they would likely have different RFs and perform different computations. In contrast, this study shows that similar RFs are used, providing a strong argument that eye movements and visual perception could be controlled by the same brain areas, or that information is shared across areas. This interesting new result complements previous studies (cited above) providing evidence for similar outputs of the perceptual and motor systems by showing that the visual information in the input is also shared across the two systems, at least for simple visual tasks. It could be argued that in both the eye-movement and perception-only conditions, participants had to provide a motor response based on a visual discrimination. Therefore, an interesting next step might be to confirm these results with a perceptual task that does not require a motor response such as a button press.
Footnotes
-
Editor's Note: These short reviews of a recent paper in the Journal, written exclusively by graduate students or postdoctoral fellows, are intended to mimic the journal clubs that exist in your own departments or institutions. For more information on the format and purpose of the Journal Club, please see http://www.jneurosci.org/misc/ifa_features.shtml.
-
I thank the participants of the Perception Journal Club of Paris, France for a lively discussion on a previous draft.
- Correspondence should be addressed to Thérèse Collins, Cognitive Psychology and Neurosciences Laboratory, Paris Descartes University and National Center for Scientific Research, 71 avenue Edouard Vaillant, 92774 Boulogne-Billancourt cedex, France. therese.collins{at}univ-paris5.fr