Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Journal Club

Predictive Coding Explains Auditory and Tactile Influences on Vision during Binocular Rivalry

Bradley N. Jack and Graeme Hacker
Journal of Neuroscience 7 May 2014, 34 (19) 6423-6424; https://doi.org/10.1523/JNEUROSCI.1040-14.2014
Bradley N. Jack
Discipline of Psychology, School of Health and Human Sciences, Southern Cross University, Coffs Harbour, Australia
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Graeme Hacker
Discipline of Psychology, School of Health and Human Sciences, Southern Cross University, Coffs Harbour, Australia
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Info & Metrics
  • eLetters
  • PDF
Loading

When the two eyes are presented with dissimilar images, such as a picture of a face to the left eye and a picture of a house to the right eye, something remarkable happens. Observers report seeing the picture of the face for a few seconds, then the picture of the house, then the face again, and so on, for as long as he or she cares to look. This phenomenon is called binocular rivalry (Alais and Blake, 2004; Miller, 2013), and although the neurocognitive mechanism that explains this visual effect remains unsolved, there is a trend toward approaches that combine top-down and bottom-up processes in the brain (Blake and Logothetis, 2002; Tong et al., 2006).

Despite there being a rich literature on the topic of binocular rivalry, only a handful of studies have investigated the influences of multisensory stimuli on vision during binocular rivalry (Zhou et al., 2010). This is surprising, because perception is almost always multisensory. That is, a stable representation of the environment often consists of a combination of visual, auditory, tactile, gustatory, and/or olfactory inputs. Recently, a study in The Journal of Neuroscience sought to investigate the influences of multisensory stimuli on vision during binocular rivalry. In particular, Lunghi et al. (2014) set out to understand whether auditory, tactile, and auditory/tactile stimuli could influence percept selection and percept alternation during binocular rivalry.

To instigate binocular rivalry, Lunghi et al. (2014) presented spatially identical, contrast-modulating, random-noise patterns to the two eyes. The stimuli differed only in the temporal frequency in which they modulated in contrast, one at 3.75 Hz and the other at 15 Hz. This type of binocular-rivalry stimulus is known to produce alternations in vision that are comparable to alternations caused by spatially dissimilar stimuli, such as those of a face and a house (Alais and Parker, 2012). During dichoptic viewing of the stimuli, Lunghi et al. (2014) presented auditory, tactile, or auditory/tactile stimuli intermittently. Their auditory and tactile stimuli consisted of amplitude-modulated sine waves at either 3.75 or 15 Hz—these frequencies are the same as the ones used by the visual stimuli, and were presented through headphones (auditory stimuli) or as vibrations (tactile stimuli). Finally, Lunghi et al. (2014) manipulated the intensity of the auditory and tactile stimuli, resulting in modulation depths of 33%, 66%, and 100%. They did this to assess the effects of auditory and tactile stimulus strength on vision during binocular rivalry.

Lunghi et al. (2014) found that presenting auditory and tactile inputs that were congruent with an observer's percept, as reported by key presses, during binocular rivalry resulted in an increased probability of maintaining that percept, compared with when there was no auditory or tactile input, whereas inputs that were incongruent resulted in an increased probability of switching that percept to the congruent alternative. Furthermore, the probability of percept switching or percept maintaining was influenced by auditory and tactile stimulus strength: inputs with a modulation depth of 100% resulted in a higher probability of percept switching (∼0.67) or percept maintaining (∼0.4) than inputs with a modulation depth of 33% (∼0.5 for percept switching and ∼0.3 for percept maintaining). Lunghi et al. (2014) also found that when auditory and tactile inputs were presented simultaneously at a modulation depth of 33% each, the probability of percept switching or percept maintaining was comparable to the results obtained for auditory only or tactile only inputs with a modulation depth of 100%. Finally, when auditory and tactile inputs were presented simultaneously at a modulation depth of 100% each but in anti-phase (i.e., the auditory input was congruent with one visual stimulus whereas the tactile input was congruent with the other), there was no effect on the probability of percept switching or percept maintaining (for more details, see Lunghi et al., 2014, their Fig. 1). Lunghi et al. (2014) concluded that there must be a common, bottom-up, neural mechanism for the integration of visual, auditory, and tactile inputs, and that this mechanism is used to deliver a stable representation of the environment.

We agree that there must be a common mechanism for the integration of multisensory inputs, but we suspect this mechanism alone would be insufficient to govern visual perception, because it does not account for top-down influences on vision during binocular rivalry (Blake and Logothetis, 2002; Tong et al., 2006). Below, we use predictive coding to provide an epistemological explanation for why, when presented with auditory, tactile, or auditory/tactile inputs during binocular rivalry, visual awareness or one of the other visual input is maintained or switched.

Predictive coding is a framework for understanding how the brain delivers a stable representation of the environment. Theories of predictive coding advocate that the cognitive system is structured in a hierarchical manner, that all perception is a testable hypothesis, and that this hypothesis is a compromise between top-down predictions and bottom-up sensory input. According to predictive coding, the brain delivers a stable representation of the environment by using top-down predictions to reduce redundant processing of repetitive stimuli from bottom-up sensory input and by processing only what is not predicted—prediction-error (Rao and Ballard, 1999; Friston, 2005). Indeed, the concepts of predictive coding have also been used to explain binocular rivalry (Hohwy et al., 2008).

Hohwy et al. (2008) argue that the selection problem in binocular rivalry, why one stimulus is selected for perception rather than the other or both, occurs because one of the images (i.e., a face) has a higher prior, or likelihood, than the other image (i.e., a house) or than both images superimposed (i.e., a face-house). In addition, Hohwy et al. (2008) solve the alternation problem, why the images alternate in visual perception over time, using prediction-error. If one of the images accounts for approximately half of the total visual input, then the other half of visual input is unexplained and is therefore prediction-error. According to theories of predictive coding, prediction-error is passed up the hierarchy until it is resolved by higher-level neurons. When this happens, there is a change in visual perception, such that the image that was previously suppressed from awareness is now perceived and the image that was previously perceived is now suppressed. Furthermore, the image that is now suppressed is the new source of prediction-error in sensory input. This pattern repeats itself for as long as one cares to look at binocular-rivalry stimuli. We think this explanation of binocular rivalry could be extended to account for the findings reported by Lunghi et al. (2014).

Lunghi et al. (2014) found that presenting auditory and tactile inputs that were congruent with an observer's percept during binocular rivalry resulted in an increased probability of maintaining that percept. This is consistent with predictive coding, because more than half of the total sensory input (visual and auditory or tactile vs visual only) is congruent with the brain's hypothesis about the visual environment; thus, the current predictive model is a good representation of sensory input, and is maintained. Of course, this percept will not be maintained forever, because the brain needs to eventually account for prediction-error from the suppressed visual stimulus. Lunghi et al. (2014) also found that presenting auditory and tactile inputs that were incongruent with an observer's percept resulted in an increased probability of switching that percept. Again, this is consistent with predictive coding, because more than half of the total sensory input is prediction-error; thus, the current predictive model is a poor representation of sensory input, and it is replaced by another model that better accounts for sensory input, forcing a change in visual perception.

Predictive coding can also explain why the probability of percept switching or percept maintaining was influenced by auditory and tactile stimulus strength. If percept maintaining and percept switching are the result of decreased and increased prediction-error, respectively, then reducing the modulation depth of auditory and tactile inputs from 100% to 33% decreases the amount of total sensory input that the current predictive model accounts for. Thus, one would expect a lower probability of percept switching or percept maintaining for auditory and tactile inputs at a lower modulation depth than at a higher modulation depth. This reasoning can also be used to explain why auditory and tactile inputs presented simultaneously at a modulation depth of 33% each yield results that are much higher than those obtained for auditory-only or tactile-only inputs with a modulation depth of 33%, as well as explain why auditory and tactile inputs presented simultaneously at a modulation depth of 100% each but in anti-phase have no effect on vision.

Of course, the concept of predictive coding do not detract from Lunghi et al.'s (2014) conclusion: that perception of one or the other stimulus during binocular rivalry is driven by a common, bottom-up, multisensory mechanism. Rather, predictive coding complements Lunghi et al.'s (2014) bottom-up explanation by acknowledging top-down influences on vision during binocular rivalry. Indeed, any explanation of percept selection and percept alternation during binocular rivalry is likely to be incomplete without including top-down mechanisms of perception (Blake and Logothetis, 2002; Tong et al., 2006). In conclusion, predictive coding is a useful framework for understanding the interaction between different levels of the visual hierarchy that allow our brains to deliver a stable representation of the environment.

Footnotes

  • Editor's Note: These short, critical reviews of recent papers in the Journal, written exclusively by graduate students or postdoctoral fellows, are intended to summarize the important findings of the paper and provide additional insight and commentary. For more information on the format and purpose of the Journal Club, please see http://www.jneurosci.org/misc/ifa_features.shtml.

  • Correspondence should be addressed to Bradley N. Jack, Discipline of Psychology, Southern Cross University, Hogbin Drive, Coffs Harbour NSW 2450, Australia. bradley.jack{at}scu.edu.au

References

  1. ↵
    1. Alais D,
    2. Parker A
    (2012) Binocular rivalry produced by temporal frequency differences. Front Hum Neurosci 6:227, doi:10.3389/fnhum.2012.00227, pmid:22866033.
    OpenUrlCrossRefPubMed
  2. ↵
    1. Alais D,
    2. Blake R
    (2004) Binocular rivalry (MIT, Cambridge, MA).
  3. ↵
    1. Blake R,
    2. Logothetis N
    (2002) Visual competition. Nat Rev Neurosci 3:13–21, doi:10.1038/nrn701, pmid:11823801.
    OpenUrlCrossRefPubMed
  4. ↵
    1. Friston K
    (2005) A theory of cortical responses. Philos Trans R Soc Lond B Biol Sci 360:815–836, doi:10.1098/rstb.2005.1622, pmid:15937014.
    OpenUrlAbstract/FREE Full Text
  5. ↵
    1. Hohwy J,
    2. Roepstorff A,
    3. Friston K
    (2008) Predictive coding explains binocular rivalry: an epistemological review. Cognition 108:687–701, doi:10.1016/j.cognition.2008.05.010, pmid:18649876.
    OpenUrlCrossRefPubMed
  6. ↵
    1. Lunghi C,
    2. Morrone MC,
    3. Alais D
    (2014) Auditory and tactile signals combine to influence vision during binocular rivalry. J Neurosci 34:784–792, doi:10.1523/JNEUROSCI.2732-13.2014, pmid:24431437.
    OpenUrlAbstract/FREE Full Text
  7. ↵
    1. Miller SM
    (2013) The constitution of visual consciousness: Lessons from binocular rivalry (John Benjamin's Publishing, Amsterdam).
  8. ↵
    1. Rao RP,
    2. Ballard DH
    (1999) Predictive coding in the visual cortex: A functional interpretation of some extra-classical receptive-field effects. Nat Neurosci 2:79–87, pmid:10195184.
    OpenUrlCrossRefPubMed
  9. ↵
    1. Tong F,
    2. Meng M,
    3. Blake R
    (2006) Neural bases of binocular rivalry. Trends in Cognitive Sciences 10:502–511, doi:10.1016/j.tics.2006.09.003, pmid:16997612.
    OpenUrlCrossRefPubMed
  10. ↵
    1. Zhou W,
    2. Jiang Y,
    3. He S,
    4. Chen D
    (2010) Olfaction modulates visual perception in binocular rivalry. Curr Biol 20:1356–1358, doi:10.1016/j.cub.2010.05.059, pmid:20598540.
    OpenUrlCrossRefPubMed
Back to top

In this issue

The Journal of Neuroscience: 34 (19)
Journal of Neuroscience
Vol. 34, Issue 19
7 May 2014
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Advertising (PDF)
  • Ed Board (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Predictive Coding Explains Auditory and Tactile Influences on Vision during Binocular Rivalry
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Predictive Coding Explains Auditory and Tactile Influences on Vision during Binocular Rivalry
Bradley N. Jack, Graeme Hacker
Journal of Neuroscience 7 May 2014, 34 (19) 6423-6424; DOI: 10.1523/JNEUROSCI.1040-14.2014

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Predictive Coding Explains Auditory and Tactile Influences on Vision during Binocular Rivalry
Bradley N. Jack, Graeme Hacker
Journal of Neuroscience 7 May 2014, 34 (19) 6423-6424; DOI: 10.1523/JNEUROSCI.1040-14.2014
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Footnotes
    • References
  • Info & Metrics
  • eLetters
  • PDF

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

  • Early Neural Development of Social Interaction Perception in the Superior Temporal Sulcus
  • Universal Coding for Uncertainty?
  • Beyond Motor Control: Diffusion MRI Reveals Associations between the Cerebello-VTA Pathway and Socio-affective Behaviors in Humans
Show more Journal Club
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.