Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
  • EDITORIAL BOARD
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
  • SUBSCRIBE

User menu

  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
  • EDITORIAL BOARD
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
  • SUBSCRIBE
PreviousNext
Research Articles, Behavioral/Cognitive

Predicting Identity-Preserving Object Transformations across the Human Ventral Visual Stream

Viola Mocz, Maryam Vaziri-Pashkam, Marvin M. Chun and Yaoda Xu
Journal of Neuroscience 1 September 2021, 41 (35) 7403-7419; DOI: https://doi.org/10.1523/JNEUROSCI.2137-20.2021
Viola Mocz
1Visual Cognitive Neuroscience Lab, Department of Psychology, Yale University, New Haven, Connecticut 06520
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Maryam Vaziri-Pashkam
3Laboratory of Brain and Cognition, National Institute of Mental Health, Bethesda, Maryland 20892
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Maryam Vaziri-Pashkam
Marvin M. Chun
1Visual Cognitive Neuroscience Lab, Department of Psychology, Yale University, New Haven, Connecticut 06520
2Department of Neuroscience, Yale School of Medicine, New Haven, Connecticut 06520
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Yaoda Xu
1Visual Cognitive Neuroscience Lab, Department of Psychology, Yale University, New Haven, Connecticut 06520
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Yaoda Xu
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Article Figures & Data

Figures

  • Tables
  • Figure 1.
    • Download figure
    • Open in new tab
    • Download powerpoint
    Figure 1.

    Possible neural representational space structures, experimental details, and analyses used. A, A schematic illustration of how object identity and nonidentity features may be represented together in a high-dimensional neural representational space, using size as a nonidentity feature. Left, Completely orthogonal representations of these two types of features, with the object responses across the two states of the size transformation being equidistant for each object in the representational space. Right, Near orthogonal representations of these two types of features, with the object responses across the two states of the size transformation for each object being different in the representational space. B, The eight natural object categories used. Each category contained 10 different exemplars varying in identity, pose (for the animate categories only), and viewing angle/orientation to minimize the low-level image similarities among them. C, The four types of nonidentity transformations examined, including position, size, image stats, and spatial frequency. Each transformation included two states. D, An illustration of the block design paradigm used. Participants performed a one-back repetition detection task on the images. An actual block in the experiment contained 10 images with two repetitions per block. E, Inflated brain surfaces from a representative participant showing the ROIs examined. They included topographically defined early visual areas V1 to V4 and functionally defined higher object processing regions LOT and VOT. F, G, Illustrations of the analyses performed to evaluate the predicted pattern. To evaluate pattern predictability, the predicted and the true patterns were directly correlated; to evaluate pattern selectivity, the correlation between the predicted and the true patterns from the same category was compared with the correlation between the predicted and the true patterns from different categories (F). To evaluate the preservation of the category representational structure, the RDM derived from the predicted patterns was vectorized and correlated with that from the true patterns (G). See Materials and Methods.

  • Figure 2.
    • Download figure
    • Open in new tab
    • Download powerpoint
    Figure 2.

    Correlations between the predicted and the corresponding true patterns for each transformation for the 75 most reliable voxels included in each ROI. The correlations are compared with two performance ceiling measurements (see Materials and Methods). A, Results plotted by ROI and transformation type showing pattern prediction for categories included in the training data (Trained Categories) and those that were not (Untrained Categories) as a function of the number of categories included in the training data. Significant values for pairwise t tests against averaged-run Ceiling and Single-run Ceiling for each condition and for the difference between the trained and untrained categories for each training set size are marked with asterisks at the top of each plot. All t tests were corrected for multiple comparisons using the Benjamini–Hochberg method across the ROIs and training set sizes within each transformation type, totaling 42 or 30 comparisons (6 ROIs × 7 or 5 training set sizes). B, Results normalized by dividing the correlations by Averaged-run Ceiling and then plotted by transformation type showing pattern prediction as a function of ROI including only the smallest and the largest training set sizes. The light ribbon around each plot line represents the between-subjects 95% confidence interval of the mean. ∼0.05 < p < 0.10, *p <0.05, **p < 0.01, ***p < 0. 001.

  • Figure 3.
    • Download figure
    • Open in new tab
    • Download powerpoint
    Figure 3.

    Schematic illustrations of the pattern predictability and pattern selectivity results when object identity and nonidentity features are represented in a near-orthogonal manner. A, Top, Predictability results of training only one category (e.g., cat). The mapping derived is specific to the one trained category. As such, the correlation between the true and predicted patterns would be higher for the category included in training than that not included in training. Bottom, Predictability results of training two categories. Because the mapping derived is not specific to any one trained category, the correlation between the true and predicted patterns could decrease for the category included in training. B, Predictability and selectivity results in lower and higher visual regions. Because objects are more correlated in lower than in higher visual regions to begin with, even when both regions could equally well predict an object category response pattern after training, category selectivity would be greater for higher than lower visual regions. Category selectivity is defined as the difference in the correlation of a predicted pattern with the true pattern of the matching category and the correlation of a predicted pattern with the true pattern of mismatching categories (i.e., Is the predicted pattern significantly closer to that of the matching category than mismatching categories?).

  • Figure 4.
    • Download figure
    • Open in new tab
    • Download powerpoint
    Figure 4.

    Category selectivity for the predicted patterns for the 75 most reliable voxels included in each ROI. Category selectivity was calculated as the correlation difference between the predicted and true patterns of the same category and that of different categories. A, Results plotted by ROI and transformation type showing category selectivity for categories included in the training data (Trained Categories) and those that were not (Untrained Categories) as a function of the number of categories included in the training data. Significant values for pairwise t tests against zero for each condition and for the difference between the trained and untrained categories for each training set size are marked with asterisks at the top of each plot. All t tests were corrected for multiple comparisons using the Benjamini–Hochberg method across the ROIs and training set sizes within each transformation type, totaling 42 or 30 comparisons (6 ROIs × 7 or 5 training set sizes). B, Results normalized by dividing the correlations by Averaged-run Ceiling and plotted by transformation type showing category selectivity as a function of the ROI including only the smallest and the largest training set sizes. The light ribbon around each plot line represents the between-subjects 95% confidence interval of the mean. ∼p < 0.10, *p < 0.05, **p < 0.01, ***p < 0.001.

  • Figure 5.
    • Download figure
    • Open in new tab
    • Download powerpoint
    Figure 5.

    Category RDM correlation between the predicted and true patterns for the 75 most reliable voxels included in each ROI. The correlations are compared with two performance ceiling measurements (see above, Materials and Methods). A, Results plotted by ROI and transformation type showing category structure preservation for categories included in the training data (Trained Categories) and those that were not (Untrained Categories) as a function of the number of categories included in the training data. Significant values for pairwise t tests against Averaged-run Ceiling and Single-run Ceiling for each condition and for the difference between the trained and untrained categories for each training set size are marked with asterisks at the top of each plot. All t tests were corrected for multiple comparisons using the Benjamini–Hochberg method across the ROIs and training set sizes within each transformation type, totaling 42 or 30 comparisons (6 ROIs × 7 or 5 training set sizes). B, Results normalized by dividing the correlations by Averaged-run Ceiling and plotted by transformation type showing category structure preservation as a function of the ROI including only the smallest and the largest training set sizes. The light ribbon around each plot line represents the between-subjects 95% confidence interval of the mean. ∼p < 0.10, *p < 0.05, **p < 0.01, ***p < 0.001.

  • Figure 6.
    • Download figure
    • Open in new tab
    • Download powerpoint
    Figure 6.

    The effect of category similarity on pattern prediction generalizability. A, A schematic illustration of how object identity and nonidentity features may be represented together in the high-dimensional neural representational space, using size as a nonidentity feature. Left, Coupled identity and nonidentity representations. Categories that are closer in the identity dimension (e.g., the cat and elephant pair, and the chair and car pair) would also be similarly apart across the two states of the size transformation in the representational space. Right, Decoupled identity and nonidentity representations. Categories that are similar in the identity dimension are not necessarily similar across the two states of the size transformation. B, Correlations between the prediction similarity matrix and category similarity matrix for the 75 most reliable voxels included in each ROI. Results are plotted by transformation type showing the effect of category similarity on pattern prediction similarity as a function of the ROI. Significant values for pairwise t tests against one for each ROI are marked with asterisks at the top of each plot. All t tests were corrected for multiple comparisons using the Benjamini–Hochberg method across the ROIs within each transformation type, totaling six comparisons. The light ribbon around each plot line represents the between-subjects 95% confidence interval of the mean. ∼p < 0.10, *p < 0.05, **p < 0.01, ***p < 0.001.

Tables

  • Figures
    • View popup
    Table 1.

    Summary of statistical results of the predicted fMRI response patterns

    Training categoryTraining set sizeInteraction
    AnalysisTransformationV1V2V3V4LOTVOTV1V2V3V4LOTVOTV1V2V3V4LOTVOT
    PatternPosition**
    0.754
    **
    0.806
    ***
    0.931
    **
    0.782
    ***
    0.948
    ***
    0.878
    ***
    0.989
    ***
    0.942
    ***
    0.921
    ***
    958
    ***
    0.762
    ***
    0.887
    ***
    0.755
    ***
    0.859
    ***
    0.830
    ***
    0.807
    ***
    0.937
    ***
    0.850
    Size**
    0.812
    **
    0.774
    **
    0.798
    **
    0.701
    **
    0.764
    **
    0.820
    ***
    0.966
    ***
    0.928
    ***
    0.937
    ***
    0.877
    ***
    0.481
    ***
    0.737
    ***
    0.861
    ***
    0.810
    ***
    0.923
    ***
    0.779
    ***
    0.882
    ***
    0.891
    Image stats***
    0.952
    ***
    0.913
    **
    0.896
    **
    0.826
    **
    0.900
    **
    0.900
    ***
    0.962
    ***
    0.905
    ***
    0.960
    ***
    0.929
    ***
    0.644
    ***
    0.765
    ***
    0.935
    ***
    0.946
    ***
    0.879
    ***
    0.867
    ***
    0.895
    ***
    0.923
    SF***
    0.785
    **
    0.709
    ***
    0.803
    ***
    0.830
    ***
    0.926
    ***
    0.904
    ***
    0.970
    ***
    0.931
    ***
    0.916
    ***
    0.946
    ***
    0.856
    ***
    0.877
    ***
    0.795
    ***
    0.720
    ***
    0.796
    ***
    0.847
    ***
    0.914
    ***
    0.868
    Category
    selectivity
    Position***
    0.883
    ***
    0.875
    **
    0.707
    **
    0.823
    ***
    0.886
    **
    0.696
    ***
    0.674
    ***
    0.721
    ***
    0.490
    **
    0.455
    ***
    0.515
    ns***
    0.883
    ***
    0.799
    ***
    0.613
    ***
    0.749
    ***
    0.902
    ***
    0.530
    Size***
    0.959
    ***
    0.900
    ***
    0.917
    ***
    0.881
    ***
    0.907
    ***
    0.884
    ***
    0.899
    ***
    0.606
    ***
    0.753
    *
    0.291
    ***
    0.606
    ***
    0.606
    ***
    0.951
    ***
    0.941
    ***
    0.902
    ***
    0.862
    ***
    0.838
    ***
    0.849
    Image stats***
    0.906
    **
    0.868
    *
    0.742
    **
    0.769
    ***
    0.888
    **
    0.868
    **
    0.486
    ***
    0.546
    ***
    0.530
    *
    0.342
    ***
    0.774
    ***
    0.756
    ***
    0.906
    ***
    0.908
    ***
    0.743
    ***
    0.696
    ***
    0.876
    ***
    0.868
    SF***
    0.865
    ***
    0.762
    ***
    0.806
    ***
    0.867
    ***
    0.796
    ***
    0.774
    ***
    0.436
    ***
    0.402
    ***
    0.582
    ***
    0.568
    *
    0.236
    *
    0.282
    ***
    0.885
    ***
    0.797
    ***
    0.837
    ***
    0.889
    ***
    0.821
    ***
    0.756
    Category
    representational
    structure
    Position*
    0.572
    *
    0.675
    ***
    0.885
    **
    0.737
    ***
    0.858
    **
    0.759
    ***
    0.655
    ***
    0.853
    ***
    0.738
    ***
    0.740
    ***
    0.836
    ***
    772
    ***
    0.498
    ***
    0.609
    ***
    0.785
    ***
    0.746
    ***
    0.729
    ***
    0.786
    Size*
    0.679
    *
    0.648
    *
    0.648
    *
    0.646
    **
    0.778
    ***
    0.871
    ***
    0.813
    ***
    0.785
    ***
    0.745
    ***
    0.554
    ***
    0.670
    ***
    0.853
    ***
    0.810
    ***
    0.868
    ***
    0.524
    ***
    0.619
    ***
    0.755
    ***
    0.674
    Image stats**
    0.820
    **
    0.897
    *
    0.699
    **
    0.782
    **
    0.877
    **
    0.850
    ***
    0.851
    ***
    0.882
    ***
    0.837
    ***
    0.587
    ***
    0.756
    ***
    0.867
    ***
    0.772
    ***
    0.786
    ***
    0.743
    ***
    0.750
    ***
    0.772
    ***
    0.793
    SF***
    0.720
    **
    0.565
    ***
    0.723
    ***
    0.743
    ***
    0.869
    ***
    0.953
    ***
    0.677
    ***
    0.766
    ***
    0.718
    ***
    0.659
    ***
    0.758
    ***
    0.820
    **
    0.339
    ***
    0.510
    ***
    0.437
    ***
    0.589
    ***
    0.629
    ***
    0.595
    • The predicted patterns were evaluated in terms of their pattern correlation with the true response patterns, their category selectivity (i.e., testing whether the predicted pattern was more similar to the true pattern of the same than different categories), and category representational structure. In all three analyses, the effects of training category (i.e., whether a category was included in the training data), training set size (i.e., the number of categories included in the training data), and their interactions were examined. The ROIs with significant effects are marked with asterisks. Additionally, the partial eta-squared measure of effect size is reported for all ROIs. All tests were corrected for multiple comparisons using the Benjamini–Hochberg method across the ROIs and transformation types within each of the three analyses, totaling 24 comparisons (6 ROIs × 4 transformation types). ns: non-significant,

    • ↵*p < 0.05,

    • ↵**p < 0.01,

    • ↵***p < 0.001.

Back to top

In this issue

The Journal of Neuroscience: 41 (35)
Journal of Neuroscience
Vol. 41, Issue 35
1 Sep 2021
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Ed Board (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Predicting Identity-Preserving Object Transformations across the Human Ventral Visual Stream
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Predicting Identity-Preserving Object Transformations across the Human Ventral Visual Stream
Viola Mocz, Maryam Vaziri-Pashkam, Marvin M. Chun, Yaoda Xu
Journal of Neuroscience 1 September 2021, 41 (35) 7403-7419; DOI: 10.1523/JNEUROSCI.2137-20.2021

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Predicting Identity-Preserving Object Transformations across the Human Ventral Visual Stream
Viola Mocz, Maryam Vaziri-Pashkam, Marvin M. Chun, Yaoda Xu
Journal of Neuroscience 1 September 2021, 41 (35) 7403-7419; DOI: 10.1523/JNEUROSCI.2137-20.2021
del.icio.us logo Digg logo Reddit logo Twitter logo Facebook logo Google logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • object invariance
  • object recognition
  • perception
  • transformation
  • visual representations

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Research Articles

  • Vigilance and behavioral state-dependent modulation of cortical neuronal activity throughout the sleep/wake cycle
  • Brain functional connectivity mapping of behavioral flexibility in rhesus monkeys
  • Accumulation System: Distributed Neural Substrates of Perceptual Decision Making Revealed by fMRI Deconvolution
Show more Research Articles

Behavioral/Cognitive

  • Brain functional connectivity mapping of behavioral flexibility in rhesus monkeys
  • Accumulation System: Distributed Neural Substrates of Perceptual Decision Making Revealed by fMRI Deconvolution
  • Recruitment of control and representational components of the semantic system during successful and unsuccessful access to complex factual knowledge
Show more Behavioral/Cognitive
  • Home
  • Alerts
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Policy
  • Contact
(JNeurosci logo)
(SfN logo)

Copyright © 2022 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.