Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Research Articles, Behavioral/Cognitive

Causal Influence of Linguistic Learning on Perceptual and Conceptual Processing: A Brain-Constrained Deep Neural Network Study of Proper Names and Category Terms

Phuc T. U. Nguyen, Malte R. Henningsen-Schomers and Friedemann Pulvermüller
Journal of Neuroscience 28 February 2024, 44 (9) e1048232023; https://doi.org/10.1523/JNEUROSCI.1048-23.2023
Phuc T. U. Nguyen
1Brain Language Laboratory, Department of Philosophy and Humanities, Freie Universität Berlin, Berlin 14195, Germany
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Phuc T. U. Nguyen
Malte R. Henningsen-Schomers
1Brain Language Laboratory, Department of Philosophy and Humanities, Freie Universität Berlin, Berlin 14195, Germany
2Cluster of Excellence “Matters of Activity Image Space Material”, Humboldt-Universität zu Berlin, Berlin 10099, Germany
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Malte R. Henningsen-Schomers
Friedemann Pulvermüller
1Brain Language Laboratory, Department of Philosophy and Humanities, Freie Universität Berlin, Berlin 14195, Germany
2Cluster of Excellence “Matters of Activity Image Space Material”, Humboldt-Universität zu Berlin, Berlin 10099, Germany
3Berlin School of Mind and Brain, Berlin 10099, Germany
4Einstein Center for Neurosciences, Berlin D-10117, Germany
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Friedemann Pulvermüller
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Language influences cognitive and conceptual processing, but the mechanisms through which such causal effects are realized in the human brain remain unknown. Here, we use a brain-constrained deep neural network model of category formation and symbol learning and analyze the emergent model’s internal mechanisms at the neural circuit level. In one set of simulations, the network was presented with similar patterns of neural activity indexing instances of objects and actions belonging to the same categories. Biologically realistic Hebbian learning led to the formation of instance-specific neurons distributed across multiple areas of the network, and, in addition, to cell assembly circuits of “shared” neurons responding to all category instances—the network correlates of conceptual categories. In two separate sets of simulations, the network learned the same patterns together with symbols for individual instances [“proper names” (PN)] or symbols related to classes of instances sharing common features [“category terms” (CT)]. Learning CT remarkably increased the number of shared neurons in the network, thereby making category representations more robust while reducing the number of neurons of instance-specific ones. In contrast, proper name learning prevented a substantial reduction of instance-specific neurons and blocked the overgrowth of category general cells. Representational similarity analysis further confirmed that the neural activity patterns of category instances became more similar to each other after category-term learning, relative to both learning with PN and without any symbols. These network-based mechanisms for concepts, PN, and CT explain why and how symbol learning changes object perception and memory, as revealed by experimental studies.

  • category learning
  • concept formation
  • deep neural network
  • Hebbian associative learning
  • instance representation
  • verbal symbol learning

Significance Statement

How do verbal symbols for specific individuals (Micky Mouse) and object categories (house mouse) causally influence conceptual representation and processing? Category terms and proper names (PN) have been shown to promote category formation and instance learning, potentially by directing attention to category critical and object-specific features, respectively. Yet the mechanisms underlying these observations at the neural circuit level remained unknown. Using a mathematically precise deep neural network model constrained by properties of the human brain, we show category-term learning strengthens and solidifies conceptual representations, whereas PN support object-specific mechanisms. Based on network internal mechanisms and unsupervised correlation-based learning, this work offers neurobiological explanations for the causal effects of symbol learning on concept formation, category building, and instance representation in the human brain.

Introduction

Most signs and symbols are used to speak about objects and actions. This led philosophers and logicians to propose that the referential link between symbol and world is essential for meaning and semantics (Wittgenstein, 1922; Frege, 1948). Yet there are quite different relationships between symbols and their related real-world entities. One most essential difference exists between “proper names” (PN) used to speak about a single object or individual (e.g., “Mickey Mouse”) and “category terms” (CT), which can refer to members of an entire class or conceptual category (e.g., “house mouse”). Such differences between referential symbols are well-described at the semantic level, but not understood in terms of their underlying mechanisms in the mind and brain.

The need for mechanistic neurobiological models of symbols and their meaning comes from reports about the causal influences of language on perception, attention, and memory. It had long been speculated and recently been confirmed that, when human subjects learn words for objects, language may help humans to attend to and distinguish between them (Majid et al., 2004; Whorf and Carroll, 2007; Miller et al., 2018; Vanek et al., 2021). Experimental research in infants showed that learning “labels” for objects increases their attention to these objects (Baldwin and Markman, 1989), which further establishes an attention-catching function of language. However, this general insight requires further specification to capture the different effects of CT and PN. In particular, learning a new symbol for a category of objects makes infants attend to the shared features of these objects and facilitates their learning of the conceptual category (Gelman and Markman, 1986, 1987; Plunkett et al., 2008); the latter even holds if the objects show little perceptual similarity (Graham et al., 2013). On the other hand, the category building function of language is absent when object-specific PN are learned. In this case, the infant's attention is directed not toward the common category features of objects but to idiosyncratic and object-specific features instead (Scott and Monesson, 2009; LaTourrette and Waxman, 2020). In summary, category-term learning directs attention to shared features of objects (Waxman and Booth, 2001; Dewar and Xu, 2007; Althaus and Mareschal, 2014; Althaus and Plunkett, 2016), whereas unique proper name learning highlights idiosyncratic and object-specific features (Best et al., 2010; Barnhart et al., 2018; Pickron et al., 2018; LaTourrette and Waxman, 2020). These specific and replicable effects of PN and CT on perception and attention have been explained in terms of different “strategies” applied by the learner. A neurobiological explanation of why these specific effects occur is still missing.

Why and how can PN and CT direct attention to specific versus shared features of category members? To develop a mechanistic explanation, we used a brain-constrained deep neural network designed according to the area structure and connectivity of major areas relevant to language and conceptual processing (Garagnani et al., 2007; Tomasello et al., 2018; Pulvermüller et al., 2021). Six “areas” of the model simulated processes in superior temporal and inferior frontal perisylvian language areas and six extrasylvian model areas simulated inferior temporo-occipital visual “where” processing stream and dorsolateral prefrontal and motor cortices (Fig. 1A). In the no-symbol (NoS) condition, the model learned activity patterns each representing 1 of 60 instances of objects or actions belonging to 10 different categories. In learning-with-symbols conditions, the model learned additional activity patterns representing word forms of PN or CT (Figs. 1B,C, 2A). After learning, the model was tested by activating previously trained instance patterns of each category and, in addition, new patterns for novel instances belonging to the same categories (Fig. 2B). We documented the neural and cognitive effects of PN and CT on instance and category learning in the model. In-depth analyses of the emerging activation patterns and representations were provided by using representational similarity analysis (RSA; Kriegeskorte et al., 2008) and by classifying neurons into instance-specific and category general ones.

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

A, Area structure and between-area connectivity of the neural network model. Left: The network model's 12 cortical areas in the left fronto-temporo-occipital lobes—inferior frontal articulatory (red) and superior temporal auditory systems (blue) of the perisylvian areas and the lateral frontal hand motor system (yellow/orange/brown) and visual “what” stream (green) in the extrasylvian cortex. Right: Connections among the 12 modeled brain areas—direct connections between adjacent areas (black arrows), second nearest-neighbor areas (blue arrows), and long-distant links (purple arrows). Figure modified from Tomasello et al. (2018). B, Schematic illustrations of activity patterns for instances of two categories. The categories are illustrated with images of robots and cat faces but note that this is for illustrative purposes. The actual input to the model was not images, but grounding patterns consisting of sets of activated neurons (see main text for details). Active neurons of given activity patterns were either shared among instances of the same category (black) or unique to each instance (color). Each model area included 25×25 excitatory neurons, i.e., 625 cells. Left: In grounding patterns (i) presented to *V1/*M1L, six shared active neurons (black) code for the common perceptual–semantic features of the category “a,” and six unique neurons (color) represent instance-specific perceptuomotor features from each of the category members. Member instances of one category activated the same six shared neurons while the instance from another category activated a different set of six shared neurons; each instance also activated six unique neurons. Middle: 12 neurons (black) make up word form pattern for the category term; in the category term condition, member instances coactivated with the same word form pattern (ii) in *A1/*M1i. Right: 12 unique neurons (color) represent each proper name of an individual instance, which are activated 1-to-1 with these instances in the proper name condition. Instances were coactivated with distinct different word form patterns (iii) in *A1/*M1i regardless of category. C, Simulating no-symbol learning (top), category term learning (bottom-left), and proper name learning (bottom-right) where no word form pattern, word form patterns (ii), and word form pattern (iii) were presented to *A1/*M1i, respectively.

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Experimental design used for instance learning and conceptual grounding. A, Training phase with 30 object instances from ten categories. The categories are illustrated with images of robots and cat faces, but note that this is for illustrative purposes. The actual input to the model was not images, but grounding patterns consisting of sets of activated neurons (see main text for details). For each trained instance, the grounding pattern (i) was either presented to the network on its own (no symbol) or combined with a “word form pattern” of type (ii, category term) or type (iii, proper name). B, Testing phase with a collection of the initially trained 30 instances and 30 novel instances from the 10 original categories, resulting in 60 testing instances (i.e., 6 per category). C, Training conditions in the main simulations (top) and control simulations (bottom) differ in the number of training trials (tt) to match the number of instance representations and the number of word form representations, respectively.

Materials and Methods

Participants

The current work does not contain experiments with human participants or animal subjects.

Neurobiological constraints

In contrast to many neural network models, the brain-constrained model aimed at biological plausibility by applying a range of structural and functional constraints (used in these studies Pulvermüller and Garagnani, 2014; Tomasello et al., 2018; Henningsen-Schomers and Pulvermüller, 2022; for review, see Pulvermüller et al., 2021) realizing:

  1. neurophysiological dynamics of spiking pyramidal cells (Connors et al., 1982; Matthews, 2001),

  2. synaptic weights under the modification of unsupervised Hebbian-type learning (i.e., synaptic plasticity and learning were modified according to the biologically plausible unsupervised Hebbian principles that incorporated both long-term potentiation and long-term depression; Artola and Singer, 1993),

  3. local and global activity regulation (Braitenberg, 1978; Yuille and Geiger, 1995) based on local and area-specific inhibition mechanisms (Knoblauch and Palm, 2002),

  4. excitatory and inhibitory within-area local connectivity (including sparse, random, and initially weak excitatory links whose probability falls off with distance; Kaas, 1997; Braitenberg and Schüz, 1998),

  5. between-area global connectivity built on neuroanatomical evidence, and

  6. built-in uncorrelated white noise in neurons of (1) all areas during training and testing mimicked spontaneous baseline neuronal firing and (2) additional noise in neurons of areas not stimulated by patterns during training, which simulated uncorrelated sensory or motor activity unrelated to instances or symbols (Rolls and Deco, 2010).

Table 2 supplies the model specifications and parameters chosen in this current work.

Model description

We applied a brain-constrained deep neural network model including spiking model neurons and 12 model areas to model sensorimotor, conceptual, and linguistic mechanisms in the left-hemispheric language-dominant fronto-temporo-occipital regions of the human brain, as described in previous studies by Tomasello et al. (2018) and Henningsen-Schomers and Pulvermüller (2022).

Anatomical architecture and connectivity

To distinguish between subparts of neural networks from their target cortical structures of the real human brain, all model areas are marked by an asterisk before (e.g., *A1, *V1). The architecture modeled three areas representing the ventral visual system [i.e., primary visual cortex (*V1), temporo-occipital area (*TO), anterior–temporal area (*AT)] and three areas representing the dorsolateral action system [i.e., dorsolateral fronto-central motor (*M1L), premotor cortex (*PML), prefrontal cortex (*PFL)]. These formed the extrasylvian region for sensorimotor processing where semantic information was stored. Another six areas of the perisylvian region for word form processing housed articulatory–phonological and acoustic–phonological information. These areas involved the three areas of the auditory system [i.e., primary auditory cortex (*A1), auditory belt (*AB), parabelt areas (*PB)] and three inferior frontal articulatory and prefrontal areas [i.e., inferior primary motor cortex (*M1i), premotor cortex (*PMi), prefrontal cortex (*PFi)], respectively. Between-area connections were reciprocal and connected next-neighbor areas, second next neighbors (Schomers et al., 2017), and long-distance corticocortical links supported by neuroanatomical evidence in the literature (Table 1).

View this table:
  • View inline
  • View popup
Table 1.

Connectivity structure of the modeled cortical areas with neuroanatomical evidence

In the current neural network model, the fundamental information processing units are artificial neuron-like elements or cells. Each modeled area comprised two layers of 625 e-cells and 625 i-cells that mimicked an (excitatory) pyramidal spiking neuron and a cluster of (inhibitory) interneurons hosted within the same cortical column in the cortical area. A more elaborate description of the firing behavior of such neurons can be found in the studies of Garagnani et al. (2017), Tomasello et al. (2018), and Henningsen-Schomers and Pulvermüller (2022).

Activity patterns applied to the networks

A total of 60 “grounding patterns” were defined as sensorimotor activation patterns thought to represent specific sensory-motor experiences of 60 different objects or “instances.” Groups of six instances overlapped in their neuronal grounding patterns and were taken as representations of different instances of the same concept (e.g., different robots). Note that the images of robots and cat faces for category members are to be taken purely for illustrative purposes here—the actual training patterns of the models consisted of sets of activated neurons with no systematic relationship to images of robots or cat faces. A category comprised three trained instances and three novel instances not presented during training; all six instance patterns were used for network testing (Fig. 2A,B). Each category instance was neuronally coded as a set of perceptual and motor neuron activations in the primary visual and hand motor areas of the brain-constrained network. These instance-related grounding patterns were activated either on their own or together with additional patterns of neuronal activation in the network's articulatory and auditory cortices, which were thought to implement symbol forms, that is, verbal labels or spoken word forms. These “word form patterns” were used either as PN and therefore specifically with only one grounding pattern or as CT, and therefore the same word form pattern co-occurred with all three trained grounding patterns of one category. To control the effect of nonlinguistic factors, a third class of trained grounding patterns was learned without concordant auditory–articulatory activation. Thus, we generated three classes of simulated stimulation patterns: (i) instance-related grounding patterns applied to *V1/*M1L (Fig. 1B, left), (ii) category term patterns to *A1/*M1i (Fig. 1B, middle), and (iii) proper name patterns to *A1/*M1i (Fig. 1B, right). Sensorimotor experiences of instances were simulated with conceptual grounding patterns (i), and symbol-related auditory–articulatory activity was simulated using word form patterns (ii and iii).

For visualization and a better conceptual understanding of the use of activity patterns, see Figure 1B,C. Instances belonging to the same category were simulated by similar grounding patterns, following Henningsen-Schomers and Pulvermüller (2022): within-category instances had grounding patterns that shared 50% of their feature neurons and differed from each other in the other half; grounding patterns simulating instances from different categories had no neuronal overlap. For each grounding pattern (i), a subset of 12 out of 625 potential cells per area was randomly chosen, consisting of 6 unique neurons and 6 shared neurons. Shared neurons simulated features characterizing all instances patterns of a category; they simulated shared conceptual features of all category members (category-critical feature, e.g., members of the first category are robots in the same height and are equipped with one camera, one speaker, two antennae, a power button, two metal legs, and a pair of shoes; members of the second category are cats and have round-shaped head, eyes, nose, mouth, ears, and whiskers; Fig. 1B, left). Unique neurons simulated the “idiosyncratic”, fully instance-specific visuomotor features; each of the corresponding feature neurons was only available in one instance pattern (e.g., robots vary in the body shape and color, the orientation of antennas, leg forms, the position of the power button, and shoe color). In sum, each category possessed 36 unique neurons from its 6 exemplars and 6 shared neurons. For word form patterns, category term patterns (ii) of within-category instances consisted of the same twelve neurons, which were coactivated with each of the three learnt grounding patterns of a category (e.g., to simulate the artificial words fos for all instances of the robot category, and coxt for all instances of the cat category; Fig. 1B, middle); each proper name pattern (iii) comprised twelve neurons, which were coactivated with one specific grounding pattern (e.g., xub, vit, and hek for the three instances of the robot category, respectively; Fig. 1B, right). The choice of cells for pattern generation was pseudorandomized and constrained by the following criteria. First, within-category neurons had to be nonadjacent to each other. This prevented coactivation merely due to close distance. Second, no grounding patterns from two different categories shared any neuron. Last, for each instance, the grounding patterns in *V1 and *M1L followed the same principles but were not identical. The same rules applied to the grounding patterns in *A1 and *M1i.

Experimental design

The current simulations involved three phases, model initialization, training phase, and testing phase, which were carried out on the high-performance computing system of Freie Universität Berlin (Bennett et al., 2020). During training, there were three different stimulation conditions, (1) where grounding patterns were learnt without symbol (no-symbol or control condition), (2) where all grounding patterns of each category were presented together with the same word form pattern (category term condition), and (3) where each grounding pattern was copresented with its own specific word form pattern (proper name condition). Thus, during learning, a stimulation pattern included two activation patterns (to *V1 and *PFL) when it was learned outside symbol context (Fig. 1C, top) or a quadruplet including the two instance-related patterns plus two-word form-related ones (to A1 and PFi) when learned in symbol context (Fig. 1C, bottom). Each test trial began with the presentation of a grounding pattern of an instance (projected to the two sensorimotor model areas V1 and M1L).

Model initialization

One crucial step prior to training was model initialization, which randomized all synaptic links (and their corresponding weights) between within-area cells and between cells from connected areas. Twelve sets of such synaptic links and weights (i.e., 12 different instantiations of the randomly initialized neural network) were chosen, each set was then triplicated (cf. Schomers et al., 2017), and each of these three copies entered one of the three training conditions—either no symbol, category term, or proper name. The use of distinct model instantiations can be seen as analogous to a within-subject study design with 12 subjects. We chose to implement three separate sets of simulations for the three conditions to avoid any possible interference effects between concepts and symbols that may emerge during training. Note, for example, that the relatively large representations that formed for CT might have interfered with further learning or may even have suppressed the activation of conceptual representations without symbols. This configuration yielded a controlled “within-subject” design with the training condition being a three-level repeated measure factor (no symbol, category term, and proper name). For the additional simulations performed to balance the number of word form presentations, there were four levels.

Training phase

The neural network model was repeatedly presented with 30 instances from ten categories. To mimic visuomotor percepts associated with an instance, the extrasylvian primary sensorimotor areas, *V1 and *M1L, were each presented with their grounding pattern (i) for 16 time steps. Following the experiment by LaTourrette and Waxman (2020) where instances were called either by a consistent label or by distinct labels each, our within-category trained instances were either paired with the same category term, by their distinct PN, or they were not labeled at all. To mimic symbols in the category term and proper name conditions, we presented to the primary perisylvian areas *A1 and *M1i word form pattern (ii and iii), respectively, for 16 time steps (Fig. 1C, bottom, 2A). Hence, in different “learning trials,” the word form patterns of CT were copresented with one of three different grounding patterns from one category, whereas those of PN co-occurred with only one specific grounding pattern. There were no word form patterns presented in the baseline no-symbol condition to control for the effect of either type of linguistic label compared with learning without one (Fig. 1C, top, 2A).

Because activity at the end of a trial might affect learning in the next trial, the network was allowed to deactivate after each stimulated learning trial. To this end, we separated every two consecutive pattern stimulations by a waiting interval during which only the uncorrelated white noise mimicking spontaneous baseline neuronal firing was supplied to all areas (see Principle 6 in Model description—Neurobiological constraints). The goal was to reset the global network (i.e., all excitatory and inhibitory cells displayed a membrane potential of zero) before a new grounding pattern was inputted into the neural network model. This interstimulus interval was terminated only after the network activity had returned to its baseline value (thresh = 0.18, Table 2). As a result, the training order was not influential in this experiment.

View this table:
  • View inline
  • View popup
Table 2.

Parameter values used in the simulations

To balance learning conditions (NoS, CT, PN), each experiential grounding pattern representing an instance was presented 2,000 times in one set of simulations. However, because each category term pattern was copresented with three different instance patterns, whereas proper name patterns co-occurred with only one, this design leads to an imbalance of the number of learning trials during which individual word form patterns were presented (three times higher for category term than for proper name presentations; Fig. 2C, top). Therefore, a second evaluation of learning trials was performed and analyzed for which the number of word form pattern activations was balanced. In this case, there were 1,000 learning trials in the category term condition (CL_1x; each instance was presented together with a category term in 1,000 training trials, resulting in a total of 3,000 training trials per CT) and 3,000 trials in the proper name condition (PN_3x; each instance was presented together with a proper name in 3,000 training trials, resulting in a total of 3,000 training trials per proper name). For the control no-symbol conditions, two comparison values were calculated, after 1,000 (NoS_1x) and 3,000 (NoS_3x) trials (i.e., each instance was presented without symbol in 1,000 and 3,000 training trials, respectively; Fig. 2C, bottom). These different subdesigns are summarized graphically in Figure 2C.

Testing phase

In the current experiment, we implemented a version of an old-new recognition task with the use of new instances. For each of the ten categories, we presented to the neural network six testing instances: three trained instances and three novel instances (Fig. 2B). In total, we used 30 previously learnt instances and 30 new instances. However, no actual old-new pairing took place because we presented trained and novel instances to the neural network in separate test trials.

Memory performance of the network model was assessed in the absence of linguistic cues, i.e., without stimulating the perisylvian primary areas *A1 or *M1i. To stimulate the experience of individual instances, the extrasylvian primary areas *V1 and *M1L were activated for two time steps with pure (i.e., free of any white noise) grounding patterns (i) and subsequentially deactivated toward the baseline for 28 time steps. We recorded network responses 30 time steps from the onset of this stimulation. Global resetting between two consecutive trials was conducted in the same manner as the training phase. Hence, the test order was not of interest.

Data analysis

Grounding pattern production, data processing, and data analysis were performed using Python 3.9.7, matplotlib 3.4.3 (Hunter, 2007), NumPy 1.20.3 (Harris et al., 2020), pandas 1.3.4 (Reback et al., 2022), SciPy 1.7.1 (Virtanen et al., 2020), and seaborn 0.11.2 (Waskom, 2021). In the current work, statistical significances were based on a conservative p value threshold of 0.005 suggested by Di Leo and Sardanelli (2020). We used rstatix 0.7.0 (Kassambara, 2021) in the R software environment (R Core Team, 2021) for statistical analyses.

When testing stimuli were presented to the primary sensorimotor areas, some of the 625 excitatory neurons per area fired in response to their conceptual grounding patterns. As described in the procedure, we recorded all their responses during 30 time steps from stimulation. Let ϕ(et) denote the output of an excitatory cell e at time t, such that ϕ only takes up the value 0 or 1 and t only allows discrete values up to 30 (corresponding to thirty possible simulation time steps); let τFavg=5 be a time constant, and the estimated instantaneous firing rate ωE(et) of cell e at time t can be calculated based on the following equation:τFavg⋅dωE(et)dt=−ωE(et)+ϕ(et) (1)Solving Equation 1 for ωE(et) returns the cell's latest spiking activity (firing rate). We estimated the mean firing rate based on t=t30 and used this value for the subsequent RSAs. For details about relevant calculation steps, see the Appendix in the study of Henningsen-Schomers and Pulvermüller (2022).

Previous research found that several of the extrasylvian areas targeted by the deep neural model (including, for example, *V1 and *AT) are important for processing instance- and concept-related information (Binder et al., 2005; Martin, 2007; Ralph et al., 2017; Henningsen-Schomers et al., 2022). Therefore, the current data analyses and statistical testing focused on the extrasylvian region of the deep neural network. This decision was motivated by the main aim of addressing possible causal influences of symbol learning on the perceptual processing of instances of concepts and on conceptual processing itself.

RSA

The estimated mean firing rate of 625 neurons in response to a testing instance reflected how this instance was represented in a neural network. To understand how differently the neural network represented within- and between-category instances, we calculated the dissimilarity in firing patterns for every pair of the 60 instances. Pairwise dissimilarities computed in terms of Euclidean distance were organized in a 60×60 representational dissimilarity matrix (RDM; Fig. 3A). Each cell in the matrix reflected the dissimilarity between the firing patterns of two instances. In total, there were 36 RDMs across 3 training conditions and 12 areas.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

A, Schematic extraction of a 60×60 RDM, which represents 12 instances from two different categories and the similarities between any instance pair. For illustration, we once again use the categories of robots and cat faces. The schematic dissimilarity matrix illustrates how between-category (cells outside the red boundaries) and within-category dissimilarities (cells within the red boundaries) were calculated. Of interest are the (1) within-category dissimilarity among trained instances (DissimW−TT , lightest blue shade), (2) within-category dissimilarity between a trained and a novel instance (DissimW−TN , intermediate blue shade), and (3) between-category dissimilarity of two trained instances (DissimB−TT , darkest blue shade). The RDM is symmetric about its diagonal (gray) of zeros (representing the nondissimilarity of each of the instances to itself). Only the upper half of the RDM is used for analysis, and the lower half could be abandoned (black). B, RDMs for each of the twelve model areas in three main simulations: no symbol (top row), category term (middle row), and proper name (bottom row). The squares indicate the degrees to which network activity in the 12 network areas elicited by (12 out of 60) grounding patterns in the three learning conditions differed between each other within and between categories and are color-coded from turquoise (no dissimilarity, Dissim=0 ), blue, pink, and to dark red (high dissimilarity, Dissim>3 ).

We defined two classes of pairwise dissimilarities, including between-category dissimilarity (DissimB) and within-category dissimilarity (DissimW) . A second way to define similarity types is based on the type of instances under study, that is, the dissimilarity between two trained instances (DissimTT) , between two novel instances (DissimNN) , and between a trained and a novel instance (DissimTN) . For example, within-category dissimilarity could be classified as either dissimilarity among trained instances 1–3 (DissimW−TT) , among novel instances 4–6 (DissimW−NN) , or between trained and novel instances (DissimW−TN) (Fig. 3A).

Category learning

Category learning was evaluated through the ability to (1) distinguish differences between categories and (2) group together category members. We assessed how different types of symbols impacted upon category learning performance based on (1) the dissimilarity between two between-category trained instances (DissimB−TT) and (2) the dissimilarity between two within-category trained instances (DissimW−TT) (Fig. 3A). Successful category learning occurred when two instances from two distinct categories were considered as dissimilar (high DissimB−TT ) and/or when two within-category instances were considered as similar (low DissimW−TT ). If, as previously claimed, applying CT invites one to encode the commonalities among instances and thereby facilitates categorization, the deep neural network should represent within-category instances similarly while highlighting the dissimilarities between instances of different categories. In the category term condition, we expected between-category dissimilarities to be greater than within-category dissimilarities DissimB−TTCT>DissimW−TTCT . In contrast, we proposed two scenarios for the proper name condition. In the first scenario, if PN focus the neural network models on encoding only unique features and inhibit the encoding of category-critical features, no traces of category learning will be observable, and the representations of individual instances will be highly dissimilar regardless of their categorical membership (DissimB−TTPN≈DissimW−TTPN) . However, because within-category instances shared 50% of their activated neurons in the extrasylvian primary areas *V1 and *M1L, the neural network could base on such similarities to form category representation. In this second scenario, PN are not sufficient to override category learning; the neural network would house not only the unique representations of the instances but also the commonalities of those belonging to the same category. Like the category term condition, the test data would also show signs of category learning (DissimB−TTPN>DissimW−TTPN) . Taking into account such intrinsic perceptuomotor similarities among instances from the same category, category learning was evaluated not only across symbol (i.e., category term or proper name) learning conditions but also in control conditions (i.e., training without symbols). For example, a superior causal influence of CT on category learning performance would be expressed through a significantly higher DissimB−TTCT and lower DissimW−TTCT relative to training with PN and also relative to training without symbols.

Generalization

Assuming the neural network had encoded the commonalities between within-category trained instances and formed category knowledge with the help of these shared features, they might have as well represented novel instances as members of that category when exposed to the category-critical features in these novel instances. Generalization performance would then be reflected by how similarly within-category trained instances and within-category novel instances stimulated the deep neural network. To evaluate the generalization performance of the neural network on novel instances, pairwise dissimilarities between two trained instances (DissimW−TT) as well as between a trained and a novel instance (DissimW−TN) were extracted. In the testing phase, the chance was low that the neural network readily applied category knowledge earned from thousands of training trials onto a novel instance in the first and only exposure. In the case of poor generalization performance, the activation pattern of within-category novel instances would be dissimilar from that of the within-category trained instances (i.e., increasing DissimW−TN ). Our criterion for a successful generalization after learning with symbols was that DissimW−TN should be as low as DissimW−TT (DissimW−TN≈DissimW−TT) . In other words, their absolute dissimilarity difference DissimDiff=|DissimW−TN−DissimW−TT| must remain lower than when the deep neural network was trained without symbols.

Cell assembly analysis

Motivated by the notion of cell assemblies (CAs; Hebb, 1949; Braitenberg, 1978; Fuster, 2005), that is, strongly interlinked sets of neurons forming as a consequence of correlated neuronal activity and potentially carrying a main role in cognitive brain processing, we conducted cell assembly analyses to discover possible neuronal correlates of grounding instances, concepts and symbols along with instance-specific and category-critical neurons after repeated exposure to instances and their CT or PN. We extracted CAs activated by each of the 60 grounding patterns used as testing instances based on the criterion described in previous work (Garagnani and Pulvermüller, 2016; Henningsen-Schomers and Pulvermüller, 2022). Grounding patterns in the testing phase tended to coactivate several excitatory neurons (e-cells) in an area, with at least one being maximally responsive (nonresponse was under the threshold of 0.01). To be part of a CA, the firing rate of a given e-cell had to exceed 75% of the firing rate of the maximally responsive cell of the same area. We then computed the number of unique, instance-specific and overlapping, and conceptual neurons among CAs for trained instances of the same category: neurons were classified according to whether they were activated by just one grounding pattern or whether they responded to two or three instances (thus being pair or triple-shared between the learnt instances of a concept). Unique neurons were conceptualized as neurons that encoded specific, “idiosyncratic” features of an instance; shared neurons could be understood as those that encoded common features shared by at least two instances and thus characteristic of their category. The specialized encoding of category-critical features could be indicated by a higher proportion of shared neurons per area, while traces of instance-specific features would be reflected by a larger proportion of unique neurons.

Representations are transformed through different levels of processing, i.e., from the primary areas to secondary areas, and the central “connector hub” areas of the model. We quantified such transformation as the change (i.e., gain/loss) in the number of unique and shared CA cells in the extrasylvian central areas (AT, PFL) comparative to the extrasylvian primary areas (V1, M1L). Gains in a type of neuron, for example, shared neurons, are indicative of intensive encoding of concept-related commonalities on the course of processing, while loss of shared neurons in the central areas implies reduced encoding of idiosyncratic features and hence instance-related information. Percentage gain was calculated as the difference between the number of neurons in the central and primary areas, as a percentage with respect to the number of neurons in the primary areas:Gain=ncentral−nprimarynprimary×100

Representations of category-critical features

A range of previous neurocomputational studies show that, when brain-like networks learn concepts and word meanings, they form CAs that are spread out across sensorimotor and more central areas of the network. The density of shared semantic neurons in the most central connector hubs is greatest due to their high connectivity degree and thus ample convergence of activity in these areas, resulting in especially strong activation, in particular for shared semantic neurons (for discussion, see Garagnani et al., 2017; Tomasello et al., 2018). Relative to instance-specific neurons, shared semantic neurons are activated more frequently during semantic learning, which predicts that these will recruit the largest number of additional cell assembly; these would therefore be semantic, too, and primarily located in the central hub regions. If a labeling condition specifically invites the neural network to encode category-relevant features, we expect (1) more shared neurons than unique neurons in the extrasylvian areas and (2) a greater gain in shared neurons in the central semantic areas compared with the primary areas. Category learning might still occur even in the presence of PN because within-category similarities also characterize sensorimotor experiences. If such information is sufficient, there should be traces of shared neurons in the central, multimodal areas as well. Additionally, CT should activate shared neurons more than PN.

Representations of instance-specific features

When a neural network represents instances as unique entities, it shall reveal specific traces of each instance in the extrasylvian areas, especially in the semantic hubs. In an extreme case where category learning is hindered and the neural network only encodes the uniqueness of instances, there should be (1) more unique than shared neurons in the extrasylvian areas and (2) a gain only in unique neurons in the central areas with respect to the primary areas. Importantly, instances with PN are expected to activate significantly more unique neurons than categorically labeled instances.

We gather from all 12 model instantiations the CAs in response to all 30 trained instances of 10 categories and classify CA cells by their uniqueness to each instance (vs sharedness). To facilitate readers’ understanding about the results, we offer an interactive illustration of these CAs on our web application at (https://phucthuun.shinyapps.io/CL_PN/). This web application enables one to compare the differential effects of CT versus PN in representing category-critical and instance-specific features of within-category and across-category instances.

Results

RSA

Figure 3B gives a first impression of the instance and category learning performance after 2,000 training trials. In the category term condition, instances from the same category activated the neural network similarly, whereas instances from different categories led to substantially more dissimilar activation patterns across the different areas of the network (i.e., firing patterns were highly dissimilar, as color-coded by dark blue and pink). Category knowledge was reflected in a relatively reduced dissimilarity (light blues), which appears as homogenous within each category, contrasting with those between categories, especially in the central areas (semantic hubs). Training the deep neural network without the aid of symbols or with PN reduced the networks’ ability to distinguish instances between categories: activity pattern dissimilarities between instances from different categories were much more substantial in the category term condition than in the proper name condition (color-coded with shades of intermediate blue). In contrast, within-category similarities and generalization performance in the category term condition were superior, as indicated by the more homogeneous (light) blue shade across all six instances (trained and not trained) from the same category, relative to the other two conditions, where different shades of light blue are visible.

Category learning

To evaluate category learning performance after 2,000 learning trials, within-category dissimilarity (DissimW−TT) and between-category dissimilarity between activity patterns elicited by grounding patterns of trained instances (DissimB−TT) were used. Figure 4A describes a global tendency of the deep neural network, across its twelve areas and three training conditions, to identify within-category instances as more similar and between-category instances as more dissimilar to each other. This feature is explained by the grounding patterns presented, which were similar across category instances, but not between. However, between-category dissimilarity is relatively enhanced in central areas, a feature not explained by the stimulations. In the next step, dissimilarity values were averaged for the six extrasylvian areas. The two-factorial repeated measure (3×2) – ANOVA with training condition (no symbol/category term/proper name) and dissimilarity type (DissimW−TT/DissimB−TT) confirmed the main effect of both factors (F(222)=2777.647 , p<0.001η2=0.982 and F(111)=11155.611p<0.001η2=0.996 , respectively) as well as their interaction effect (F(222)=6113.987 , p<0.001η2=0.986 ) on the dissimilarity between instances within these extrasylvian areas. Figure 4B illustrates category-related activation performance of the deep neural network in the extrasylvian areas of the three learning conditions: the neural network successfully grouped together instances from the same category while distinguishing between instances from the same versus from two different categories. Pairwise comparisons with Bonferroni’s correction were computed to observe the effect of training conditions on each level of dissimilarity type and vice versa. The results showed that DissimW−TT was significantly lower than DissimB−TT in all three conditions (ps<0.001) ; same category membership was thus manifest as relatively enhanced activation similarity in all conditions and across areas. The (DissimW−TT) in the category term condition (M = 0.229, SD = 0.005) and the proper name condition (M = 0.264, SD = 0.004) was significantly smaller (i.e., greater similarity) than that in the control no-symbol condition (M = 0.29, SD = 0.006), and they were also significantly different from each other, with greatest similarities after category term labeling (ps<0.001) . Relative to the control no-symbol condition, the deep neural network responded similarly to trained instances coming from the same category when it was trained with symbols and such performance was above baseline. Importantly, the benefit of CT was superior to both training without symbols and with PN. Likewise, the deep neural network returned the highest DissimB−TT (M = 1.48, SD = 0.018) for the category term condition (ps<0.001) , while DissimB−TT in the proper name condition (M = 0.706, SD = 0.01) was not significantly different from that in the no-symbol condition (M = 0.749, SD = 0.045) (p=0.01) , after application of the Bonferroni’s-corrected significance threshold of 0.005. Compared to the no-symbol condition, training with PN only gradually hindered the discrimination of between-category instances but left the separation of within-category instances unaffected. In contrast, both aspects of category learning were present with the aid of CT, reduced within- and enhanced between-category similarities.

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

Bar charts depicting dissimilarities between network activity elicited by trained grounding patterns after learning for each of the three training conditions. A, Main simulation: within-category (W-TT) and between-category (B-TT) dissimilarity values across all 30 trained activity patterns were averaged for each of the twelve model areas. B,C, Within-category (W-TT) and between-category (B-TT) dissimilarities across the 30 trained items were averaged for extrasylvian model areas. The three training conditions of the main simulations (B) were no symbol (NoS, gray), category term (CT, blue), and proper name (PN, pink). The four training conditions of the control simulation (C) were no symbol with each instance presented over 1,000 (NoS_1x, blue-striped gray) or 3,000 trials (NoS_3x, pink-striped gray), Category term where each instance presented over 1,000 trials (CT_1x, blue) and proper name where each instance presented over 3,000 trials (PN_3x, pink). The error bars represent 95% confidence intervals of the mean. The circles above the bars represent post hoc pairwise comparisons between a reference (circles with filled colors) and a corresponding mean (unfilled circles) after Bonferroni’s correction (critical p value = 0.005). Ten comparisons relevant to the main effects of training condition and dissimilarity type and their interaction are illustrated. The asterisks represent two-tailed p values: **p < 0.005, and ***p < 0.001, ns, not significant. The results were replicated in the whole model architecture (six extrasylvian and six perisylvian model areas); see Extended Data Figure 4-1 and Extended Data Table 4-1.

Figure 4-1

Bar charts depicting dissimilarities between network activity elicited by trained grounding patterns after learning for each of the three training conditions. Within-category (W-TT) and between-category (B-TT) dissimilarities across the 30 trained items were averaged for A-B) all 12 model areas and C-D) extrasylvian model areas. For further explanation, see Figure 4. Download Figure 4-1, TIF file.

Table 4-1

ANOVA table reporting significant effects of training condition (No symbol/Category term/Proper name) and dissimilarity type (DissimW−TT/DissimB−TT) on averaged dissimilarity across 12 model areas. Download Table 4-1, DOCX file.

The simulations performed to control for the number of word form presentations during learning were evaluated using a two-factorial repeated measure (4×2) ANOVA with training condition (now four levels, NoS_1x/NoS_3x/CT_1x/PN_3x) and dissimilarity type (DissimW−TT/DissimB−TT) . This confirmed the main effect of both factors (F(1.6718.35)=1113.758 , p<0.001 , η2=0.964 and F(111)=7485.295 , p<0.001 , η2=0.993 , respectively) as well as their interaction effect (F(1.6518.10)=1961.497 , p<0.001η2=0.973) on the dissimilarity between instances within extrasylvian areas. Pairwise comparisons with Bonferroni’s correction were computed to observe the effect of training conditions on each level of dissimilarity type and vice versa. In essence, DissimB−TT in the category term condition was significantly higher than that in the proper name and both no-symbol control conditions (ps<0.001) (Fig. 4C); category term learning increased the dissimilarity across conceptual categories relative to no-symbol learning and proper name learning. The reverse effect, greater dissimilarity values for PN than CT, was found within categories. These observations were therefore valid even when PN were “shown” to the model three times more than CT during learning.

Generalization

To evaluate the generalization performance of the deep neural network on novel instances, pairwise dissimilarities between two trained instances (DissimW−TT) as well as between a trained and a novel instance (DissimW−TN) were used. Figure 5A illustrates the tendency of the deep neural network to represent two trained instances of the same category as more dissimilar, whereas the representations of a novel and a trained instance from the same category were less dissimilar (lighter-shaded columns were mostly higher than darker-shaded columns). In the six extrasylvian areas, a 3×2 ANOVA was computed with training condition (no symbol/category term/proper name) and type of within-category dissimilarity (DissimW−TT/DissimW−TN) as repeated measure factors. Both the main effects of training condition (F(222)=465.217p<0.001η2=0.956) and dissimilarity type (F(111)=7711.618p<0.001η2=0.939) were significant. For these two factors, there was also a significant interaction (F222=635.788p<0.001η2=0.707) (Fig. 5B). The Greenhouse–Geisser sphericity correction to the violated sphericity assumption (p=0.024) for training conditions (p[GG]=2.38×10−11) confirmed this result. Two-sided pairwise comparisons with Bonferroni’s correction showed that DissimW−TN in the category term (M=0.214SD=0.004) and in the proper name conditions (M=0.220SD=0.003) were significantly lower than that in the control no-symbol condition (M=0.249SD=0.004) (ps<0.001) , but they did not differ significantly from each other (p=0.01) (Fig. 5B). DissimW−TN was significantly lower than DissimW−TT in all three conditions (ps<0.001) (Fig. 5B), which means that within-category trained instances were represented as less similar to each other than when each of them was compared with a novel instance from the same category. In other words, trained instances resulted in neuronal response patterns that were more similar to those caused by novel instances than those caused by trained instances from the same category, a finding easily explained by the lack of learning of the idiosyncratic features of novel instances. A further set of pairwise comparisons using Bonferroni’s correction revealed that the absolute DissimDiff in the no-symbol condition (M=0.041 , SD=0.016) was significantly higher than DissimDiff in the category term condition (M=0.016SD=0.012) (p<0.001) but not significantly different from that in the proper name condition (M=0.044SD=0.02) (p=0.009) . In other words, category term learning resulted in the most similar processing of learnt and not-learnt instances and thus to the greatest degree of generalization.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Bar charts depicting dissimilarities between network activity elicited by trained novel grounding patterns after learning for each of the three training conditions. A, Main simulation: within-category dissimilarity values between any two trained instances (W-TT) and between trained and novel instances were averaged for each of the twelve model areas. B,C Within-category dissimilarities between any two trained instances (W-TT) and between trained and novel instances (W-TN) were averaged for extrasylvian model areas. The three training conditions of the main simulations (B) were no symbol (NoS, gray), category term (CT, blue), and proper name (PN, pink). The four training conditions of the control simulation (C) were NoS_1x (blue-striped gray) or NoS_3x (pink-striped gray), CT_1x (blue) and PN_3x (pink). For further explanation, see Figure 4. The results were replicated in the whole model architecture (six extrasylvian and six perisylvian model areas); see Extended Data Figure 5-1 and Extended Data Table 5-1.

Figure 5-1

Bar charts depicting dissimilarities between network activity areas elicited by trained novel grounding patterns after learning for each of the three training conditions. Within-category dissimilarities between any two trained instances (W-TT) and between trained and novel instances (W-TN) were averaged for A&B) all 12 model areas and C&D) perisylvian model areas. For further explanation, see Figure 4. Download Figure 5-1, TIF file.

Table 5-1

ANOVA table reporting significant effects of training condition (No symbol/Category term/Proper name) and dissimilarity type (DissimW−TT/DissimW−TN) on averaged dissimilarity across 12 model areas. Download Table 5-1, DOCX file.

The results from the additional simulations controlling for the number of word form presentations during learning (i.e., four training conditions NoS_1x, NoS_3x, CT_1x, PN_3x, see Materials and Methods) also confirmed that generalization was maximal for novel members of categories for which category term had been learned (Fig. 5C). The mere exposure to instances or learning PN showed little generalization relative to category learning.

These results investigating brain-constrained neural network correlates of conceptual generalization sit well with well-known observations that language-learning children often generalize—or even overcategorize—CT to novel items. In case of overgeneralization of an item, subsequent learning may establish a novel category to which the item belongs. While our results offer a mechanistic perspective on generalization, a detailed simulation of overgeneralization and reclassification learning is left for future study.

Cell assembly analysis

Figure 6A illustrates the tendency of the deep neural network to encode fewer unique neurons (U-shaped function across areas) and more shared neurons (inverted U-shaped function) in the extrasylvian central areas than in the extrasylvian primary areas. In the first step, the number of unique neurons and shared neurons activated by each instance were calculated and averaged across two training conditions. The repeated measure 3×2 ANOVA with training condition (no symbol/category term/proper name) and neuron type (unique/shared) confirmed the significant main effects (F(222)=902.098 , p<0.001η2=0.926 and F(111)=13966.410p<0.001η2=0.998 , respectively) and a significant interaction involving both factors (F222=5027.907p<0.001η2=0.985) . The supplementary 2×2 ANOVA with training condition with symbols (category term/proper name) and neuron type (unique/shared) returned comparable results with two significant main effects (F(111)=1009.255 , p<0.001η2=0.951 and F(111)= 23994.328, p<0.001η2=0.998 , respectively) and a significant interaction involving both factors (F(111)= 4593.789p<0.001η2=0.986) . Pairwise comparisons with Bonferroni’s correction revealed that CT made the neural network reactivate more shared neurons (M=11.242 , SD=0.127) than unique neurons (M=2.861SD=0.051) (p<0.001) . This also applied for training with PN (shared neurons, M=7.963SD=0.222 ; unique neurons, M=3.89 , SD=0.064 ) and training without symbols (shared neurons, M=8.029 , SD=0.194 ; unique neurons, M=4.493 , SD=0.08 ) (ps<0.001) (Fig. 6B). Compared to this control condition, the number of unique instance-specific neurons was moderately reduced by PN but radically so by CT (p<0.001) , whereas the number of shared, conceptual category neurons remained unchanged after proper name learning (p=0.447) but increased dramatically with category term acquisition (p<0.001) . The latter is clear evidence for a facilitatory effect of language, more specifically, of category term learning, on conceptual category formation in brain-constrained deep neural networks.

Figure 6.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 6.

Bar charts depicting average numbers of instance-specific (“unique”) and category general (“shared”) neurons activated by grounding patterns of instances learnt in the three training conditions, no symbol (gray), category term (blue), and proper name (pink). A, Main simulation: The number of activated unique (U) and shared (S) neurons in response to each of the 30 trained instances was averaged across all 12 model areas. B,C, The number of activated neurons in response to the 30 trained grounding patterns was averaged for each of the six extrasylvian areas. D,E, Changes in neuronal activation seen between extrasylvian primary areas, where stimulation was given, and the “higher” more central connector hub areas central to the architecture. Changes in the number of activated neurons in response to trained grounding patterns are shown for the three training conditions. Unique neurons are shown by solid lines with crossed ends and shared ones by broken lines with triangular ends. The three training conditions of the main simulations (B,D) were no symbol (NoS, gray), category term (CT, blue), and proper name (PN, pink). The four training conditions of the control simulation (C) were NoS_1x (blue-striped gray) or NoS_3x (pink-striped gray), CT_1x (blue) and PN_3x (pink). For further explanations, see Figure 4. The results were replicated in the whole model architecture (six extrasylvian and six perisylvian model areas); see Extended Data Figure 6-1 and Extended Data Table 6-1.

Figure 6-1

A &B) The number of activated neurons in response to the 30 trained grounding patterns was averaged for each of the 12 model areas. D&E) Changes in neuronal activation seen between primary areas, where stimulation was given, and the ‘higher’ more central connector hub areas central to the architecturssssse. For further explanations see Figure 4. Download Figure 6-1, TIF file.

Table 6-1

ANOVA table reporting significant effects (across 12 area model) of training condition (No symbol/Category term/Proper name) and neuron type (shared/unique) on the number of activated neurons (left) and on the gain/loss of unique/shared neurons from primary areas to the connector hub (right). Download Table 6-1, DOCX file.

With respect to the gain/loss of neurons in the extrasylvian central areas relative to the primary ones, our repeated-measure 3×2 ANOVA with two factors training condition (no symbol/category term/proper name) and neuron type (unique/shared) confirmed both main effects on the percentage change of neurons and their interaction to be significant (F222=55.17837 , p<0.001 , η2=0.5519424 , F(111)=6471.54090 , p<0.001 , η2=0.9954 , and F(222)=1484.43893 , p<0.001 , η2=0.966 , respectively). According to the subsequent pairwise t tests, the deep neural networks gained shared neurons but lost unique neurons in the central areas, which held true for all conditions (ps<0.001) (Fig. 6D, upward dotted lines represent positive gains in shared neurons and downward solid lines mean negative gains in unique neurons). On the three levels of training condition, the gain in shared neurons and the loss in unique neurons in the category term condition were significantly larger than that in the proper name and no-symbol conditions (ps<0.001) (Fig. 6D). PN did not significantly increase the gain in shared neurons (p=0.1) but led only to a moderate loss of unique neurons, as compared with the control training condition (ps<0.001) . These results further confirm that training with CT magnified both the gain in shared semantic neurons in central areas and the loss of unique instance-specific neurons there. The simulations performed for balancing the number of word form presentations during proper name and category term learning also confirmed these observations (Fig. 6C,E). Therefore, the overgrowth of shared neurons in category term learning does not depend on an abundant number of word form presentations and cannot be explained by adding word form information to instance-related information.

Both RSA and CA analyses were also conducted for the whole model architecture (six extrasylvian and six perisylvian model areas). The findings replicated previous results, indicating category learning (Extended Data Figure 4-1, Extended Data Table 4-1), generalization (Extended Data Figure 5-1, Extended Data Table 5-1), and representations of category critical as well as instance-specific features (Extended Data Figure 6-1, Extended Data Table 6-1).

Discussion

When sensorimotor patterns simulating the processing of similar objects or actions from different categories were presented, the brain-constrained network applied in the current study showed successful conceptual category learning. Category learning outside symbol context was manifested in greater similarities of activity patterns elicited by different instances of the same category as compared with between-category pattern similarities. Importantly, compared with the training of instances per se, concurrent learning of category instances and symbols had a substantial effect on both categorial and instance-specific processes. Category term learning led to an additional increase in dissimilarities between activity patterns across conceptual categories, while making category members substantially more similar to each other. In contrast, proper name learning did not change between-category similarities and led to a relatively minor similarity increase between members of the same category. The model gave evidence of generalization to novel members of learned categories and showed that such generalization was maximal for novel members of categories for which CT had been learned. Meticulous analyses of neuronal activity patterns suggest that the enhancement of within-category similarities and between-category dissimilarities in the context of category symbols is due to an increase in the number of cells responding to all category members. Likewise, the relative persistence of instance-specific neurons with proper name learning underlies the maintained activation differences between category instances observed in this case. All observed effects regarding pattern dissimilarities and neuronal microstructure were greatly pronounced in the central “connector hub” areas of the brain-constrained model applied, as compared with primary areas. Table 3 summarizes major observations in the current data and the corresponding learning aspects these observations reflect.

View this table:
  • View inline
  • View popup
Table 3.

Critical and significant observations and the corresponding aspects of learning

Relationship to experimental and neurocomputational research

Our results can be used to address observations delivered by neurocognitive and neurobehavioral experiments. Neuropsychological evidence highlights the role of the prefrontal cortex in categorical representation (for review, see Kéri, 2003). Prefrontal areas (PFi and PFl) are part of the four central areas of our model, where conceptual neurons constituting category representations emerged most numerously. This is explained by the high degree of convergence of neural activity in these areas, which are not only located in the center of the model architecture but also show the highest connectivity degrees. Due to ample activity converging on these connector hub areas, their frequently activated shared semantic neurons can most efficiently recruit other neurons, which therefore take on similar response properties (Doursat and Bienenstock, 2006). This mechanism may contribute to why these areas act as “semantic hubs” and house neurons reflecting category membership (e.g., PF and AT, see Miller et al., 2002; Seger and Miller, 2010; Garagnani and Pulvermüller, 2016; Tomasello et al., 2017). On the other hand, the higher density of instance-specific neurons in the primary visual/motor model area relative to the centre is evidence for exemplar learning in the sensorimotor cortices (Kéri, 2003; Bowman et al., 2020)—a type of category learning that is based on the representations of specific category instances (Nosofsky, 1988) and should be independent of signs and symbols. Here, solid evidence for category formation was obtained even in the control condition where only sensorimotor patterns were presented to the model without symbols. In line with neural data (Freedman et al., 2001; Seger and Miller, 2010), experimental evidence shows that perceptuomotor similarities among category members are sufficient to trigger category learning in preverbal infants (Sloutsky and Fisher, 2004; de Heering and Rossion, 2015) and animals (Güntürkün et al., 2018; Pusch et al., 2023).

When learning conceptual instances in the context of CT, infants show the most pronounced category building and an attention bias toward shared features of category members (Waxman and Markow, 1995; Dewar and Xu, 2007; Althaus and Mareschal, 2014). In contrast, encountering PN for individual instances focuses their attention relatively more on object-specific features (Barnhart et al., 2018; Pickron et al., 2018; La Tourette and Waxman, 2020). In the current network model, symbol association raises the number of neurons involved in the processing of a given sensorimotor pattern. This can be interpreted as biased attention to the object or action for which the pattern codes and thus explains why label learning generally increases attention to object features. Furthermore, as category term learning increases the number of category-critical shared semantic neurons in the network, at the cost of reducing the number of instance-specific ones, the preobserved greater attention to shared features has a direct model correlate, along with the label-related tendency to build stronger category representations. Infants’ attentional focus on instance-specific features of objects is in line with the relative preservation of instance-specific neurons in the model of proper name learning. Thus, the opposing effects of proper name and category term learning, which, respectively, drive attention toward instance-specific and category general features of objects, are captured by the current model.

A range of neurocomputational studies previously explored the putative brain basis of cognitive processes (Deco and Rolls, 2005; Rolls and Deco, 2015; Palm, 2016), including conceptual category learning and the influence of language on object perception (Rogers and McClelland, 2014; Henningsen-Schomers and Pulvermüller, 2022). For example, Westermann and Mareschal (2014) demonstrated, using a fully distributed parallel processing model, that learning a category label made the neural patterns of category members more similar to each other, whereas different categories moved away from each other in representational space. Our RSA in models mimicking cortical area structure and connectivity, along with within-area excitatory and inhibitory connectivity, achieved the same result. In addition, we determined the neuron-level mechanisms and contributions of different model areas to this result and, in particular, revealed the model–central connector hub areas as the loci where the differences between categorical and instance-specific mechanisms as well as those between the shared- versus specific-feature promoting roles of instance-specific and category labels are most pronounced. As to our knowledge, the contrast between activity patterns and neuronal correlates of PN and CT has not been addressed by previous computational work.

Model explanation

The present simulations offer explanations of the observed phenomena based on neuroscience principles. Of special relevance here are the biological learning mechanisms applied, which include unsupervised Hebbian synaptic strengthening of connections between coactivated neurons and weakening of links between cells firing independently of each other. This principle explains why category labels primarily interlink with the shared neurons of instance representations belonging to the same category. The reason lies in the highest correlation values, as instance-specific neurons are silent when the category term is used together with other category instances. This implies some weakening of connections between the CT’ and the instance-specific neurons, based on the “anti-Hebbian” “neurons out-of-sync delink” rule. The opposite difference applies to PN, whose neural correlates strongly connect to instance-specific neurons but weaken their links with the category-critical shared neurons whenever a different category member co-occurs with its own and thus different name. Effects are most clearly present in the central areas of the network where the neural correlates of words and entities are equally manifest so that their correlation structure can easily be mapped.

Limitations and future direction

The current simulations use idealized instance and category learning conditions. The activation patterns representing conceptual instances and word forms were chosen to be nonoverlapping, except for the neurons coding for shared features. These are idealizations considering both the features of word forms and those of objects and actions could be shared across categories (compare phonological, e.g., “cat”-“hat” or perceptual color/shape similarities). Such similarities are irrelevant to category membership and hence were omitted to keep the simulation well-controlled. Secondly, only a small number of conceptual features were realized, and a small set of shared features determined concept membership. This situation may hold for some concrete terms but not for others and certainly not for abstract concepts (Henningsen-Schomers et al., 2022). Furthermore, PN and CT were acquired by different networks to allow straightforward separation and evaluation of the mechanistic side of different label types—although label types are normally copresent in the same mind and brain. In the future, it is desirable to complement this work with simulations of more realistic conceptual categories and to build one model in which interaction/interference effects between different learning conditions are possible.

Conclusion

The current study strived to meet the need for a mechanistic model of symbols and their meaning within a neurobiological computational framework by addressing specific features of PN (Mickey Mouse) and category symbols (house mouse). Developmentalists and linguists have long been proposing that CT and PN distinctively impact infants’ locus of attention toward category-shared and instance-specific object and action features, respectively. By simulating concept and instance learning in a deep neural network with neurobiologically realistic architecture and brain-like connectivity, we demonstrate that learning these two different symbol types had opposing effects on the emergent neuronal CAs representing and processing instances of a category and the shared conceptual features of that category, which can explain preobserved differences in perceptual, attentive, and memory processes related to the specific and shared features of category instances. These explanations were based on unsupervised Hebbian associative learning mechanism binding neurons involved in correlated processing of instance-specific category general information. The current work could thus not only replicate but also offer underlying neuronal mechanisms and causal neurobiological explanations for well-established observations in cognitive science.

Footnotes

  • We thank Thomas Wennekers, Rosario Tomasello, Luigi Grisoni, Laura Ciaccio, Maxime Carrière, Fynn Dobler, and the other members of the MatCo research group for their insightful opinions and brilliant discussion on theoretical and practical questions and for their help and suggestions at different stages of this work. In addition, we thank the high-performance computing services of the Freie Universität Berlin and Martin Freyer and Philip Krause for their technical support. We thank Ngan Nguyen for her support in image creation and visualization. This work was supported by the European Research Council (ERC) through the Advanced Grant “Material constraints enabling human cognition, MatCo” (ERC-2019-ADG 883811) and by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) under Germany's Excellence Strategy through the Cluster of Excellence “Matters of Activity. Image Space Material” (DFG EXC 2025/1-390648296).

  • The authors declare no competing financial interests.

  • Correspondence should be addressed to Phuc T. U. Nguyen at phuc.thu.uyen.nguyen{at}gmail.com or Friedemann Pulvermüller at friedemann.pulvermuller{at}fu-berlin.de.

This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which permits unrestricted use, distribution and reproduction in any medium provided that the original work is properly attributed.

References

  1. ↵
    1. Althaus N,
    2. Mareschal D
    (2014) Labels direct infants’ attention to commonalities during novel category learning. PLoS One 9:e99670. doi:10.1371/journal.pone.0099670
    OpenUrlCrossRef
  2. ↵
    1. Althaus N,
    2. Plunkett K
    (2016) Categorization in infancy: labeling induces a persisting focus on commonalities. Dev Sci 19:770–780. doi:10.1111/desc.12358
    OpenUrlCrossRef
    1. Arikuni T,
    2. Watanabe K,
    3. Kubota K
    (1988) Connections of area 8 with area 6 in the brain of the macaque monkey. J Comp Neurol 277:21–40. doi:10.1002/cne.902770103
    OpenUrlCrossRefPubMed
  3. ↵
    1. Artola A,
    2. Singer W
    (1993) Long-term depression of excitatory synaptic transmission and its relationship to long-term potentiation. Trends Neurosci 16:480–487. doi:10.1016/0166-2236(93)90081-V
    OpenUrlCrossRefPubMed
  4. ↵
    1. Baldwin DA,
    2. Markman EM
    (1989) Establishing word-object relations: a first step. Child Dev 60:381–398. doi:10.1111/j.1467-8624.1989.tb02723.x
    OpenUrlCrossRefPubMed
  5. ↵
    1. Barnhart WR,
    2. Rivera S,
    3. Robinson CW
    (2018) Effects of linguistic labels on visual attention in children and young adults. Front Psychol 9:358. doi:10.3389/fpsyg.2018.00358
    OpenUrlCrossRef
    1. Bauer RH,
    2. Fuster JM
    (1978) The effect of ambient illumination on delayed-matching and delayed-response deficits from cooling dorsolateral prefrontal cortex. Behav Biol 22:60–66. doi:10.1016/S0091-6773(78)92019-9
    OpenUrlCrossRefPubMed
    1. Bauer RH,
    2. Jones CN
    (1976) Feedback training of 36–44 Hz EEG activity in the visual cortex and hippocampus of cats: evidence for sensory and motor involvement. Physiol Behav 17:885–890. doi:10.1016/0031-9384(76)90003-2
    OpenUrlCrossRefPubMed
  6. ↵
    1. Bennett L,
    2. Melchers B,
    3. Proppe B
    (2020) Curta: a general-purpose high-performance computer at ZEDAT, Freie Universität Berlin. 5 S.
  7. ↵
    1. Best C,
    2. Robinson C,
    3. Sloutsky V
    (2010) The effect of labels on visual attention: an eye tracking study. Proceedings of the Annual Meeting of the Cognitive Science Society 32. Available at: https://escholarship.org/uc/item/0wn1j6px [Accessed Oct. 9, 2022].
  8. ↵
    1. Binder JR,
    2. Westbury CF,
    3. McKiernan KA,
    4. Possing ET,
    5. Medler DA
    (2005) Distinct brain systems for processing concrete and abstract concepts. J Cogn Neurosci 17:905917. doi:10.1162/0898929054021102
    OpenUrlCrossRefPubMed
  9. ↵
    1. Bowman CR,
    2. Iwashita T,
    3. Zeithamova D
    (2020) Tracking prototype and exemplar representations in the brain across learning Behrens TE, Barense M, Barense M, Tompary A, eds. Elife 9:e59360. doi:10.7554/eLife.59360
    OpenUrlCrossRef
  10. ↵
    1. Braitenberg V
    (1978) Cell assemblies in the cerebral Cortex. In: Theoretical approaches to complex systems. Lecture notes in biomathematics (Heim R, Palm G, eds), pp 171–188. Berlin, Heidelberg: Springer.
  11. ↵
    1. Braitenberg V,
    2. Schüz A
    (1998) Cortex: statistics and geometry of neuronal connectivity. Berlin, Heidelberg: Springer Berlin Heidelberg.
    1. Bressler SL,
    2. Coppola R,
    3. Nakamura R
    (1993) Episodic multiregional cortical coherence at multiple frequencies during visual task performance. Nature 366:153–156. doi:10.1038/366153a0
    OpenUrlCrossRefPubMed
    1. Catani M,
    2. Jones DK,
    3. Donato R,
    4. Ffytche DH
    (2003) Occipito-temporal connections in the human brain. Brain 126:2093–2107. doi:10.1093/brain/awg203
    OpenUrlCrossRefPubMed
    1. Catani M,
    2. Jones DK,
    3. Ffytche DH
    (2005) Perisylvian language networks of the human brain. Ann Neurol 57:8–16. doi:10.1002/ana.20319
    OpenUrlCrossRefPubMed
    1. Chafee MV,
    2. Goldman-Rakic PS
    (2000) Inactivation of parietal and prefrontal cortex reveals interdependence of neural activity during memory-guided saccades. J Neurophysiol 83:1550–1566. doi:10.1152/jn.2000.83.3.1550
    OpenUrlCrossRefPubMed
  12. ↵
    1. Connors BW,
    2. Gutnick MJ,
    3. Prince DA
    (1982) Electrophysiological properties of neocortical neurons in vitro. J Neurophysiol 48:1302–1320. doi:10.1152/jn.1982.48.6.1302
    OpenUrlCrossRefPubMed
  13. ↵
    1. de Heering A,
    2. Rossion B
    (2015) Rapid categorization of natural face images in the infant right hemisphere Culham JC, ed. Elife 4:e06564. doi:10.7554/eLife.06564
    OpenUrlCrossRefPubMed
    1. Deacon TW
    (1992) Cortical connections of the inferior arcuate sulcus cortex in the macaque brain. Brain Res 573:8–26. doi:10.1016/0006-8993(92)90109-M
    OpenUrlCrossRefPubMed
  14. ↵
    1. Deco G,
    2. Rolls ET
    (2005) Neurodynamics of biased competition and cooperation for attention: a model with spiking neurons. J Neurophysiol 94:295–313. doi:10.1152/jn.01095.2004
    OpenUrlCrossRefPubMed
  15. ↵
    1. Dewar K,
    2. Xu F
    (2007) Do 9-month-old infants expect distinct words to refer to kinds? Dev Psychol 43:1227–1238. doi:10.1037/0012-1649.43.5.1227
    OpenUrlCrossRefPubMed
  16. ↵
    1. Di Leo G,
    2. Sardanelli F
    (2020) Statistical significance: p value, 0.05 threshold, and applications to radiomics—reasons for a conservative approach. Eur Radiol Exp 4:18. doi:10.1186/s41747-020-0145-y
    OpenUrlCrossRefPubMed
    1. Distler C,
    2. Boussaoud D,
    3. Desimone R,
    4. Ungerleider LG
    (1993) Cortical connections of inferior temporal area TEO in macaque monkeys. J Comp Neurol 334:125–150. doi:10.1002/cne.903340111
    OpenUrlCrossRefPubMed
  17. ↵
    1. Doursat R,
    2. Bienenstock E
    (2006) Neocortical self-structuration as a basis for learning. 5th International Conference on Development and Learning (ICDL 2006).
    1. Dum RP,
    2. Strick PL
    (2002) Motor areas in the frontal lobe of the primate. Physiol Behav 77:677–682. doi:10.1016/S0031-9384(02)00929-0
    OpenUrlCrossRefPubMed
    1. Dum RP,
    2. Strick PL
    (2005) Frontal lobe inputs to the digit representations of the motor areas on the lateral surface of the hemisphere. J Neurosci 25:1375–1386. doi:10.1523/JNEUROSCI.3902-04.2005
    OpenUrlAbstract/FREE Full Text
    1. Eacott MJ,
    2. Gaffan D
    (1992) Inferotemporal-frontal disconnection: the uncinate fascicle and visual associative learning in monkeys. Eur J Neurosci 4:1320–1332. doi:10.1111/j.1460-9568.1992.tb00157.x
    OpenUrlCrossRefPubMed
  18. ↵
    1. Freedman DJ,
    2. Riesenhuber M,
    3. Poggio T,
    4. Miller EK
    (2001) Categorical representation of visual stimuli in the primate prefrontal cortex. Science 291:312–316. doi:10.1126/science.291.5502.312
    OpenUrlAbstract/FREE Full Text
  19. ↵
    1. Frege G
    (1948) Sense and reference. Philos Rev 57:209. doi:10.2307/2181485
    OpenUrlCrossRef
  20. ↵
    1. Fuster JM
    (2005) Cortex and mind. New York: Oxford University Press.
    1. Fuster JM,
    2. Jervey JP
    (1981) Inferotemporal neurons distinguish and retain behaviorally relevant features of visual stimuli. Science 212:952–955. doi:10.1126/science.7233192
    OpenUrlAbstract/FREE Full Text
    1. Fuster JM,
    2. Bauer RH,
    3. Jervey JP
    (1985) Functional interactions between inferotemporal and prefrontal cortex in a cognitive task. Brain Res 330:299–307. doi:10.1016/0006-8993(85)90689-4
    OpenUrlCrossRefPubMed
  21. ↵
    1. Garagnani M,
    2. Pulvermüller F
    (2016) Conceptual grounding of language in action and perception: a neurocomputational model of the emergence of category specificity and semantic hubs Barbas H, ed. Eur J Neurosci 43:721–737. doi:10.1111/ejn.13145
    OpenUrlCrossRefPubMed
  22. ↵
    1. Garagnani M,
    2. Wennekers T,
    3. Pulvermüller F
    (2007) A neuronal model of the language cortex. Neurocomputing 70:1914–1919. doi:10.1016/j.neucom.2006.10.076
    OpenUrlCrossRef
  23. ↵
    1. Garagnani M,
    2. Lucchese G,
    3. Tomasello R,
    4. Wennekers T,
    5. Pulvermüller F
    (2017) A spiking neurocomputational model of high-frequency oscillatory brain responses to words and pseudowords. Front Comput Neurosci 10:145. doi:10.3389/fncom.2016.00145
    OpenUrlCrossRef
  24. ↵
    1. Gelman SA,
    2. Markman EM
    (1986) Categories and induction in young children. Cognition 23:183–209. doi:10.1016/0010-0277(86)90034-X
    OpenUrlCrossRefPubMed
  25. ↵
    1. Gelman SA,
    2. Markman EM
    (1987) Young children’s inductions from natural kinds: the role of categories and appearances. Child Dev 58:1532–1541. doi:10.2307/1130693
    OpenUrlCrossRefPubMed
    1. Gierhan SME
    (2013) Connections for auditory language in the human brain. Brain Lang 127:205–221. doi:10.1016/j.bandl.2012.11.002
    OpenUrlCrossRef
  26. ↵
    1. Graham S,
    2. Keates J,
    3. Vukatana E,
    4. Khu M
    (2013) Distinct labels attenuate 15-month-olds’ attention to shape in an inductive inference task. Front Psychol 3:586. doi:10.3389/fpsyg.2012.00586
    OpenUrlCrossRef
  27. ↵
    1. Güntürkün O,
    2. Koenen C,
    3. Iovine F,
    4. Garland A,
    5. Pusch R
    (2018) The neuroscience of perceptual categorization in pigeons: a mechanistic hypothesis. Learn Behav 46:229–241. doi:10.3758/s13420-018-0321-6
    OpenUrlCrossRef
    1. Guye M,
    2. Parker GJM,
    3. Symms M,
    4. Boulby P,
    5. Wheeler-Kingshott CAM,
    6. Salek-Haddadi A,
    7. Barker GJ,
    8. Duncan JS
    (2003) Combined functional MRI and tractography to demonstrate the connectivity of the human primary motor cortex in vivo. NeuroImage 19:1349–1360. doi:10.1016/S1053-8119(03)00165-4
    OpenUrlCrossRefPubMed
  28. ↵
    1. Harris CR
    , et al. (2020) Array programming with NumPy. Nature 585:357–362. doi:10.1038/s41586-020-2649-2
    OpenUrlCrossRefPubMed
  29. ↵
    1. Hebb DO
    (1949) The organization of behavior: a neuropsychological theory. Oxford, England: Wiley.
  30. ↵
    1. Henningsen-Schomers MR,
    2. Pulvermüller F
    (2022) Modelling concrete and abstract concepts using brain-constrained deep neural networks. Psychol Res 86:2533–2559. doi:10.1007/s00426-021-01591-6
    OpenUrlCrossRef
  31. ↵
    1. Henningsen-Schomers MR,
    2. Garagnani M,
    3. Pulvermüller F
    (2022) Influence of language on perception and concept formation in a brain-constrained deep neural network model. Phil Trans R Soc B 378:20210373. doi:10.1098/rstb.2021.0373
    OpenUrlCrossRef
  32. ↵
    1. Hunter JD
    (2007) Matplotlib: a 2D graphics environment. Comput Sci Eng 9:90–95. doi:10.1109/MCSE.2007.55
    OpenUrlCrossRefPubMed
  33. ↵
    1. Kaas JH
    (1997) Topographic maps are fundamental to sensory processing. Brain Res Bull 44:107–112. doi:10.1016/S0361-9230(97)00094-4
    OpenUrlCrossRefPubMed
    1. Kaas JH,
    2. Hackett TA
    (2000) Subdivisions of auditory cortex and processing streams in primates. Proc Natl Acad Sci U S A 97:11793–11799. doi:10.1073/pnas.97.22.11793
    OpenUrlAbstract/FREE Full Text
  34. ↵
    1. Kassambara A
    (2021) rstatix: pipe-friendly framework for basic statistical tests. Available at: https://CRAN.R-project.org/package=rstatix [Accessed July 25, 2022].
  35. ↵
    1. Kéri S
    (2003) The cognitive neuroscience of category learning. Brain Res Rev 43:85–109. doi:10.1016/S0165-0173(03)00204-2
    OpenUrlCrossRefPubMed
  36. ↵
    1. Knoblauch A,
    2. Palm G
    (2002) Scene segmentation by spike synchronization in reciprocally connected visual areas. I. Local effects of cortical feedback. Biol Cybern 87:151–167. doi:10.1007/s00422-002-0331-4
    OpenUrlCrossRefPubMed
  37. ↵
    1. Kriegeskorte N,
    2. Mur M,
    3. Bandettini P
    (2008) Representational similarity analysis - connecting the branches of systems neuroscience. Front Syst Neurosci 2:4. doi:10.3389/neuro.01.016.2008
    OpenUrlCrossRefPubMed
  38. ↵
    1. LaTourrette AS,
    2. Waxman SR
    (2020) Naming guides how 12-month-old infants encode and remember objects. Proc Natl Acad Sci U S A 117:21230–21234. doi:10.1073/pnas.2006608117
    OpenUrlAbstract/FREE Full Text
    1. Lu M-T,
    2. Preston JB,
    3. Strick PL
    (1994) Interconnections between the prefrontal cortex and the premotor areas in the frontal lobe. J Comp Neurol 341:375–392. doi:10.1002/cne.903410308
    OpenUrlCrossRefPubMed
  39. ↵
    1. Majid A,
    2. Bowerman M,
    3. Kita S,
    4. Haun DBM,
    5. Levinson SC
    (2004) Can language restructure cognition? The case for space. Trends Cogn Sci 8:108–114. doi:10.1016/j.tics.2004.01.003
    OpenUrlCrossRefPubMed
    1. Makris N,
    2. Pandya DN
    (2009) The extreme capsule in humans and rethinking of the language circuitry. Brain Struct Funct 213:343–358. doi:10.1007/s00429-008-0199-8
    OpenUrlCrossRefPubMed
  40. ↵
    1. Martin A
    (2007) The representation of object concepts in the brain. Annu Rev Psychol 58:25–45. doi:10.1146/annurev.psych.57.102904.190143
    OpenUrlCrossRefPubMed
  41. ↵
    1. Matthews GG
    (2001) Neurobiology: molecules, cells, and systems, 2nd ed. Malden, MA: Blackwell Science.
    1. Meyer JW,
    2. Makris N,
    3. Bates JF,
    4. Caviness VS,
    5. Kennedy DN
    (1999) MRI-based topographic parcellation of human cerebral white matter. Neuroimage 9:1–17. doi:10.1006/nimg.1998.0383
    OpenUrlCrossRefPubMed
  42. ↵
    1. Miller EK,
    2. Freedman DJ,
    3. Wallis JD
    (2002) The prefrontal cortex: categories, concepts and cognition Parker A, Derrington A, Blakemore C, eds. Phil Trans R Soc Lond B 357:1123–1136. doi:10.1098/rstb.2002.1099
    OpenUrlCrossRefPubMed
  43. ↵
    1. Miller TM,
    2. Schmidt TT,
    3. Blankenburg F,
    4. Pulvermüller F
    (2018) Verbal labels facilitate tactile perception. Cognition 171:172–179. doi:10.1016/j.cognition.2017.10.010
    OpenUrlCrossRef
  44. ↵
    1. Nosofsky RM
    (1988) Exemplar-based accounts of relations between classification, recognition, and typicality. J Exp Psychol Learn Mem Cogn 14:700–708. doi:10.1037/0278-7393.14.4.700
    OpenUrlCrossRef
  45. ↵
    1. Palm G
    (2016) Neural information processing in cognition: we start to understand the orchestra, but where is the conductor? Front Comput Neurosci 10:3. doi:10.3389/fncom.2016.00003
    OpenUrlCrossRef
    1. Pandya DN
    (1995) Anatomy of the auditory cortex. Rev Neurol 151:486–494.
    OpenUrlPubMed
    1. Pandya DN,
    2. Barnes CL
    (1987) Architecture and connections of the frontal lobe. In: The frontal lobes revisited (Perecman E, ed), pp 41–72. New York, NY, US: The IRBN Press.
    1. Pandya DN,
    2. Yeterian EH
    (1985) Architecture and connections of cortical association areas. In: Association and auditory cortices. Cerebral cortex (Peters A, Jones EG, eds), pp 3–61. Boston, MA: Springer US.
    1. Parker A,
    2. Gaffan D
    (1998) Interaction of frontal and perirhinal cortices in visual object recognition memory in monkeys. Eur J Neurosci 10:3044–3057. doi:10.1046/j.1460-9568.1998.00306.x
    OpenUrlCrossRefPubMed
    1. Parker GJM,
    2. Luzzi S,
    3. Alexander DC,
    4. Wheeler-Kingshott CAM,
    5. Ciccarelli O,
    6. Lambon Ralph MA
    (2005) Lateralization of ventral and dorsal auditory-language pathways in the human brain. NeuroImage 24:656–666. doi:10.1016/j.neuroimage.2004.08.047
    OpenUrlCrossRefPubMed
    1. Paus T,
    2. Castro-Alamancos MA,
    3. Petrides M
    (2001) Cortico-cortical connectivity of the human mid-dorsolateral frontal cortex and its modulation by repetitive transcranial magnetic stimulation. Eur J Neurosci 14:1405–1411. doi:10.1046/j.0953-816x.2001.01757.x
    OpenUrlCrossRefPubMed
    1. Petrides M,
    2. Pandya DN
    (2009) Distinct parietal and temporal pathways to the homologues of Broca’s area in the monkey Ungerleider L, ed. PLoS Biol 7:e1000170. doi:10.1371/journal.pbio.1000170
    OpenUrlCrossRefPubMed
  46. ↵
    1. Pickron CB,
    2. Iyer A,
    3. Fava E,
    4. Scott LS
    (2018) Learning to individuate: the specificity of labels differentially impacts infant visual attention. Child Dev 89:698–710. doi:10.1111/cdev.13004
    OpenUrlCrossRef
  47. ↵
    1. Plunkett K,
    2. Hu J-F,
    3. Cohen LB
    (2008) Labels can override perceptual categories in early infancy. Cognition 106:665–681. doi:10.1016/j.cognition.2007.04.003
    OpenUrlCrossRefPubMed
  48. ↵
    Pulvermüller F, Garagnani M (2014) From sensorimotor learning to memory cells in prefrontal and temporal association cortex: a neurocomputational study of disembodiment. Cortex 57:1–21. doi:10.1016/j.cortex.2014.02.015
    OpenUrlCrossRefPubMed
  49. ↵
    1. Pulvermüller F,
    2. Tomasello R,
    3. Henningsen-Schomers MR,
    4. Wennekers T
    (2021) Biological constraints on neural network models of cognitive function. Nat Rev Neurosci 22:488–502. doi:10.1038/s41583-021-00473-5
    OpenUrlCrossRef
  50. ↵
    1. Pusch R,
    2. Clark W,
    3. Rose J,
    4. Güntürkün O
    (2023) Visual categories and concepts in the avian brain. Anim Cogn 26:153–173. doi:10.1007/s10071-022-01711-8
    OpenUrlCrossRef
  51. ↵
    R Core Team (2021) R: a language and environment for statistical computing. Available at: https://www.R-project.org/ [Accessed July 25, 2022].
  52. ↵
    1. Ralph MAL,
    2. Jefferies E,
    3. Patterson K,
    4. Rogers TT
    (2017) The neural and computational bases of semantic cognition. Nat Rev Neurosci 18:42–55. doi:10.1038/nrn.2016.150
    OpenUrlCrossRefPubMed
    1. Rauschecker JP,
    2. Tian B
    (2000) Mechanisms and streams for processing of “what” and “where” in auditory cortex. Proc Natl Acad Sci U S A 97:11800–11806. doi:10.1073/pnas.97.22.11800
    OpenUrlAbstract/FREE Full Text
    1. Rauschecker JP,
    2. Scott SK
    (2009) Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing. Nat Neurosci 12:718–724. doi:10.1038/nn.2331
    OpenUrlCrossRefPubMed
  53. ↵
    1. Reback J
    , et al. (2022) pandas-dev/pandas: pandas 1.4.3. Available at: https://zenodo.org/record/3509134 [Accessed July 25, 2022].
    1. Rilling JK
    (2014) Comparative primate neuroimaging: insights into human brain evolution. Trends Cogn Sci 18:46–55. doi:10.1016/j.tics.2013.09.013
    OpenUrlCrossRefPubMed
    1. Rilling JK,
    2. van den Heuvel MP
    (2018) Comparative primate connectomics. Brain Behav Evol 91:170–179. doi:10.1159/000488886
    OpenUrlCrossRef
    1. Rilling JK,
    2. Glasser MF,
    3. Preuss TM,
    4. Ma X,
    5. Zhao T,
    6. Hu X,
    7. Behrens TEJ
    (2008) The evolution of the arcuate fasciculus revealed with comparative DTI. Nat Neurosci 11:426–428. doi:10.1038/nn2072
    OpenUrlCrossRefPubMed
    1. Rilling JK,
    2. Glasser MF,
    3. Jbabdi S,
    4. Andersson J,
    5. Preuss TM
    (2012) Continuity, divergence, and the evolution of brain language pathways. Front Evol Neurosci 3:11. doi:10.3389/fnevo.2011.00011
    OpenUrlCrossRefPubMed
    1. Rizzolatti G,
    2. Luppino G
    (2001) The cortical motor system. Neuron 31:889–901. doi:10.1016/S0896-6273(01)00423-8
    OpenUrlCrossRefPubMed
  54. ↵
    1. Rogers TT,
    2. McClelland JL
    (2014) Parallel distributed processing at 25: further explorations in the microstructure of cognition. Cogn Sci 38:1024–1077. doi:10.1111/cogs.12148
    OpenUrlCrossRefPubMed
  55. ↵
    1. Rolls ET,
    2. Deco G
    (2010) The noisy brain stochastic dynamics as a principle of brain function. New York: Oxford University Press.
  56. ↵
    1. Rolls ET,
    2. Deco G
    (2015) Networks for memory, perception, and decision-making, and beyond to how the syntax for language might be implemented in the brain. Brain Res 1621:316–334. doi:10.1016/j.brainres.2014.09.021
    OpenUrlCrossRef
    1. Romanski LM
    (2007) Representation and integration of auditory and visual stimuli in the primate ventral lateral prefrontal cortex. Cereb Cortex 17:i61–i69. doi:10.1093/cercor/bhm099
    OpenUrlCrossRefPubMed
    1. Romanski LM,
    2. Bates JF,
    3. Goldman-Rakic PS
    (1999a) Auditory belt and parabelt projections to the prefrontal cortex in the rhesus monkey. J Comp Neurol 403:141–157. doi:10.1002/(SICI)1096-9861(19990111)403:2<141::AID-CNE1>3.0.CO;2-V
    OpenUrlCrossRefPubMed
    1. Romanski LM,
    2. Tian B,
    3. Fritz J,
    4. Mishkin M,
    5. Goldman-Rakic PS,
    6. Rauschecker JP
    (1999b) Dual streams of auditory afferents target multiple domains in the primate prefrontal cortex. Nat Neurosci 2:1131–1136. doi:10.1038/16056
    OpenUrlCrossRefPubMed
    1. Saur D
    , et al. (2008) Ventral and dorsal pathways for language. Proc Natl Acad Sci U S A 105:18035–18040. doi:10.1073/pnas.0805234105
    OpenUrlAbstract/FREE Full Text
  57. ↵
    1. Schomers MR,
    2. Garagnani M,
    3. Pulvermüller F
    (2017) Neurocomputational consequences of evolutionary connectivity changes in perisylvian language cortex. J Neurosci 37:3045–3055. doi:10.1523/JNEUROSCI.2693-16.2017
    OpenUrlAbstract/FREE Full Text
  58. ↵
    1. Scott LS,
    2. Monesson A
    (2009) The origin of biases in face perception. Psychol Sci 20:676–680. doi:10.1111/j.1467-9280.2009.02348.x
    OpenUrlCrossRefPubMed
  59. ↵
    1. Seger CA,
    2. Miller EK
    (2010) Category learning in the brain. Annu Rev Neurosci 33:203–219. doi:10.1146/annurev.neuro.051508.135546
    OpenUrlCrossRefPubMed
    1. Seltzer B,
    2. Pandya DN
    (1989) Intrinsic connections and architectonics of the superior temporal sulcus in the rhesus monkey. J Comp Neurol 290:451–471. doi:10.1002/cne.902900402
    OpenUrlCrossRefPubMed
  60. ↵
    1. Sloutsky VM,
    2. Fisher AV
    (2004) Induction and categorization in young children: a similarity-based model. J Exp Psychol Gen 133:166–188. doi:10.1037/0096-3445.133.2.166
    OpenUrlCrossRef
    1. Thiebaut de Schotten M,
    2. Dell’Acqua F,
    3. Valabregue R,
    4. Catani M
    (2012) Monkey to human comparative anatomy of the frontal lobe association tracts. Cortex 48:82–96. doi:10.1016/j.cortex.2011.10.001
    OpenUrlCrossRefPubMed
  61. ↵
    1. Tomasello R,
    2. Garagnani M,
    3. Wennekers T,
    4. Pulvermüller F
    (2017) Brain connections of words, perceptions and actions: a neurobiological model of spatio-temporal semantic activation in the human cortex. Neuropsychologia 98:111–129. doi:10.1016/j.neuropsychologia.2016.07.004
    OpenUrlCrossRefPubMed
  62. ↵
    1. Tomasello R,
    2. Garagnani M,
    3. Wennekers T,
    4. Pulvermüller F
    (2018) A neurobiologically constrained cortex model of semantic grounding with spiking neurons and brain-like connectivity. Front Comput Neurosci 12:88. doi:10.3389/fncom.2018.00088
    OpenUrlCrossRef
    1. Ungerleider LG,
    2. Gaffan D,
    3. Pelak VS
    (1989) Projections from inferior temporal cortex to prefrontal cortex via the uncinate fascicle in rhesus monkeys. Exp Brain Res 76:473–484. doi:10.1007/BF00248903
    OpenUrlCrossRefPubMed
  63. ↵
    1. Vanek N,
    2. Sóskuthy M,
    3. Majid A
    (2021) Consistent verbal labels promote odor category learning. Cognition 206:104485. doi:10.1016/j.cognition.2020.104485
    OpenUrlCrossRef
  64. ↵
    1. Virtanen P
    , et al. (2020) Scipy 1.0: fundamental algorithms for scientific computing in python. Nat Methods 17:261–272. doi:10.1038/s41592-019-0686-2
    OpenUrlCrossRefPubMed
    1. Wakana S,
    2. Jiang H,
    3. Nagae-Poetscher LM,
    4. van Zijl PCM,
    5. Mori S
    (2004) Fiber tract–based atlas of human white matter anatomy. Radiology 230:77–87. doi:10.1148/radiol.2301021640
    OpenUrlCrossRefPubMed
  65. ↵
    1. Waskom M
    (2021) Seaborn: statistical data visualization. J Open Source Softw 6:3021. doi:10.21105/joss.03021
    OpenUrlCrossRef
  66. ↵
    1. Waxman SR,
    2. Booth AE
    (2001) Seeing pink elephants: fourteen-month-olds’ interpretations of novel nouns and adjectives. Cogn Psychol 43:217–242. doi:10.1006/cogp.2001.0764
    OpenUrlCrossRefPubMed
  67. ↵
    1. Waxman SR,
    2. Markow DB
    (1995) Words as invitations to form categories: evidence from 12- to 13-month-old infants. Cogn Psychol 29:257–302. doi:10.1006/cogp.1995.1016
    OpenUrlCrossRefPubMed
    1. Webster MJ,
    2. Bachevalier J,
    3. Ungerleider LG
    (1994) Connections of inferior temporal areas TEO and TE with parietal and frontal cortex in macaque monkeys. Cereb Cortex 4:470–483. doi:10.1093/cercor/4.5.470
    OpenUrlCrossRefPubMed
  68. ↵
    1. Westermann G,
    2. Mareschal D
    (2014) From perceptual to language-mediated categorization. Philos Trans R Soc Lond B Biol Sci 369:20120391. doi:10.1098/rstb.2012.0391
    OpenUrlCrossRefPubMed
  69. ↵
    1. Whorf BL,
    2. Carroll JB
    (2007) Language, thought, and reality: selected writings. Cambridge, Mass: The MIT Press.
  70. ↵
    1. Wittgenstein L
    (1922) Tractatus logico-philosophicus. London: Routledge & Kegan Paul.
    1. Yeterian EH,
    2. Pandya DN,
    3. Tomaiuolo F,
    4. Petrides M
    (2012) The cortical connectivity of the prefrontal cortex in the monkey brain. Cortex 48:58–81. doi:10.1016/j.cortex.2011.03.004
    OpenUrlCrossRefPubMed
    1. Young MP,
    2. Scanneil JW,
    3. Burns GAPC,
    4. Blakemore C
    (1994) Analysis of connectivity: neural systems in the cerebral cortex. Rev Neurosci 5:227–250. doi:10.1515/REVNEURO.1994.5.3.227
    OpenUrlCrossRefPubMed
    1. Young MP,
    2. Scannell JW,
    3. Burns G
    (1995a) The analysis of cortical connectivity. New York; Austin: Springer; R.G. Landes.
    1. Young MP,
    2. Scannell JW,
    3. Burns G
    (1995b) The analysis of cortical connectivity, 1st ed. Berlin, Heidelberg: Springer.
  71. ↵
    1. Yuille AL,
    2. Geiger D
    (1995) Winner-Take-All mechanisms. In: Handbook of brain theory and neural networks (Arbib MA, eds), pp 1–1056: MIT Press.
Back to top

In this issue

The Journal of Neuroscience: 44 (9)
Journal of Neuroscience
Vol. 44, Issue 9
28 Feb 2024
  • Table of Contents
  • About the Cover
  • Index by author
  • Masthead (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Causal Influence of Linguistic Learning on Perceptual and Conceptual Processing: A Brain-Constrained Deep Neural Network Study of Proper Names and Category Terms
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Causal Influence of Linguistic Learning on Perceptual and Conceptual Processing: A Brain-Constrained Deep Neural Network Study of Proper Names and Category Terms
Phuc T. U. Nguyen, Malte R. Henningsen-Schomers, Friedemann Pulvermüller
Journal of Neuroscience 28 February 2024, 44 (9) e1048232023; DOI: 10.1523/JNEUROSCI.1048-23.2023

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Causal Influence of Linguistic Learning on Perceptual and Conceptual Processing: A Brain-Constrained Deep Neural Network Study of Proper Names and Category Terms
Phuc T. U. Nguyen, Malte R. Henningsen-Schomers, Friedemann Pulvermüller
Journal of Neuroscience 28 February 2024, 44 (9) e1048232023; DOI: 10.1523/JNEUROSCI.1048-23.2023
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Significance Statement
    • Introduction
    • Materials and Methods
    • Results
    • Discussion
    • Conclusion
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • category learning
  • concept formation
  • deep neural network
  • Hebbian associative learning
  • instance representation
  • verbal symbol learning

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Research Articles

  • NEOCORTICAL AND HIPPOCAMPAL THETA OSCILLATIONS TRACK AUDIOVISUAL INTEGRATION AND REPLAY OF SPEECH MEMORIES
  • Anchoring functional connectivity to individual sulcal morphology yields insights in a pediatric study of reasoning
  • Chemogenetic disruption of monkey perirhinal neurons projecting to rostromedial caudate impairs associative learning
Show more Research Articles

Behavioral/Cognitive

  • NEOCORTICAL AND HIPPOCAMPAL THETA OSCILLATIONS TRACK AUDIOVISUAL INTEGRATION AND REPLAY OF SPEECH MEMORIES
  • Anchoring functional connectivity to individual sulcal morphology yields insights in a pediatric study of reasoning
  • Electrophysiological Correlates of Lucid Dreaming: Sensor and Source Level Signatures
Show more Behavioral/Cognitive
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.