Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Symposium and Mini-Symposium

Recent Advances at the Interface of Neuroscience and Artificial Neural Networks

Yarden Cohen, Tatiana A. Engel, Christopher Langdon, Grace W. Lindsay, Torben Ott, Megan A. K. Peters, James M. Shine, Vincent Breton-Provencher and Srikanth Ramaswamy
Journal of Neuroscience 9 November 2022, 42 (45) 8514-8523; https://doi.org/10.1523/JNEUROSCI.1503-22.2022
Yarden Cohen
1Department of Brain Sciences, Weizmann Institute of Science, Rehovot, 76100, Israel
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Tatiana A. Engel
2Cold Spring Harbor Laboratory, Cold Spring Harbor, New York, NY 11724
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Tatiana A. Engel
Christopher Langdon
2Cold Spring Harbor Laboratory, Cold Spring Harbor, New York, NY 11724
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Grace W. Lindsay
3Department of Psychology, Center for Data Science, New York University, New York, NY 10003
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Torben Ott
4Bernstein Center for Computational Neuroscience Berlin, Institute of Biology, Humboldt University of Berlin, 10117, Berlin, Germany
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Megan A. K. Peters
5Department of Cognitive Sciences, University of California–Irvine, Irvine, CA 92697
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Megan A. K. Peters
James M. Shine
6Brain and Mind Centre, University of Sydney, Sydney, NSW 2006, Australia
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for James M. Shine
Vincent Breton-Provencher
7Département de psychiatrie et neurosciences, Université Laval, Quebec City, Québec, G1J 2G3, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Vincent Breton-Provencher
Srikanth Ramaswamy
8Biosciences Institute, Newcastle University, Newcastle upon Tyne, NE2 4HH, United Kingdom
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Srikanth Ramaswamy
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Biological neural networks adapt and learn in diverse behavioral contexts. Artificial neural networks (ANNs) have exploited biological properties to solve complex problems. However, despite their effectiveness for specific tasks, ANNs are yet to realize the flexibility and adaptability of biological cognition. This review highlights recent advances in computational and experimental research to advance our understanding of biological and artificial intelligence. In particular, we discuss critical mechanisms from the cellular, systems, and cognitive neuroscience fields that have contributed to refining the architecture and training algorithms of ANNs. Additionally, we discuss how recent work used ANNs to understand complex neuronal correlates of cognition and to process high throughput behavioral data.

  • plasticity
  • artificial neural networks
  • neuromodulators
  • behavior
  • vision
  • cognition

Introduction

Recent technological advances have transformed our access to the fine-grain spatiotemporal organization of the anatomy and physiology of biological neural networks. Over the years, big data on an astounding diversity of genes, proteins, neurons and glia, dendrites, synapses, and neural network functions have transformed our understanding of the brain (Sejnowski et al., 2014). On the other hand, brain-inspired implementations of artificial neural networks (ANNs), the perceptron model (McCulloch and Pitts, 1943; Rosenblatt, 1958), Boltzmann machines (Ackley et al., 1985), and Hopfield networks (Hopfield, 1982), have had profound implications for biological research and computational problems. These ANN architectures have been foundational for applications in pattern completion, attractor networks, dynamical systems, and diverse algorithmic capabilities in modern convolutional, multilayer, and recurrent neural network (RNN) models.

Although biological neural networks have continued to guide the development of their artificial counterparts, the beacon has been held by mathematics and statistical physics to develop efficient models of optimization functions (Sutskever et al., 2011; Cox and Dean, 2014). ANNs have leapfrogged from nonlinear systems and networks (Minsky and Papert, 1972; Haykin, 1994) to deep and recurrent networks (LeCun et al., 2015; Schmidhuber, 2015). More recently, backpropagation of error (Werbos, 1974, 1982; Rumelhart et al., 1986) has enabled the efficient training of neural networks, by computing gradients with respect to the weights of a multilayer network. Although methods to train ANNs have evolved to include improved weight initializations, optimization, and gradient descent algorithms, they do not appear to have any analogous neurobiological principles (Marblestone et al., 2016).

Here, we review the current state of the art of ANN models in terms of “biological realism,” their applications and limitations, with the ultimate aim of identifying the operational principles and functional settings through which biological neural networks and ANNs can inform each other toward synergistic development. With this background, the review focuses on four distinct aspects.

  1. How can biological intelligence guide the refinement of ANN architectures?

  2. How can ANNs drive a better understanding of cognition?

  3. What are the limitations of ANNs with respect to modeling human cognition?

  4. What are the recent advances in applying ANNs to quantify complex behavior?

Refining ANNs with biological precision

ANNs share many interesting features in common with biological neural networks. This is, of course, no accident, as the original ANN algorithms were in part inspired by the anatomy of the cerebral cortex (Sejnowski, 2020). The successful use of ANNs to model computations has forced a recalibration of our working models of the nervous system, leading to the embrace of dynamical models of computation that incorporate distributed computations across widespread, ever-changing networks (Sohn et al., 2019). Advances in cellular neuroscience, neuroimaging, and computational modeling are enabling the integration of new details into advanced versions of ANNs that will, hopefully, bring us closer to the goal of understanding how the brain works, while simultaneously refining artificial intelligence (AI).

A prime example of an emerging tension between neuroscience and AI is the recognition that pyramidal neurons, the workhorse of the cerebral cortex and the primary feature mimicked by ANNs, have highly nonlinear operating modes (Larkum, 2013). Traditional models of pyramidal neurons assumed that the dendrites of pyramidal neurons linearly summed their action potentials within a given window, and only spiked when the inputs exceeded a certain threshold (Larkum, 2022). In stark contrast, recent work has clearly demonstrated that many pyramidal neurons in the cerebral cortex have distinct modes of operation, sometimes firing linearly with inputs and other times ignoring inputs altogether (Fig. 1A,B) (Ramaswamy and Markram, 2015; Roelfsema and Holtmaat, 2018; Richards et al., 2019). Rather than reflecting passive integrative inputs, the active dendrites of pyramidal neurons have been shown to underpin striking computational complexity (Johnston and Narayanan, 2008; Spruston, 2008; Poirazi and Papoutsi, 2020; Larkum, 2022). Indeed, deep neural networks with at least 5-8 layers are needed to model the complex input/output functions of pyramidal cells (Beniaguev et al., 2021). The ability to convert the presumed integrator-like dynamics of neurons and their dendrites to coincident detectors (or resonators) is an important function that specific ion channels perform (Rudolph and Destexhe, 2003; Ratté et al., 2013). The identity of individual neurons (integrators vs resonators) has important implications for connectivity, computation, and information coding (Rudolph and Destexhe, 2003; Ratté et al., 2013). Such features have been recently incorporated into ANNs, toward solving the so-called credit-assignment problem (Payeur et al., 2021) using single-phase learning (Fig. 1C,D) (Greedy et al., 2022).

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Dendritic integration of inputs and neuromodulation-aware deep learning. A, How a pyramidal neuron responds to an input depends on dendritic location. Feedforward inputs located near the soma directly drive the firing rate of the neuron, whereas feedback inputs on apical dendrites affect burst firing (P(Burst)). B, The firing rate of presynaptic and postsynaptic neurons and P(Burst) control plasticity long-term potentiation - LTP; long-term depression - LTD. C, The dendritic integration of feedback and feedforward inputs by cortical neurons could solve the credit assignment problem in hierarchical ANNs. D, Diagram of how neuromodulation can be integrated by ANNs. Left, Error signal of a network perturbation is carried through a global neuromodulatory influence. Middle, Error signals are carried through node-specific neuromodulatory inputs. Right, Various neuromodulatory inputs could take part in signaling distinct error functions. A–C, Adapted from Payeur et al. (2021).

Nonetheless, plasticity, as a biological mechanism, is not limited to synaptic contacts. With emerging roles attributed to entire neurons (engram cells) in the physiology of memory, learning theories that focus exclusively on synaptic plasticity appear to be inadequate as a premise for models of ANNs (Titley et al., 2017; Lisman et al., 2018; Josselyn and Tonegawa, 2020). Plasticity is a ubiquitous phenomenon, which spans multiple scales of organization in biological neural networks: from synapses and dendritic branches to neurons and microcircuits (Le Bé and Markram, 2006; Branco and Häusser, 2010; Titley et al., 2017; Mishra and Narayanan, 2021). Incorporating a broad repertoire of plasticity mechanisms, such as those available to biological neural networks, is an essential step in refining ANN architectures and extending their utility. We are only just scratching the surface of the potential for biological insights to suggest novel algorithmic solutions to problems that have been trained on classical network architectures, such as RNNs.

The cerebral cortex is also deeply embedded within a web of dense interconnections with a number of highly conserved subcortical structures whose functional importance to the working of the nervous system should not be understated. One particular structure that is often overlooked in ANNs is the thalamus, a bilateral structure in the diencephalon that is densely (and heterogeneously) interconnected with the cerebral cortex (Jones, 2001). Although the functional benefits of one class of corticothalamic cell is relatively well understood (the so-called “core” regions) (Crandall et al., 2015), the more diffusely projecting “matrix” cells remain more enigmatic. Recently, a neural mass model of the corticothalamic system was created to investigate the impact of this topological projection on emergent whole-brain dynamics (Shine, 2021). In brief, the model found that the matrix cells tuned the functional repertoire of the brain, providing a flexible, yet robust, platform for instantiating an array of novel combinations of cortical coalitions. Others have shown that these same cells can alter information flow in neural circuits (Anastasiades et al., 2021; Mukherjee et al., 2021) and are crucial sites for behaviorally relevant plasticity (Williams and Holtmaat, 2019). It would be interesting to note how these circuit-level features could inform future implementations of ANNs, such as models that mimic the interactions between the cerebellum and cortex (Pemberton et al., 2021; Boven et al., 2022).

The operating mode of the cerebral cortex (along with the rest of the brain) is also fundamentally altered by the presence (or absence) of neuromodulatory chemicals, such as noradrenaline, acetylcholine, and serotonin. By interacting with GPCRs on target neurons and glia, these ligands can alter the excitability and receptivity of the network (Shine et al., 2021), facilitating different information processing regimens that shift neural populations between information storage and information transfer (Li et al., 2019). These changes in gain, while relatively low-dimensional, can substantially impact the functional outputs of ANNs (Stroud et al., 2018), suggesting that their incorporation into modern deep learning architectures could be quite informative (Mei et al., 2022). In addition, by combining these approaches with sophisticated, high-resolution recordings of the neuromodulatory system in vivo (Breton-Provencher et al., 2022), we can also simultaneously test hypotheses regarding the functional operation of the brain as well.

A prominent example of biologically inspired ANNs that has gained considerable interest in machine vision is the convolutional neural network (CNN). CNNs are extensions of ANNs with an architecture inspired by that of the mammalian visual system, with convolutions representing the function of simple cells and the pooling operations of complex cells (Lindsay, 2021). When trained appropriately, these models can produce representations that match those of biological visual systems better than previous models (Khaligh-Razavi and Kriegeskorte, 2014; Yamins et al., 2014). Traditionally CNNs are strictly feedforward; that is, they do not include lateral or feedback recurrent connections (Fig. 2). Yet, it is known that visual systems of humans contain many such connections, and these connections are implicated in important computations, such as object recognition (Wyatte et al., 2012). Previous work has shown how these connections can make models better at visual tasks and better match biological processing (Fig. 2) (Spoerer et al., 2017; Linsley et al., 2018; Kubilius et al., 2019; Nayebi et al., 2021). An unmet potential of these models, however, is to use them as an idealized experimental setup to analyze the computational role that recurrence plays. Promising work in this direction has shown that recurrence can help object classification by carrying information about unrelated, auxiliary variables (Thorat et al., 2021).

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Feedforward versus bio-inspired CNNs. By adding connections inspired by the anatomy and physiology of the visual system, such as lateral (e.g., center surround suppression) or feedback (e.g., top-down predictions), CNNs with recurrent connections show improved accuracy. Black and red arrows represent feedforward and recurrent connections, respectively. Adapted from Lindsay et al. (2022).

In a recent study (Lindsay et al., 2022), four different kinds of recurrence were added to a feedforward CNN: feedback connections that implement predictive processing to one network, lateral connections that implement surround suppression to another, and two more networks with feedback and lateral connections trained directly to classify degraded images. This choice of task, wherein the network must classify images of digits that are degraded by one of several types of noise, such as occlusion and blur, was chosen to capture some of the functions believed to be performed by recurrence. Counterintuitively, recurrence added to the CNN was not related to its function in these models: both forms of task-trained recurrence (feedback and lateral connections) change neural activity and behavior similarly to each other and differently from their bio-inspired anatomic counterparts. Specifically, in the case of feedback, predictive feedback denoises the representation of noisy images at the first layer of the network, leading to an expected increase in classification performance. In the task-trained networks, representations are not denoised over time at the first layer (indeed, they become “noisier”), yet these dynamics do lead to denoising at later layers and increased performance. We analyzed an open fMRI dataset (Abdelhack and Kamitani, 2018) using the same tools, such as dimensionality reduction, activity correlations, and representational geometry analysis, applied to the models and found weak support for the predictive feedback model. Such analysis of artificial networks provides an opportunity to test the tools of systems neuroscience (Lindsay, 2022).

Decisions, artificial RNNs, and functional neuron types

Many ingredients make up our decisions, a rich stream of sensory information, a lifetime of memories, long-term goals, and current mood or emotions. This poses a challenge in identifying the neural processes of decision formation: The activity of cortical neurons, for instance, reflects an equally large complexity of decision-related features, from sensory and spatial information (Rao et al., 1997), to short-term memory (Funahashi et al., 1989), economic value (Padoa-Schioppa and Assad, 2006), risk (Ogawa et al., 2013) and confidence (Kepecs et al., 2008), or abstract rules (Wallis et al., 2001). Furthermore, single neurons often demonstrate mixtures of these features (Mante et al., 2013; Rigotti et al., 2013; Fusi et al., 2016), precluding straightforward functional interpretations of the signal they carry (Fig. 3). How can we identify any organizational principles by which cortical neurons or neural networks take part in decision-making?

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

Using RNNs to study neuronal correlates of complex tasks. In an example task where various contexts (u1) and sensory cues (u2) guide task outputs (z) or decisions, recordings of neurons from associative brain areas (e.g., pre-frontal cortex - PFC) show multidimensional encoding of task variables by individual neurons. Representing population dynamics in neural state space where each point in space represents a unique pattern of neuronal activity that is useful to dissect how the correlated activity of a large number of neurons represents task variables. To model physiological dynamics, RNNs are trained to perform a similar task. Key features of physiological dynamics of neuronal populations are reproduced by RNNs. Complex perturbation studies can thus be performed with these trained RNNs to test causality. In a recent study, Langdon and Engel (2022) found that low-dimensional latent circuits can be extracted from high-dimensional RNN dynamics and used to perform patterned connectivity perturbations. Adapted from Mante et al. (2013) and Langdon and Engel (2022).

Recent approaches have focused on neural population as the primary computational units for cognition (Pandarinath et al., 2018; Saxena and Cunningham, 2019; Vyas et al., 2020; Barack and Krakauer, 2021; Duncker and Sahani, 2021; Ebitz and Hayden, 2021; Jazayeri and Ostojic, 2021). Population approaches identify low-dimensional patterns in neural population data, describing the subspaces or manifolds in which neural trajectories move (Cunningham and Yu, 2014). Applied to neural population recordings during flexible decision-making as a prime example, information about sensory information, choice, and rules can be reliably separated at the level of neural populations (Mante et al., 2013; Malagon-Vina et al., 2018; Sohn et al., 2019; Aoi et al., 2020; Ebitz and Hayden, 2021).

RNNs, which are able to mimic the complexity of real cortical responses, have served as a valuable model for understanding computation in large heterogeneous neural populations. For instance, RNNs suggest dynamic network mechanisms by which decision rules are flexibly applied to determine a decision (Mante et al., 2013). Hopfield networks and restricted Boltzmann machines provide valuable insight into the storage and retrieval of associative memories via unsupervised learning rules (Marullo and Agliari, 2020). Recently, supervised learning approaches have been used to train RNNs to perform the same cognitive tasks as behaving animals. This approach provides a powerful alternative for studying how neural computations underlying cognitive tasks are distributed across heterogeneous populations (Fig. 3) (Mante et al., 2013; Song et al., 2016; Wang et al., 2018; Yang et al., 2019) and how networks leverage previously learned tasks for continual learning (Driscoll et al., 2022). Because of the complexity of their connectivity and dynamics, reverse engineering trained RNNs mimic the challenges faced when analyzing real neural data. This observation has motivated their use as a testbed for candidate dimensionality reduction methods aimed at uncovering low-dimensional latent dynamics. Such methods model heterogeneous neural responses as linear mixing of task-relevant variables and can uncover neural mechanisms which exist only at the population level (Cunningham and Yu, 2014; Kobak et al., 2016). The ability to perform precise perturbation tests in RNNs (Yang et al., 2019) offers the possibility of validating the causal role of neural representations revealed by candidate dimensionality reduction strategies (Mante et al., 2013; Song et al., 2016; Wang et al., 2018; Yang et al., 2019).

In a recent study, RNNs were trained on cognitive tasks to develop and validate latent circuit models of heterogeneous neural responses (Langdon and Engel, 2022) (Fig. 3). The latent circuit model uncovers low-dimensional task-relevant representations together with recurrent circuit interactions between these representations. To validate this method, RNNs were trained on a motion-color discrimination task in which the subject must flexibly discriminate either the motion or color of random dot stimulus depending on a contextual cue (Langdon and Engel, 2022) (Fig. 3). Fitting a latent circuit model to the responses of this RNN revealed a latent inhibitory mechanism in which contextual representations inhibit irrelevant stimulus representations, allowing the network to flexibly select the correct stimulus–response association (Langdon and Engel, 2022). This inhibitory mechanism is mirrored in dynamics as a suppression of irrelevant stimulus representations.

Despite the success of population-centered analysis, recent studies have discovered groups of single neurons with prototypical dynamic activity and encoding of decision variables. For instance, Hirokawa et al. (2019) started from neural population activity but considered the possibility that the neurons' dynamic activity as well as its tuning to decision variables in a combined sensory and value-based decision task clustered into distinct groups of neurons with distinct dynamic and tuning profiles. Unsupervised clustering revealed dedicated groups of single neurons in rat orbitofrontal cortex that were tuned to canonical decision variables, that is, combinations of task features that explained the animals' decision behavior, such as reward likelihood, integrated value, and choice outcome. A dedicated group of neurons in the orbitofrontal cortex carried information about the certainty that a decision was correct (i.e., decision confidence) (Masset et al., 2020). These neurons predicted subsequent confidence-guided behavior: the variable time rats invested into their decision to obtain an uncertain, delayed reward before abandoning their investment (Lak et al., 2014; Ott et al., 2018). These groups of neurons might constitute functional neuron types, characterized by assuming specific algorithmic roles to realize decision computations (Christensen et al., 2022). Similarly, functional clusters were found in the orbitofrontal cortex during value-based decision tasks in rats (Hocker et al., 2021) and primates (Onken et al., 2019), and in mice using calcium imaging during associative learning (Namboodiri et al., 2019).

When and why might we expect to find functional neuron types? Recent computational studies using RNNs suggest that neural subpopulations with distinct dynamics or categorical representations arise in trained networks that are required for flexible decision-making, such as context-dependent decision tasks (Dubreuil et al., 2022; Flesch et al., 2022; Langdon and Engel, 2022). Functional neuron types might thus be a feature shared by biological neural networks and ANNs to provide a robust computational solution for flexible decision-making. On the other hand, these interpretations are limited, since it is unclear what the biological counterpart to an ANN unit might precisely be. While many approaches interpret RNN units as candidates for single neurons (Barrett et al., 2019), the complex computations performed by single neurons outperform simple RNN units and can only be described by deep networks themselves (Beniaguev et al., 2021). Specifically, the functional coupling between neuronal compartments (dendrites and soma, compare with previous section) can be controlled by thalamic input (Aru et al., 2020), and depends on learning (d'Aquin et al., 2022) further suggesting that RNN units might correspond to computations of neuronal compartments or biophysical processes. Categorical representations in RNNs might thus shed light onto the functions performed by biophysical elements of neurons.

Functional neuron types might emerge as a result of the cortical microcircuit structure. Emerging evidence suggests that cortical cell types, defined by distinct gene expression or connectivity patterns (Tasic et al., 2018; Winnubst et al., 2019), assume specialized functions during decision-making. For example, orbitofrontal cortex neurons that project to the striatum predominantly carry sustained task-related signals (Bari et al., 2019; Terra et al., 2020), such as information about choice outcome (Hirokawa et al., 2019) (whether the animal was rewarded or not), and projection-defined neurons in motor cortex signal movement onset or choice signals, respectively (Economo et al., 2018). Cell type identity might thus be a structural constraint on the dynamic decision algorithms in biological neural networks that could inform the design of ANNs (Sacramento et al., 2018; Greedy et al., 2022).

Uncertainty and decisions: can insights from human (and animal) cognition contribute to AI development?

Humans make decisions based on perceptual, value-based, or other information. Such decisions are accompanied by a sense of confidence. That is, our brains seem to compute not only the best decisional outcome, but also estimates related to the probability that the decision is correct (Pouget et al., 2016; Mamassian, 2022; Peters, 2022). This sense of uncertainty accompanies the moment-to-moment information processing across many perceptual and cognitive domains, and can help any organism decide whether to update their internal models of the world, how to allocate resources, or how to sample new information. Importantly, computing decisional confidence can also help us to better learn from erroneous predictions (Guggenmos et al., 2016; Stolyarova et al., 2019; Ptasczynski et al., 2022).

An argument can thus be made that getting artificial systems to also compute such a confidence judgment could lead not only to better decision-making under uncertainty, but also better and more self-directed learning. A foundational goal of AI research is to build systems that not only behave adaptively in the world, but which “know” when they have made correct or erroneous decisions, or when they have such a high level of uncertainty that they should sample more information before committing to a decision at all. Thus far, most “confidence” type signals in artificial systems typically compute uncertainty estimates according to probabilistic inference: for example, the variance of a (posterior) probability distribution, or entropy of an outcome distribution, can potentially be reasonable proxies for confidence in biological systems (Li and Ma, 2020). This is because these quantities reflect the relative evidence in favor of multiple possible decisional outcomes. However, there are a number of problems with these approaches for uncertainty estimation in artificial systems.

First, it is not clear that humans and other animals rate confidence according to optimal inference, as implemented in ANNs or similar; instead, a large body of work suggests that other influences on decisional confidence are likely, ranging from motor preparation and execution (Fleming et al., 2015) to detectability heuristics (Maniscalco et al., 2016, 2021; Rausch et al., 2018). While these contributions to uncertainty/confidence estimates may seem suboptimal or even random, we also must note that our systems have been optimized through millennia of evolution, such that apparent “biases” in confidence judgments may actually reflect some optimal behavior where the cost function remains unknown to us as researchers (Michel and Peters, 2021).

Second, current implementations of confidence in multidecision alternatives often do not have an option for artificial systems to “opt out” of the decision, and instead decide to sample more information. Current AI does not have agency in such a fashion. However, we know that biological observers use confidence to guide their decision-making behavior, including decisions about whether and how to continue sampling their environments (Kepecs and Mainen, 2012; Guggenmos et al., 2016; Stolyarova et al., 2019; Ptasczynski et al., 2022). One area in which uncertainty-based self-directed information sampling is likely to be of utility is in meta-learning, wherein an artificial system must learn which weights to update based on an inferred context (and thus may avoid catastrophic forgetting when trained on multiple tasks). Several architectures implementing explicit metacognition or confidence have been proposed (Griffiths et al., 2019). For example, in rats trained to report confidence by placing a wager on difficult decisions, single neurons in the frontal cortex encode uncertainty information across multiple sensory modalities, and predict both confidence-scaled time investment in learning (Lak et al., 2014; Masset et al., 2020). These results suggest a generalized representation of confidence as a “summary scalar” that could provide a robust uncertainty signal used for subsequent decisions or learning processes and therefore constitute a precursor of metacognitive signals. The field is ripe for more exploration whether such biological implementations could inform uncertainty predictions in ANNs (Griffiths et al., 2019; Gawlikowski et al., 2021).

Here, we have discussed one among many examples of how artificial system development may benefit from the study of metacognition and confidence in biological systems, and vice versa. Future work may also examine how cooperation between humans and artificial systems may be optimized through confidence-weighted communication, as it is between dyads or small groups of human deciders (Bahrami et al., 2010). Efforts to align vocabularies, literatures, and concepts across the fields of cognitive science and AI will assuredly benefit both fields.

Development of deep learning algorithms for high throughput processing of complex behavior

Studying natural behaviors affords new understanding of how the brain controls motion (Krakauer et al., 2017) and processes sensory inputs (Hayhoe, 2018). But two major characteristics of natural behaviors challenge their use in neuroscience experiments: dynamic properties that are often difficult to quantify and rich repertoires that require processing datasets much larger than tractable with traditional manual or semimanual methods. Modern machine learning offers both unsupervised and supervised approaches to meet these challenges.

Unsupervised algorithms help researchers identify structure in high-dimensional behavior data (Gilpin et al., 2020; McCullough and Goodhill, 2021). For example, researchers applied unsupervised dimensionality reduction, linear projections, or nonlinear embeddings (Tenenbaum et al., 2000; van der Maaten and Hinton, 2008; McInnes et al., 2018) to videos of freely moving worms (Stephens et al., 2008), fish (Mearns et al., 2020), flies (Berman et al., 2014), rodents (Stringer et al., 2019), and primates (Yan et al., 2020). Then, unsupervised clustering could discover robust behavior states in the resulting low-dimensional representations. Additionally, clustering was replaced by hidden Markov models (HMMs), capturing sequences of behavior states (Wiltschko et al., 2015).

Successful approaches using ANNs to process textual, auditory, and visual data, not as models in neuroscience, were recently harnessed and applied to quantify complex behavior in tandem or replacing these machine learning methods. Unsupervised ANNs were used because they better capture various data distributions (Graving and Couzin, 2020); convolutional and variational autoencoders (Batty et al., 2019; Graving and Couzin, 2020; Luxem et al., 2022) performed dimensionality reduction before clustering or fitting HMMs and generative adversarial networks (Goodfellow et al., 2014; Radford et al., 2015) improved the interpolation between low-dimensional representations (Sainburg et al., 2018).

Supervised ANNs are extremely useful in automating manual processing where humans can identify which data features to track and provide labeled training examples. Many striking examples come from pose estimation in movies of freely behaving animals. These algorithms learn to track human-defined anatomic features, such as joints and locations on the body of single or multiple animals, in videos captured from single (Mathis et al., 2018; Graving et al., 2019; Pereira et al., 2019) or multiple (Marshall et al., 2021) cameras. Mostly using deep convolutional ANNs, these supervised methods, like the unsupervised ones, extended and improved previous methods based on supervised classifier models (Dankert et al., 2009; Kabra et al., 2013; Machado et al., 2015).

Together, these ANN-based algorithms ushered the field of computational neuroethology (Anderson and Perona, 2014; Datta et al., 2019; Pereira et al., 2020). However, while many species naturally vocalize and offer a rich window onto complex social interactions, fewer works developed audio analysis methods comparable with those created for video analysis (Sainburg and Gentner, 2021).

Audio analyses predominantly start by converting sound signals to spectrograms, a two-dimensional representation in the time and frequency domains. This “image of sound,” like visual data, was used to extract low-dimensional representations. For example, human-defined features, such as pitch, entropy, and amplitude, were continuously extracted from spectrograms and automated measuring similarities between juvenile and tutor zebra finch songs (Tchernichovski et al., 2000). Unsupervised variational autoencoders were also used for continuous low-dimensional embedding of spectrograms (Goffinet et al., 2021). Still, rather than working on continuous signals, most machine learning tools for bioacoustics were developed for analyzing audio segments, thought to represent basic vocal units or syllables.

Segmenting vocal communication allows creating models of syntax (Berwick et al., 2011; Jin and Kozhevnikov, 2011; Markowitz et al., 2013; Hedley, 2016) and motor learning (Sober and Brainard, 2009, 2012), and to relate syllable acoustics and sequence to neural activity (Leonardo and Fee, 2005; Sober et al., 2008; Wohlgemuth et al., 2010). Researchers used unsupervised similarity metrics (Mets and Brainard, 2018), clustering (Daou et al., 2012; Burkett et al., 2015), embedding (Morfi et al., 2021; Sainburg et al., 2021), variational autoencoders (Kohlsdorf et al., 2020), and other generative deep networks (Pagliarini et al., 2021) to assist human identification of vocal units, visualize repertoire structures (Sainburg et al., 2020), and study their dynamics (Mets and Brainard, 2018; Kollmorgen et al., 2020). When human annotators created training sets of labeled audio segments, those segments were used to train supervised algorithms (Nicholson, 2016), support vector machines (Tachibana et al., 2014), template matching (Anderson et al., 1996), HMMs (Kogan and Margoliash, 1998), and k-nearest neighbors (Nicholson, 2016, 2021), that allowed scaling up analyses on annotated syllables.

Still, these methods require the audio to be a priori well segmented. Traditional segmentation techniques hence create a bottleneck, limiting the questions researchers can answer. Using supervised deep ANNs introduces various solutions to this problem in rodents (Coffey et al., 2019) and songbirds (Koumura and Okanoya, 2016; Steinfath et al., 2021; Cohen et al., 2022). For example, TweetyNet (Cohen et al., 2022) is a supervised deep ANN that leverages temporal vocal dependencies to achieve high-precision annotation of multiple species. TweetyNet offers a powerful tool to study the neuronal encoding of bird song syntax (Cohen et al., 2020) and demonstrates how development in modern machine learning opens new boundaries in the study of natural behavior.

Finally, as different research laboratories, developing and using various ANNs to analyze behavior, also develop their own data formats and algorithms, it is of uttermost importance for our community to develop and foster an ecosystem of interoperable methods to increase reproducibility and access.

In conclusion, we have briefly reviewed the state of the art of ANNs and how their development has been inspired by biological neural networks. Although ANNs are remarkably effective at solving specific tasks, they lack the ability of biological neural networks to generalize robustly across tasks (but see Reed et al., 2022). We suggest that future implementations of ANNs should incorporate some of the intricate multiscale organizing features of biological neural networks to generalize as well as they do and learn continually over a lifetime of experience.

Neuromodulatory systems endow biological neural networks with the ability to learn and adapt to constantly changing behavioral demands. Neuromodulators, such as dopamine, serotonin, noradrenaline, and acetylcholine, play crucial roles in modulating a repertoire of brain states from reward assessment, motivation, patience, arousal, and attention. The diverse phenomenology of neuromodulatory function is yet to be fully explored in ANNs, and their implementation has so far been mostly restricted to models of reinforcement learning (Shine et al., 2021; Mei et al., 2022).

Neurotransmitter and neuromodulator receptors are thought to modulate perceptual processes by activating receptor “hotspots” on the distal apical dendrites of neocortical layer 5 pyramidal cells (Takahashi et al., 2016). Recent implementations of ANNs have incorporated dendritic mechanisms to address the credit-assignment problem (Sacramento et al., 2018). Incorporating neurotransmitter receptor clusters on dendrites in ANNs could help unravel their role in gating perceptual processes, for example, NMDA receptors that are distributed nonlinearly on the dendrites of most neuron types (Chavlis and Poirazi, 2021).

Biological neural networks promote renormalization and homeostasis of synaptic strength during different states of sleep, and facilitate learning and memory through replay. Replay enables the brain to consolidate memory and overcome forgetting of acquired knowledge, also referred to as “catastrophic forgetting” in machine learning. Implementing “sleep-like states” in deep neural networks could mimic biological replay mechanisms and prevent catastrophic forgetting (Roscow et al., 2021; Kudithipudi et al., 2022; Mei et al., 2022; Tsuda et al., 2022).

Recent findings demonstrate that metabolic state dynamically governs coding precision in biological neural networks. Metabolic scarcity in the brain inactivates biological neural networks required for long-term memory to preserve energy (Padamsey et al., 2022). Biological neural networks regulate energy use through intrinsic mechanisms that determine the degree of energy consumption by reducing the impact of subthreshold variability on information coding. Therefore, biological neural networks dynamically adapt their coding precision and energy expenditure in a context-dependent manner (Padamsey et al., 2022). Most ANN architectures are energetically expensive. How could metabolic principles controlling coding precision inform implementations of energy efficient ANNs?

Neuroscience is witnessing significant advances in our understanding of biological learning mechanisms that can continue to inform new avenues for ANNs. We suggest that the machine learning community could adopt these ideas and integrate them into standard ANN frameworks to build a solid foundation, and develop the next generation of ANNs informed by the multiscale organizing features of their biological analogs.

Footnotes

  • This work was supported by Brain and Behavior Research Foundation National Alliance for Research on Schizophrenia and Depression Young Investigator Award to V.B.-P.; Natural Sciences and Engineering Research Council Discovery Grants DGECR-2021-00293 and RGPIN-2021-03284 to V.B.-P.; Canadian Institute for Advanced Research Azrieli Global Scholar Fellowship to M.A.K.P.; Fonds de Recherche du Québec - Santé salary award 311492 to V.B.-P.; Marie Skłodowska-Curie Global Fellowship Agreement 842492 to S.R.; Newcastle University Academic Track Fellowship to S.R.; Fulbright Research Scholarship to S.R.; German Research Foundation OT562/1-1 and OT562/2-1 to T.O.; Marie Curie Individual Fellowship 844003 to G.W.L.; National Health and Medical Research Council Fellowship GNT1193857 to J.M.S.; Swartz Foundation to C.L.; National Institutes of Health Grants RF1DA055666 to T.A.E. and S10OD028632-01 to T.A.E. and C.L.; and Alfred P. Sloan Foundation Research Fellowship to T.A.E.

  • The authors declare no competing financial interests.

  • Correspondence should be addressed to Vincent Breton-Provencher at vincent.breton-provencher{at}cervo.ulaval.ca or Srikanth Ramaswamy at Srikanth.Ramaswamy{at}newcastle.ac.uk

SfN exclusive license.

References

  1. ↵
    1. Abdelhack M,
    2. Kamitani Y
    (2018) Sharpening of hierarchical visual feature representations of blurred images. eNeuro 5:ENEURO.0443-17.2018. doi:10.1523/ENEURO.0443-17.2018
    OpenUrlAbstract/FREE Full Text
  2. ↵
    1. Ackley DH,
    2. Hinton GE,
    3. Sejnowski TJ
    (1985) A learning algorithm for Boltzmann machines. Cogn Sci 9:147–169. doi:10.1207/s15516709cog0901_7
    OpenUrlCrossRef
  3. ↵
    1. Anastasiades PG,
    2. Collins DP,
    3. Carter AG
    (2021) Mediodorsal and ventromedial thalamus engage distinct L1 circuits in the prefrontal cortex. Neuron 109:314–330.e5. doi:10.1016/j.neuron.2020.10.031 pmid:33188733
    OpenUrlCrossRefPubMed
  4. ↵
    1. Anderson DJ,
    2. Perona P
    (2014) Toward a science of computational ethology. Neuron 84:18–31. doi:10.1016/j.neuron.2014.09.005 pmid:25277452
    OpenUrlCrossRefPubMed
  5. ↵
    1. Anderson SE,
    2. Dave AS,
    3. Margoliash D
    (1996) Template-based automatic recognition of birdsong syllables from continuous recordings. J Acoust Soc Am 100:1209–1219. doi:10.1121/1.415968 pmid:8759970
    OpenUrlCrossRefPubMed
  6. ↵
    1. Aoi MC,
    2. Mante V,
    3. Pillow JW
    (2020) Prefrontal cortex exhibits multidimensional dynamic encoding during decision-making. Nat Neurosci 23:1410–1420. doi:10.1038/s41593-020-0696-5 pmid:33020653
    OpenUrlCrossRefPubMed
  7. ↵
    1. Aru J,
    2. Suzuki M,
    3. Larkum ME
    (2020) Cellular mechanisms of conscious processing. Trends Cogn Sci 24:814–825. doi:10.1016/j.tics.2020.07.006 pmid:32855048
    OpenUrlCrossRefPubMed
  8. ↵
    1. Bahrami B,
    2. Olsen K,
    3. Latham PE,
    4. Roepstorff A,
    5. Rees G,
    6. Frith CD
    (2010) Optimally interacting minds. Science 329:1081–1085. doi:10.1126/science.1185718 pmid:20798320
    OpenUrlAbstract/FREE Full Text
  9. ↵
    1. Barack DL,
    2. Krakauer JW
    (2021) Two views on the cognitive brain. Nat Rev Neurosci 22:359–371. doi:10.1038/s41583-021-00448-6 pmid:33859408
    OpenUrlCrossRefPubMed
  10. ↵
    1. Bari BA,
    2. Grossman CD,
    3. Lubin EE,
    4. Rajagopalan AE,
    5. Cressy JI,
    6. Cohen JY
    (2019) Stable representations of decision variables for flexible behavior. Neuron 103:922–933.e7. doi:10.1016/j.neuron.2019.06.001 pmid:31280924
    OpenUrlCrossRefPubMed
  11. ↵
    1. Barrett DG,
    2. Morcos AS,
    3. Macke JH
    (2019) Analyzing biological and artificial neural networks: challenges with opportunities for synergy? Curr Opin Neurobiol 55:55–64. doi:10.1016/j.conb.2019.01.007 pmid:30785004
    OpenUrlCrossRefPubMed
  12. ↵
    1. Batty E,
    2. Whiteway M,
    3. Saxena S,
    4. Biderman D,
    5. Abe T,
    6. Musall S,
    7. Gillis W,
    8. Markowitz J,
    9. Churchland A,
    10. Cunningham JP,
    11. Datta SR,
    12. Linderman S,
    13. Paninski L
    (2019) BehaveNet: nonlinear embedding and Bayesian neural decoding of behavioral videos. Advances in Neural Information Processing Systems 32.
  13. ↵
    1. Beniaguev D,
    2. Segev I,
    3. London M
    (2021) Single cortical neurons as deep artificial neural networks. Neuron 109:2727–2739.e3. doi:10.1016/j.neuron.2021.07.002 pmid:34380016
    OpenUrlCrossRefPubMed
  14. ↵
    1. Berman GJ,
    2. Choi DM,
    3. Bialek W,
    4. Shaevitz JW
    (2014) Mapping the stereotyped behaviour of freely moving fruit flies. J R Soc Interface 11:20140672. doi:10.1098/rsif.2014.0672
    OpenUrlCrossRefPubMed
  15. ↵
    1. Berwick RC,
    2. Okanoya K,
    3. Beckers GJ,
    4. Bolhuis JJ
    (2011) Songs to syntax: the linguistics of birdsong. Trends Cogn Sci 15:113–121. doi:10.1016/j.tics.2011.01.002 pmid:21296608
    OpenUrlCrossRefPubMed
  16. ↵
    1. Boven E,
    2. Pemberton J,
    3. Chadderton P,
    4. Apps R,
    5. Costa RP
    (2022) Cerebro-cerebellar networks facilitate learning through feedback decoupling. bioRxiv 2022.01.28.477827.
  17. ↵
    1. Branco T,
    2. Häusser M
    (2010) The single dendritic branch as a fundamental functional unit in the nervous system. Curr Opin Neurobiol 20:494–502. doi:10.1016/j.conb.2010.07.009 pmid:20800473
    OpenUrlCrossRefPubMed
  18. ↵
    1. Breton-Provencher V,
    2. Drummond GT,
    3. Feng J,
    4. Li Y,
    5. Sur M
    (2022) Spatiotemporal dynamics of noradrenaline during learned behaviour. Nature 606:732–738. doi:10.1038/s41586-022-04782-2 pmid:35650441
    OpenUrlCrossRefPubMed
  19. ↵
    1. Burkett ZD,
    2. Day NF,
    3. Peñagarikano O,
    4. Geschwind DH,
    5. White SA
    (2015) VoICE: a semi-automated pipeline for standardizing vocal analysis across models. Sci Rep 5:10237. doi:10.1038/srep10237 pmid:26018425
    OpenUrlCrossRefPubMed
  20. ↵
    1. Chavlis S,
    2. Poirazi P
    (2021) Drawing inspiration from biological dendrites to empower artificial neural networks. Curr Opin Neurobiol 70:1–10. doi:10.1016/j.conb.2021.04.007 pmid:34087540
    OpenUrlCrossRefPubMed
  21. ↵
    1. Christensen AJ,
    2. Ott T,
    3. Kepecs A
    (2022) Cognition and the single neuron: How cell types construct the dynamic computations of frontal cortex. Curr Opin Neurobiol 77:102630.
    OpenUrl
  22. ↵
    1. Coffey KR,
    2. Marx RG,
    3. Neumaier JF
    (2019) DeepSqueak: a deep learning-based system for detection and analysis of ultrasonic vocalizations. Neuropsychopharmacology 44:859–868. doi:10.1038/s41386-018-0303-6 pmid:30610191
    OpenUrlCrossRefPubMed
  23. ↵
    1. Cohen Y,
    2. Nicholson DA,
    3. Sanchioni A,
    4. Mallaber EK,
    5. Skidanova V,
    6. Gardner TJ
    (2022) Automated annotation of birdsong with a neural network that segments spectrograms. Elife 11:e63853. doi:10.7554/Elife.63853
    OpenUrlCrossRef
  24. ↵
    1. Cohen Y,
    2. Shen J,
    3. Semu D,
    4. Leman DP,
    5. Liberti WA,
    6. Perkins LN,
    7. Liberti DC,
    8. Kotton DN,
    9. Gardner TJ
    (2020) Hidden neural states underlie canary song syntax. Nature 582:539–544. doi:10.1038/s41586-020-2397-3 pmid:32555461
    OpenUrlCrossRefPubMed
  25. ↵
    1. Cox DD,
    2. Dean T
    (2014) Neural networks and neuroscience-inspired computer vision. Curr Biol 24:R921–R929. doi:10.1016/j.cub.2014.08.026 pmid:25247371
    OpenUrlCrossRefPubMed
  26. ↵
    1. Crandall SR,
    2. Cruikshank SJ,
    3. Connors BW
    (2015) A corticothalamic switch: controlling the thalamus with dynamic synapses. Neuron 86:768–782. doi:10.1016/j.neuron.2015.03.040 pmid:25913856
    OpenUrlCrossRefPubMed
  27. ↵
    1. Cunningham JP,
    2. Yu BM
    (2014) Dimensionality reduction for large-scale neural recordings. Nat Neurosci 17:1500–1509. doi:10.1038/nn.3776 pmid:25151264
    OpenUrlCrossRefPubMed
  28. ↵
    1. d'Aquin S,
    2. Szonyi A,
    3. Mahn M,
    4. Krabbe S,
    5. Gründemann J,
    6. Lüthi A
    (2022) Compartmentalized dendritic plasticity during associative learning. Science 376:eabf7052. doi:10.1126/science.abf7052 pmid:35420958
    OpenUrlCrossRefPubMed
  29. ↵
    1. Dankert H,
    2. Wang L,
    3. Hoopfer ED,
    4. Anderson DJ,
    5. Perona P
    (2009) Automated monitoring and analysis of social behavior in Drosophila. Nat Methods 6:297–303. doi:10.1038/nmeth.1310 pmid:19270697
    OpenUrlCrossRefPubMed
  30. ↵
    1. Daou A,
    2. Johnson F,
    3. Wu W,
    4. Bertram R
    (2012) A computational tool for automated large-scale analysis and measurement of bird-song syntax. J Neurosci Methods 210:147–160. doi:10.1016/j.jneumeth.2012.07.020 pmid:22890237
    OpenUrlCrossRefPubMed
  31. ↵
    1. Datta SR,
    2. Anderson DJ,
    3. Branson K,
    4. Perona P,
    5. Leifer A
    (2019) Computational neuroethology: a call to action. Neuron 104:11–24. doi:10.1016/j.neuron.2019.09.038 pmid:31600508
    OpenUrlCrossRefPubMed
  32. ↵
    1. Driscoll L,
    2. Shenoy K,
    3. Sussillo D
    (2022) Flexible multitask computation in recurrent networks utilizes shared dynamical motifs. bioRxiv 2022.08.15.503870.
  33. ↵
    1. Dubreuil A,
    2. Valente A,
    3. Beiran M,
    4. Mastrogiuseppe F,
    5. Ostojic S
    (2022) The role of population structure in computations through neural dynamics. Nat Neurosci 25:783–794. doi:10.1038/s41593-022-01088-4 pmid:35668174
    OpenUrlCrossRefPubMed
  34. ↵
    1. Duncker L,
    2. Sahani M
    (2021) Dynamics on the manifold: identifying computational dynamical activity from neural population recordings. Curr Opin Neurobiol 70:163–170. doi:10.1016/j.conb.2021.10.014 pmid:34837752
    OpenUrlCrossRefPubMed
  35. ↵
    1. Ebitz RB,
    2. Hayden BY
    (2021) The population doctrine in cognitive neuroscience. Neuron 109:3055–3068. doi:10.1016/j.neuron.2021.07.011 pmid:34416170
    OpenUrlCrossRefPubMed
  36. ↵
    1. Economo MN,
    2. Viswanathan S,
    3. Tasic B,
    4. Bas E,
    5. Winnubst J,
    6. Menon V,
    7. Graybuck LT,
    8. Nguyen TN,
    9. Smith KA,
    10. Yao Z,
    11. Wang L,
    12. Gerfen CR,
    13. Chandrashekar J,
    14. Zeng H,
    15. Looger LL,
    16. Svoboda K
    (2018) Distinct descending motor cortex pathways and their roles in movement. Nature 563:79–84. doi:10.1038/s41586-018-0642-9 pmid:30382200
    OpenUrlCrossRefPubMed
  37. ↵
    1. Fleming SM,
    2. Maniscalco B,
    3. Ko Y,
    4. Amendi N,
    5. Ro T,
    6. Lau H
    (2015) Action-specific disruption of perceptual confidence. Psychol Sci 26:89–98. doi:10.1177/0956797614557697 pmid:25425059
    OpenUrlCrossRefPubMed
  38. ↵
    1. Flesch T,
    2. Juechems K,
    3. Dumbalska T,
    4. Saxe A,
    5. Summerfield C
    (2022) Orthogonal representations for robust context-dependent task performance in brains and neural networks. Neuron 110:1258–1270.e11. doi:10.1016/j.neuron.2022.01.005 pmid:35085492
    OpenUrlCrossRefPubMed
  39. ↵
    1. Funahashi S,
    2. Bruce CJ,
    3. Goldman-Rakic PS
    (1989) Mnemonic coding of visual space in the monkey's dorsolateral prefrontal cortex. J Neurophysiol 61:331–349. doi:10.1152/jn.1989.61.2.331 pmid:2918358
    OpenUrlCrossRefPubMed
  40. ↵
    1. Fusi S,
    2. Miller EK,
    3. Rigotti M
    (2016) Why neurons mix: high dimensionality for higher cognition. Curr Opin Neurobiol 37:66–74. doi:10.1016/j.conb.2016.01.010 pmid:26851755
    OpenUrlCrossRefPubMed
  41. ↵
    1. Gawlikowski J,
    2. Tassi CR,
    3. Ali M,
    4. Lee J,
    5. Humt M,
    6. Feng J,
    7. Kruspe AM,
    8. Triebel R,
    9. Jung P,
    10. Roscher R,
    11. Shahzad M,
    12. Yang W,
    13. Bamler R,
    14. Zhu XX
    (2021) A survey of uncertainty in deep neural networks. arXiv:2107.03342.
  42. ↵
    1. Gilpin W,
    2. Huang Y,
    3. Forger DB
    (2020) Learning dynamics from large biological data sets: machine learning meets systems biology. Curr Opin Syst Biol 22:1–7. doi:10.1016/j.coisb.2020.07.009
    OpenUrlCrossRef
  43. ↵
    1. Goffinet J,
    2. Brudner S,
    3. Mooney R,
    4. Pearson J
    (2021) Low-dimensional learned feature spaces quantify individual and group differences in vocal repertoires. Elife 10:e67855. doi:10.7554/Elife.67855
    OpenUrlCrossRef
  44. ↵
    1. Goodfellow IJ,
    2. Pouget-Abadie J,
    3. Mirza M,
    4. Xu B,
    5. Warde-Farley D,
    6. Ozair S,
    7. Courville A,
    8. Bengio Y
    (2014) Generative adversarial networks. arXiv:2107.03342.
  45. ↵
    1. Graving JM,
    2. Chae D,
    3. Naik H,
    4. Li L,
    5. Koger B,
    6. Costelloe BR,
    7. Couzin ID
    (2019) DeepPoseKit, a software toolkit for fast and robust animal pose estimation using deep learning. Elife 8:e47994. doi:10.7554/Elife.47994
    OpenUrlCrossRef
  46. ↵
    1. Graving JM,
    2. Couzin ID
    (2020) VAE-SNE: a deep generative model for simultaneous dimensionality reduction and clustering. bioRxiv 207993. doi:10.1101/2020.07.17.207993.
    OpenUrlCrossRef
  47. ↵
    1. Greedy W,
    2. Zhu HW,
    3. Pemberton J,
    4. Mellor J,
    5. Costa RP
    (2022) Single-phase deep learning in cortico-cortical networks. arXiv:2206.11769.
  48. ↵
    1. Griffiths TL,
    2. Callaway F,
    3. Chang MB,
    4. Grant E,
    5. Krueger PM,
    6. Lieder F
    (2019) Doing more with less: meta-reasoning and meta-learning in humans and machines. Curr Opin Behav Sci 29:24–30. doi:10.1016/j.cobeha.2019.01.005
    OpenUrlCrossRef
  49. ↵
    1. Guggenmos M,
    2. Wilbertz G,
    3. Hebart MN,
    4. Sterzer P
    (2016) Mesolimbic confidence signals guide perceptual learning in the absence of external feedback. Elife 5:e13388. doi:10.7554/Elife.13388
    OpenUrlCrossRefPubMed
  50. ↵
    1. Hayhoe MM
    (2018) Davida Teller Award Lecture 2017: what can be learned from natural behavior? J Vis 18:10. doi:10.1167/18.4.10 pmid:29710300
    OpenUrlCrossRefPubMed
  51. ↵
    1. Haykin SS
    (1994) Neural networks: a comprehensive foundation. New York: Maxwell Macmillan.
  52. ↵
    1. Hedley RW
    (2016) Complexity, predictability and time homogeneity of syntax in the songs of Cassin's vireo (Vireo cassinii). PLoS One 11:e0150822. doi:10.1371/journal.pone.0150822 pmid:27050537
    OpenUrlCrossRefPubMed
  53. ↵
    1. Hirokawa J,
    2. Vaughan A,
    3. Masset P,
    4. Ott T,
    5. Kepecs A
    (2019) Frontal cortex neuron types categorically encode single decision variables. Nature 576:446–451. doi:10.1038/s41586-019-1816-9 pmid:31801999
    OpenUrlCrossRefPubMed
  54. ↵
    1. Hocker DL,
    2. Brody CD,
    3. Savin C,
    4. Constantinople CM
    (2021) Subpopulations of neurons in lOFC encode previous and current rewards at time of choice. Elife 10:e70129. doi:10.7554/Elife.70129
    OpenUrlCrossRef
  55. ↵
    1. Hopfield JJ
    (1982) Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA 79:2554–2558. doi:10.1073/pnas.79.8.2554 pmid:6953413
    OpenUrlAbstract/FREE Full Text
  56. ↵
    1. Jazayeri M,
    2. Ostojic S
    (2021) Interpreting neural computations by examining intrinsic and embedding dimensionality of neural activity. Curr Opin Neurobiol 70:113–120. doi:10.1016/j.conb.2021.08.002 pmid:34537579
    OpenUrlCrossRefPubMed
  57. ↵
    1. Jin DZ,
    2. Kozhevnikov AA
    (2011) A compact statistical model of the song syntax in Bengalese finch. PLoS Comput Biol 7:e1001108. doi:10.1371/journal.pcbi.1001108 pmid:21445230
    OpenUrlCrossRefPubMed
  58. ↵
    1. Johnston D,
    2. Narayanan R
    (2008) Active dendrites: colorful wings of the mysterious butterflies. Trends Neurosci 31:309–316. doi:10.1016/j.tins.2008.03.004 pmid:18471907
    OpenUrlCrossRefPubMed
  59. ↵
    1. Jones EG
    (2001) The thalamic matrix and thalamocortical synchrony. Trends Neurosci 24:595–601. doi:10.1016/s0166-2236(00)01922-6 pmid:11576674
    OpenUrlCrossRefPubMed
  60. ↵
    1. Josselyn SA,
    2. Tonegawa S
    (2020) Memory engrams: recalling the past and imagining the future. Science 367:eaaw4325. doi:10.1126/science.aaw4325
    OpenUrlAbstract/FREE Full Text
  61. ↵
    1. Kabra M,
    2. Robie AA,
    3. Rivera-Alba M,
    4. Branson S,
    5. Branson K
    (2013) JAABA: interactive machine learning for automatic annotation of animal behavior. Nat Methods 10:64–67. doi:10.1038/nmeth.2281 pmid:23202433
    OpenUrlCrossRefPubMed
  62. ↵
    1. Kepecs A,
    2. Mainen ZF
    (2012) A computational framework for the study of confidence in humans and animals. Philos Trans R Soc Lond B Biol Sci 367:1322–1337. doi:10.1098/rstb.2012.0037 pmid:22492750
    OpenUrlCrossRefPubMed
  63. ↵
    1. Kepecs A,
    2. Uchida N,
    3. Zariwala HA,
    4. Mainen ZF
    (2008) Neural correlates, computation and behavioural impact of decision confidence. Nature 455:227–231. doi:10.1038/nature07200 pmid:18690210
    OpenUrlCrossRefPubMed
  64. ↵
    1. Khaligh-Razavi SM,
    2. Kriegeskorte N
    (2014) Deep supervised, but not unsupervised, models may explain IT cortical representation. PLoS Comput Biol 10:e1003915. doi:10.1371/journal.pcbi.1003915 pmid:25375136
    OpenUrlCrossRefPubMed
  65. ↵
    1. Kobak D,
    2. Brendel W,
    3. Constantinidis C,
    4. Feierstein CE,
    5. Kepecs A,
    6. Mainen ZF,
    7. Qi XL,
    8. Romo R,
    9. Uchida N,
    10. Machens CK
    (2016) Demixed principal component analysis of neural population data. Elife 5:e10989. doi:10.7554/Elife.10989
    OpenUrlCrossRefPubMed
  66. ↵
    1. Kogan JA,
    2. Margoliash D
    (1998) Automated recognition of bird song elements from continuous recordings using dynamic time warping and hidden Markov models: a comparative study. J Acoust Soc Am 103:2185–2196. doi:10.1121/1.421364 pmid:9566338
    OpenUrlCrossRefPubMed
  67. ↵
    1. Kohlsdorf D,
    2. Herzing D,
    3. Starner T
    (2020) An auto encoder for audio dolphin communication. International Joint Conference on Neural Networks IJCNN, pp 1–7. IEEE.
  68. ↵
    1. Kollmorgen S,
    2. Hahnloser RH,
    3. Mante V
    (2020) Nearest neighbours reveal fast and slow components of motor learning. Nature 577:526–530. doi:10.1038/s41586-019-1892-x pmid:31915383
    OpenUrlCrossRefPubMed
  69. ↵
    1. Koumura T,
    2. Okanoya K
    (2016) Automatic recognition of element classes and boundaries in the birdsong with variable sequences. PLoS One 11:e0159188. doi:10.1371/journal.pone.0159188 pmid:27442240
    OpenUrlCrossRefPubMed
  70. ↵
    1. Krakauer JW,
    2. Ghazanfar AA,
    3. Gomez-Marin A,
    4. MacIver MA,
    5. Poeppel D
    (2017) Neuroscience needs behavior: correcting a reductionist bias. Neuron 93:480–490. doi:10.1016/j.neuron.2016.12.041 pmid:28182904
    OpenUrlCrossRefPubMed
  71. ↵
    1. Kubilius J,
    2. Schrimpf M,
    3. Kar K,
    4. Hong H,
    5. Majaj NJ,
    6. Rajalingham R,
    7. Issa EB,
    8. Bashivan P,
    9. Prescott-Roy J,
    10. Schmidt K,
    11. Nayebi A,
    12. Bear D,
    13. Yamins DL,
    14. DiCarlo JJ
    (2019) Brain-like object recognition with high-performing shallow recurrent ANNs. Advances in neural information processing systems, 32.
  72. ↵
    1. Kudithipudi D, et al
    . (2022) Biological underpinnings for lifelong learning machines. Nat Mach Intell 4:196–210. doi:10.1038/s42256-022-00452-0
    OpenUrlCrossRef
  73. ↵
    1. Lak A,
    2. Costa GM,
    3. Romberg E,
    4. Koulakov AA,
    5. Mainen ZF,
    6. Kepecs A
    (2014) Orbitofrontal cortex is required for optimal waiting based on decision confidence. Neuron 84:190–201. doi:10.1016/j.neuron.2014.08.039 pmid:25242219
    OpenUrlCrossRefPubMed
  74. ↵
    1. Langdon C,
    2. Engel TA
    (2022) Latent circuit inference from heterogeneous neural responses during cognitive tasks. bioRxiv 2022.01.23.477431.
  75. ↵
    1. Larkum M
    (2013) A cellular mechanism for cortical associations: an organizing principle for the cerebral cortex. Trends Neurosci 36:141–151. doi:10.1016/j.tins.2012.11.006 pmid:23273272
    OpenUrlCrossRefPubMed
  76. ↵
    1. Larkum ME
    (2022) Are dendrites conceptually useful? Dendritic Contrib Biol Artif Comput 489:4–14. doi:10.1016/j.neuroscience.2022.03.008 pmid:35288178
    OpenUrlCrossRefPubMed
  77. ↵
    1. Le Bé JV,
    2. Markram H
    (2006) Spontaneous and evoked synaptic rewiring in the neonatal neocortex. Proc Natl Acad Sci USA 103:13214–13219. doi:10.1073/pnas.0604691103 pmid:16924105
    OpenUrlAbstract/FREE Full Text
  78. ↵
    1. LeCun Y,
    2. Bengio Y,
    3. Hinton G
    (2015) Deep learning. Nature 521:436–444. doi:10.1038/nature14539 pmid:26017442
    OpenUrlCrossRefPubMed
  79. ↵
    1. Leonardo A,
    2. Fee MS
    (2005) Ensemble coding of vocal control in birdsong. J Neurosci 25:652–661. doi:10.1523/JNEUROSCI.3036-04.2005 pmid:15659602
    OpenUrlAbstract/FREE Full Text
  80. ↵
    1. Li HH,
    2. Ma WJ
    (2020) Confidence reports in decision-making with multiple alternatives violate the Bayesian confidence hypothesis. Nat Commun 11:2004. doi:10.1038/s41467-020-15581-6 pmid:32332712
    OpenUrlCrossRefPubMed
  81. ↵
    1. Li M,
    2. Han Y,
    3. Aburn MJ,
    4. Breakspear M,
    5. Poldrack RA,
    6. Shine JM,
    7. Lizier JT
    (2019) Transitions in information processing dynamics at the whole-brain network level are driven by alterations in neural gain. PLoS Comput Biol 15:e1006957. doi:10.1371/journal.pcbi.1006957 pmid:31613882
    OpenUrlCrossRefPubMed
  82. ↵
    1. Lindsay GW
    (2021) Convolutional neural networks as a model of the visual system: past, present, and future. J Cogn Neurosci 33:2017–2031. doi:10.1162/jocn_a_01544 pmid:32027584
    OpenUrlCrossRefPubMed
  83. ↵
    1. Lindsay GW
    (2022) Testing the tools of systems neuroscience on artificial neural networks. arXiv:2202.07035.
  84. ↵
    1. Lindsay GW,
    2. Mrsic-Flogel TD,
    3. Sahani M
    (2022) Bio-inspired neural networks implement different recurrent visual processing strategies than task-trained ones do. bioRxiv 483196. doi:10.1101/2022.03.07.483196.
    OpenUrlCrossRef
  85. ↵
    1. Linsley D,
    2. Kim J,
    3. Veerabadran V,
    4. Windolf C,
    5. Serre T
    (2018) Learning long-range spatial dependencies with horizontal gated recurrent units. Advances in neural information processing systems, 31.
  86. ↵
    1. Lisman J,
    2. Cooper K,
    3. Sehgal M,
    4. Silva AJ
    (2018) Memory formation depends on both synapse-specific modifications of synaptic strength and cell-specific increases in excitability. Nat Neurosci 21:309–314. doi:10.1038/s41593-018-0076-6 pmid:29434376
    OpenUrlCrossRefPubMed
  87. ↵
    1. Luxem K,
    2. Mocellin P,
    3. Fuhrmann F,
    4. Kürsch J,
    5. Remy S,
    6. Bauer P
    (2022) Identifying behavioral structure from deep variational embeddings of animal motion. bioRxiv 095430. doi:10.1101/2020.05.14.095430.
    OpenUrlAbstract/FREE Full Text
  88. ↵
    1. Machado AS,
    2. Darmohray DM,
    3. Fayad J,
    4. Marques HG,
    5. Carey MR
    (2015) A quantitative framework for whole-body coordination reveals specific deficits in freely walking ataxic mice. Elife 4:e07892. doi:10.7554/Elife.07892
    OpenUrlCrossRefPubMed
  89. ↵
    1. Malagon-Vina H,
    2. Ciocchi S,
    3. Passecker J,
    4. Dorffner G,
    5. Klausberger T
    (2018) Fluid network dynamics in the prefrontal cortex during multiple strategy switching. Nat Commun 9:309. doi:10.1038/s41467-017-02764-x pmid:29358717
    OpenUrlCrossRefPubMed
  90. ↵
    1. Mamassian P
    (2022) Uncertain perceptual confidence. Nat Hum Behav 6:179–180. doi:10.1038/s41562-021-01248-9 pmid:35058642
    OpenUrlCrossRefPubMed
  91. ↵
    1. Maniscalco B,
    2. Peters MA,
    3. Lau H
    (2016) Heuristic use of perceptual evidence leads to dissociation between performance and metacognitive sensitivity. Atten Percept Psychophys 78:923–937. doi:10.3758/s13414-016-1059-x pmid:26791233
    OpenUrlCrossRefPubMed
  92. ↵
    1. Maniscalco B,
    2. Odegaard B,
    3. Grimaldi P,
    4. Cho SH,
    5. Basso MA,
    6. Lau H,
    7. Peters MA
    (2021) Tuned inhibition in perceptual decision-making circuits can explain seemingly suboptimal confidence behavior. PLoS Comput Biol 17:e1008779. doi:10.1371/journal.pcbi.1008779
    OpenUrlCrossRef
  93. ↵
    1. Mante V,
    2. Sussillo D,
    3. Shenoy KV,
    4. Newsome WT
    (2013) Context-dependent computation by recurrent dynamics in prefrontal cortex. Nature 503:78–84. doi:10.1038/nature12742 pmid:24201281
    OpenUrlCrossRefPubMed
  94. ↵
    1. Marblestone AH,
    2. Wayne G,
    3. Kording KP
    (2016) Toward an integration of deep learning and neuroscience. Front Comput Neurosci 10:94. doi:10.3389/fncom.2016.00094 pmid:27683554
    OpenUrlCrossRefPubMed
  95. ↵
    1. Markowitz JE,
    2. Ivie E,
    3. Kligler L,
    4. Gardner TJ
    (2013) Long-range order in canary song. PLoS Comput Biol 9:e1003052. doi:10.1371/journal.pcbi.1003052 pmid:23658509
    OpenUrlCrossRefPubMed
  96. ↵
    1. Marshall JD,
    2. Aldarondo DE,
    3. Dunn TW,
    4. Wang WL,
    5. Berman GJ,
    6. Ölveczky BP
    (2021) Continuous whole-body 3D kinematic recordings across the rodent behavioral repertoire. Neuron 109:420–437.e8. doi:10.1016/j.neuron.2020.11.016 pmid:33340448
    OpenUrlCrossRefPubMed
  97. ↵
    1. Marullo C,
    2. Agliari E
    (2020) Boltzmann machines as generalized Hopfield networks: a review of recent results and outlooks. Entropy Basel Switz 23:34. doi:10.3390/e23010034
    OpenUrlCrossRef
  98. ↵
    1. Masset P,
    2. Ott T,
    3. Lak A,
    4. Hirokawa J,
    5. Kepecs A
    (2020) Behavior- and modality-general representation of confidence in orbitofrontal cortex. Cell 182:112–126.e18. doi:10.1016/j.cell.2020.05.022 pmid:32504542
    OpenUrlCrossRefPubMed
  99. ↵
    1. Mathis A,
    2. Mamidanna P,
    3. Cury KM,
    4. Abe T,
    5. Murthy VN,
    6. Mathis MW,
    7. Bethge M
    (2018) DeepLabCut: markerless pose estimation of user-defined body parts with deep learning. Nat Neurosci 21:1281–1289. doi:10.1038/s41593-018-0209-y pmid:30127430
    OpenUrlCrossRefPubMed
  100. ↵
    1. McCullough MH,
    2. Goodhill GJ
    (2021) Unsupervised quantification of naturalistic animal behaviors for gaining insight into the brain. Curr Opin Neurobiol 70:89–100. doi:10.1016/j.conb.2021.07.014 pmid:34482006
    OpenUrlCrossRefPubMed
  101. ↵
    1. McCulloch WS,
    2. Pitts W
    (1943) A logical calculus of the ideas immanent in nervous activity. Bull Math Biophys 5:115–133. doi:10.1007/BF02478259
    OpenUrlCrossRef
  102. ↵
    1. McInnes L,
    2. Healy J,
    3. Melville J
    (2018) Umap: Uniform manifold approximation and projection for dimension reduction. arXiv:1802.03426.
  103. ↵
    1. Mearns DS,
    2. Donovan JC,
    3. Fernandes AM,
    4. Semmelhack JL,
    5. Baier H
    (2020) Deconstructing hunting behavior reveals a tightly coupled stimulus-response loop. Curr Biol 30:54–69.e9. doi:10.1016/j.cub.2019.11.022 pmid:31866365
    OpenUrlCrossRefPubMed
  104. ↵
    1. Mei J,
    2. Muller E,
    3. Ramaswamy S
    (2022) Informing deep neural networks by multiscale principles of neuromodulatory systems. Trends Neurosci 45:237–250. doi:10.1016/j.tins.2021.12.008 pmid:35074219
    OpenUrlCrossRefPubMed
  105. ↵
    1. Mets DG,
    2. Brainard MS
    (2018) An automated approach to the quantitation of vocalizations and vocal learning in the songbird. PLoS Comput Biol 14:e1006437. doi:10.1371/journal.pcbi.1006437 pmid:30169523
    OpenUrlCrossRefPubMed
  106. ↵
    1. Michel M,
    2. Peters MA
    (2021) Confirmation bias without rhyme or reason. Synthese 199:2757–2772. doi:10.1007/s11229-020-02910-x
    OpenUrlCrossRef
  107. ↵
    1. Minsky M,
    2. Papert S
    (1972) Artificial Intelligence Progress Report. AI Memo 252. Cambridge, MA.
  108. ↵
    1. Mishra P,
    2. Narayanan R
    (2021) Stable continual learning through structured multiscale plasticity manifolds. Curr Opin Neurobiol 70:51–63. doi:10.1016/j.conb.2021.07.009 pmid:34416674
    OpenUrlCrossRefPubMed
  109. ↵
    1. Morfi V,
    2. Lachlan RF,
    3. Stowell D
    (2021) Deep perceptual embeddings for unlabelled animal sound events. J Acoust Soc Am 150:2–11. doi:10.1121/10.0005475 pmid:34340499
    OpenUrlCrossRefPubMed
  110. ↵
    1. Mukherjee A,
    2. Lam NH,
    3. Wimmer RD,
    4. Halassa MM
    (2021) Thalamic circuits for independent control of prefrontal signal and noise. Nature 600:100–104. doi:10.1038/s41586-021-04056-3 pmid:34614503
    OpenUrlCrossRefPubMed
  111. ↵
    1. Namboodiri VM,
    2. Otis JM,
    3. van Heeswijk K,
    4. Voets ES,
    5. Alghorazi RA,
    6. Rodriguez-Romaguera J,
    7. Mihalas S,
    8. Stuber GD
    (2019) Single-cell activity tracking reveals that orbitofrontal neurons acquire and maintain a long-term memory to guide behavioral adaptation. Nat Neurosci 22:1110–1121. doi:10.1038/s41593-019-0408-1 pmid:31160741
    OpenUrlCrossRefPubMed
  112. ↵
    1. Nayebi A,
    2. Sagastuy-Brena J,
    3. Bear DM,
    4. Kar K,
    5. Kubilius J,
    6. Ganguli S,
    7. Sussillo D,
    8. DiCarlo JJ,
    9. Yamins DL
    (2021) Goal-driven recurrent neural network models of the ventral visual stream. bioRxiv 431717. doi:10.1101/2021.02.17.431717.
    OpenUrlAbstract/FREE Full Text
  113. ↵
    1. Nicholson D
    (2016) Comparison of machine learning methods applied to birdsong element classification. In: Proceedings of the 15th Python in Science Conference (Benthall S, Rostrup S, eds), pp 57–61. doi:10.25080/Majora-629e541a-008
    OpenUrlCrossRef
  114. ↵
    1. Nicholson D
    (2021) NickleDave/hybrid-vocal-classifier. 0.3.0Github. Available at https://github.com/NickleDave/hybrid-vocal-classifier.git.
  115. ↵
    1. Ogawa M,
    2. van der Meer MA,
    3. Esber GR,
    4. Cerri DH,
    5. Stalnaker TA,
    6. Schoenbaum G
    (2013) Risk-responsive orbitofrontal neurons track acquired salience. Neuron 77:251–258. doi:10.1016/j.neuron.2012.11.006 pmid:23352162
    OpenUrlCrossRefPubMed
  116. ↵
    1. Onken A,
    2. Xie J,
    3. Panzeri S,
    4. Padoa-Schioppa C
    (2019) Categorical encoding of decision variables in orbitofrontal cortex. PLoS Comput Biol 15:e1006667. doi:10.1371/journal.pcbi.1006667 pmid:31609973
    OpenUrlCrossRefPubMed
  117. ↵
    1. Ott T,
    2. Masset P,
    3. Kepecs A
    (2018) The neurobiology of confidence: from beliefs to neurons. Cold Spring Harb Symp Quant Biol 83:9–16. doi:10.1101/sqb.2018.83.038794 pmid:31270145
    OpenUrlAbstract/FREE Full Text
  118. ↵
    1. Padamsey Z,
    2. Katsanevaki D,
    3. Dupuy N,
    4. Rochefort NL
    (2022) Neocortex saves energy by reducing coding precision during food scarcity. Neuron 110:280–296.e10. doi:10.1016/j.neuron.2021.10.024 pmid:34741806
    OpenUrlCrossRefPubMed
  119. ↵
    1. Padoa-Schioppa C,
    2. Assad JA
    (2006) Neurons in the orbitofrontal cortex encode economic value. Nature 441:223–226. doi:10.1038/nature04676 pmid:16633341
    OpenUrlCrossRefPubMed
  120. ↵
    1. Pagliarini S,
    2. Trouvain N,
    3. Leblois A,
    4. Hinaut X
    (2021) What does the canary say? Low-dimensional GAN applied to birdsong. hal-03244723v2.
  121. ↵
    1. Pandarinath C,
    2. Ames KC,
    3. Russo AA,
    4. Farshchian A,
    5. Miller LE,
    6. Dyer EL,
    7. Kao JC
    (2018) Latent factors and dynamics in motor cortex and their application to brain–machine interfaces. J Neurosci 38:9390–9401. doi:10.1523/JNEUROSCI.1669-18.2018 pmid:30381431
    OpenUrlAbstract/FREE Full Text
  122. ↵
    1. Payeur A,
    2. Guerguiev J,
    3. Zenke F,
    4. Richards BA,
    5. Naud R
    (2021) Burst-dependent synaptic plasticity can coordinate learning in hierarchical circuits. Nat Neurosci 24:1010–1019. doi:10.1038/s41593-021-00857-x pmid:33986551
    OpenUrlCrossRefPubMed
  123. ↵
    1. Pemberton JO,
    2. Boven E,
    3. Apps R,
    4. Costa RP
    (2021) Cortico-cerebellar networks as decoupling neural interfaces. Advances in neural information processing systems 34:7745–7759.
    OpenUrl
  124. ↵
    1. Pereira TD,
    2. Aldarondo DE,
    3. Willmore L,
    4. Kislin M,
    5. Wang SS,
    6. Murthy M,
    7. Shaevitz JW
    (2019) Fast animal pose estimation using deep neural networks. Nat Methods 16:117–125. doi:10.1038/s41592-018-0234-5
    OpenUrlCrossRefPubMed
  125. ↵
    1. Pereira TD,
    2. Shaevitz JW,
    3. Murthy M
    (2020) Quantifying behavior to understand the brain. Nat Neurosci 23:1537–1549. doi:10.1038/s41593-020-00734-z pmid:33169033
    OpenUrlCrossRefPubMed
  126. ↵
    1. Peters MA
    (2022) Confidence in Decision-Making. Oxford Research Encyclopedia of Neuroscience.
  127. ↵
    1. Poirazi P,
    2. Papoutsi A
    (2020) Illuminating dendritic function with computational models. Nat Rev Neurosci 21:303–321. doi:10.1038/s41583-020-0301-7 pmid:32393820
    OpenUrlCrossRefPubMed
  128. ↵
    1. Pouget A,
    2. Drugowitsch J,
    3. Kepecs A
    (2016) Confidence and certainty: distinct probabilistic quantities for different goals. Nat Neurosci 19:366–374. doi:10.1038/nn.4240 pmid:26906503
    OpenUrlCrossRefPubMed
  129. ↵
    1. Ptasczynski LE,
    2. Steinecker I,
    3. Sterzer P,
    4. Guggenmos M
    (2022) The value of confidence: confidence prediction errors drive value-based learning in the absence of external feedback. PLOS Comput Biol 18:e1010580.
    OpenUrl
  130. ↵
    1. Radford A,
    2. Metz L,
    3. Chintala S
    (2015) Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv:1511.06434.
  131. ↵
    1. Ramaswamy S,
    2. Markram H
    (2015) Anatomy and physiology of the thick-tufted layer 5 pyramidal neuron. Front Cell Neurosci 9:233. doi:10.3389/fncel.2015.00233 pmid:26167146
    OpenUrlCrossRefPubMed
  132. ↵
    1. Rao SC,
    2. Rainer G,
    3. Miller EK
    (1997) Integration of what and where in the primate prefrontal cortex. Science 276:821–824. doi:10.1126/science.276.5313.821 pmid:9115211
    OpenUrlAbstract/FREE Full Text
  133. ↵
    1. Ratté S,
    2. Hong S,
    3. De Schutter E,
    4. Prescott SA
    (2013) Impact of neuronal properties on network coding: roles of spike initiation dynamics and robust synchrony transfer. Neuron 78:758–772. doi:10.1016/j.neuron.2013.05.030 pmid:23764282
    OpenUrlCrossRefPubMed
  134. ↵
    1. Rausch M,
    2. Hellmann S,
    3. Zehetleitner M
    (2018) Confidence in masked orientation judgments is informed by both evidence and visibility. Atten Percept Psychophys 80:134–154. doi:10.3758/s13414-017-1431-5 pmid:29043657
    OpenUrlCrossRefPubMed
  135. ↵
    1. Reed S, et al
    . (2022) A generalist agent. arXiv:2205.06175.
  136. ↵
    1. Richards BA, et al
    . (2019) A deep learning framework for neuroscience. Nat Neurosci 22:1761–1770. doi:10.1038/s41593-019-0520-2 pmid:31659335
    OpenUrlCrossRefPubMed
  137. ↵
    1. Rigotti M,
    2. Barak O,
    3. Warden MR,
    4. Wang XJ,
    5. Daw ND,
    6. Miller EK,
    7. Fusi S
    (2013) The importance of mixed selectivity in complex cognitive tasks. Nature 497:585–590. doi:10.1038/nature12160 pmid:23685452
    OpenUrlCrossRefPubMed
  138. ↵
    1. Roelfsema PR,
    2. Holtmaat A
    (2018) Control of synaptic plasticity in deep cortical networks. Nat Rev Neurosci 19:166–180. doi:10.1038/nrn.2018.6 pmid:29449713
    OpenUrlCrossRefPubMed
  139. ↵
    1. Roscow EL,
    2. Chua R,
    3. Costa RP,
    4. Jones MW,
    5. Lepora N
    (2021) Learning offline: memory replay in biological and artificial reinforcement learning. Trends Neurosci 44:808–821. doi:10.1016/j.tins.2021.07.007 pmid:34481635
    OpenUrlCrossRefPubMed
  140. ↵
    1. Rosenblatt F
    (1958) The perceptron: a probabilistic model for information storage and organization in the brain. Psychol Rev 65:386–408. doi:10.1037/h0042519 pmid:13602029
    OpenUrlCrossRefPubMed
  141. ↵
    1. Rudolph M,
    2. Destexhe A
    (2003) Tuning neocortical pyramidal neurons between integrators and coincidence detectors. J Comput Neurosci 14:239–251. doi:10.1023/a:1023245625896 pmid:12766426
    OpenUrlCrossRefPubMed
  142. ↵
    1. Rumelhart DE,
    2. Hinton GE,
    3. Williams RJ
    (1986) Learning representations by back-propagating errors. Nature 323:533–536. doi:10.1038/323533a0
    OpenUrlCrossRef
  143. ↵
    1. Sacramento J,
    2. Ponte Costa R,
    3. Bengio Y,
    4. Senn W
    (2018) Dendritic cortical microcircuits approximate the backpropagation algorithm. Advances in neural information processing systems, 31.
  144. ↵
    1. Sainburg T,
    2. Gentner TQ
    (2021) Toward a computational neuroethology of vocal communication: from bioacoustics to neurophysiology, emerging tools and future directions. Front Behav Neurosci 15:811737. doi:10.3389/fnbeh.2021.811737 pmid:34987365
    OpenUrlCrossRefPubMed
  145. ↵
    1. Sainburg T,
    2. Thielk M,
    3. Theilman B,
    4. Migliori B,
    5. Gentner T
    (2018) Generative adversarial interpolative autoencoding: adversarial training on latent space interpolations encourage convex latent distributions. arXiv:1807.06650.
  146. ↵
    1. Sainburg T,
    2. Thielk M,
    3. Gentner TQ
    (2020) Finding, visualizing, and quantifying latent structure across diverse animal vocal repertoires. PLoS Comput Biol 16:e1008228. doi:10.1371/journal.pcbi.1008228 pmid:33057332
    OpenUrlCrossRefPubMed
  147. ↵
    1. Sainburg T,
    2. McInnes L,
    3. Gentner TQ
    (2021) Parametric UMAP embeddings for representation and semisupervised learning. Neural Comput 33:2881–2907. doi:10.1162/neco_a_01434 pmid:34474477
    OpenUrlCrossRefPubMed
  148. ↵
    1. Saxena S,
    2. Cunningham JP
    (2019) Towards the neural population doctrine. Curr Opin Neurobiol 55:103–111. doi:10.1016/j.conb.2019.02.002 pmid:30877963
    OpenUrlCrossRefPubMed
  149. ↵
    1. Schmidhuber J
    (2015) Deep learning in neural networks: an overview. Neural Netw 61:85–117. doi:10.1016/j.neunet.2014.09.003 pmid:25462637
    OpenUrlCrossRefPubMed
  150. ↵
    1. Sejnowski TJ
    (2020) The unreasonable effectiveness of deep learning in artificial intelligence. Proc Natl Acad Sci USA 117:30033–30038. doi:10.1073/pnas.1907373117 pmid:31992643
    OpenUrlAbstract/FREE Full Text
  151. ↵
    1. Sejnowski TJ,
    2. Churchland PS,
    3. Movshon JA
    (2014) Putting big data to good use in neuroscience. Nat Neurosci 17:1440–1441. doi:10.1038/nn.3839 pmid:25349909
    OpenUrlCrossRefPubMed
  152. ↵
    1. Shine JM
    (2021) The thalamus integrates the macrosystems of the brain to facilitate complex, adaptive brain network dynamics. Prog Neurobiol 199:101951. doi:10.1016/j.pneurobio.2020.101951 pmid:33189781
    OpenUrlCrossRefPubMed
  153. ↵
    1. Shine JM,
    2. Müller EJ,
    3. Munn B,
    4. Cabral J,
    5. Moran RJ,
    6. Breakspear M
    (2021) Computational models link cellular mechanisms of neuromodulation to large-scale neural dynamics. Nat Neurosci 24:765–776. doi:10.1038/s41593-021-00824-6 pmid:33958801
    OpenUrlCrossRefPubMed
  154. ↵
    1. Sober SJ,
    2. Brainard MS
    (2009) Adult birdsong is actively maintained by error correction. Nat Neurosci 12:927–931. doi:10.1038/nn.2336 pmid:19525945
    OpenUrlCrossRefPubMed
  155. ↵
    1. Sober SJ,
    2. Brainard MS
    (2012) Vocal learning is constrained by the statistics of sensorimotor experience. Proc Natl Acad Sci USA 109:21099–21103. doi:10.1073/pnas.1213622109 pmid:23213223
    OpenUrlAbstract/FREE Full Text
  156. ↵
    1. Sober SJ,
    2. Wohlgemuth MJ,
    3. Brainard MS
    (2008) Central contributions to acoustic variation in birdsong. J Neurosci 28:10370–10379. doi:10.1523/JNEUROSCI.2448-08.2008 pmid:18842896
    OpenUrlAbstract/FREE Full Text
  157. ↵
    1. Sohn H,
    2. Narain D,
    3. Meirhaeghe N,
    4. Jazayeri M
    (2019) Bayesian computation through cortical latent dynamics. Neuron 103:934–947.e5. doi:10.1016/j.neuron.2019.06.012 pmid:31320220
    OpenUrlCrossRefPubMed
  158. ↵
    1. Song HF,
    2. Yang GR,
    3. Wang XJ
    (2016) Training excitatory-inhibitory recurrent neural networks for cognitive tasks: a simple and flexible framework. PLoS Comput Biol 12:e1004792. doi:10.1371/journal.pcbi.1004792 pmid:26928718
    OpenUrlCrossRefPubMed
  159. ↵
    1. Spoerer CJ,
    2. McClure P,
    3. Kriegeskorte N
    (2017) Recurrent convolutional neural networks: a better model of biological object recognition. Front Psychol 8:1551. pmid:28955272
    OpenUrlCrossRefPubMed
  160. ↵
    1. Spruston N
    (2008) Pyramidal neurons: dendritic structure and synaptic integration. Nat Rev Neurosci 9:206–221. doi:10.1038/nrn2286 pmid:18270515
    OpenUrlCrossRefPubMed
  161. ↵
    1. Steinfath E,
    2. Palacios-Muñoz A,
    3. Rottschäfer JR,
    4. Yuezak D,
    5. Clemens J
    (2021) Fast and accurate annotation of acoustic signals with deep neural networks. Elife 10:e68837. doi:10.7554/Elife.68837
    OpenUrlCrossRef
  162. ↵
    1. Stephens GJ,
    2. Johnson-Kerner B,
    3. Bialek W,
    4. Ryu WS
    (2008) Dimensionality and dynamics in the behavior of C. elegans. PLoS Comput Biol 4:e1000028. doi:10.1371/journal.pcbi.1000028
    OpenUrlCrossRefPubMed
  163. ↵
    1. Stolyarova A,
    2. Rakhshan M,
    3. Hart EE,
    4. O'Dell TJ,
    5. Peters MA,
    6. Lau H,
    7. Soltani A,
    8. Izquierdo A
    (2019) Contributions of anterior cingulate cortex and basolateral amygdala to decision confidence and learning under uncertainty. Nat Commun 10:4704. doi:10.1038/s41467-019-12725-1 pmid:31624264
    OpenUrlCrossRefPubMed
  164. ↵
    1. Stringer C,
    2. Pachitariu M,
    3. Steinmetz N,
    4. Reddy CB,
    5. Carandini M,
    6. Harris KD
    (2019) Spontaneous behaviors drive multidimensional, brainwide activity. Science 364:eaav7893. doi:10.1126/science.aav7893
    OpenUrlAbstract/FREE Full Text
  165. ↵
    1. Stroud JP,
    2. Porter MA,
    3. Hennequin G,
    4. Vogels TP
    (2018) Motor primitives in space and time via targeted gain modulation in cortical networks. Nat Neurosci 21:1774–1783. doi:10.1038/s41593-018-0276-0 pmid:30482949
    OpenUrlCrossRefPubMed
  166. ↵
    1. Sutskever I,
    2. Martens J,
    3. Hinton G
    (2011) Generating text with recurrent neural networks. In: Proceedings of the 28th International Conference on International Conference on Machine Learning, pp 1017–1024. Madison, WI: Omnipress.
  167. ↵
    1. Tachibana RO,
    2. Oosugi N,
    3. Okanoya K
    (2014) Semi-automatic classification of birdsong elements using a linear support vector machine. PLoS One 9:e92584. doi:10.1371/journal.pone.0092584 pmid:24658578
    OpenUrlCrossRefPubMed
  168. ↵
    1. Takahashi N,
    2. Oertner TG,
    3. Hegemann P,
    4. Larkum ME
    (2016) Active cortical dendrites modulate perception. Science 354:1587–1590. doi:10.1126/science.aah6066 pmid:28008068
    OpenUrlAbstract/FREE Full Text
  169. ↵
    1. Tasic B, et al
    . (2018) Shared and distinct transcriptomic cell types across neocortical areas. Nature 563:72–78. doi:10.1038/s41586-018-0654-5 pmid:30382198
    OpenUrlCrossRefPubMed
  170. ↵
    1. Tchernichovski O,
    2. Nottebohm F,
    3. Ho CE,
    4. Pesaran B,
    5. Mitra PP
    (2000) A procedure for an automated measurement of song similarity. Anim Behav 59:1167–1176. doi:10.1006/anbe.1999.1416 pmid:10877896
    OpenUrlCrossRefPubMed
  171. ↵
    1. Tenenbaum JB,
    2. Silva VD,
    3. Langford JC
    (2000) A global geometric framework for nonlinear dimensionality reduction. Science 290:2319–2323. doi:10.1126/science.290.5500.2319 pmid:11125149
    OpenUrlAbstract/FREE Full Text
  172. ↵
    1. Terra H,
    2. Bruinsma B,
    3. de Kloet SF,
    4. van der Roest M,
    5. Pattij T,
    6. Mansvelder HD
    (2020) Prefrontal cortical projection neurons targeting dorsomedial striatum control behavioral inhibition. Curr Biol 30:4188–4200.e5. doi:10.1016/j.cub.2020.08.031 pmid:32888489
    OpenUrlCrossRefPubMed
  173. ↵
    1. Thorat S,
    2. Aldegheri G,
    3. Kietzmann TC
    (2021) Category-orthogonal object features guide information processing in recurrent neural networks trained for object categorization. arXiv:2111.07898.
  174. ↵
    1. Titley HK,
    2. Brunel N,
    3. Hansel C
    (2017) Toward a neurocentric view of learning. Neuron 95:19–32. doi:10.1016/j.neuron.2017.05.021 pmid:28683265
    OpenUrlCrossRefPubMed
  175. ↵
    1. Tsuda B,
    2. Pate SC,
    3. Tye KM,
    4. Siegelmann HT,
    5. Sejnowski TJ
    (2022) Neuromodulators generate multiple context-relevant behaviors in a recurrent neural network by shifting activity hypertubes. bioRxiv 446462. doi:10.1101/2021.05.31.446462.
    OpenUrlAbstract/FREE Full Text
  176. ↵
    1. van der Maaten L,
    2. Hinton G
    (2008) Visualizing data using t-SNE. J Mach Learn Res 9:2579–2605.
    OpenUrlCrossRefPubMed
  177. ↵
    1. Vyas S,
    2. Golub MD,
    3. Sussillo D,
    4. Shenoy KV
    (2020) Computation through neural population dynamics. Annu Rev Neurosci 43:249–275. doi:10.1146/annurev-neuro-092619-094115 pmid:32640928
    OpenUrlCrossRefPubMed
  178. ↵
    1. Wallis JD,
    2. Anderson KC,
    3. Miller EK
    (2001) Single neurons in prefrontal cortex encode abstract rules. Nature 411:953–956. doi:10.1038/35082081 pmid:11418860
    OpenUrlCrossRefPubMed
  179. ↵
    1. Wang J,
    2. Narain D,
    3. Hosseini EA,
    4. Jazayeri M
    (2018) Flexible timing by temporal scaling of cortical responses. Nat Neurosci 21:102–110. doi:10.1038/s41593-017-0028-6 pmid:29203897
    OpenUrlCrossRefPubMed
  180. ↵
    1. Werbos PJ
    (1974) Beyond regression: new tools for prediction and analysis in the behavioral sciences. PhD dissertation, Harvard University.
  181. ↵
    1. Werbos PJ
    (1982) Applications of advances in nonlinear sensitivity analysis. In: System modeling and optimization (Drenick RF, Kozin F, eds), pp 762–770. Berlin: Springer.
  182. ↵
    1. Williams LE,
    2. Holtmaat A
    (2019) Higher-order thalamocortical inputs gate synaptic long-term potentiation via disinhibition. Neuron 101:91–102.e4. doi:10.1016/j.neuron.2018.10.049 pmid:30472077
    OpenUrlCrossRefPubMed
  183. ↵
    1. Wiltschko AB,
    2. Johnson MJ,
    3. Iurilli G,
    4. Peterson RE,
    5. Katon JM,
    6. Pashkovski SL,
    7. Abraira VE,
    8. Adams RP,
    9. Datta SR
    (2015) Mapping sub-second structure in mouse behavior. Neuron 88:1121–1135. doi:10.1016/j.neuron.2015.11.031 pmid:26687221
    OpenUrlCrossRefPubMed
  184. ↵
    1. Winnubst J, et al
    . (2019) Reconstruction of 1,000 projection neurons reveals new cell types and organization of long-range connectivity in the mouse brain. Cell 179:268–281.e13. pmid:31495573
    OpenUrlCrossRefPubMed
  185. ↵
    1. Wohlgemuth MJ,
    2. Sober SJ,
    3. Brainard MS
    (2010) Linked control of syllable sequence and phonology in birdsong. J Neurosci 30:12936–12949. doi:10.1523/JNEUROSCI.2690-10.2010 pmid:20881112
    OpenUrlAbstract/FREE Full Text
  186. ↵
    1. Wyatte D,
    2. Curran T,
    3. O'Reilly R
    (2012) The limits of feedforward vision: recurrent processing promotes robust object recognition when objects are degraded. J Cogn Neurosci 24:2248–2261. doi:10.1162/jocn_a_00282 pmid:22905822
    OpenUrlCrossRefPubMed
  187. ↵
    1. Yamins DL,
    2. Hong H,
    3. Cadieu CF,
    4. Solomon EA,
    5. Seibert D,
    6. DiCarlo JJ
    (2014) Performance-optimized hierarchical models predict neural responses in higher visual cortex. Proc Natl Acad Sci USA 111:8619–8624. doi:10.1073/pnas.1403112111 pmid:24812127
    OpenUrlAbstract/FREE Full Text
  188. ↵
    1. Yan Y,
    2. Goodman JM,
    3. Moore DD,
    4. Solla SA,
    5. Bensmaia SJ
    (2020) Unexpected complexity of everyday manual behaviors. Nat Commun 11:3564. doi:10.1038/s41467-020-17404-0 pmid:32678102
    OpenUrlCrossRefPubMed
  189. ↵
    1. Yang GR,
    2. Joglekar MR,
    3. Song HF,
    4. Newsome WT,
    5. Wang XJ
    (2019) Task representations in neural networks trained to perform many cognitive tasks. Nat Neurosci 22:297–306. doi:10.1038/s41593-018-0310-2 pmid:30643294
    OpenUrlCrossRefPubMed
Back to top

In this issue

The Journal of Neuroscience: 42 (45)
Journal of Neuroscience
Vol. 42, Issue 45
9 Nov 2022
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Masthead (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Recent Advances at the Interface of Neuroscience and Artificial Neural Networks
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Recent Advances at the Interface of Neuroscience and Artificial Neural Networks
Yarden Cohen, Tatiana A. Engel, Christopher Langdon, Grace W. Lindsay, Torben Ott, Megan A. K. Peters, James M. Shine, Vincent Breton-Provencher, Srikanth Ramaswamy
Journal of Neuroscience 9 November 2022, 42 (45) 8514-8523; DOI: 10.1523/JNEUROSCI.1503-22.2022

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Recent Advances at the Interface of Neuroscience and Artificial Neural Networks
Yarden Cohen, Tatiana A. Engel, Christopher Langdon, Grace W. Lindsay, Torben Ott, Megan A. K. Peters, James M. Shine, Vincent Breton-Provencher, Srikanth Ramaswamy
Journal of Neuroscience 9 November 2022, 42 (45) 8514-8523; DOI: 10.1523/JNEUROSCI.1503-22.2022
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Introduction
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Keywords

  • plasticity
  • artificial neural networks
  • neuromodulators
  • behavior
  • vision
  • cognition

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

  • Unraveling the Link between Olfactory Deficits and Neuropsychiatric Disorders
  • Cognitive-Affective Functions of the Cerebellum
  • Time for Memories
Show more Symposium and Mini-Symposium
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.