Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Viewpoints

On the Role of Theory and Modeling in Neuroscience

Daniel Levenstein, Veronica A. Alvarez, Asohan Amarasingham, Habiba Azab, Zhe S. Chen, Richard C. Gerkin, Andrea Hasenstaub, Ramakrishnan Iyer, Renaud B. Jolivet, Sarah Marzen, Joseph D. Monaco, Astrid A. Prinz, Salma Quraishi, Fidel Santamaria, Sabyasachi Shivkumar, Matthew F. Singh, Roger Traub, Farzan Nadim, Horacio G. Rotstein and A. David Redish
Journal of Neuroscience 15 February 2023, 43 (7) 1074-1088; https://doi.org/10.1523/JNEUROSCI.1179-22.2022
Daniel Levenstein
1Montreal Neurological Institute, McGill University, Montreal, Quebec H3A 2B4, Canada
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Veronica A. Alvarez
2Laboratory on Neurobiology of Compulsive Behaviors, National Institute on Alcohol Abuse and Alcoholism, National Institutes of Health, Bethesda, Maryland 20892
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Veronica A. Alvarez
Asohan Amarasingham
3Departments of Mathematics and Biology, City College and the Graduate Center, City University of New York, New York, New York 10032
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Habiba Azab
4Department of Neuroscience, Center for Magnetic Resonance Research, University of Minnesota, Minneapolis, Minnesota 55455
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Habiba Azab
Zhe S. Chen
5Department of Psychiatry, Neuroscience & Physiology, New York University School of Medicine, New York, New York, 10016
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Zhe S. Chen
Richard C. Gerkin
6School of Life Sciences, Arizona State University, Tempe, Arizona 85281
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Andrea Hasenstaub
7Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, California 94115
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Andrea Hasenstaub
Ramakrishnan Iyer
8Allen Institute for Brain Science, Seattle, Washington 98109
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Renaud B. Jolivet
9Maastricht Centre for Systems Biology, Maastricht University, Maastricht, The Netherlands
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Sarah Marzen
10W. M. Keck Science Department, Pitzer, Scripps, and Claremont McKenna Colleges, Claremont, California 91711
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Joseph D. Monaco
11Department of Biomedical Engineering, Johns Hopkins University School of Medicine, Baltimore, Maryland 21218
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Joseph D. Monaco
Astrid A. Prinz
12Department of Biology, Emory University, Atlanta, Georgia 30322
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Astrid A. Prinz
Salma Quraishi
13Neuroscience, Developmental and Regnerative Biology Department, University of Texas at San Antonio, San Antonio, Texas 78249
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Fidel Santamaria
13Neuroscience, Developmental and Regnerative Biology Department, University of Texas at San Antonio, San Antonio, Texas 78249
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Sabyasachi Shivkumar
14Brain and Cognitive Sciences, University of Rochester, Rochester, New York 14627
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Sabyasachi Shivkumar
Matthew F. Singh
15Department of Psychological & Brain Sciences, Department of Electrical & Systems Engineering, Washington University in St. Louis, St. Louis, Missouri 63112
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Roger Traub
16IBM T.J. Watson Research Center, AI Foundations, Yorktown Heights, New York 10598
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Roger Traub
Farzan Nadim
1Montreal Neurological Institute, McGill University, Montreal, Quebec H3A 2B4, Canada
7Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, California 94115
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Farzan Nadim
Horacio G. Rotstein
1Montreal Neurological Institute, McGill University, Montreal, Quebec H3A 2B4, Canada
7Department of Otolaryngology-Head and Neck Surgery, University of California San Francisco, San Francisco, California 94115
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for Horacio G. Rotstein
A. David Redish
18Department of Neuroscience, University of Minnesota, Minneapolis, Minnesota 55455
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • ORCID record for A. David Redish
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

In recent years, the field of neuroscience has gone through rapid experimental advances and a significant increase in the use of quantitative and computational methods. This growth has created a need for clearer analyses of the theory and modeling approaches used in the field. This issue is particularly complex in neuroscience because the field studies phenomena that cross a wide range of scales and often require consideration at varying degrees of abstraction, from precise biophysical interactions to the computations they implement. We argue that a pragmatic perspective of science, in which descriptive, mechanistic, and normative models and theories each play a distinct role in defining and bridging levels of abstraction, will facilitate neuroscientific practice. This analysis leads to methodological suggestions, including selecting a level of abstraction that is appropriate for a given problem, identifying transfer functions to connect models and data, and the use of models themselves as a form of experiment.

Introduction

Recent technological advances in neuroscience have prompted the growth of new experimental approaches and subfields that investigate phenomena from single neurons to social behavior. However, rapid growth has also revealed a need to develop new theoretical frameworks (Phillips, 2015) that integrate the growing quantities of data and to establish relationships between their underlying processes. While neuroscience has a strong history of interactions between experimental and theoretical approaches (Hodgkin and Huxley, 1952; O'Keefe and Nadel, 1978; Marr, 1982, 1991), there is still disagreement as to the nature of theory and its role in neuroscience, including how it should be developed, used, and evaluated by the community (Bialek, 2018; Goldstein, 2018).

We argue that an idealized view of scientific progress, in which science is a problem-solving enterprise that strives to explain phenomena, is well suited to inform scientific practice. In neuroscience, the phenomena of interest are those that pertain to neurons, the nervous system, and its contribution to cognition and behavior. Because these phenomena span a wide range of spatiotemporal scales, their explanations often require a “multilevel” approach that combines data from dramatically different modalities. Descriptive, mechanistic, and normative explanations each play distinct roles in building a multilevel account of neural phenomena: descriptive explanations delineate an abstract characterization of a phenomenon, while mechanistic and normative explanations bridge abstractions of different levels. Collectively, these operations unify scientific theories across disparate experimental approaches and fields. We show how this view facilitates the bidirectional interaction between theory and experimentation as well as theory development.

What is a theory and what is it good for?

Theories are the primary tools by which scientists make sense of observations and make predictions. Given this central role, it is surprising how little methodological attention is given in scientific training to the general nature of theories. Traditional descriptions of science tend to be based on the processes of theory identification and falsification, in which theories are proposed as universal truths about the world, tested, provisionally accepted if found to be compatible with experimental data, and rejected when found to be incompatible (Popper, 1959). According to these traditional descriptions, when theories are incompatible with experimental data, the conceptual framework on which they are based is called into question and a new framework is found that can better account for the data (Popper, 1959; Lakatos, 1980; Kuhn, 2012). However, historical, philosophical, and sociological analyses argue that these views do not account for how theory is used in practice (Lakatos, 1980; Feyerabend, 1993; Godfrey-Smith, 2003; Ben-Ari, 2011; Kaiser, 2014; Firestein, 2015; Laplane et al., 2019). For example, theories are rarely, if ever, decisively testable, scientists can have a variety of attitudes toward a theory rather than to simply accept or reject it (Lakatos, 1978; van Fraassen, 1980; Mermin, 1989; Ben-Ari, 2011; Kaiser, 2014); and although new discoveries can provide answers to open questions, the new questions they prompt may be more consequential (Firestein, 2012).

A pragmatic view: science as problem-solving

We propose that a pragmatic view of the scientific enterprise (James, 1907; Laudan, 1978; Ben-Ari, 2011; Douglas, 2014) is better suited to inform scientific practice. In this view, science is a process through which we solve empirical problems and answer questions about observable phenomena (Laudan, 1978; Nickles, 1981; Haig, 1987; Douglas, 2014; Firestein, 2015; Redish et al., 2018). Empirical problems can range from matters of basic scientific interest (for example, “How does the brain process visual signals?” or “How does an animal select between alternative choices?”), to those with more obvious applications (such as “Which brain functions are disrupted in schizophrenia?”). Like any other problem, a scientific problem can be seen as a search to achieve a desired goal, which is specified by the statement of the problem (Newell and Simon, 1972). However, scientific problems are often ill defined (Bechtel and Richardson, 2010), in part because the search space and solution criteria are not always explicitly stated and in part because they evolve with additional discoveries (Firestein, 2012). For example, the discovery of multiple memory and decision-making systems raises further questions of how those systems interact (Scoville and Milner, 1957; O'Keefe and Nadel, 1978; Squire, 1987; Nadel, 1994; Balleine and Dickinson, 1998; Schacter, 2001; Daw et al., 2005; Redish, 2013), while the question “How does the pineal gland generate consciousness?” (Descartes, 1637) is now considered outdated. Further, scientific problems are never definitively solved, but are only deemed “adequately solved” by a research community. What is seen as an adequate solution in one socio-historical context may not be in another: as new data become available, standards change, or alternative solutions are presented. While a continuously evolving landscape of problems and proposed solutions might seem to counter a notion of progress in science, scientific theories have been used to explain and control progressively more phenomena over the course of the scientific record (Laudan, 1978; Douglas, 2014). According to the pragmatic view, this progress results from community-maintained standards of explanation, under an overarching drive to better predict and control natural phenomena of potential relevance to society (Hacking, 1983; Douglas, 2014).

We can thus define a scientific explanation (Hempel and Oppenheim, 1948; Woodward, 2019) as a proposed solution to an empirical problem, and scientific theories to be the ideas we use to form explanations. Where traditional views have tried to specify the form theories take, the pragmatic view sees theory structure as closely tied to its function and context. As a result, a theory can include a wide and complex range of structural elements, including those that are not formalized (Winther, 2021). While theories may be spelled out in the scientific literature, they are more often used implicitly in the explanation of phenomena and design of experiments. By shifting theories from “proposals of truth to be falsified” to “proposed problem-solving tools,” the pragmatic view prompts us to assess a theory by its utility: what empirical problems it can solve, how easily it can be used to solve them, and how good its solutions are. It also requires criteria to evaluate the quality of solutions to a problem and a set of standards by which we measure the utility of the theory, such as accuracy, simplicity, falsifiability, generalizability, and reproducibility (Laudan, 1978; van Fraassen, 1980; Chang, 2007b, 2011; Schindler, 2018). Through competition to solve empirical problems, theories become more precise, provide clearer and more concise explanations, can be used to make more reliable and accurate predictions, and can be applied to larger domains.

Conceptual frameworks provide constructs and constraints

Assessing scientific explanations inevitably involves considerations that are not directly related to solution quality but are instead constraints on the form solutions can take. These constraints constitute a conceptual framework (Table 1): a language within which explanations are proposed. In effect, a conceptual framework is a set of foundational theories that provide a conceptual structure on which further theories within that program are built (Lakatos, 1978; Laudan, 1978; Kuhn, 2012).

View this table:
  • View inline
  • View popup
Table 1.

Terminology used in this manuscript: three neuroscience examples

The stability of such a framework allows its component theories to change without rebuilding their conceptual foundations. For example, under the modern framework of neuropsychiatry, psychiatric disorders are framed in terms of biophysical dysfunctions in neural structure. Current debates about the underpinnings of schizophrenia include hypotheses of dysfunction within dopaminergic or glutamatergic systems, dysfunctional pruning of dendrites, and dysfunctional oscillatory dynamics (Moghaddam and Javitt, 2012; Glausier and Lewis, 2013; Uhlhaas and Singer, 2015; Howes et al., 2017). However, they all lie within a general framework of biophysical changes in neural processes. The consistency of this founding idea allows us to modify theories without disrupting the foundational premise, which allows them to be directly compared and contrasted.

While explanations are naturally comparable within a framework, theories under different frameworks are composed of fundamentally different objects and describe the world in different terms, which makes them difficult to compare. For example, explanations under the traditional psychoanalytic framework (Luyten et al., 2015) are fundamentally different from those under the modern neuropsychiatry framework (American Psychiatric Association, 2013; Cuthbert and Insel, 2013; Insel and Cuthbert, 2015; World Health Organization, 2021). The two frameworks are composed of fundamentally different objects and are described in different terms: in contrast to the neuropsychiatric framework, psychoanalytic explanations for schizophrenia invoke unconscious conflicts and/or distorted ego functions as the key factors underlying psychosis (Luyten et al., 2015). Even the categorizations of psychiatric phenomena are different under these frameworks, making direct comparisons of explanations for the same phenomena across frameworks difficult (Feyerabend, 1993).

Despite the difficulties in directly comparing theories across frameworks, all frameworks are not equivalent. One can compare conceptual frameworks by asking how well their theories allow us to predict and control our environment (Lakatos, 1978). This is not to say that all research requires a direct application, but rather that consideration of practical components is necessary for a complete understanding of scientific progress (Laudan, 1978; Douglas, 2014). For example, the psychoanalytic framework implies treatment-based analytic therapy, while the modern neuropsychiatry framework suggests medication as a key component. Furthermore, under the new framework known as computational psychiatry, psychiatric disorders are attributed to computational “vulnerabilities” in the systems architecture of the brain (Redish, 2004; Redish et al., 2008; Montague et al., 2012; Huys et al., 2016; Redish and Gordon, 2016). Theories in this new framework suggest that such disorders would be treatable by changing information processing: by modifying the physical substrate (e.g., through electrical stimulation or pharmacological changes), enhancing compensation processes (e.g., through cognitive training), or changing the environment (e.g., by giving a student with ADHD extra time on a test). The pragmatic view suggests that the ultimate adoption (or not) of this framework will come down to how successfully it can be applied to unsolved problems.

Models as the interface between theories and phenomena

While “theoretical” work may appear further from “applied” science than its experimental counterpart, models can act as an interface between theory and phenomena. A model consists of a structure and an interpretation of how that structure relates to its target phenomena (Frigg and Hartmann, 2006), also known as the model's “construal” (Weisberg, 2013). For example, the equation τdV/dt=−V+Vrest is a mathematical structure that is interpreted to represent the temporal dynamics of the membrane potential, V, of a passive cell with time constant, τ, and resting potential, Vrest (Hodgkin and Huxley, 1952; Koch and Segev, 1989; Rall, 1992; Hille, 2001; Gerstner et al., 2014). Models whose structure consists of mathematical equations or computational processes are amenable to simulation and analytical treatment. Models can be constructed from many different kinds of interpreted structures, such as physical structures that are interpreted to represent the double helix of DNA (Watson and Crick, 1953) or diagrammatic structures that are interpreted to represent protein interactions involved in signaling cascades (Alon, 2006). Many “animal models” used in experimental neuroscience are physical structures interpreted to represent other phenomena, such as the 6-OHDA rat or the MPTP monkey, which are interpreted to represent the pathology of Parkinson's disease (Schultz et al., 1989; Dorval and Grill, 2014).

In creating a model, a researcher has to make foundational assumptions in the terms they use, the form those terms take, and the relationships between them. These assumptions instantiate aspects of a theory in an explicit expression with a well-defined form. The voltage equation above instantiates the theory that a neuron's electrical properties arise from a semipermeable membrane (Hodgkin and Huxley, 1952; Rall, 1992), while the 6-OHDA model instantiates the theory that Parkinson's disease arises from dopaminergic dysfunction (Langston and Palfreman, 2013). This explicit formulation of theories can force us to confront hidden assumptions (Marder, 2000), and provide useful insights for the design of experiments or potential engineering applications.

Further, in selecting some aspects of a phenomenon to include, and others to ignore, creating a model abstracts a multi-faceted phenomenon into a concise, but inevitably simplified, representation. Thus, models simultaneously act as an instantiation of a theory and an abstraction of a phenomenon (Rosenblueth and Wiener, 1945; Stafford, 2009). This dual role of models is the foundation of their use in explanation (Cartwright, 1997).

Descriptive, mechanistic, and normative explanation

The terms “descriptive,” “mechanistic,” and “normative” are widely used in neuroscience to describe various models. A pragmatic view prompts us to consider how these terms relate to the type of problem they are used to solve (Kording et al., 2020). In doing so, we find that these labels correspond to three different explanatory approaches in neuroscience, which are used to solve three different types of problems: “what” problems, “how” problems, and “why” problems (Dayan and Abbott, 2001) (Fig. 1).

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

The three explanatory processes that underlie scientific explanations. Descriptive theories address the question of “what is the phenomenon?” and identify the repeatable characteristics of that phenomenon. Mechanistic theories address the question of “how does the phenomenon arise?” and explains the phenomenon in terms of the parts and interactions of other phenomena at lower levels of abstraction. Normative theories address the question of “why do the phenomena exist?” and allow a comparison of the phenomenon to an identified function or goal. Normative theories allow the determination of whether a process is achieving its goal; inadequacies generally imply an incomplete understanding of the limitations engendered by processes at a lower level of abstraction.

Descriptive explanations

The first problem often encountered in scientific research is: What is the phenomenon? Phenomena are not divided into discrete entities a priori, but instead appear as a continuous multifaceted stream with many possible methods of observation and many aspects that could be observed. Thus, the set of characteristics that define a phenomenon are often unclear. This problem is addressed with a descriptive explanation (Kaplan and Bechtel, 2011). For example, to explain the spikes observed from a hippocampal neuron we could use a theory of “place cells” (O'Keefe and Nadel, 1978): a collection of ideas that defines the relationship between neural activity in the hippocampus and an animal's position in an environment, which can be instantiated in a model that specifies that relationship in an equation (O'Keefe and Nadel, 1978; Redish, 1999; Colgin, 2020). Descriptive models are founded on basic assumptions of which variables to observe and how to relate them. At its heart, a descriptive explanation is simply a selective account of phenomenological data; indeed, descriptive models are often called phenomenological models (Craver, 2007; Kaplan, 2011) or, when they are well established, phenomenological laws (Cartwright, 1997).

Mechanistic explanations

After addressing the “what” question, one might ask: How does the phenomenon arise? This problem is addressed with a mechanistic explanation, which explains a phenomenon in terms of its component parts and their interactions (Machamer et al., 2000; Craver, 2007; Bechtel and Richardson, 2010). For example, to explain the activity of place cells, we can create explanations based on afferent information from other structures, internal connectivity patterns, and intraneuronal processing, which can be instantiated in a model that specifies how they interact to produce neural firing (Redish, 1999; Hartley et al., 2000; Barry et al., 2006; Fuhs and Touretzky, 2006; Solstad et al., 2006; Giocomo et al., 2011; Sanders et al., 2015). A mechanistic model is founded on an assumption of which parts and processes are relevant, and illustrates how their interaction can produce a phenomenon or, equivalently, how the phenomenon can emerge from these parts. Often these parts are considered to be causally relevant to the phenomenon, and a mechanistic explanation is often also referred to as a causal explanation (Machamer et al., 2000; Craver, 2007; Bechtel and Richardson, 2010).

Mathematical mechanistic models in neuroscience often take the form of a dynamical system (Koch and Segev, 1989; Ellner and Guckenheimer, 2006; Izhikevich, 2007; Ermentrout and Terman 2010; Gerstner et al., 2014; Börgers, 2017; Gabbiani and Cox, 2017), in which a set of variables represent the temporal evolution of component processes or their equilibrium conditions. For example, the classic Hodgkin–Huxley model uses a set of four coupled differential equations to represent the dynamics of membrane potential and voltage-dependent conductances, and shows how an action potential can emerge from their interaction by producing a precise prediction of the progression of the membrane potential in time (Hodgkin and Huxley, 1952). However, qualitative mechanistic models, in which complex processes are summarized in schematic or conceptual structures that represent general properties of components and their interactions, are also commonly used. For example, Hebb considered a conceptualization of neural processing in which coincident firing of synaptically connected neurons strengthened the coupling between them. From this model, Hebb was able to propose how memories could be retrieved by the completion of partial patterns and how these processes could emerge from synaptic plasticity, as cells that were coactive during a particular stimulus or event would form assemblies with the ability to complete partially activated patterns (Hebb, 1949).

Mechanistic models represent the (assumed) underlying processes that produce the phenomenon (Craver, 2007; Kaplan and Bechtel, 2011). They can be used to make predictions about situations where the same processes are presumed to operate (Ellner and Guckenheimer, 2006). This includes the effects of manipulations to component parts, and circumstances beyond the scope of data used to calibrate the model.

Normative explanations

In addition to the mechanistic question of “how,” we can also ask the question: Why does the phenomenon exist? This kind of problem is addressed with a normative explanation, which is used to explain a phenomenon in terms of its function (Barlow, 1961; Kording et al., 2007; Bialek, 2012). A normative explanation of place cells would appeal to an animal's need to accurately encode its location, and could instantiate that need in a model of a navigation task (O'Keefe and Nadel, 1978; McNaughton and Nadel, 1990; Redish, 1999; Zilli and Hasselmo, 2008). Appealing to a system's function serves as a guiding concept that can be a powerful heuristic to explain its behavior based on what it ought to do to perform its function (Dennett, 1989). This kind of explanation has a long history in the form of teleological explanation, which explains a thing by its “purpose” (Aristotle, no date), and is often used implicitly in biological sciences, for example, stating that the visual system is “for” processing visual information. In neuroscience, functions often come in the form of cognitive, computational, or behavioral goals.

When quantified, normative models formalize the goal of the phenomenon in an objective function (also known as a utility or cost function), which defines what it means for a system to perform “well.” These models are founded on an assumed statement of a goal and the constraints under which the system operates. For example, models of retinal function formalize the goal of visual processing using equations that represent the ability to reconstruct a sensory signal from neural responses, under the constraints of sensory degradation and a limited number of noisy neurons (Rieke et al., 1997; Field and Rieke, 2002; Doi and Lewicki, 2014). Such an approach also relies on an assumption of an underlying optimization process. This assumption is often justified by appealing to evolution, which might be expected to optimize systems (Barlow, 1961; Parker and Smith, 1990; Bialek and Setayeshgar, 2008). However, evolution does not guarantee optimality because of limitations of genetic search (Gould and Lewontin, 1979; Gould, 1983). Moreover, there are numerous processes in physical, biological, neurologic, and social systems that can drive phenomena toward a state that maximizes or minimizes some objective function; however, these processes each also have their own unique limitations. For example, physical processes that minimize surface-to-volume ratio create hexagonal tessellations in beehives, but this process is limited by the physical properties of construction (Thompson, 1992; Smith et al., 2021). Economic markets might be expected to optimize the balance between offer and selling price, but are limited by imperfect and unbalanced information and the limited decision-making abilities of agents (Akerlof, 1978; Kahneman et al., 1991; Shleifer, 2000; Fox, 2009; Gigerenzer and Gaissmaier, 2011). Similarly, supervised learning might be expected to optimize object discrimination, but its implementation in the brain would be limited by constraints, such as synaptic locality and the availability of credit signals and training data (McNaughton et al., 1978; Häusser and Mel, 2003; Takeuchi et al., 2014; Richards et al., 2019; Hamrick et al., 2020; Hunt et al., 2021). Where each of these processes might be expected to bring systems toward an optimal solution, the constraints under which they operate may themselves impose distinct signatures on the systems they optimize.

The descriptive/mechanistic/normative classification depends on context

Theories and models do not exist in isolation, but are embedded in scientific practice. As the descriptive/mechanistic/normative categorization reflects the problem being solved, it can be applied to both theories and models depending on the context (i.e., kind of explanation) in which they are being used. In general, this categorization is independent of whether an explanation is accepted by the scientific community. For instance, a mechanistic explanation does not cease to be mechanistic if it is not adopted (e.g., because some of its predictions are not experimentally corroborated). Further, models with the same structure can be used for different purposes, and can thus be assigned to a different category in different contexts. For example, the integrate-and-fire model can be used as a descriptive model for membrane potential dynamics, or as a mechanistic model for the neuronal input-output transformation; and while the Hodgkin–Huxley model was discussed above as a mechanistic model for the problem of spike generation, it was originally proposed to be “an empirical description of the time course of the changes in permeability to sodium and potassium” (Hodgkin and Huxley, 1952). Indeed, theories often start as an effort to solve one class of problem, and over time develop aspects to address related problems of different classes, resulting in a theory with descriptive, mechanistic, and normative aspects.

Levels of abstraction

In selecting some aspects of a phenomenon to include, and others to ignore, a model abstracts a multifaceted phenomenon into a more concise, but inevitably simplified, representation. That is, in making a model, we replace a part of the universe with a simpler structure with arguably similar properties (Rosenblueth and Wiener, 1945; Weisberg, 2013). It could be argued that abstraction is detrimental to model accuracy (i.e., that “The best material model for a cat is another, or preferably the same cat”), and is only necessary in light of practical and cognitive limitations (Rosenblueth and Wiener, 1945). However, abstraction is important in scientific practice, and its role extends beyond addressing those limitations (Potochnik, 2017).

Box 1: levels of abstraction

An illustrative example of levels of abstraction comes from computer science (Colburn and Shute, 2007; Wing, 2008), in which higher-level languages abstract the details specified in lower-level languages by concealing detailed code in a single function that provides the same relationship. Computational abstraction simplifies a process, such that it is independent of its component processes or even its physical substrate. For example, there are many algorithms that sort a list of numbers, but any computational sort command produces the same output regardless of the algorithm used. Computational abstraction is used in neuroscience, for example, when we simplify the molecular process of synaptic transmission in a more abstract model that represents its net effect as an increased firing rate of a postsynaptic neuron. This simplification is akin to conceptual abstraction (O'Leary et al., 2015), by which more abstract, or idealized, models aim to capture general properties of a process rather than the specific details of any one event or dataset. In neuroscience, computational abstraction is often discussed in terms of David Marr's three levels of analysis (Marr, 1982; Pylyshyn, 1984): the implementational level is a low-level, concrete statement of a phenomenon, the algorithmic level is an abstraction of the implementational level, explaining the process by which the phenomenon occurs, and the computational level is a high-level (normative) statement of the goal of the process.

Distinct levels of abstraction also arise in neuroscience when considering problems at different spatiotemporal scales (Churchland and Sejnowski, 1994). For example, we might consider synaptic transmission in terms of the interactions of various proteins at nanometer to micrometer scales, or we might consider a more abstract model in which neural activity is propagated across the cortex at scales of millimeters or centimeters. When we model phenomena at a given spatiotemporal scale, we make an abstraction that prioritizes organizational details at that scale (e.g., cellular), while further simplifying details at others (e.g., subcellular and network) (Eronen and Brooks, 2018). One promising perspective on the emergence of spatiotemporal levels suggests that models at higher levels of abstraction arise from their lower level counterparts via a natural dimensionality reduction of the parameter space (Machta et al., 2013; Transtrum et al., 2015). Such a reduction is possible because models of complex systems are “sloppy”: they have a large number of dimensions in parameter space along which model parameters can vary without affecting relevant macroscopic observables (i.e., the microscopic parameters are degenerate with respect to macroscopic behavior) (Gutenkunst et al., 2007; for examples in neuroscience, see, e.g., Prinz et al., 2004; Panas et al., 2015). Thus, abstraction from lower to higher spatiotemporal scales can be seen as a reduction of the lower-level parameter space that removes sloppy dimensions but preserves “stiff” dimensions that have strong influence on observable properties at the higher level. The appropriate dimensionality reduction could be as simple as taking the mean or asymptote of some parameter over a population (Wilson and Cowan, 1973; Pinto et al., 1996; Destexhe and Sejnowski, 2009), or the set of microscopic parameters needed to produce the same macroscopic behavior might be nonlinear and complex (Prinz et al., 2004; Rotstein et al., 2006; Jalics et al., 2010; Transtrum et al., 2015).

Descriptive models define abstractions at different levels

Abstraction is most obvious when we consider the construction of descriptive explanations. First, abstractions are made when researchers decide which aspects of a phenomenon not to include. For example, the cable equation which describes the relationship between axonal conductance and membrane potential (Rinzel and Ermentrout, 1989; Rall, 1992; Gerstner et al., 2014) does not include details about intracellular organelles, the dynamics of individual ion channels, or the impact of nearby neurons on the extracellular potential. Importantly, these models do not include many larger-scale effects (e.g., the neuron's embedding in a circuit, or the social dynamics of the agent) as well as smaller scale factors (Vinogradov et al., 2022). The process of abstraction thus applies to both phenomena at smaller scales (organelles) and at larger scales (social interactions of the agent) that are hypothesized to be unnecessary to address the question at hand. Each of these factors are abstracted away, leaving only the features chosen to be represented in a model's structure.

Second, the aspects that are included must be represented in an idealized form. For instance, ionic flux through the cell membrane is not a strictly linear current function of voltage and conductance, but we often idealize it as such for tractability (Koch and Segev, 1989; Rall, 1992; Hille, 2001). These idealizations are assumptions about a phenomenon which are, strictly speaking, false, but are used because they serve some purpose in creating the model (Potochnik, 2017).

Classic accounts of neuroscience emphasize analysis at different levels of abstraction (Wimsatt, 1976; Marr, 1982; Sejnowski et al., 1988; Churchland and Sejnowski, 1994; Shepherd, 1994; Craver, 2007) (Box 1). However, despite the ubiquity of level-based views of neuroscience and a number of proposed schemes, no consensus can be found on what the relevant levels of abstraction are, or even what defines a level (Guttinger and Love, 2019). Suggestions of different level schemes range from those of computational abstraction (Colburn and Shute, 2007; Wing, 2008), which simplifies a process to be independent of its specific implementation or physical substrate, to levels of conceptual abstraction, which delineate the degree of idealization versus relatability to data (O'Leary et al., 2015), and levels of physical abstraction, which are used to deal with different spatiotemporal scales (Churchland and Sejnowski, 1994). However, recent analyses suggest that natural phenomena are not organized into levels in a universally coherent manner (Potochnik and McGill, 2012; Potochnik, 2017, 2020). From a pragmatic view, levels of abstraction need not reflect discrete “levels” in nature but are indicative of our problem-solving strategies and constraints. Because different abstractions can facilitate different research aims (Potochnik, 2017), multiple descriptive models are needed to represent the same phenomenon but abstract different features to different degrees.

Mechanistic and normative models connect levels of abstraction

Without links between them, we would be left with a hodgepodge of different descriptions. However, unification has been noted as a strong desideratum for scientific theories (Keas, 2018; Schindler, 2018). The relationship between different descriptions of the same phenomena can often be expressed in terms of a mechanistic explanation. For example, we might describe single-neuron activity in terms of membrane currents, or by listing a set of spike times: a natural reduction in the dimensionality that can result from many possible combinations of currents (Golowasch et al., 2002; Prinz et al., 2004). A mechanistic model (e.g., Hodgkin and Huxley, 1952) that demonstrates how spike times emerge from currents connects the descriptions at the two levels and, in addition, does so asymmetrically, as it does not claim to be a mechanism by which currents emerge from spike times. By bridging descriptions that each abstract different features to different degrees, mechanistic explanations create a multilevel “mosaic unity” in neuroscience (Craver, 2007), in which descriptions are grounded through their interconnections, and more abstract features are grounded in their emergence from less abstract counterparts (Oppenheim and Putnam, 1958; Craver, 2002, 2007; Bechtel, 2008; Kaplan and Bechtel, 2011).

In contrast, a normative explanation connects descriptions by appealing to the ability of less abstract features to satisfy a description of more abstract goals. For example, the mammalian hypothalamus could be described as maintaining body temperature like a thermostat (Morrison and Nakamura, 2011; Tan and Knight, 2018) or as a circuit of interconnected neurons. A normative model connects the two descriptions by explaining the negative feedback loop in the circuit through its ability to achieve those thermostatic functions. Because functions exist over a range of levels, from cellular to behavioral or computational, we could imagine a “multilevel” approach to understanding the mammalian hypothalamus that in turn uses the goal of a negative feedback loop to explain the developmental processes that establish hypothalamic connectivity. Like their mechanistic counterparts, normative explanations establish links between descriptions, which each have their own utility for different problems by virtue of their unique abstractions.

Thus, the threefold division of explanatory labor in neuroscience falls naturally into the different roles a model can play in terms of levels of abstraction. Descriptive explanations define abstractions of phenomena at different levels, while mechanistic and normative explanations bridge levels of abstraction. Descriptive models, rather than “mere” descriptions of phenomena (as they are sometimes dismissed), are the necessary foundation of both normative and mechanistic models. In turn, mechanistic and normative explanations connect a description at a “source” level to a description at a higher or lower “target” level (Fig. 2). Each of the terms that represent the components of mechanistic models and the constraints of normative models are descriptive models at a lower level of abstraction, while those that represent the emergent properties of mechanistic models and the goals of normative models are descriptive models at a higher level of abstraction. Given their multilevel nature, a dialogue between descriptive, normative, and mechanistic models is needed for a theoretical account of any neuroscientific phenomenon.

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Interactions between three explanatory processes and levels of abstraction. Descriptive explanations define an idealized abstraction of specific aspects of a phenomenon for discussion, measurement, and repeatability. Mechanistic explanations account for properties of a phenomenon by their emergence from less abstract phenomena, while normative explanations account for those properties by appealing to their ability to perform more abstract goals.

At what level of abstraction should a model be built?

As different abstractions trade-off advantages and disadvantages, the selection of which abstraction to use is highly dependent on the problem at hand (Herz et al., 2006). Current neuroscientific practice generally attempts two approaches for selecting the appropriate level of abstraction, which serve different purposes. The first approach is to try to find as low a level as possible that still includes experimentally supported details and accounts for the phenomenon. For example, one might explain the phenomenon of associative memories using compartmental models of pyramidal cell networks, including specific active conductances, dendritic compartments, pharmacological effects on different inputs arriving at different compartments and identifying the consequences for learning and recall (Hasselmo, 1993). The multiplicity of parameters and variables used in this approach provides many details that can be matched to observable features of a phenomenon and can capture unexpected properties that emerge from their interaction. However, these details need to be extensively calibrated to ensure the model is accurate, and can be very sensitive to missing, degenerate, or improperly tuned parameters (Traub et al., 1991, 1999). The second approach is to try to find the most abstract level that can still account for the phenomenon. For example, we might instead appeal to the classic Hopfield network, in which units are binary (+1, −1), connections are symmetrical, and are updated using a very simple asynchronous rule (Hopfield, 1982; Hertz et al., 1991). While more abstract models sacrifice the ability to make predictions about lower-level details, their insights are often more robust to specific (e.g., unobserved) physiological details, and by reducing a complicated system to a small number of effective parameters, they allow for powerful analysis on the influences to the system properties. Further, abstract models can provide conceptual benefits, such as intuition for how the system works and the ability to generalize to other systems that can be similarly abstracted (Gilead et al., 2012, 2019; O'Leary et al., 2015).

Another important consideration is the ability of models at different levels to interface with different experimental modalities or scientific fields. Every measurement is itself an abstraction, in that it is a reduced description of the part of the universe corresponding to the measurement (Chang, 2007b). For example, fMRI measures blood flow across wide swaths of cortex but abstracts away the interactions between individual neurons, while silicon probes measure extracellular voltage but abstract away intracellular processes, and calcium imaging measures neuronal calcium levels, but abstracts away the electrophysiology of neuronal spiking. All of these are discussed as “neural activity,” but they likely reflect different aspects of learning, performance, and dynamics. Moreover, subsequent processing abstracts these signals even further, such as correlation (functional connectomics) in fMRI, sorting voltage signals into putative cell “spiking” from silicon probes, and treating calcium transients as “events” from calcium imaging. The abstraction made by one measurement device might lend itself to explanations at a given level, but not others, and the measurements available are important considerations when selecting which abstractions to make in our models.

Similarly, models at different levels are often used by distinct scientific fields or communities. The existence of a literature with a rich body of relevant work can provide details and support for components of a model outside of the immediate problem of interest. Integrating theories and models across these different fields can be particularly beneficial for scientific progress (Grim et al., 2013; Wu et al., 2019); however, crossing levels can be a sociological problem as well as a methodological one because different fields of study often use different languages and operate under different conceptual frameworks.

In general, it is important that researchers spell out the abstractions being made in their models, including their purposes as well as their limitations. By being concrete about the abstractions made, researchers can increase the reliability of their theories. Importantly, as noted above, it is useful to acknowledge not only the simplifications made about smaller-scale phenomena, but also the simplifications made as to larger-scale interactions that have been abstracted away from a theory.

Theory development and experimentation

Traditional views emphasize the use of experiments to test proposed theories (Popper, 1959), and even consider an interplay in which theories suggest new experiments and unexpected experimental results reveal the need for new theories (Laudan, 1978; Firestein, 2015). However, theories do not arise fully formed but are developed over time through an interaction with experimentation (Laudan, 1978; Hacking, 1983; Bechtel, 2013; Douglas, 2014; Firestein, 2015). We now consider two crucial pieces of that dialogue: the domain of a theory, or phenomena it is intended to pertain to, and a translation function, which specifies how it should relate to phenomena in its domain. Experimentation plays two key roles in relation to theory: (1) grounding model assumptions and (2) assessing the quality of model-based explanations. We then discuss an often underappreciated form of experimentation, in which models themselves are the experimental subjects. These modeling experiments allow us to explore the (sometimes hidden or unexpected) implications of a theory itself, identify its underlying inconsistencies, and can be used to predict novel phenomena. Together, this reveals a picture in which theory development is not relegated to simply proposing theories to be tested, but instead entails a complex experimental paradigm in which models play an active role in the simultaneous development, assessment, and utilization of theories within explicit conceptual frameworks.

Linking theory and phenomena

The domain of a theory is the set of phenomena that it purports to explain (Mitchell et al., 1986; Redish, 1997; Kuhn, 2011, 2012). The domain is therefore a set of data-imposed constraints, and the theory should provide an explanation consistent with those constraints. Theoretical studies should be explicit about what phenomena do and do not lie in their intended domain. In practice, nascent theories are often evaluated not only by their ability to explain data in their proposed domain (Laudan, 1978; Feyerabend, 1993) but also by their potential to expand beyond the initial domain with further development (Lakatos, 1978). For example, the theory that action potentials arise from voltage-dependent changes in ionic permeability (Hodgkin and Huxley, 1952; Goldman and Morad, 1977; Katz, 1993; Hille, 2001) should apply to the domain of all action potentials in all neurons. Early theories of action potential function identified voltage-gated sodium currents as the primary depolarizing component and formalized their action in models that developed into the Hodgkin–Huxley framework (Hodgkin and Huxley, 1952). When some action potentials were later found to be independent of sodium concentrations, it was straightforward to incorporate other voltage-gated channels within the same framework (Koch and Segev, 1989; Hille, 2001; Gerstner et al., 2014).

By instantiating a theory in a specific structure (Rosenblueth and Wiener, 1945; Stafford 2009), models play a key role in connecting a theory to phenomena in its domain. However, no model is directly comparable to experimental data by virtue of its structure alone. As noted above, a model also consists of an interpretation of how that structure relates to its target phenomena (Weisberg, 2013). This interpretation is specified by a translation function: a statement of how the model's components map onto its target phenomena. A translation function may be as straightforward as “variable V represents the membrane potential in millivolts,” but it can also be less constrained, e.g., “variable V describes the slow changes in the membrane potential and ignores all spiking activity.” In other cases, the translation function can be complex, as parts of the model can have a loose correspondence to general features of large classes of data, and can represent highly abstract effective parameters or qualitative behaviors. For example, the units in Hopfield's attractor network models (Hopfield, 1982; Hopfield and Tank, 1985; Hertz et al., 1991) are not meant to directly correspond to measurable properties of biological neurons but are instead intended to reflect qualitative features, namely, that neural populations are “active” or not. In effect, the translation function spells out the abstractions made by the model. Specifying the translation function of a model is as important as defining its structure (Weisberg, 2013). While these descriptions are often provided for highly abstract models, models that describe finer spatiotemporal scales (e.g., detailed compartmental models of neurons) are often considered to be “biologically realistic” and assume a simple or obvious translation function. However, it is important to remember that these models are also abstractions, albeit at a different level, and a proper description of the abstractions made will help clarify both the uses and the limitations of such models. By specifying the intended correspondence between model terms and phenomena, the translation function operationalizes the concepts associated with those terms in the theory (Bridgman, 1927; Chang, 2007a).

Experiments ground model assumptions

With a well-defined translation function in hand, we can consider the ways in which models are informed by experimental data. As outlined above, the components of descriptive, mechanistic, and normative models are each based on a different set of foundational assumptions. These assumptions are not generally arbitrary but are informed by experimental observations and results.

Descriptive models are founded on an assumed relationship between variables, which is generally formulated to capture an observed regularity in experimental data. These initial observations often rely on “exploratory” experiments, which attempt to identify empirical regularities and the constructs with which to describe them (Steinle, 1997). In specifying the characteristic properties of a phenomenon, descriptive explanations delineate the attributes that are expected to be replicable in future experiments and play a foundational role in subsequent mechanistic and normative models. This is extremely important for the current replication controversy (Baker, 2016; Goodman et al., 2016; Fanelli, 2018; Redish et al., 2018). A recent National Academy report (National Academies of Sciences and Medicine, 2019) characterizes replicability as the ability to obtain consistent results across multiple studies, and contrasts it with reproducibility, defined as the ability to get the same results when applying the same analyses to the same data. Several authors have suggested that the replication crisis is indeed a crisis of theory development, as it is the scientific claims (not data) that should be replicable (Drugmonkey, 2018; Redish et al., 2018; Smaldino, 2019). We suggest that this crisis stems from three sources: (1) a failure to define domains correctly, assuming that limited observations correspond to a much larger range of phenomena than they actually do; (2) a failure to formalize observations in adequate descriptive models (e.g., an overreliance on correlation, or assumed simple relationships); and (3) a failure to connect those descriptive models with mechanistic or normative models that integrate descriptions at different levels of abstraction.

Mechanistic models are founded on a set of parts and interactions that are assumed to be relevant to a target phenomenon. The existence of candidate parts/interactions can be informed by experimental observations, and their relevance (or irrelevance) to a given target phenomenon is often derived from experimental or natural interventions (Pearl, 2009). Once the decision is made to include a part/interaction in a mechanistic model, its corresponding terms can be parameterized by virtue of the descriptive models at their source level of abstraction. For example, when trying to explain the phenomenon of burst spiking in thalamocortical neurons, we might observe the presence of a hyperpolarization-activated current (Ih) which, when blocked, disrupts burst spiking (McCormick and Pape, 1990). We can then calibrate the parameters used to model Ih with values acquired through slice experiments.

Experimental data can also inform the founding assumptions (goal/constraints) of normative models. For example, when trying to explain the responses of visual neurons, we might parameterize the constraints of an efficient coding model with data from retinal photoreceptors (Field and Rieke, 2002). As with mechanistic models, these normative parameters rely on the descriptive models we have for photoreceptor properties. However, grounding an assumed function (e.g., “vision”) in experimental data can be more challenging. This arises from a notable asymmetry between mechanistic and normative approaches: while the founding assumptions of a mechanistic model (parts/interactions) are less abstract than their target phenomena, the founding assumptions of normative approaches (a function/goal) are generally more abstract than the phenomenon they are used to explain. This often results in normative approaches being termed “top-down,” in contrast to “bottom-up” mechanistic modeling. In practice, functions are often operationalized via performance on a specified task, rendering them groundable in experimental data. For example, the assumed goal of primate facial recognition areas is grounded in the change in facial recognition abilities when those neural systems are manipulated or absent, neural responses to facial stimuli, and in the coupling of those areas with sensory and motor areas providing a behavioral circuit (Gross et al., 1969; Tsao et al., 2006; Grimaldi et al., 2016; Moeller et al., 2017).

Experiments assess solution quality

As has been noted by many previous authors, we cannot definitively “confirm” theories (Popper, 1959), nor can we definitively test/falsify the validity of a theory in isolation (Lakatos, 1980; Duhem, 1991). However, a theory's utility does not require absolute confidence in its validity, but only a track record of solving problems in its domain. By instantiating theories in a model with a well-defined translation function, we can assess the quality of solutions proposed with a given theory by comparing the behavior of those models to experimental observations.

In the case of descriptive models, model fitting can estimate confidence intervals and goodness of fit for the best-fitting parameter values, and can even be used to quantitatively compare candidate models to determine which can best explain experimental data with the fewest parameters. A researcher might build a mechanistic model with terms that correspond to the proposed parts to see if they are able to reproduce features of the data, or test the model's ability to predict the effect of experimental manipulations. Alternatively, a researcher can hypothesize that the system is performing some function, make a normative model that instantiates the goal, and see if properties of the data match those expected from a system optimizing that goal. In these “confirmatory” (theory-driven/hypothesis-testing) experiments, models are used to apply existing theories to account for observed phenomena, compare possible instantiations of a theory, or even compare theories with overlapping domains to see which better accounts for the phenomenon. In each case, the assumptions of the model act as a hypothesis and the degree of similarity between model and experimental data are used to assess the sufficiency of a theory (and its specific model instantiation) to account for a phenomenon.

However, the value of modeling is often in its ability to show insufficiency of a theory/model to account for experimental data. Rather than invalidating the theory, this can often prompt updates to the theory or a search for yet-unobserved relevant phenomena. For example, early models of head-direction tuning found that a mechanism based on attractor networks required recurrent connections not supported by anatomic data (Redish et al., 1996). This incompatibility led to subsequent analyses which found that the tuning curves were more complicated than originally described, matching those seen in the model without the recurrent connections (Blair et al., 1997). Similarly, the usefulness of normative models often lies in their ability to identify when a system is performing suboptimally (Parker and Smith, 1990). Such a finding can provide additional information about unexpected functions or constraints. When there is a mismatch between a normative model and observed phenomena, one could hypothesize that the agent is optimizing a different goal (Fehr and Schmidt, 1999; Binmore, 2005), new constraints that limit the processes available (Simon, 1972; Mullainathan, 2002), historical processes that could limit the optimization itself (Gould and Lewontin, 1979; Gould, 1983), or computational processes that limit the calculations available to the system (Nadel, 1994; Schacter, 2001; Redish, 2013; Webb et al., 2021). For instance, several studies have found that foraging subjects tend to remain at reward sites longer than needed (Camerer, 1997; Nonacs, 2001; Carter and Redish, 2016) and accept longer-delay offers than would be expected to maximize total reward (Wikenheiser et al., 2013; Sweis et al., 2018a,b; Schmidt et al., 2019). However, optimality could be restored by assuming an additional factor in the cost function (Simon, 1972) subsequently characterized as “regret”: an increased cost of making a mistake of one's own agency compared with equivalently poor outcomes that were not because of recognizable mistakes (Zeelenberg et al., 2000; Schmidt et al., 2005; Wikenheiser et al., 2013; Steiner and Redish, 2014; Sweis et al., 2018b, Coricelli et al., 2005). Similarly, Fehr and colleagues have found that normative explanations of behavior in a multiplayer game require an additional component with information about one's companion's success in addition to one's own, to account for the observed behavior (Fehr and Schmidt, 1999; Binmore, 2005; Fehr and Krajbich, 2014).

Modeling experiments explore theory implications

Confirmatory experiments can even be conducted without direct comparison to data, as phenomena at both the target and source levels of abstraction can be pure theoretical entities. Similar to their benchtop counterparts, we can treat different parameters or model instantiations as independent variables in the experiment, and test their sufficiency to account for different aspects of the phenomenon as the dependent variables (Omar et al., 2014; Gerkin et al., 2018). One can use these models as experiments to test the feasibility of theoretical claims in tractable idealized systems. For example, Hopfield's attractor network models (Hopfield, 1982; Hopfield and Tank, 1985) provided strong support for Hebb's theory (Hebb, 1949) that increased connectivity from coactive firing could create associative memory, by showing that strong connections between simple neuron-like entities were sufficient to produce cell assemblies that could be accessed through a pattern-completion process (Hertz et al., 1991).

Like their physical analogues (e.g., the 6-OHDA rat or the MPTP monkey), models can be used for exploratory experiments as well. Exploration of the Hopfield model (Kohonen, 1980, 1984; Hopfield, 1982; Hopfield and Tank, 1985) revealed novel properties of categorization, tuning curves, and pattern completion in the neuron-like entities, which were later identified experimentally (Rosch, 1983; Lakoff, 1990; Obermayer et al., 1992; Swindale and Bauer, 1998; Freedman et al., 2001, 2003; Obermayer and Sejnowski, 2001; Swindale, 2004; Wills et al., 2005; Yang and Shadlen, 2007; Colgin et al., 2010; de Villers-Sidani and Merzenich, 2011; Jezek et al., 2011; Nahum et al., 2013; Kelemen and Fenton, 2016). Exploratory modeling experiments can instantiate idealized aspects of a theory to help build intuition for the theory itself. Hopfield's model and its subsequent derivatives have provided researchers with a deeper understanding of how memories can be accessed by content through pattern-completion processes and given rise to concepts, such as “basins of attraction” (Hopfield, 1982; Hertz et al., 1991). These computational discoveries can help build understanding of the theory, and lead to predictions and ideas for new experiments.

Modeling experiments are especially useful in the context of theory development (Guest and Martin, 2020). When a phenomenon cannot be readily explained using an existing theory, assumptions can be made as the basis of a modeling experiment. The behavior of this model can then be used to evaluate the sufficiency of these assumptions to account for the phenomenon. Often, these modeling experiments precede a well-formed theory, and a theorist will perform numerous experiments with different models in the process of developing a theory (van Rooij and Baggio, 2020). Over time, specific successful model formulations can become closely associated with the theory and develop into its canonical instantiations that make the theory applicable to a wider range of problems and give more precise solutions.

Box 2: What makes a good neuroscientific theory? What makes a good model?

Be specific. A theory should be specific, particularly in terms of what the theory is attempting to explain and the strategies for doing so. The theory should define what problems it is trying to solve, and provide the criteria for an adequate solution. It is important to define the descriptive, mechanistic, and normative components of the theory and the rationale behind their selection.

Identify the domain. The theory should define the set of questions and problems that it is trying to solve. Importantly, this definition should be a reasonable space of phenomena such that it is easy for someone to determine whether a new experiment falls within the domain of the theory or not.

Specify which aspects of the theory are instantiated in each model, and how. As shown in Table 1, models instantiate theories, enabling them to be compared with data. It is important to specify which aspects of the theory are instantiated in the model and how those aspects are instantiated. It is also important to identify how those aspects were chosen, whether from experimental measurements, theoretical assumptions, or best-fit solutions (or arbitrarily).

Specify the translation function for all models. All models require translation to be compared with data. While sometimes those translations will be straightforward, they usually are not. However, even in situations where the translation is straightforward, being explicit about the translation function will make clear what data it explains and what experimental predictions it makes.

Identify the abstractions. Models at all levels can be useful, but to be useful, one must identify what aspects of the world are being abstracted away. It is important to include both abstractions of low-level phenomena and what additional (potentially higher-level) complexities are being ignored.

Define which aspects of the research are exploratory and which are confirmatory. The fact that models are a form of experiment creates a way forward for theoretical grant proposals. For example, a researcher can propose to build a model that crosses levels to address the question of theoretical viability. Such a proposal may have preliminary data to show that one can build models at each level, even if the researcher has not yet put those levels together. Similarly, a grant proposal can define the domain, even if the literature review is incomplete. One can also identify how one is going to explore the parameter space of a set of models to determine how those parameters affect phenomena across levels.

By being explicit about the scientific question being addressed, about the assumptions of the theory, the domain the theory is purporting to address, and the process of building and testing models underlying that theory, grant proposals could be viable even if the theory itself remains incomplete. We call on funding agencies and reviewers to recognize that theory is the foundation of any science, and that construction of rigorous theory and systematic computational modeling are time-consuming processes that require dedicated personnel with extensive training. Our hope is that the framework and associated language outlined in this document can be used to specify deliverables that can be understood by both funders and investigators.

In conclusion, a scientific theory is a thinking tool: a set of ideas used to solve specific problems. We can think of theoretical neuroscience as a field which approaches problems in neuroscience with the following problem-solving methodology: theories exist within conceptual frameworks and are instantiated in models which, by virtue of a translation function, can be used to assess a theory's ability to account for phenomena in the theory's domain or explore its further implications (Fig. 3).

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

How the various components discussed in this manuscript interact. The domain of a theory is the set of phenomena which it purports to explain. Theories are instantiated in models, which are an abstraction of phenomena in the domain, as specified by a translation function. By constraining the form solutions can take, a conceptual framework defines a way of looking at a problem, within which models and theories can be proposed. A given model can instantiate more than one theory, and a theory can be instantiated by more than one model.

We identified three kinds of explanations that play distinct roles in this process: those in which descriptive theories and models are used to define the abstractions by which we describe a phenomenon; those in which mechanistic theories and models are used to explain phenomena in terms of lower-level parts and their interactions; and those in which normative theories and models are used to explain phenomena in terms of a function at a higher level of abstraction.

These considerations lead to a more concrete view of theory in neuroscience under the pragmatic view: a theory is a set of assumptions available to be instantiated in models, whose adequacy for problems in their domain has been vetted via experimentation, and with a well-established translation function that defines their connection to phenomena. Over time and through the development of canonical model formulations, theories become more rigorous, such that researchers agree on how they should be implemented to explain specific domains. A theory in this sense is not a formal set of laws, but a continuously developing body of canonical models and model-phenomenon correspondences, bound together partly by history and partly by shared problem-solving methods and standards (Bechtel, 1993).

What recommendations can we take away from this perspective? First and foremost, scientists should be explicit about the underlying components of their theory. Reliability of theoretical work depends on being explicit about the domain that the theory purports to cover, the abstractions used (what has been ignored and left out), and the translation function to connect the theory to actual measurements. Furthermore, thinking of the pragmatic aspects suggests being explicit about what problems the work proposes to solve, what conceptual frameworks the theory fits in, and what the founding assumptions of the models are.

Finally, it is interesting to consider that we might apply our taxonomy to our own framework. The concept that “the ultimate goal of a theory is to provide tools that allow one to better explain and control one's environment” is a normative theory of the goal of scientific theories; the concept that “models instantiate theories and allow one to test their viability and their relationship to phenomena” is a mechanistic theory of how those theories achieve that goal; and the concept that “theories live within a framework that a community applies to them” is a descriptive theory of theories. One could imagine a metascientific research program which studies the available phenomena (e.g., the scientific literature) to test and further develop those theories, and even the use of models of the scientific process itself (e.g., Devezer et al., 2019). The benefits of such a research program could prove as impactful for scientific practice as other theories have proven for manipulation of phenomena in their domain.

Footnotes

  • This paper is the result of discussions as part of the workshop “Theoretical and Future Theoretical Frameworks in Neuroscience” (San Antonio, Feb 4-8, 2019) supported by the National Science Foundation Grants DBI-1820631 to H.G.R. and IOS-1516648 to F.S. This work was supported by National Institutes of Health Grant T90DA043219 and the Samuel J. and Joan B. Williamson Fellowship to D.L.; IRP-National Institutes of Health ZIA-AA000421 and DDIR Innovation Award, National Institutes of Health to V.A.A.; MH118928 to Z.S.C.; National Institute of Neurological Disorders and Stroke 1U19NS112953, National Institute on Deafness and Other Communication Disorders 1R01DC018455, National Institute of Mental Health 1R01MH106674, and NIBIB 1R01EB021711 to R.C.G.; National Institute on Deafness and Other Communication Disorders R01DC014101 and Hearing Research Incorporated, Sandler Foundation to A.H.; H2020 GAMMA-MRI (964644) and H2020 IN-FET (862882) to R.B.J.; National Institute of Neurological Disorders and Stroke 1R03NS109923 and National Science Foundation/NCS-FO 1835279 to J.D.M.; National Institute of Mental Health-NIBIB BRAIN Theories 1R01EB026939 to F.S.; IBM Exploratory Research Councils to R.T.; DOD ARO W911F-15-1-0426 to A.A.; National Science Foundation CRCNS-DMS-1608077 and National Science Foundation-IOS 2002863 to H.G.R.; National Institutes of Health MH060605 to F.N.; and National Institutes of Health MH080318, MH119569, and MH112688 to A.D.R. The authors further acknowledge the University of Texas at San Antonio Neuroscience Institute and the New Jersey Institute of Technology Department of Biological Sciences and Institute for Brain and Neuroscience Research for technical support in the organization of the workshop, as well as all of the participants in the workshop. We thank Erich Kummerfeld, Hal Greenwald, Kathryn McClain, Simón(e) Sun, and György Buzsáki for comments on parts of the manuscript; and Matt Chafee and Sophia Vinogradov for help with citations.

  • The authors declare no competing financial interests.

  • Correspondence should be addressed to A. David Redish at redish{at}umn.edu

SfN exclusive license.

References

  1. ↵
    1. Akerlof GA
    (1978) The market for 'lemons': quality uncertainty and the market mechanism. In: Uncertainty in economics, pp 235–251. Amsterdam: Elsevier.
  2. ↵
    1. Alon U
    (2006) An introduction to systems biology: design principles of biological circuits. London: Chapman and Hall/CRC.
  3. ↵
    American Psychiatric Association (2013) Diagnostic and Statistical Manual of Mental Disorders (DSM-5®). Washington, DC: American Psychiatric Association.
  4. ↵
    Aristotle (no date) Physics. Translated by R. P. Hardie and R. K. Gaye. Internet Classics Archive.
  5. ↵
    1. Baker M
    (2016) 1,500 scientists lift the lid on reproducibility. Nature 533:452–454. doi:10.1038/533452a pmid:27225100
    OpenUrlCrossRefPubMed
  6. ↵
    1. Balleine BW,
    2. Dickinson A
    (1998) Goal-directed instrumental action: contingency and incentive learning and their cortical substrates. Neuropharmacology 37:407–419. doi:10.1016/S0028-3908(98)00033-1 pmid:9704982
    OpenUrlCrossRefPubMed
  7. ↵
    1. Barlow HB
    (1961) Possible principles underlying the transformation of sensory messages. Sensory Commun 1:217–234.
    OpenUrl
  8. ↵
    1. Barry C,
    2. Lever C,
    3. Hayman R,
    4. Hartley T,
    5. Burton S,
    6. O'Keefe J,
    7. Jeffery K,
    8. Burgess N
    (2006) The boundary vector cell model of place cell firing and spatial memory. Rev Neurosci 17:71–97. doi:10.1515/revneuro.2006.17.1-2.71 pmid:16703944
    OpenUrlCrossRefPubMed
  9. ↵
    1. Bechtel W
    (1993) Integrating sciences by creating new disciplines: the case of cell biology. Biol Philos 8:277–299. doi:10.1007/BF00860430
    OpenUrlCrossRef
  10. ↵
    1. Bechtel W
    (2008) Mechanisms in cognitive psychology: what are the operations? Philos Sci 75:983–994.
    OpenUrlCrossRef
  11. ↵
    1. Bechtel W
    (2013) Philosophy of science: an overview for cognitive science. London: Psychology.
  12. ↵
    1. Bechtel W,
    2. Richardson RC
    (2010) Discovering complexity: decomposition and localization as strategies in scientific research. Cambridge, MA: Massachusetts Institute of Technology.
  13. ↵
    1. Ben-Ari M
    (2011) Just a theory: exploring the nature of science. Buffalo: Prometheus.
  14. ↵
    1. Bialek W
    (2012) Biophysics: searching for principles. Princeton, NJ: Princeton UP.
  15. ↵
    1. Bialek W
    (2018) Perspectives on theory at the interface of physics and biology. Rep Prog Phys 81:012601.
  16. ↵
    1. Bialek W,
    2. Setayeshgar S
    (2008) Cooperativity, sensitivity, and noise in biochemical signaling. Phys Rev Lett 100:258101. doi:10.1103/PhysRevLett.100.258101 pmid:18643705
    OpenUrlCrossRefPubMed
  17. ↵
    1. Binmore K
    (2005) Natural justice. Oxford: Oxford UP.
  18. ↵
    1. Blair HT,
    2. Lipscomb BW,
    3. Sharp PE
    (1997) Anticipatory time intervals of head-direction cells in the anterior thalamus of the rat, implications for path integration in the head-direction circuit. J Neurophysiol 78:145–159. doi:10.1152/jn.1997.78.1.145 pmid:9242269
    OpenUrlCrossRefPubMed
  19. ↵
    1. Börgers C
    (2017) An introduction to modeling neuronal dynamics. New York: Springer.
  20. ↵
    1. Bridgman PW
    (1927) The logic of modern physics. New York: Macmillan.
  21. ↵
    1. Camerer CF
    (1997) Taxi drivers and beauty contests. Eng Sci 60:10–19.
    OpenUrl
  22. ↵
    1. Carter EC,
    2. Redish AD
    (2016) Rats value time differently on equivalent foraging and delay-discounting tasks. J Exp Psychol Gen 145:1093–1101. pmid:27359127
    OpenUrlCrossRefPubMed
  23. ↵
    1. Cartwright N
    (1997) Models: the blueprints for laws. Philos Sci 64:S292–S303. doi:10.1086/392608
    OpenUrlCrossRef
  24. ↵
    1. Chang H
    (2007a) Inventing temperature: measurement and scientific progress. Oxford: Oxford UP.
  25. ↵
    1. Chang H
    (2007b) Scientific progress: beyond foundationalism and coherentism 1. R Inst Philos Suppl 61:1–20.
    OpenUrl
  26. ↵
    1. Chang H
    (2011) The persistence of epistemic objects through scientific change. Int J Analytic Philos 75:413–429.
    OpenUrl
  27. ↵
    1. Churchland P,
    2. Sejnowski TJ
    (1994) The computational brain. Cambridge, MA: Massachusetts Institute of Technology.
  28. ↵
    1. Colburn T,
    2. Shute G
    (2007) Abstraction in computer science. Minds Machines 17:169–184. doi:10.1007/s11023-007-9061-7
    OpenUrlCrossRef
  29. ↵
    1. Colgin LL
    (2020) Five decades of hippocampal place cells and EEG rhythms in behaving rats. J Neurosci 40:54–60. doi:10.1523/JNEUROSCI.0741-19.2019
    OpenUrlAbstract/FREE Full Text
  30. ↵
    1. Colgin LL,
    2. Leutgeb S,
    3. Jezek K,
    4. Leutgeb JK,
    5. Moser EI,
    6. McNaughton BL,
    7. Moser MB
    (2010) Attractor-map versus autoassociation based attractor dynamics in the hippocampal network. J Neurophysiol 104:35–50. doi:10.1152/jn.00202.2010
    OpenUrlCrossRefPubMed
  31. ↵
    1. Coricelli G,
    2. Critchley HD,
    3. Joffily M,
    4. O'Doherty JP,
    5. Sirigu A,
    6. Dolan RJ
    (2005) Regret and its avoidance: a neuroimaging study of choice behavior. Nat Neurosci 8:1255–1262. doi:10.1038/nn1514 pmid:16116457
    OpenUrlCrossRefPubMed
  32. ↵
    1. Craver CF
    (2002) Interlevel experiments and multilevel mechanisms in the neuroscience of memory. Philos Sci 69:S83–S97. doi:10.1086/341836
    OpenUrlCrossRef
  33. ↵
    1. Craver CF
    (2007) Explaining the brain: mechanisms and the mosaic unity of neuroscience. Oxford: Clarendon.
  34. ↵
    1. Cuthbert BN,
    2. Insel TR
    (2013) Toward the future of psychiatric diagnosis: the seven pillars of RDoC. BMC Med 11:126. doi:10.1186/1741-7015-11-126 pmid:23672542
    OpenUrlCrossRefPubMed
  35. ↵
    1. Daw ND,
    2. Niv Y,
    3. Dayan P
    (2005) Uncertainty-based competition between prefrontal and dorsolateral striatal systems for behavioral control. Nat Neurosci 8:1704–1711. doi:10.1038/nn1560 pmid:16286932
    OpenUrlCrossRefPubMed
  36. ↵
    1. Dayan P,
    2. Abbott LF
    (2001) Theoretical neuroscience. Cambridge, MA: Massachusetts Institute of Technology.
  37. ↵
    1. Dennett DC
    (1989) The intentional stance. Cambridge, MA: Massachusetts Institute of Technology.
  38. ↵
    1. Descartes R
    (1637) Discours de La Méthode Pour Bien Conduire Sa Raison, et Chercher La Vérité Dans Les Sciences.
  39. ↵
    1. Destexhe A,
    2. Sejnowski TJ
    (2009) The Wilson–Cowan model, 36 years later. Biol Cybern 101:1–2. doi:10.1007/s00422-009-0328-3 pmid:19662434
    OpenUrlCrossRefPubMed
  40. ↵
    1. Devezer B,
    2. Nardin LG,
    3. Baumgaertner B,
    4. Buzbas EO
    (2019) Scientific discovery in a model-centric framework: reproducibility, innovation, and epistemic diversity. PLoS One 14:e0216125. doi:10.1371/journal.pone.0216125
    OpenUrlCrossRefPubMed
  41. ↵
    1. Doi E,
    2. Lewicki M
    (2014) Optimal retinal population coding predicts inhomogeneous light adaptation and contrast sensitivity across the visual field. J Vis 14:1188. doi:10.1167/14.10.1188
    OpenUrlAbstract
  42. ↵
    1. Dorval AD,
    2. Grill WM
    (2014) Deep brain stimulation of the subthalamic nucleus reestablishes neuronal information transmission in the 6-OHDA rat model of parkinsonism. J Neurophysiol 111:1949–1959. doi:10.1152/jn.00713.2013 pmid:24554786
    OpenUrlCrossRefPubMed
  43. ↵
    1. Douglas H
    (2014) Pure science and the problem of progress. Stud Hist Philos Sci 46:55–63. doi:10.1016/j.shpsa.2014.02.001 pmid:25051872
    OpenUrlCrossRefPubMed
  44. ↵
    Drugmonkey (2018) Generalization, Not 'reproducibility.' Drugmonkey February 27, 2018.
  45. ↵
    1. Duhem PM
    (1991) The aim and structure of physical theory. Princeton, NJ: Princeton UP.
  46. ↵
    1. Ellner SP,
    2. Guckenheimer J
    (2006) Dynamic models in biology. Princeton, NJ: Princeton UP.
  47. ↵
    1. Ermentrout GB,
    2. Terman DH
    (2010) Mathematical foundations of neuroscience. New York: Springer Science and Business Media.
  48. ↵
    1. Eronen MI,
    2. Brooks DS
    (2018) Levels of organization in biology. In: The Stanford encyclopedia of philosophy (Zalta E, ed). Stanford, CA: Metaphysics Research Lab, Stanford University.
  49. ↵
    1. Fanelli D
    (2018) Opinion: is science really facing a reproducibility crisis, and do we need it to? Proc Natl Acad Sci USA 115:2628–2631. doi:10.1073/pnas.1708272114 pmid:29531051
    OpenUrlAbstract/FREE Full Text
  50. ↵
    1. Fehr E,
    2. Krajbich I
    (2014) Social preferences and the brain. In: Neuroeconomics, Ed 2, pp 193–218. Amsterdam: Elsevier.
  51. ↵
    1. Fehr E,
    2. Schmidt KM
    (1999) A theory of fairness, competition, and cooperation. Q J Econ 114:817–868. doi:10.1162/003355399556151
    OpenUrlCrossRef
  52. ↵
    1. Feyerabend P
    (1993) Against method. London: Verso.
  53. ↵
    1. Field GD,
    2. Rieke F
    (2002) Nonlinear signal transfer from mouse rods to bipolar cells and implications for visual sensitivity. Neuron 34:773–785. doi:10.1016/s0896-6273(02)00700-6 pmid:12062023
    OpenUrlCrossRefPubMed
  54. ↵
    1. Firestein S
    (2015) Failure: why science is so successful. Oxford: Oxford UP.
  55. ↵
    1. Fox J
    (2009) The myth of the rational market: a history of risk, reward, and delusion on Wall Street. New York: HarperCollins.
  56. ↵
    1. Freedman DJ,
    2. Riesenhuber M,
    3. Poggio T,
    4. Miller EK
    (2001) Categorical representation of visual stimuli in the primate prefrontal cortex. Science 291:312–316. doi:10.1126/science.291.5502.312 pmid:11209083
    OpenUrlAbstract/FREE Full Text
  57. ↵
    1. Freedman DJ,
    2. Riesenhuber M,
    3. Poggio T,
    4. Miller EK
    (2003) A comparison of primate prefrontal and inferior temporal cortices during visual categorization. J Neurosci 23:5235–5246.
    OpenUrlAbstract/FREE Full Text
  58. ↵
    1. Frigg R,
    2. Hartmann S
    (2006) Models in science. In Stanford Encyclopedia of Philsophy (Zalta EN, ed). Stanford CA: Metaphysics Research Lab.
  59. ↵
    1. Fuhs MC,
    2. Touretzky DS
    (2006) A spin glass model of path integration in rat medial entorhinal cortex. J Neurosci 26:4266–4276. doi:10.1523/JNEUROSCI.4353-05.2006 pmid:16624947
    OpenUrlAbstract/FREE Full Text
  60. ↵
    1. Gabbiani F,
    2. Cox SJ
    (2017) Mathematics for neuroscientists. San Diego: Academic.
  61. ↵
    1. Gerkin RC,
    2. Jarvis RJ,
    3. Crook SM
    (2018) Towards systematic, data-driven validation of a collaborative, multi-scale model of Caenorhabditis elegans. Philos Trans R Soc Lond B Biol Sci 373:20170381.
    OpenUrlCrossRefPubMed
  62. ↵
    1. Gerstner W,
    2. Kistler WM,
    3. Naud R,
    4. Paninski L
    (2014) Neuronal Dynamics: From single neurons to networks and models of cognition. Cambridge UK: Cambridge University Press.
  63. ↵
    1. Gigerenzer G,
    2. Gaissmaier W
    (2011) Heuristic decision making. Annu Rev Psychol 62:451–482. doi:10.1146/annurev-psych-120709-145346 pmid:21126183
    OpenUrlCrossRefPubMed
  64. ↵
    1. Gilead M,
    2. Liberman N,
    3. Maril A
    (2012) Construing counterfactual worlds: the role of abstraction. Eur J Soc Psychol 42:391–397. doi:10.1002/ejsp.1862
    OpenUrlCrossRef
  65. ↵
    1. Gilead M,
    2. Trope Y,
    3. Liberman N
    (2019) Above and beyond the concrete: the diverse representational substrates of the predictive brain. Behav Brain Sci 43:e121.
    OpenUrl
  66. ↵
    1. Giocomo LM,
    2. Moser MB,
    3. Moser EI
    (2011) Computational models of grid cells. Neuron 71:589–603. doi:10.1016/j.neuron.2011.07.023 pmid:21867877
    OpenUrlCrossRefPubMed
  67. ↵
    1. Glausier JR,
    2. Lewis DA
    (2013) Dendritic spine pathology in schizophrenia. Neuroscience 251:90–107. doi:10.1016/j.neuroscience.2012.04.044 pmid:22546337
    OpenUrlCrossRefPubMed
  68. ↵
    1. Godfrey-Smith P
    (2003) An introduction to the philosophy of science: theory and reality. Chicago: University of Chicago.
  69. ↵
    1. Goldman Y,
    2. Morad M
    (1977) Ionic membrane conductance during the time course of the cardiac action potential. J Physiol 268:655–695. doi:10.1113/jphysiol.1977.sp011876 pmid:560474
    OpenUrlCrossRefPubMed
  70. ↵
    1. Goldstein RE
    (2018) Are theoretical results 'results?' Elife 7:e40018. doi:10.7554/Elife.40018
    OpenUrlCrossRef
  71. ↵
    1. Golowasch J,
    2. Goldman MS,
    3. Abbott LF,
    4. Marder E
    (2002) Failure of averaging in the construction of a conductance-based neuron model. J Neurophysiol 87:1129–1131. doi:10.1152/jn.00412.2001 pmid:11826077
    OpenUrlCrossRefPubMed
  72. ↵
    1. Goodman SN,
    2. Fanelli D,
    3. Ioannidis JPA
    (2016) What does research reproducibility mean? Sci Transl Med 8:341ps12. doi:10.1126/scitranslmed.aaf5027 pmid:27252173
    OpenUrlFREE Full Text
  73. ↵
    1. Gould SJ
    (1983) Hen's teeth and horse's toes. New York: Norton.
  74. ↵
    1. Gould SJ,
    2. Lewontin RC
    (1979) The Spandrels of San Marco and the Panglossian paradigm: a critique of the adaptationist programme. Proc R Soc Lond B 205:581–598.
    OpenUrlCrossRef
  75. ↵
    1. Grimaldi P,
    2. Saleem KS,
    3. Tsao D
    (2016) Anatomical connections of the functionally defined 'face patches' in the macaque monkey. Neuron 90:1325–1342. doi:10.1016/j.neuron.2016.05.009 pmid:27263973
    OpenUrlCrossRefPubMed
  76. ↵
    1. Grim P,
    2. Singer DJ,
    3. Fisher S,
    4. Bramson A,
    5. Berger WJ,
    6. Reade C,
    7. Flocken C,
    8. Sales A
    (2013) Scientific networks on data landscapes: question difficulty, epistemic success, and convergence. Episteme 10:441–464. doi:10.1017/epi.2013.36
    OpenUrlCrossRef
  77. ↵
    1. Gross CG,
    2. Bender DB,
    3. Rocha-Miranda CE
    (1969) Visual receptive fields of neurons in inferotemporal cortex of the monkey. Science 166:1303–1306. doi:10.1126/science.166.3910.1303 pmid:4982685
    OpenUrlAbstract/FREE Full Text
  78. ↵
    1. Guest O,
    2. Martin AE
    (2020) How computational modeling can force theory building in psychological science. Perspect Psychol Sci 16:789–802.
    OpenUrl
  79. ↵
    1. Gutenkunst RN,
    2. Waterfall JJ,
    3. Casey FP,
    4. Brown KS,
    5. Myers CR,
    6. Sethna JP
    (2007) Universally sloppy parameter sensitivities in systems biology models. PLoS Comput Biol 3:1871–1878. doi:10.1371/journal.pcbi.0030189 pmid:17922568
    OpenUrlCrossRefPubMed
  80. ↵
    1. Guttinger S,
    2. Love AC
    (2019) Characterizing scientific failure: putting the replication crisis in context. EMBO Rep 20:e48765. doi:10.15252/embr.201948765 pmid:31379051
    OpenUrlCrossRefPubMed
  81. ↵
    1. Hacking I
    (1983) Representing and intervening: introductory topics in the philosophy of natural science. Cambridge: Cambridge UP.
  82. ↵
    1. Haig BD
    (1987) Scientific problems and the conduct of research. Educ Philos Theory 19:22–32.
    OpenUrl
  83. ↵
    1. Hamrick JB,
    2. Friesen AL,
    3. Behbahani F,
    4. Guez A,
    5. Viola F,
    6. Witherspoon S,
    7. Anthony T,
    8. Buesing L,
    9. Veličković P,
    10. Weber T
    (2020) On the role of planning in model-based deep reinforcement learning. arXiv:2011.04021. doi:10.48550/arXiv.2011.04021.
    OpenUrlCrossRef
  84. ↵
    1. Hartley T,
    2. Burgess N,
    3. Lever C,
    4. Cacucci F,
    5. O'Keefe J
    (2000) Modeling place fields in terms of the cortical inputs to the hippocampus. Hippocampus 10:369–379. doi:10.1002/1098-1063(2000)10:4<369::AID-HIPO3>3.0.CO;2-0
    OpenUrlCrossRefPubMed
  85. ↵
    1. Hasselmo ME
    (1993) Acetylcholine and learning in a cortical associative memory. Neural Comput 5:32–44. doi:10.1162/neco.1993.5.1.32
    OpenUrlCrossRef
  86. ↵
    1. Häusser M,
    2. Mel B
    (2003) Dendrites: bug or feature? Curr Opin Neurobiol 13:372–383. doi:10.1016/s0959-4388(03)00075-8 pmid:12850223
    OpenUrlCrossRefPubMed
  87. ↵
    1. Hebb DO
    (1949) The organization of behavior. New York: Wiley.
  88. ↵
    1. Hempel CG,
    2. Oppenheim P
    (1948) Studies in the logic of explanation. Philos Sci 15:135–175. doi:10.1086/286983
    OpenUrlCrossRef
  89. ↵
    1. Hertz J,
    2. Krogh A,
    3. Palmer RG
    (1991) Introduction to the theory of neural computation. Reading MA: Addison-Wesley.
  90. ↵
    1. Herz AV,
    2. Gollisch T,
    3. Machens CK,
    4. Jaeger D
    (2006) Modeling single-neuron dynamics and computations: a balance of detail and abstraction. Science 314:80–85. doi:10.1126/science.1127240 pmid:17023649
    OpenUrlAbstract/FREE Full Text
  91. ↵
    1. Hille B
    (2001) Ion channels of excitable membranes. Sunderland MA: Sinauer.
  92. ↵
    1. Hodgkin AL,
    2. Huxley AF
    (1952) A quantitative description of membrane current and its application to conduction and excitation in nerve. J Physiol 117:500–544. doi:10.1113/jphysiol.1952.sp004764 pmid:12991237
    OpenUrlCrossRefPubMed
  93. ↵
    1. Hopfield JJ
    (1982) Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA 79:2554–2558. doi:10.1073/pnas.79.8.2554 pmid:6953413
    OpenUrlAbstract/FREE Full Text
  94. ↵
    1. Hopfield JJ,
    2. Tank D
    (1985) 'Neural' computation of decisions in optimization problems. Biol Cybern 52:141–152. doi:10.1007/BF00339943 pmid:4027280
    OpenUrlCrossRefPubMed
  95. ↵
    1. Howes OD,
    2. McCutcheon R,
    3. Owen MJ,
    4. Murray RM
    (2017) The role of genes, stress, and dopamine in the development of schizophrenia. Biol Psychiatry 81:9–20. doi:10.1016/j.biopsych.2016.07.014 pmid:27720198
    OpenUrlCrossRefPubMed
  96. ↵
    1. Hunt LT,
    2. Daw ND,
    3. Kaanders P,
    4. MacIver MA,
    5. Mugan U,
    6. Procyk E,
    7. Redish AD,
    8. Russo E,
    9. Scholl J,
    10. Stachenfeld K,
    11. Wilson CR,
    12. Kolling N
    (2021) Formalizing planning and information search in naturalistic decision-making. Nat Neurosci 24:1051–1064. doi:10.1038/s41593-021-00866-w
    OpenUrlCrossRef
  97. ↵
    1. Huys QJ,
    2. Maia TV,
    3. Frank MJ
    (2016) Computational psychiatry as a bridge from neuroscience to clinical applications. Nat Neurosci 19:404–413. doi:10.1038/nn.4238 pmid:26906507
    OpenUrlCrossRefPubMed
  98. ↵
    1. Insel TR,
    2. Cuthbert BN
    (2015) Medicine. Brain disorders? Precisely. Science 348:499–500. doi:10.1126/science.aab2358 pmid:25931539
    OpenUrlAbstract/FREE Full Text
  99. ↵
    1. Izhikevich EM
    (2007) Dynamical systems in neuroscience. Cambridge, MA: Massachusetts Institute of Technology.
  100. ↵
    1. Jalics J,
    2. Krupa M,
    3. Rotstein HG
    (2010) A novel mechanism for mixed-mode oscillations in a neuronal model. Dynamic Syst 4:445–482.
    OpenUrl
  101. ↵
    1. James W
    (1907) Pragmatism: a new name for some old ways of thinking. London: Longmans, Green, and Co.
  102. ↵
    1. Jezek K,
    2. Henriksen EJ,
    3. Treves A,
    4. Moser EI,
    5. Moser MB
    (2011) Theta-paced flickering between place-cell maps in the hippocampus. Nature 478:246–249. doi:10.1038/nature10439 pmid:21964339
    OpenUrlCrossRefPubMed
  103. ↵
    1. Kahneman D,
    2. Knetsch JL,
    3. Thaler RH
    (1991) The endowment effect, loss aversion, and status quo bias. J Econ Perspect 5:193–206. doi:10.1257/jep.5.1.193
    OpenUrlCrossRef
  104. ↵
    1. Kaiser D
    (2014) History: shut up and calculate! Nature 505:153–155. doi:10.1038/505153a pmid:24409511
    OpenUrlCrossRefPubMed
  105. ↵
    1. Kaplan DM
    (2011) Explanation and description in computational neuroscience. Synthese 183:339–373. doi:10.1007/s11229-011-9970-0
    OpenUrlCrossRef
  106. ↵
    1. Kaplan DM,
    2. Bechtel W
    (2011) Dynamical models: an alternative or complement to mechanistic explanations? Top Cogn Sci 3:438–444. doi:10.1111/j.1756-8765.2011.01147.x
    OpenUrlCrossRef
  107. ↵
    1. Katz AM
    (1993) Cardiac ion channels. N Engl J Med 328:1244–1251. doi:10.1056/NEJM199304293281707 pmid:7681934
    OpenUrlCrossRefPubMed
  108. ↵
    1. Keas MN
    (2018) Systematizing the theoretical virtues. Synthese 195:2761–2793. doi:10.1007/s11229-017-1355-6
    OpenUrlCrossRef
  109. ↵
    1. Kelemen E,
    2. Fenton AA
    (2016) Coordinating different representations in the hippocampus. Neurobiol Learn Mem 129:50–59. doi:10.1016/j.nlm.2015.12.011 pmid:26748023
    OpenUrlCrossRefPubMed
  110. ↵
    1. Koch C,
    2. Segev I
    , eds (1989) Methods in neuronal modeling. Cambridge, MA: Massachusetts Institute of Technology.
  111. ↵
    1. Kohonen T
    (1980) Content-addressable memories. New York: Springer.
  112. ↵
    1. Kohonen T
    (1984) Self-organization and associative memory. New York: Springer.
  113. ↵
    1. Kording KP,
    2. Tenenbaum JB,
    3. Shadmehr R
    (2007) The dynamics of memory as a consequence of optimal adaptation to a changing body. Nat Neurosci 10:779–786. doi:10.1038/nn1901 pmid:17496891
    OpenUrlCrossRefPubMed
  114. ↵
    1. Kording KP,
    2. Blohm G,
    3. Schrater P,
    4. Kay K
    (2020) Appreciating the variety of goals in computational neuroscience. arXiv:2002.03211. doi:10.48550/arXiv.2002.03211.
    OpenUrlCrossRef
  115. ↵
    1. Kuhn TS
    (2011) The essential tension. Chicago: University of Chicago.
  116. ↵
    1. Kuhn TS
    (2012) The structure of scientific revolutions: 50th anniversary edition. Chicago: University of Chicago.
  117. ↵
    1. Lakatos I
    (1978) Science and pseudoscience. Philosophical papers. Cambridge: Cambridge UP.
  118. ↵
    1. Lakatos I
    (1980) The methodology of scientific research programmes: Volume 1. Philosophical papers. Cambridge: Cambridge UP.
  119. ↵
    1. Lakoff G
    (1990) Women, fire, and dangerous things. Chicago: University of Chicago.
  120. ↵
    1. Langston JW,
    2. Palfreman J
    (2013) The case of the frozen addicts: how the solution of a medical mystery revolutionized the understanding of Parkinson's disease. New York: Pantheon books.
  121. ↵
    1. Laplane L,
    2. Mantovani P,
    3. Adolphs R,
    4. Chang H,
    5. Mantovani A,
    6. McFall-Ngai M,
    7. Rovelli C,
    8. Sober E,
    9. Pradeu T
    (2019) Why science needs philosophy. Proc Natl Acad Sci USA 116:3948–3952. doi:10.1073/pnas.1900357116
    OpenUrlFREE Full Text
  122. ↵
    1. Laudan L
    (1978) Progress and its problems: towards a theory of scientific growth. Los Angeles: University of California.
  123. ↵
    1. Luyten P,
    2. Mayes LC,
    3. Fonagy P,
    4. Target M,
    5. Blatt SJ
    (2015) Handbook of psychodynamic approaches to psychopathology. New York: Guilford.
  124. ↵
    1. Machamer P,
    2. Darden L,
    3. Craver CF
    (2000) Thinking about mechanisms. Philos Sci 67:1–25. doi:10.1086/392759
    OpenUrlCrossRef
  125. ↵
    1. Machta BB,
    2. Chachra R,
    3. Transtrum MK,
    4. Sethna JP
    (2013) Parameter space compression underlies emergent theories and predictive models. Science 342:604–607. doi:10.1126/science.1238723 pmid:24179222
    OpenUrlAbstract/FREE Full Text
  126. ↵
    1. Marder E
    (2000) Models identify hidden assumptions. Nat Neurosci 3:1198–1198. doi:10.1038/81477
    OpenUrlCrossRef
  127. ↵
    1. Marr D
    (1982) Vision. New York: Freeman.
  128. ↵
    1. Marr D
    (1991) From the Retina to the Neocortex. Boston MA: Birkhauser.
  129. ↵
    1. McCormick DA,
    2. Pape HC
    (1990) Properties of a hyperpolarization-activated cation current and its role in rhythmic oscillation in thalamic relay neurones. J Physiol 431:291–318. doi:10.1113/jphysiol.1990.sp018331 pmid:1712843
    OpenUrlCrossRefPubMed
  130. ↵
    1. McNaughton BL,
    2. Douglas RM,
    3. Goddard GV
    (1978) Synaptic enhancement in fascia dentata: cooperativity among coactive afferents. Brain Res 157:277–293. doi:10.1016/0006-8993(78)90030-6 pmid:719524
    OpenUrlCrossRefPubMed
  131. ↵
    1. McNaughton BL,
    2. Nadel L
    (1990) Hebb-Marr networks and the neurobiological representation of action in space. In: Neuroscience and connectionist theory (Gluck MA, Rumelhart DE, eds) pp 1–63. Hillsdale NJ: Erlbaum.
  132. ↵
    1. Mermin ND
    (1989) What's wrong with this pillow? Physics Today 42:9–11. doi:10.1063/1.2811173
    OpenUrlCrossRef
  133. ↵
    1. Mitchell TM,
    2. Keller RM,
    3. Kedar-Cabelli ST
    (1986) Explanation-based generalization: A unifying view. Machine learning. 1:47–80. doi:10.1023/A:1022691120807
    OpenUrlCrossRef
  134. ↵
    1. Moeller S,
    2. Crapse T,
    3. Chang L,
    4. Tsao DY
    (2017) The effect of face patch microstimulation on perception of faces and objects. Nat Neurosci 20:743–752. doi:10.1038/nn.4527 pmid:28288127
    OpenUrlCrossRefPubMed
  135. ↵
    1. Moghaddam B,
    2. Javitt D
    (2012) From revolution to evolution: the glutamate hypothesis of schizophrenia and its implication for treatment. Neuropsychopharmacology 37:4–15. doi:10.1038/npp.2011.181 pmid:21956446
    OpenUrlCrossRefPubMed
  136. ↵
    1. Montague PR,
    2. Dolan RJ,
    3. Friston KJ,
    4. Dayan P
    (2012) Computational psychiatry. Trends Cogn Sci 16:72–80. doi:10.1016/j.tics.2011.11.018 pmid:22177032
    OpenUrlCrossRefPubMed
  137. ↵
    1. Morrison SF,
    2. Nakamura K
    (2011) Central neural pathways for thermoregulation. Front Biosci (Landmark Ed) 16:74–104. doi:10.2741/3677 pmid:21196160
    OpenUrlCrossRefPubMed
  138. ↵
    1. Mullainathan S
    (2002) A memory-based model of bounded rationality. Q J Econ 117:735–774. doi:10.1162/003355302760193887
    OpenUrlCrossRef
  139. ↵
    1. Nadel L
    (1994) Multiple memory systems: what and why, an update. In: Memory systems (Schacter DL, Tulving E, eds), pp 39–64. Cambridge MA: Massachusetts Institute of Technology.
  140. ↵
    1. Nahum M,
    2. Lee H,
    3. Merzenich MM
    (2013) Principles of neuroplasticity-based rehabilitation. Prog Brain Res 207:141–171. doi:10.1016/B978-0-444-63327-9.00009-6 pmid:24309254
    OpenUrlCrossRefPubMed
  141. ↵
    National Academies of Sciences and Medicine (2019) Reproducibility and replicability in science. Washington, DC: National Academies of Sciences and Medicine.
  142. ↵
    1. Newell A,
    2. Simon HA
    (1972) Human problem solving. Hoboken, NJ: Prentice-Hall.
  143. ↵
    1. Nickles T
    (1981) What is a problem that we may solve it? Synthese 47:85–118. doi:10.1007/BF01064267
    OpenUrlCrossRef
  144. ↵
    1. Nonacs P
    (2001) State dependent patch use and the marginal value theorem. Behav Ecol 12:71–83. doi:10.1093/oxfordjournals.beheco.a000381
    OpenUrlCrossRef
  145. ↵
    1. Obermayer K,
    2. Sejnowski TJ
    (2001) Self-organizing map formation: foundations of neural computation. Cambridge, MA: Massachusetts Institute of Technology.
  146. ↵
    1. Obermayer K,
    2. Blasdel GG,
    3. Schulten K
    (1992) Statistical-mechanical analysis of self-organization and pattern formation during the development of visual maps. Phys Rev A 45:7568–7589. doi:10.1103/PhysRevA.45.7568
    OpenUrlCrossRefPubMed
  147. ↵
    1. O'Keefe J,
    2. Nadel L
    (1978) The hippocampus as a cognitive map. Oxford: Clarendon.
  148. ↵
    1. O'Leary T,
    2. Sutton AC,
    3. Marder E
    (2015) Computational models in the age of large datasets. Curr Opin Neurobiol 32:87–94. doi:10.1016/j.conb.2015.01.006 pmid:25637959
    OpenUrlCrossRefPubMed
  149. ↵
    1. Omar CJ,
    2. Aldrich J,
    3. Gerkin RC
    (2014) Collaborative infrastructure for test-driven scientific model validation. 36th International Conference on Software Engineering. doi:10.1145/2591062.2591129
    OpenUrlCrossRef
  150. ↵
    1. Oppenheim P,
    2. Putnam H
    (1958) Unity of science as a working hypothesis. Minn Stud Philos Sci 2:3–36.
    OpenUrl
  151. ↵
    1. Panas D,
    2. Amin H,
    3. Maccione A,
    4. Muthmann O,
    5. van Rossum M,
    6. Berdondini L,
    7. Hennig MH
    (2015) Sloppiness in spontaneously active neuronal networks. J Neurosci 35:8480–8492. doi:10.1523/JNEUROSCI.4421-14.2015 pmid:26041916
    OpenUrlAbstract/FREE Full Text
  152. ↵
    1. Parker GA,
    2. Smith JM
    (1990) Optimality theory in evolutionary biology. Nature 348:27–33. doi:10.1038/348027a0
    OpenUrlCrossRef
  153. ↵
    1. Pearl J
    (2009) Causality: models, reasoning and inference. Cambridge: Cambridge UP.
  154. ↵
    1. Phillips R
    (2015) Theory in biology: Figure 1 or Figure 7? Trends Cell Biol 25:723–729. doi:10.1016/j.tcb.2015.10.007 pmid:26584768
    OpenUrlCrossRefPubMed
  155. ↵
    1. Pinto DJ,
    2. Brumberg JC,
    3. Simons DJ,
    4. Ermentrout GB
    (1996) A quantitative population model of whisker barrels: re-examining the Wilson-Cowan equations. J Comput Neurosci 3:247–264. doi:10.1007/BF00161134 pmid:8872703
    OpenUrlCrossRefPubMed
  156. ↵
    1. Popper KR
    (1959) The logic of scientific discovery. New York: Science.
  157. ↵
    1. Potochnik A
    (2017) Idealization and the aims of science. Chicago: University of Chicago.
  158. ↵
    1. Potochnik A
    (2020) Idealization and many aims. Philos Sci 87:933–943.
    OpenUrl
  159. ↵
    1. Potochnik A,
    2. McGill B
    (2012) The limitations of hierarchical organization. Philos Sci 79:120–140. doi:10.1086/663237
    OpenUrlCrossRef
  160. ↵
    1. Prinz AA,
    2. Bucher D,
    3. Marder E
    (2004) Similar network activity from disparate circuit parameters. Nat Neurosci 7:1345–1352. doi:10.1038/nn1352
    OpenUrlCrossRefPubMed
  161. ↵
    1. Pylyshyn ZW
    (1984) Computation and cognition: toward a foundation for cognitive science. Cambridge, MA: Massachusetts Institute of Technology.
  162. ↵
    1. Rall W
    (1992) Cable theory for dendritic neurons. In: Methods in neuronal modeling (Koch C, Segev I, eds), pp 9–62. Cambridge, MA: Massachusetts Institute of Technology.
  163. ↵
    1. Redish AD
    (1997) Beyond the cognitive map: contributions to a computational neuroscience theory of rodent navigation (Touretzky DS, ed). Ann Arbor, MI: Carnegie Mellon UP.
  164. ↵
    1. Redish AD
    (1999) Beyond the cognitive map: from place cells to episodic memory. Cambridge MA: Massachusetts Institute of Technology.
  165. ↵
    1. Redish AD
    (2004) Addiction as a computational process gone awry. Science 306:1944–1947.
    OpenUrlAbstract/FREE Full Text
  166. ↵
    1. Redish AD
    (2013) The mind within the brain: how we make decisions and how those decisions go wrong. Oxford: Oxford UP.
  167. ↵
    1. Redish AD,
    2. Gordon JA
    , eds (2016) Computational psychiatry: new perspectives on mental illness. In: Strüngmann Forum Reports. Cambridge MA: Massachusetts Institute of Technology.
  168. ↵
    1. Redish AD,
    2. Elga AN,
    3. Touretzky DS
    (1996) A coupled attractor model of the rodent head direction system. Network 7:671–685. doi:10.1088/0954-898X_7_4_004
    OpenUrlCrossRef
  169. ↵
    1. Redish AD,
    2. Jensen S,
    3. Johnson A
    (2008) A unified framework for addiction: vulnerabilities in the decision process. Behav Brain Sci 31:415–487. doi:10.1017/S0140525X0800472X pmid:18662461
    OpenUrlCrossRefPubMed
  170. ↵
    1. Redish AD,
    2. Kummerfeld E,
    3. Morris RL,
    4. Love AC
    (2018) Opinion: reproducibility failures are essential to scientific inquiry. Proc Natl Acad Sci USA 115:5042–5046. doi:10.1073/pnas.1806370115 pmid:29765001
    OpenUrlFREE Full Text
  171. ↵
    1. Richards BA, et al
    . (2019) A deep learning framework for neuroscience. Nat Neurosci 22:1761–1770. doi:10.1038/s41593-019-0520-2 pmid:31659335
    OpenUrlCrossRefPubMed
  172. ↵
    1. Rieke F,
    2. Warland D,
    3. de Ruyter van Steveninck R,
    4. Bialek W
    (1997) Spikes. Cambridge MA: Massachusetts Institute of Technology.
  173. ↵
    1. Rinzel J,
    2. Ermentrout GB
    (1989) Analysis of neural excitability and oscillations. In: Methods in neuronal modeling (Koch C, Segev I, eds), pp 135–169. Cambridge, MA: Massachusetts Institute of Technology.
  174. ↵
    1. Rosch E
    (1983) Prototype classification and logical classification: the two systems. In: New trends in cognitive representation (Scholnick EF, ed), pp 73–86. Mahwah NJ: Erlbaum.
  175. ↵
    1. Rosenblueth A,
    2. Wiener N
    (1945) The role of models in science. Philos Sci 12:316–321. doi:10.1086/286874
    OpenUrlCrossRef
  176. ↵
    1. Rotstein HG,
    2. Oppermann T,
    3. White JA,
    4. Kopell N
    (2006) A reduced model for medial entorhinal cortex stellate cells: subthreshold oscillations, spiking and synchronization. J Comput Neurosci 21:271–292. doi:10.1007/s10827-006-8096-8 pmid:16927211
    OpenUrlCrossRefPubMed
  177. ↵
    1. Sanders H,
    2. Rennó-Costa C,
    3. Idiart M,
    4. Lisman J
    (2015) Grid cells and place cells: an integrated view of their navigational and memory function. Trends Neurosci 38:763–775. doi:10.1016/j.tins.2015.10.004 pmid:26616686
    OpenUrlCrossRefPubMed
  178. ↵
    1. Schacter DL
    (2001) The seven sins of memory. Orlando, FL: Houghton Mifflin.
  179. ↵
    1. Schindler S
    (2018) Theoretical virtues in science: uncovering reality through theory. Cambridge: Cambridge UP.
  180. ↵
    1. Schmidt B,
    2. Duin AA,
    3. Redish AD
    (2019) Disrupting the medial prefrontal cortex alters hippocampal sequences during deliberative decision making. J Neurophysiol 121:1981–2000. doi:10.1152/jn.00793.2018
    OpenUrlCrossRef
  181. ↵
    1. Schultz W,
    2. Studer A,
    3. Romo R,
    4. Sundstrom E,
    5. Jonsson G,
    6. Scarnati E
    (1989) Deficits in reaction times and movement times as correlates of hypokinesia in monkeys with MPTP-induced striatal dopamine depletion. J Neurophysiol 61:651–668. doi:10.1152/jn.1989.61.3.651 pmid:2785168
    OpenUrlCrossRefPubMed
  182. ↵
    1. Scoville WB,
    2. Milner B
    (1957) Loss of recent memory after bilateral hippocampal lesions. J Neurol Neurosurg Psychiatry 20:11–21. doi:10.1136/jnnp.20.1.11 pmid:13406589
    OpenUrlFREE Full Text
  183. ↵
    1. Sejnowski T,
    2. Koch C,
    3. Churchland P
    (1988) Computational neuroscience. Science 241:1299–1306. doi:10.1126/science.3045969
    OpenUrlAbstract/FREE Full Text
  184. ↵
    1. Shepherd GM
    (1994) Neurobiology. Oxford: Oxford UP.
  185. ↵
    1. Shleifer A
    (2000) Inefficient markets: an introduction to behavioral finance. Oxford: OUP.
  186. ↵
    1. Simon HA
    (1972) Theories of bounded rationality. Decis Organ 1:161–176.
    OpenUrl
  187. ↵
    1. Smaldino P
    (2019) Better methods can't make up for mediocre theory. Nature 575:9. doi:10.1038/d41586-019-03350-5
    OpenUrlCrossRef
  188. ↵
    1. Smith ML,
    2. Napp N,
    3. Petersen KH
    (2021) Imperfect comb construction reveals the architectural abilities of honeybees. Proc Natl Acad Sci USA 118:e2103605118. doi:10.1073/pnas.2103605118
    OpenUrlAbstract/FREE Full Text
  189. ↵
    1. Solstad T,
    2. Moser EI,
    3. Einevoll GT
    (2006) From grid cells to place cells: a mathematical model. Hippocampus 16:1026–1031. doi:10.1002/hipo.20244 pmid:17094145
    OpenUrlCrossRefPubMed
  190. ↵
    1. Squire LR
    (1987) Memory and brain. Oxford: Oxford UP.
  191. ↵
    1. Stafford T
    (2009) What use are computational models of cognitive processes?” In: Connectionist models of behaviour and cognition: II. Proceedings of the 11th Neural Computation and Psychology Workshop (Mayor J, Ruh N, Plunkett K, eds). Singapore: World Scientific. doi:10.1142/9789812834232_0022
  192. ↵
    1. Steiner AP,
    2. Redish AD
    (2014) Behavioral and neurophysiological correlates of regret in rat decision-making on a neuroeconomic task. Nat Neurosci 17:995–1002. doi:10.1038/nn.3740 pmid:24908102
    OpenUrlCrossRefPubMed
  193. ↵
    1. Steinle F
    (1997) Entering new fields: exploratory uses of experimentation. Philos Sci 64:S65–S74. doi:10.1086/392587
    OpenUrlCrossRef
  194. ↵
    1. Sweis BM,
    2. Abram SV,
    3. Schmidt BJ,
    4. Seeland KD,
    5. MacDonald AW,
    6. Thomas MJ,
    7. Redish DA
    (2018a) Sensitivity to 'sunk costs' in mice, rats, and humans. Science 361:178–181. doi:10.1126/science.aar8644 pmid:30002252
    OpenUrlAbstract/FREE Full Text
  195. ↵
    1. Sweis BM,
    2. Thomas MJ,
    3. Redish DA
    (2018b) Mice learn to avoid regret. PLoS Biol 16:e2005853. doi:10.1371/journal.pbio.2005853 pmid:29927938
    OpenUrlCrossRefPubMed
  196. ↵
    1. Swindale NV
    (2004) How different feature spaces may be represented in cortical maps. Network 15:217–242. pmid:15600232
    OpenUrlPubMed
  197. ↵
    1. Swindale NV,
    2. Bauer HU
    (1998) Application of Kohonen's self-organizing feature map algorithm to cortical maps of orientation and direction preference. Proc R Soc Lond B Biol Sci 265:827–838. doi:10.1098/rspb.1998.0367
    OpenUrlCrossRef
  198. ↵
    1. Takeuchi T,
    2. Duszkiewicz AJ,
    3. Morris RG
    (2014) The synaptic plasticity and memory hypothesis: encoding, storage and persistence. Philos Trans R Soc Lond B Biol Sci 369:20130288. doi:10.1098/rstb.2013.0288
    OpenUrlCrossRefPubMed
  199. ↵
    1. Tan CL,
    2. Knight ZA
    (2018) Regulation of body temperature by the nervous system. Neuron 98:31–48. doi:10.1016/j.neuron.2018.02.022 pmid:29621489
    OpenUrlCrossRefPubMed
  200. ↵
    1. Thompson DW
    (1992) On growth and form. North Chelmsford, MA: Courier.
  201. ↵
    1. Transtrum MK,
    2. Machta BB,
    3. Brown KS,
    4. Daniels BC,
    5. Myers CR,
    6. Sethna JP
    (2015) Perspective: sloppiness and emergent theories in physics, biology, and beyond. J Chem Phys 143:010901. doi:10.1063/1.4923066 pmid:26156455
    OpenUrlCrossRefPubMed
  202. ↵
    1. Traub RD,
    2. Jefferys JG,
    3. Whittington MA
    (1999) Fast oscillations in cortical circuits. Cambridge MA: Massachusetts Institute of Technology.
  203. ↵
    1. Traub RD,
    2. Wong RK,
    3. Miles R,
    4. Michelson H
    (1991) A model of a CA3 hippocampal pyramidal neuron incorporating voltage-clamp data on intrinsic conductances. J Neurophysiol 66:635–650. doi:10.1152/jn.1991.66.2.635 pmid:1663538
    OpenUrlCrossRefPubMed
  204. ↵
    1. Tsao DY,
    2. Freiwald WA,
    3. Tootell RB,
    4. Livingstone MS
    (2006) A cortical region consisting entirely of face-selective cells. Science 311:670–674. doi:10.1126/science.1119983 pmid:16456083
    OpenUrlAbstract/FREE Full Text
  205. ↵
    1. Uhlhaas PJ,
    2. Singer W
    (2015) Oscillations and neuronal dynamics in schizophrenia: the search for basic symptoms and translational opportunities. Biol Psychiatry 77:1001–1009. doi:10.1016/j.biopsych.2014.11.019
    OpenUrlCrossRefPubMed
  206. ↵
    1. van Fraassen BC
    (1980) The scientific image. Oxford: Clarendon.
  207. ↵
    1. van Rooij I,
    2. Baggio G
    (2020) Theory before the test: how to build high-verisimilitude explanatory theories in psychological science. Perspect Psychol Sci 16:682–697.
    OpenUrl
  208. ↵
    1. Villers-Sidani E,
    2. Merzenich MM
    (2011) Lifelong plasticity in the rat auditory cortex: basic mechanisms and role of sensory experience. Prog Brain Res 191:119–131. pmid:21741548
    OpenUrlCrossRefPubMed
  209. ↵
    1. Vinogradov S,
    2. Hamid A,
    3. Redish DA
    (2022) Etiopathogenic models of psychosis spectrum illnesses must resolve four key features. Biol Psychiatry 92:514–522. doi:10.1016/j.biopsych.2022.06.024 pmid:35931575
    OpenUrlCrossRefPubMed
  210. ↵
    1. Watson JD,
    2. Crick FH
    (1953) Molecular structure of nucleic acids: a structure for deoxyribose nucleic acid. Nature 171:737–738. doi:10.1038/171737a0 pmid:13054692
    OpenUrlCrossRefPubMed
  211. ↵
    1. Webb R,
    2. Glimcher PW,
    3. Louie K
    (2021) The normalization of consumer valuations: context-dependent preferences from neurobiological constraints. Manage Sci 67:93–125. doi:10.1287/mnsc.2019.3536
    OpenUrlCrossRef
  212. ↵
    1. Weisberg M
    (2013) Simulation and similarity: using models to understand the world. Oxford UK: Oxford University Press.
  213. ↵
    1. Wikenheiser A,
    2. Stephens DW,
    3. Redish AD
    (2013) Subjective costs drive overly-patient foraging strategies in rats on an intertemporal foraging task. Proc Natl Acad Sci USA 110:8308–8313. doi:10.1073/pnas.1220738110 pmid:23630289
    OpenUrlAbstract/FREE Full Text
  214. ↵
    1. Wills TJ,
    2. Lever C,
    3. Cacucci F,
    4. Burgess N,
    5. O'Keefe J
    (2005) Attractor dynamics in the hippocampal representation of the local environment. Science 308:873–876. doi:10.1126/science.1108905 pmid:15879220
    OpenUrlAbstract/FREE Full Text
  215. ↵
    1. Wilson HR,
    2. Cowan JD
    (1973) A mathematical theory of the functionaldynamics of cortical and thalamic tissue. Kybernetik 13:55–80. doi:10.1007/BF00288786 pmid:4767470
    OpenUrlCrossRefPubMed
  216. ↵
    1. Wimsatt WC
    (1976) Reductionism, levels of organization, and the mind-body problem. In: Consciousness and the brain: a scientific and philosophical inquiry (Globus GG, Maxwell G, Savodnik I, eds), pp 205–267. New York: Springer.
  217. ↵
    1. Wing JM
    (2008) Computational thinking and thinking about computing. Philos Trans A Math Phys Eng Sci 366:3717–3725. doi:10.1098/rsta.2008.0118 pmid:18672462
    OpenUrlCrossRefPubMed
  218. ↵
    1. Winther RG
    (2021) The structure of scientific theories. In: The Stanford encyclopedia of philosophy (Zalta EN, ed). Stanford, CA: Metaphysics Research Lab, Stanford University.
  219. ↵
    1. Woodward J
    (2019) Scientific explanation. In: The Stanford encyclopedia of philosophy (Zalta EN, ed). Stanford, CA: Metaphysics Research Lab, Stanford University.
  220. ↵
    World Health Organization (2021) International classification of diseases, 11th revision (ICD-11). Geneva: World Health Organization.
  221. ↵
    1. Wu L,
    2. Wang D,
    3. Evans JA
    (2019) Large teams develop and small teams disrupt science and technology. Nature 566:378–382. doi:10.1038/s41586-019-0941-9
    OpenUrlCrossRefPubMed
  222. ↵
    1. Yang T,
    2. Shadlen MN
    (2007) Probabilistic reasoning by neurons. Nature 447:1075–1080. doi:10.1038/nature05852 pmid:17546027
    OpenUrlCrossRefPubMed
  223. ↵
    1. Zeelenberg M,
    2. Van Dijk WW,
    3. Manstead AS,
    4. van der Pligt J
    (2000) On bad decisions and disconfirmed expectancies: the psychology of regret and disappointment. Cogn Emot 14:521–541. doi:10.1080/026999300402781
    OpenUrlCrossRef
  224. ↵
    1. Zilli EA,
    2. Hasselmo ME
    (2008) Modeling the role of working memory and episodic memory in behavioral tasks. Hippocampus 18:193–209. doi:10.1002/hipo.20382 pmid:17979198
    OpenUrlCrossRefPubMed
Back to top

In this issue

The Journal of Neuroscience: 43 (7)
Journal of Neuroscience
Vol. 43, Issue 7
15 Feb 2023
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Masthead (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
On the Role of Theory and Modeling in Neuroscience
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
On the Role of Theory and Modeling in Neuroscience
Daniel Levenstein, Veronica A. Alvarez, Asohan Amarasingham, Habiba Azab, Zhe S. Chen, Richard C. Gerkin, Andrea Hasenstaub, Ramakrishnan Iyer, Renaud B. Jolivet, Sarah Marzen, Joseph D. Monaco, Astrid A. Prinz, Salma Quraishi, Fidel Santamaria, Sabyasachi Shivkumar, Matthew F. Singh, Roger Traub, Farzan Nadim, Horacio G. Rotstein, A. David Redish
Journal of Neuroscience 15 February 2023, 43 (7) 1074-1088; DOI: 10.1523/JNEUROSCI.1179-22.2022

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
On the Role of Theory and Modeling in Neuroscience
Daniel Levenstein, Veronica A. Alvarez, Asohan Amarasingham, Habiba Azab, Zhe S. Chen, Richard C. Gerkin, Andrea Hasenstaub, Ramakrishnan Iyer, Renaud B. Jolivet, Sarah Marzen, Joseph D. Monaco, Astrid A. Prinz, Salma Quraishi, Fidel Santamaria, Sabyasachi Shivkumar, Matthew F. Singh, Roger Traub, Farzan Nadim, Horacio G. Rotstein, A. David Redish
Journal of Neuroscience 15 February 2023, 43 (7) 1074-1088; DOI: 10.1523/JNEUROSCI.1179-22.2022
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • Abstract
    • Introduction
    • Footnotes
    • References
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Responses to this article

Respond to this article

Jump to comment:

  • Forging Connections: Uniting Neuroscience and Philosophy of Science
    Marcin Milkowski and Daniel Kostic
    Published on: 16 June 2023
  • Published on: (16 June 2023)
    Page navigation anchor for Forging Connections: Uniting Neuroscience and Philosophy of Science
    Forging Connections: Uniting Neuroscience and Philosophy of Science
    • Marcin Milkowski, Philosopher of Science, Institute of Philosophy and Sociology, Polish Academy of Sciences
    • Other Contributors:
      • Daniel Kostic

    Levenstein et al. aptly highlight some of the foundational issues in theoretical neuroscience, such as the role of abstraction and idealization in providing scientific explanations and understanding, and distinguishing under which conditions neuroscientific models provide genuine explanations, or mere descriptions, predictions, or control.

    The authors rightly emphasize that philosophers of science can also gain valuable insights from the vast body of neuroscience literature, by employing methods of digital humanities, such as text mining, in line with the cognitive metascience approach (Miłkowski, 2023).

    Conversely, neuroscientists can benefit from the latest research in philosophy of science, particularly on the epistemic norms of diverse kinds of explanations in neuroscience, e.g., topological (Kostić, 2020), computational (Miłkowski, 2013; Chirimuuta, 2018), dynamical (Favela, 2020), and functional (Egan, 2017). Understanding the diverse goals of theories is conducive to fruitful exploratory and explanatory scientific practice. Some forms of scientific understanding are not always grounded in explanations, but also in diverse theoretical representations, e.g., diagrams, taxonomies, and diagnostic manuals (Miłkowski, 2023). Philosophical theories of understanding can also integrate different kinds of explanations in a single framework (Khalifa et al., 2022).

    The need for these interdisciplinary connections is particularly vivid in the case of net...

    Show More

    Levenstein et al. aptly highlight some of the foundational issues in theoretical neuroscience, such as the role of abstraction and idealization in providing scientific explanations and understanding, and distinguishing under which conditions neuroscientific models provide genuine explanations, or mere descriptions, predictions, or control.

    The authors rightly emphasize that philosophers of science can also gain valuable insights from the vast body of neuroscience literature, by employing methods of digital humanities, such as text mining, in line with the cognitive metascience approach (Miłkowski, 2023).

    Conversely, neuroscientists can benefit from the latest research in philosophy of science, particularly on the epistemic norms of diverse kinds of explanations in neuroscience, e.g., topological (Kostić, 2020), computational (Miłkowski, 2013; Chirimuuta, 2018), dynamical (Favela, 2020), and functional (Egan, 2017). Understanding the diverse goals of theories is conducive to fruitful exploratory and explanatory scientific practice. Some forms of scientific understanding are not always grounded in explanations, but also in diverse theoretical representations, e.g., diagrams, taxonomies, and diagnostic manuals (Miłkowski, 2023). Philosophical theories of understanding can also integrate different kinds of explanations in a single framework (Khalifa et al., 2022).

    The need for these interdisciplinary connections is particularly vivid in the case of network models that are increasingly being used in neuroscience (Kostić et al., 2020), and especially when it comes to evaluating their explanatory power (Kostić, 2020; Kostić and Khalifa, 2023).

    In conclusion, we strongly support Levenstein et al.'s call for greater collaboration between philosophers of science and neuroscientists on the foundational issues in neuroscience. Embracing this interdisciplinary approach is essential for advancing our knowledge of the brain and its function.

    Chirimuuta M (2018) Explanation in Computational Neuroscience: Causal and Non-causal. Br J Philos Sci 69:849–880.
    Egan F (2017) Function-Theoretic Explanation and the Search for Neural Mechanisms. In: Explanation and Integration in Mind and Brain Science (Kaplan DM, ed), pp 0. Oxford University Press. Available at: https://doi.org/10.1093/oso/9780199685509.003.0007 [Accessed June 16, 2023].
    Favela LH (2020) Dynamical systems theory in cognitive science and neuroscience. Philosophy Compass 15:e12695.
    Khalifa K, Islam F, Gamboa JP, Wilkenfeld DA, Kostić D (2022) Integrating Philosophy of Understanding With the Cognitive Sciences. Frontiers in Systems Neuroscience 16 Available at: https://www.frontiersin.org/articles/10.3389/fnsys.2022.764708 [Accessed June 16, 2023].
    Kostić D (2020) General theory of topological explanations and explanatory asymmetry. Philosophical Transactions of the Royal Society B: Biological Sciences 375:20190321.
    Kostić D, Hilgetag CC, Tittgemeyer M (2020) Unifying the essential concepts of biological networks: biological insights and philosophical foundations. Philosophical Transactions of the Royal Society B: Biological Sciences 375:20190314.
    Kostić D, Khalifa K (2023) Decoupling Topological Explanations from Mechanisms. Philosophy of Science 90:245–268.
    Miłkowski M (2013) Explaining the Computational Mind. Cambridge, Mass.: MIT Press.
    Miłkowski M (2023) Cognitive Metascience: A New Approach to the Study of Theories. Available at: https://psyarxiv.com/xvyz3/ [Accessed April 13, 2023].

    Show Less
    Competing Interests: None declared.

Related Articles

Cited By...

More in this TOC Section

  • Striatal Dopamine Actions and Movement: Inferences from Parkinson Disease
  • Thinking as Analogy-Making: Toward a Neural Process Account of General Intelligence
  • Close Encounters with Art of Neuroscience 14th Edition
Show more Viewpoints
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.