Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Editorial

Consideration of Sample Size in Neuroscience Studies

Journal of Neuroscience 20 May 2020, 40 (21) 4076-4077; https://doi.org/10.1523/JNEUROSCI.0866-20.2020
  • Article
  • Info & Metrics
  • eLetters
  • PDF
Loading

Reproducibility of neuroscience studies is a primary goal of The Journal of Neuroscience. There are two main reasons for problems of reproducibility in the neuroscience literature. The first is the inflated false-positive rates that result in many studies falsely rejecting their null hypotheses. This often has its roots in biases in statistical inference. These biases can be introduced by “researcher degrees of freedom,” selecting analytical procedures according to the study outcome; by “hypothesizing after results are known,” offering credibility to tests lacking a hypothesis; or by using parametric procedures when the structure of the data does not warrant them. Such procedural biases and how to minimize them were covered by a previous JNeurosci editorial on analytical transparency and reproducibility (Picciotto, 2018).

This editorial focuses on a second reason for limited reproducibility in neuroscience studies: low statistical power, frequently caused by small sample sizes. Here we provide suggestions on how to approach the determination of sample size in the context of the noisy and subtle effects often observed in neuroscience studies. We emphasize how sample size planning depends on whether the statistical goal of the study is to determine the presence of an effect or to obtain accurate estimates of the effect.

Statistical power (1 − β, where β is the false-negative rate or the probability to fail to reject the null hypothesis when an effect is present) increases with sample size. Given a true effect of a certain size, studies with smaller samples have lower power to detect it. Effects found in studies with low power are subject to the problem of low positive predictive value (Button et al., 2013). On a single test, a small sample size does not inflate the probability of falsely rejecting the null hypothesis (e.g., α = 5%). However, not all researchers are aware that low power in their study increases the probability that their estimation overestimates the true effect size, a situation aptly labeled “Winner's Curse” (Button et al., 2013).

What does this mean for neuroscience? The real effect sizes for most phenomena uncovered in exploratory studies are in fact smaller than reported, even without accounting for the procedural biases leading to inflated inferential statistics. Therefore, follow-up studies based on those estimated effect sizes should expect to find smaller effects, due to a regression toward the mean of the underlying distribution of effect size. Unfortunately, this phenomenon is often missed given that many studies still fail to report effect size.

These considerations lead us to suggest that, whenever possible, studies should accommodate and plan for two related experiments. First, an “exploratory” experiment provides provisional statistical evidence for the presence of an effect. The findings of this exploratory stage generate (likely inflated) estimates of the effect magnitude, and likely to yield confidence intervals that are wide and imprecise (Maxwell et al., 2008). Second, an “estimation” experiment provides a more precise and accurate estimate of the real sizes of those effects. The exploratory stage could be powered to detect medium to large effect sizes using intermediate sample sizes to avoid the risk of detecting biologically marginal effects when using large samples (Wilson et al., 2020). The exploratory stage should also quantify the statistical power provided by the experimental design, either a priori or with post hoc simulations. In contrast, the estimation stage should be used to optimize sample size for effect size estimation. The sample size necessary to obtain an accurate estimate of an effect size is usually larger than the sample size necessary for adequate power to detect the presence of an effect (Maxwell et al., 2008).

Procedurally, this suggestion might appear similar to the requirement of providing two independent sets of inferential statistics on the same basic effect, at the core of most replicability efforts (Lindsay, 2017). However, the estimation stage is not about confirming the “truth” of an exploratory observation, already controlled for by the nominal rate of false positives (e.g., α = 5%) or equivalently corrected when multiple tests are performed. The rationale for providing a second, independent set of observations is to increase the precision of effect size determination for a finding deemed interesting enough to justify additional and substantial measurement efforts, and to consider whether the magnitude of that more precise estimate is biologically relevant. In this context, the estimation phase would benefit from registration, since it is important to document the precise replication of the experimental protocol and analytical procedures of the exploratory stage. This is especially important if first-stage estimation experiments are not published. This procedure should lead researchers to expect, rather than criticize, smaller effect sizes in the estimation stage. Those smaller effect sizes, combined with low power in the exploratory stage, will result in many estimation studies failing to confirm the rejection of the null hypothesis. However, rather than jumping to the conclusion that the inferences of the estimation stage were “false” (Ioannidis, 2005), this two-step procedure might shift the emphasis toward precisely estimating the magnitude and direction of an effect (“how much”) and away from a dichotomous (“Does the effect exist or not?”) question (Calin-Jageman and Cumming, 2019). Put differently, this invites researchers to evaluate the biological plausibility of more precisely estimated effects, rather than use an inferential threshold as a license to suspend critical judgment (Gigerenzer, 2018).

The suggested exploration-then-estimation procedure is functionally equivalent to practices already adopted by some subfields of neuroscience. For instance, in cognitive neuroscience, it is customary to separate the estimation phase of model fitting from the validation phase of the model parameters. That approach is valid as long as the validation phase operates on independent data and does not introduce new parameters. While many of these practices typically rely on large sample sizes, some areas of neuroscience make statistical inferences on individual subjects, implementing a sort of exploration-then-estimation procedure across successive subjects (e.g., patients or nonhuman animal models in electrophysiology; machine-learning explorations of fMRI data; psychophysics and human brain lesion studies). These small-N approaches focus their statistical power on individual-level characterization of an effect; a finding is deemed present when all or a majority of a small pool of subjects show an effect, usually based on a large sample of trial-level observations (Smith and Little, 2018). It should be acknowledged that this approach only allows for statements that pertain to the existence and magnitude of effects in those subjects, rather than in the populations those subjects are drawn from. Many of the most robust findings in psychophysics have come from a small-N approach (Smith and Little, 2018), and it could be preferred ethically when animal welfare or vulnerable individuals are involved.

In addition to screening submissions for the rigor of their statistical procedures, we believe it is also important to steer the community through positive examples. JNeurosci welcomes contributions that provide a definite statement on a research question by commenting on biological plausibility and using rigorous statistical procedures, such as those discussed in this editorial.

We invite you to contribute to this discussion by emailing JNeurosci at JN_EiC{at}sfn.org or tweeting to @marinap63.

The Editorial Board of The Journal of Neuroscience

References

  1. ↵
    1. Button KS,
    2. Ioannidis JPA,
    3. Mokrysz C,
    4. Nosek BA,
    5. Flint J,
    6. Robinson ESJ,
    7. Munafò MR
    (2013) Power failure: why small sample size undermines the reliability of neuroscience. Nat Rev Neurosci 14:365–376. doi:10.1038/nrn3475 pmid:23571845
    OpenUrlCrossRefPubMed
  2. ↵
    1. Calin-Jageman RJ,
    2. Cumming G
    (2019) Estimation for better inference in neuroscience. eNeuro 6:ENEURO.0205-19.2019. doi:10.1523/ENEURO.0205-19.2019
    OpenUrlCrossRef
  3. ↵
    1. Gigerenzer G
    (2018) Statistical rituals: the replication delusion and how we got there. Adv Methods Pract Psychol Sci 1:198–218. doi:10.1177/2515245918771329
    OpenUrlCrossRef
  4. ↵
    1. Ioannidis JPA
    (2005) Why most published research findings are false. PLoS Med 2:e124.
    OpenUrlCrossRefPubMed
  5. ↵
    1. Lindsay DS
    (2017) Preregistered direct replications in psychological science. Psychol Sci 28:1191–1192. doi:10.1177/0956797617718802 pmid:28793201
    OpenUrlCrossRefPubMed
  6. ↵
    1. Maxwell SE,
    2. Kelley K,
    3. Rausch JR
    (2008) Sample size planning for statistical power and accuracy in parameter estimation. Annu Rev Psychol 59:537–563. doi:10.1146/annurev.psych.59.103006.093735 pmid:17937603
    OpenUrlCrossRefPubMed
  7. ↵
    1. Picciotto M
    (2018) Analytical transparency and reproducibility in human neuroimaging studies. J Neurosci 38:3375–3376. doi:10.1523/JNEUROSCI.0424-18.2018 pmid:29618543
    OpenUrlFREE Full Text
  8. ↵
    1. Smith PL,
    2. Little DR
    (2018) Small is beautiful: in defense of the small-N design. Psychon Bull Rev 25:2083–2101. doi:10.3758/s13423-018-1451-8 pmid:29557067
    OpenUrlCrossRefPubMed
  9. ↵
    1. Wilson BM,
    2. Harris CR,
    3. Wixted JT
    (2020) Science is not a signal detection problem. Proc Natl Acad Sci U S A 117:5559–5567. doi:10.1073/pnas.1914237117
    OpenUrlAbstract/FREE Full Text
Back to top

In this issue

The Journal of Neuroscience: 40 (21)
Journal of Neuroscience
Vol. 40, Issue 21
20 May 2020
  • Table of Contents
  • Table of Contents (PDF)
  • About the Cover
  • Index by author
  • Advertising (PDF)
  • Ed Board (PDF)
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Consideration of Sample Size in Neuroscience Studies
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Print
View Full Page PDF
Citation Tools
Consideration of Sample Size in Neuroscience Studies
Journal of Neuroscience 20 May 2020, 40 (21) 4076-4077; DOI: 10.1523/JNEUROSCI.0866-20.2020

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Consideration of Sample Size in Neuroscience Studies
Journal of Neuroscience 20 May 2020, 40 (21) 4076-4077; DOI: 10.1523/JNEUROSCI.0866-20.2020
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
    • References
  • Info & Metrics
  • eLetters
  • PDF

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

  • The Trials and Aspirations of an Early Career Researcher Advisory Board
  • Author-Centered Approach to Scientific Publishing
  • Including Early Career Researchers in Scientific Publishing
Show more Editorial

Subjects

  • Experimental Design Editorials
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.