Skip to main content

Main menu

  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE

User menu

  • Log out
  • Log in
  • My Cart

Search

  • Advanced search
Journal of Neuroscience
  • Log out
  • Log in
  • My Cart
Journal of Neuroscience

Advanced Search

Submit a Manuscript
  • HOME
  • CONTENT
    • Early Release
    • Featured
    • Current Issue
    • Issue Archive
    • Collections
    • Podcast
  • ALERTS
  • FOR AUTHORS
    • Information for Authors
    • Fees
    • Journal Clubs
    • eLetters
    • Submit
    • Special Collections
  • EDITORIAL BOARD
    • Editorial Board
    • ECR Advisory Board
    • Journal Staff
  • ABOUT
    • Overview
    • Advertise
    • For the Media
    • Rights and Permissions
    • Privacy Policy
    • Feedback
    • Accessibility
  • SUBSCRIBE
PreviousNext
Research Articles, Behavioral/Cognitive

Joint population coding and temporal coherence link an attended talker’s voice and location features in naturalistic multi-talker scenes

Kiki van der Heijden, Prachi Patel, Stephan Bickel, Jose L. Herrero, Ashesh D. Mehta and Nima Mesgarani
Journal of Neuroscience 9 October 2025, e0754252025; https://doi.org/10.1523/JNEUROSCI.0754-25.2025
Kiki van der Heijden
1Donders Institute for Brain, Cognition and Behavior, Radboud University, 6525 GD, Nijmegen, Netherlands.
2Mortimer B. Zuckerman Mind Brain Behavior Institute, Columbia University, NY 10027, New York, United States.
3Maastricht Center for Systems Biology (MaCSBio), Maastricht University, 6229 EN, Maastricht, Netherlands.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Prachi Patel
2Mortimer B. Zuckerman Mind Brain Behavior Institute, Columbia University, NY 10027, New York, United States.
4Department of Electrical Engineering, Columbia University, NY 10027, New York, United States.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Stephan Bickel
5Hofstra Northwell School of Medicine, NY 11549, New York City, United States.
6The Feinstein Institute for Medical Research, NY 11030, New York City, United States.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Jose L. Herrero
5Hofstra Northwell School of Medicine, NY 11549, New York City, United States.
6The Feinstein Institute for Medical Research, NY 11030, New York City, United States.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Ashesh D. Mehta
5Hofstra Northwell School of Medicine, NY 11549, New York City, United States.
6The Feinstein Institute for Medical Research, NY 11030, New York City, United States.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Nima Mesgarani
2Mortimer B. Zuckerman Mind Brain Behavior Institute, Columbia University, NY 10027, New York, United States.
4Department of Electrical Engineering, Columbia University, NY 10027, New York, United States.
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • For correspondence: nima{at}ee.columbia.edu
  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF
Loading

Abstract

Listeners effortlessly extract multidimensional auditory objects, such as a localized talker, from complex acoustic scenes. However, the neural mechanisms that enable simultaneous encoding and linking of distinct sound features—such as a talker’s voice and location—are not fully understood. Using invasive intracranial recordings in seven neurosurgical patients (4 male, 3 female), we investigated how the human auditory cortex processes and integrates these features during naturalistic multi-talker scenes and how attentional mechanisms modulate such feature integration. We found that cortical sites exhibit a continuum of feature sensitivity, ranging from single-feature sensitive sites (responsive primarily to voice spectral features or to location features) to dual-feature sensitive sites (responsive to both features). At the population level, neural response patterns from both single- and dual-feature sensitive sites jointly encoded the attended talker’s voice and location. Notably, single-feature sensitive sites encoded their primary feature with greater precision but also represented coarse information about the secondary feature. Sites selectively tracking a single, attended speech stream concurrently encoded both voice and location features, demonstrating a link between selective attention and feature integration. Additionally, attention selectively enhanced temporal coherence between voice- and location-sensitive sites, suggesting that temporal synchronization serves as a mechanism for linking these features. Our findings highlight two complementary neural mechanisms—joint population coding and temporal coherence—that enable the integration of voice and location features in the auditory cortex. These results provide new insights into the distributed, multidimensional nature of auditory object formation during active listening in complex environments.

Significance statement In everyday life, listeners effortlessly extract individual sound sources from complex acoustic scenes which contain multiple sound sources. Yet, how the brain links the different features of a particular sound source to each other – such as a talker’s voice characteristics and location - is poorly understood. Here, we show that two neural mechanisms contribute to encoding and integrating voice and location features in multi-talker sound scenes: (1) some neuronal sites are sensitive to both voice and location and their activity patterns encode these features jointly; (2) the responses of neuronal sites that process only one sound feature – that is, location or voice – align temporally to form a stream that is segregated from the other talker.

Footnotes

  • We thank Menoua Keshishian for sharing his expertise on STRF analysis. This study was supported by National Institute on Deafness and other Communication Disorders grant R01DC014279 (NM) and grant R01DC018805 (NM). This project has also received funding from the European Union’s Horizon 2020 Research and Innovation Program under the Marie Sklodowska-Curie grant agreement No 898134 (KH) and from the NWO Talent Program under the Veni grant agreement VI.Veni.202.184 (KH).

SfN exclusive license.

Back to top
Email

Thank you for sharing this Journal of Neuroscience article.

NOTE: We request your email address only to inform the recipient that it was you who recommended this article, and that it is not junk mail. We do not retain these email addresses.

Enter multiple addresses on separate lines or separate them with commas.
Joint population coding and temporal coherence link an attended talker’s voice and location features in naturalistic multi-talker scenes
(Your Name) has forwarded a page to you from Journal of Neuroscience
(Your Name) thought you would be interested in this article in Journal of Neuroscience.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
View Full Page PDF
Citation Tools
Joint population coding and temporal coherence link an attended talker’s voice and location features in naturalistic multi-talker scenes
Kiki van der Heijden, Prachi Patel, Stephan Bickel, Jose L. Herrero, Ashesh D. Mehta, Nima Mesgarani
Journal of Neuroscience 9 October 2025, e0754252025; DOI: 10.1523/JNEUROSCI.0754-25.2025

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Respond to this article
Request Permissions
Share
Joint population coding and temporal coherence link an attended talker’s voice and location features in naturalistic multi-talker scenes
Kiki van der Heijden, Prachi Patel, Stephan Bickel, Jose L. Herrero, Ashesh D. Mehta, Nima Mesgarani
Journal of Neuroscience 9 October 2025, e0754252025; DOI: 10.1523/JNEUROSCI.0754-25.2025
Twitter logo Facebook logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One

Jump to section

  • Article
  • Figures & Data
  • Info & Metrics
  • eLetters
  • PDF

Responses to this article

Respond to this article

Jump to comment:

No eLetters have been published for this article.

Related Articles

Cited By...

More in this TOC Section

Research Articles

  • Gene expression-based lesion-symptom mapping: FOXP2 and language impairments after stroke
  • Visual Distortions in Human Amblyopia Are Correlated with Deficits in Contrast Sensitivity
  • Distinct Portions of Superior Temporal Sulcus Combine Auditory Representations with Different Visual Streams
Show more Research Articles

Behavioral/Cognitive

  • Gene expression-based lesion-symptom mapping: FOXP2 and language impairments after stroke
  • Distinct Portions of Superior Temporal Sulcus Combine Auditory Representations with Different Visual Streams
  • Microsaccade Direction Reveals the Variation in Auditory Selective Attention Processes
Show more Behavioral/Cognitive
  • Home
  • Alerts
  • Follow SFN on BlueSky
  • Visit Society for Neuroscience on Facebook
  • Follow Society for Neuroscience on Twitter
  • Follow Society for Neuroscience on LinkedIn
  • Visit Society for Neuroscience on Youtube
  • Follow our RSS feeds

Content

  • Early Release
  • Current Issue
  • Issue Archive
  • Collections

Information

  • For Authors
  • For Advertisers
  • For the Media
  • For Subscribers

About

  • About the Journal
  • Editorial Board
  • Privacy Notice
  • Contact
  • Accessibility
(JNeurosci logo)
(SfN logo)

Copyright © 2025 by the Society for Neuroscience.
JNeurosci Online ISSN: 1529-2401

The ideas and opinions expressed in JNeurosci do not necessarily reflect those of SfN or the JNeurosci Editorial Board. Publication of an advertisement or other product mention in JNeurosci should not be construed as an endorsement of the manufacturer’s claims. SfN does not assume any responsibility for any injury and/or damage to persons or property arising from or related to any use of any material contained in JNeurosci.