Abstract
During rest, intrinsic neural dynamics manifest at multiple timescales, which progressively increase along visual and somatosensory hierarchies. Theoretically, intrinsic timescales are thought to facilitate processing of external stimuli at multiple stages. However, direct links between timescales at rest and sensory processing, as well as translation to the auditory system are lacking. Here, we measured intracranial EEG in 11 human patients with epilepsy (4 women), while listening to pure tones. We show that, in the auditory network, intrinsic neural timescales progressively increase, while the spectral exponent flattens, from temporal to entorhinal cortex, hippocampus, and amygdala. Within the neocortex, intrinsic timescales exhibit spatial gradients that follow the temporal lobe anatomy. Crucially, intrinsic timescales at baseline can explain the latency of auditory responses: as intrinsic timescales increase, so do the single-electrode response onset and peak latencies. Our results suggest that the human auditory network exhibits a repertoire of intrinsic neural dynamics, which manifest in cortical gradients with millimeter resolution and may provide a variety of temporal windows to support auditory processing.
SIGNIFICANCE STATEMENT Endogenous neural dynamics are often characterized by their intrinsic timescales. These are thought to facilitate processing of external stimuli. However, a direct link between intrinsic timing at rest and sensory processing is missing. Here, with intracranial EEG, we show that intrinsic timescales progressively increase from temporal to entorhinal cortex, hippocampus, and amygdala. Intrinsic timescales at baseline can explain the variability in the timing of intracranial EEG responses to sounds: cortical electrodes with fast timescales also show fast- and short-lasting responses to auditory stimuli, which progressively increase in the hippocampus and amygdala. Our results suggest that a hierarchy of neural dynamics in the temporal lobe manifests across cortical and limbic structures and can explain the temporal richness of auditory responses.
Introduction
The human brain gives rise to rich neural dynamics, which play a fundamental role in processing sensory information. Intrinsic dynamics of the brain operate at multiple timescales (Hasson et al., 2008; Honey et al., 2012; Murray et al., 2014; Raut et al., 2020) through oscillatory (Frauscher et al., 2018; Mahjoory et al., 2020; Vezoli et al., 2021) and nonoscillatory (Gao et al., 2020) processes. In the visual and somatosensory systems, intrinsic timescales manifest at rest, in ongoing neural activity: primary areas exhibit short timescales that may facilitate a quick reaction to incoming stimuli (Murray et al., 2014; Siegle et al., 2021). These progressively increase while advancing through the cortical hierarchy, supporting integration of information (Murray et al., 2014; Chaudhuri et al., 2015). Whether a similar hierarchy of intrinsic dynamics exists in the auditory system, and in particular within the temporal lobe, a hub for auditory processing, remains underexplored.
In the auditory system, evidence for processing of external stimuli at multiple latencies stems from studying evoked responses (Honey et al., 2012; Norman-Haignere et al., 2022). Primary auditory areas show fast- and short-lasting responses to sounds (Camalier et al., 2012). Response latencies progressively increase while advancing in a processing hierarchy, from primary to secondary areas, as, for example, the superior temporal gyrus (Nourski et al., 2014). Beyond this “classical” auditory cortex circuitry of the temporal lobe, an extensive network of adjacent cortical and deeper regions is also sensitive to auditory input and exhibits diverse response profiles and latencies. At a cortical level, the insula, for example, shows relatively fast auditory responses (Blenkmann et al., 2019), while deeper structures, such as the hippocampus and amygdala, show slower, long-lasting responses to auditory stimuli (Halgren et al., 1980), possibly mediating the integration of sensory information (Zuo et al., 2020).
This richness in auditory responses suggests that, when stimulated with sounds, the temporal lobe facilitates auditory processing at multiple timescales (Stephens et al., 2013). These are thought to reflect temporal “integration” windows that manifest in response to external stimuli (Lerner et al., 2011; Honey et al., 2012; Norman-Haignere et al., 2022). Whether a similar temporal lobe hierarchical organization also exists during rest and contributes to auditory processing remains underexplored. Importantly, there is a critical lack of studies that simultaneously assess neural timescales not only in the temporal cortex, but also in the hippocampus and amygdala, which are key, yet underexplored regions in processing of auditory information (Billig et al., 2022). The question of how these structures are positioned in a hierarchy of intrinsic timescales remains therefore open. In humans, in particular, a fine-grained measurement of neural dynamics in the temporal lobe can be challenging with noninvasive techniques (Tzovara et al., 2019; Johnson et al., 2020; Raut et al., 2020), but evidence from invasive recordings remains limited.
Here, we aimed at characterizing spontaneous intrinsic neural dynamics within cortical and limbic structures of the extended auditory system, covering the temporal lobe and insula, and their contribution to auditory processing. We focused on this network, which is relatively accessible through intracranial EEG (iEEG) recordings in patients with pharmaco-resistant epilepsies. We hypothesized that spontaneous intrinsic neural timescales, estimated via the autocorrelation function (ACF) (Golesorkhi et al., 2021b; Zeraati et al., 2022), or via the knee frequency of the power spectral density (PSD) (Gao et al., 2020) of iEEG signals, would show a hierarchical organization within an extended auditory network, which could, in turn, explain a hierarchy of neural responses to incoming auditory stimuli. We additionally hypothesized that nonoscillatory brain dynamics, characterized by the spectral exponent of aperiodic neural activity, which has been suggested to reflect a proxy of the excitation to inhibition balance (Gao et al., 2017), would also reveal a hierarchical organization across the temporal lobe.
Materials and Methods
Patients
We recorded iEEG in 11 neurosurgical patients (4 women, median age = 32 years, minimum = 27, maximum = 56 years) with drug-refractory epilepsy who had been implanted with depth electrodes to identify seizure foci (for a detailed patient description, Table 1). Electrode locations were based on clinical criteria only. Recordings took place at the EPI Clinic, Zurich, and at the Inselspital, Bern. The number of patients included in this study is following standards in the field and is in line with, or larger than, existing intracranial studies investigating intrinsic neural dynamics (Honey et al., 2012; Hullett et al., 2016; Lendner et al., 2020; Mercier et al., 2022). Patients provided written informed consent before participation in this research study, approved by institutional ethics review boards of the Canton of Zurich (PB-2016-02055), and Inselspital, Bern (#2018-01387). All experiments were performed in accordance with the 6th Declaration of Helsinki.
Experimental protocol
Patients were presented with auditory stimuli consisting of pure tones at three frequencies (500, 1250, 2500 Hz) with a random interstimulus interval between 0.9 and 19 s. Each tone had a duration of 100 ms with 5 ms on/off ramps to avoid clicks. Interstimulus interval and tone frequency were drawn from a pseudorandom distribution such that each was played 120 times per hour (in total, 360 tones per hour). Auditory stimuli were presented via in-ear headphones, and their intensity was adjusted individually for each patient at a comfortable level. Patients were instructed to relax and ignore the sounds. Some of the patients were additionally presented with the auditory stimuli during sleep, at a later session, which was not analyzed in the context of the present study.
iEEG recordings and preprocessing
Depth electrodes were used for iEEG recordings (DIXI Medical, 3 patients; Ad-Tech Medical, 8 patients) targeting different brain regions and varying from 8 to 18 platinum iEEG contacts along their shaft. Data were recorded at 4096 or 1024 Hz. Recordings with 4096 Hz sampling rate were downsampled offline to 1024 Hz.
All data were visually inspected to exclude electrodes with persistent spiking activity. Continuous data were notch filtered at ∼50 Hz and harmonics, and rereferenced with a bipolar scheme (i.e., each electrode to the closest one in the same electrode lead outwardly) to remove any source of widespread noise. This was done to retain a local signal and mitigate effects of volume conduction, following recommendations in the analysis of iEEG data (Lachaux et al., 2012; Mercier et al., 2022). Peristimulus epochs were then extracted, spanning from −5 s before the sounds' onset to 5 s after stimulus onset. Only epochs that did not overlap with another sound in this period were kept. All epochs were then visually inspected, and any epochs with remaining artifacts were rejected. The baseline period of each epoch was defined as the interval [−1, 0] s preceding the sounds. For studying auditory responses (see Responsive electrodes), the raw signal from all electrodes was additionally bandpass filtered between 1 and 40 Hz. Processing of iEEG data was performed using MNE-Python (Gramfort et al., 2013).
Electrode localization
Electrodes were localized on postimplant CT scans using the Lead-DBS toolbox (Horn and Kühn, 2015) and transformed into standard MNI coordinates for group analyses. The postimplant CT scan was registered to a preimplant structural T1-weighted MRI scan from which anatomic labels were reconstructed using the FreeSurfer toolbox and the Destrieux atlas. Subsequently, electrode coordinates identified on the postimplant CT scans were mapped to their corresponding anatomic regions identified on the preimplant MRI. Anatomical label assignment was validated for all electrodes by an expert neurologist, who verified their location and additionally ensured that none of the electrodes that were included in our analyses were in white matter. The available electrodes were divided across four regions of interest (ROIs), covering the temporal cortex, the insula because of its prominent auditory responses (included in temporal cortex), entorhinal cortex, hippocampus, and amygdala. This resulted in N = 270 electrodes in total, with a median = 25, minimum = 8 and maximum = 37 electrodes per patient (Table 1).
Intrinsic neural timescales
For estimating spontaneous intrinsic neural timescales, we first computed the ACF on each epoch during 1 s baseline period (function acf from Python's statsmodels) (Seabold and Perktold, 2010). The resulting ACFs across epochs were then averaged to yield a single ACF for each electrode. We then defined the “intrinsic timescale” of each electrode as the time lag at which the ACF reaches the value 1/e, consistent with an analytical decay of the form f(t) = exp(–t/τ). The precise time lag was computed by interpolating with a spline fit to the ACF, as in Raut et al. (2020).
To ensure that the estimation of timescales was not trivially driven by neural oscillations, we performed two additional control analyses, following previous literature (Murray et al., 2014; Chaudhuri et al., 2015; Gao et al., 2020; Zeraati et al., 2022). First, we fitted a curve of the form f(t) = a*exp(–t/τ) + (1 – a)*cos(2πft) to the ACF with (a, τ, f) as parameters to be optimized (Zeraati et al., 2022); a represents the amplitude parameter, f the putative oscillatory frequency, and τ the estimated timescale. In a second control analysis, we computed timescales as the inverse of the knee frequency in power spectra, estimated as fk = k1/exp with k being the knee parameter and exp the spectral exponent, as implemented in the specparam toolbox (Donoghue et al., 2020) in “knee” mode. We fitted power spectra from 2 to 35 Hz, to have a reliable power estimate on the lower limit and to keep consistency with the “fixed” spectral parametrization for the higher limit (for a discussion on the choice of frequency band, see PSD and spectral exponent). Electrodes where the algorithm could not find a knee frequency were excluded.
PSD and spectral exponent
For estimating the spectral exponent we computed power spectra with a Hann-windowed and detrended Fourier transform on the baseline period (function spectrogram from Python's scipy) (Virtanen et al., 2020). Power spectra were averaged using a “meanlog” approach (i.e., taking the mean of the logarithm of the power spectra across epochs) to yield a single power spectrum density for each electrode. The spectral exponent was then computed on each electrode's average power spectrum density using the standard implementation of the spectral parameterization algorithm (Donoghue et al., 2020) in the “fixed” mode (linear fit in log-log plot) in two different frequency ranges: a lower one, at 20-35 Hz, and a higher one, at 80-150 Hz. The lower range was chosen following a large body of literature to avoid low-frequency knees, high-power peaks, and spectral plateaus (Gerster et al., 2021), and has been previously linked with individual variations to excitation to inhibition balance (Gao et al., 2017; Lendner et al., 2020). Different alternative, but related, frequency ranges were tested in exploratory analyses on a subset of patients (e.g., 30-45 Hz, or 20-40 Hz). All of those gave comparable results, and we used 20-35 Hz for our analysis, as it was the band that more consistently avoided the above-mentioned problems. The higher range was chosen as a typical high-frequency range that is often computed in iEEG studies (Lachaux et al., 2012).
The spectral exponent was computed as the slope of nonperiodic parts of the power spectra observed at each electrode via a standardized approach with the specparam toolbox (Donoghue et al., 2020) (parameters for the fitting: peak_threshold = 2, min_peak_height: 0.1, peak_width_limits: [1, 10], with max_n_peaks = 2 for the lower range and 0 for the higher one). Fits for every electrode were visually inspected, and any electrodes with clear artifacts on the power spectra, or where the fit was particularly noisy, were excluded to ensure an accurate estimation of the spectral exponent. After this step, all remaining electrodes (N = 270) had fits with R2 of at least 0.8. Amygdalar electrodes from 2 patients had a prominent peak in their power spectra at ∼40 Hz (see Fig. 5A). This was found for electrodes of the amygdala only, and not other electrodes, and to the best of our knowledge was unrelated to any sources of noise, or pathologic findings in these patients. We confirmed that fitting of the spectral exponent was not affected by these peaks in any of the 2 patients, which were outside the range of our fits.
Responsive electrodes
Responsive electrodes were identified following common approaches in the field of iEEG (Dürschmid et al., 2016). Briefly, differences between the average signal in poststimulus time points
Statistical analyses
Statistical tests were conducted in R version 4.2.0 (R Development Core Team, 2020) using linear mixed-effects models with a random intercepts term corresponding to the patient. The random intercepts term captures interpatient variability, which is needed when analyzing electrodes from multiple patients together. This ensured that any identified effects were not trivially driven by the fact that the electrodes were recorded from multiple patients (Yu et al., 2022) (implemented with nlme package) (Lindstrom and Bates, 1990). The omnibus tests for the “brain region” factor were computed with F tests, while post hoc pairwise comparisons were computed with Tukey's range test, controlling for multiple comparisons (implemented with emmeans package). In the case of omnibus tests on multiple time lags (see Fig. 2A) and tests over multiple MNI coordinates, p values were Bonferroni-corrected. For regression analyses, we used linear mixed-effects models with a continuous predictor and random intercepts accounting for across-patient variability. We computed correlation values starting from R2 as described by Nakagawa and Schielzeth (2013) and took the square root, mimicking a fixed-effects-only linear model (implemented with MuMIn package) (Barton, 2020). p values were computed with F tests, correcting with Bonferroni when regressing on each level of the region factor separately (pcorr).
Data and code availability
Because of the sensitive nature of the data, data and code can be made available from the corresponding author on reasonable request.
Results
We analyzed iEEG signals in 270 electrodes from 11 epilepsy patients (median = 25, minimum = 8, maximum = 37 electrodes per patient, Table 1). In a first step, we assessed a macroscopic organization of neural dynamics by dividing electrodes into four ROIs, selected based on the most consistent implantation schemes across patients. These were targeting the entorhinal cortex (ENT), hippocampus (HIP), and amygdala (AMY) in their innermost electrodes, and had additional electrodes covering the temporal and adjacent cortices (CTX) (for an exemplar implantation, see Fig. 1A). In a second step, we grouped all available electrodes together (for full electrode coverage, see Fig. 1B), regardless of ROIs, and assessed their spatial organization at a finer level, with respect to cortical and limbic anatomies.
iEEG signals in the four ROIs present striking qualitative differences already in their ongoing neural activity before sound presentation (for exemplar iEEG recordings, see Fig. 1B). To characterize ongoing neural dynamics, we computed their intrinsic timescales before the presentation of sounds (Fig. 1A, middle). For each electrode, we computed the autocorrelation function of baseline iEEG signals, which quantifies how similar a time series is to its past values across multiple time lags. The mean autocorrelation, computed across patients and brain regions, shows a characteristic decay as the time lag increases (Fig. 2A). For short time lags, the mean autocorrelation follows an ordering: electrodes in the temporal cortex have the most rapid decay, followed by electrodes in the entorhinal cortex, the hippocampus, and last, the amygdala (Fig. 2A), with significant differences across the four regions at time lags between 10 and 80 ms (mixed-effects models, accounting for different patients, pcorr < 0.05 with Bonferroni correction) (Fig. 2A, solid horizontal line).
We next computed intrinsic neural timescales (τ). These were defined as the time lag at which the autocorrelation of each electrode decayed to a fixed value (in our case, 1/e, Fig. 2A, dashed horizontal line). The extracted intrinsic timescales τ confirm the macroscopic hierarchy observed via the autocorrelation function and show a significant difference across brain regions (F(3,256) = 27.313, p = 2.33 × 10−15, mixed-effects model with random intercepts) (Fig. 2B). The temporal cortex exhibits significantly faster intrinsic timescales, at 40.6 ms on average compared with both the hippocampus and amygdala, which have slower timescales, at 56.1 and 63.3 ms, respectively (for a detailed report of all paired statistical comparisons, based on t tests derived via the linear mixed-effects models, see Table 2). Within subregions of the cortex, intrinsic timescales tend to be slower in the pole, and faster in the transverse gyrus, while the superior, middle, and inferior temporal cortex, and the insula lie in between (Table 3). The entorhinal cortex (46.9 ms) is also significantly faster compared with other limbic areas, but not different from the temporal cortex (Table 2).
These results were confirmed with two additional control analyses, which accounted for potential biases because of oscillations. First, when estimating timescales by a direct exponential decay fit to the ACFs, similar to previous studies (Murray et al., 2014; Siegle et al., 2021), but accounting for oscillations (Zeraati et al., 2022), the same macroscopic hierarchy was observed, highlighted by a significant difference of timescales across regions (F(3,256) = 16.789, p = 5.49 × 10−10). Second, the same hierarchy was also observed when estimating timescales as the inverse of the knee frequency in power spectra, similar to Gao et al. (2020) (F(3,197) = 28.769, p = 1.78 × 10−15). Both of these control analyses replicate the same ordering of timescales as reported in Figure 2B. These findings reveal a robust macroscopic hierarchy in spontaneous neural activity, confirmed with three different methods, where the temporal cortex shows short intrinsic timescales, while limbic areas exhibit slower dynamics.
We then delved into a finer characterization of timescales by exploring their spatial organization within anatomic regions (for an overview of cortical subregions, see Table 3). Within the temporal and entorhinal cortices, intrinsic timescales show a gradient that spans the temporal lobe through the posterolateral (fast timescales) to the anteromedial (slow timescales) axis, following the temporal lobe anatomy (Fig. 2C). This gradient is particularly prominent in the Y and Z directions that mostly define the temporal lobe orientation (Fig. 2D, correlation between coordinates in MNI space and intrinsic timescales: ρX = 0.231, pX = 2.44 × 10−6; ρY = 0.292, pY = 1.83 × 10−9; ρZ = −0.377, pZ = 2.94 × 10−12, mixed-effects models and Bonferroni-corrected).
The spatial distribution of timescales in the hippocampus and amygdala, on the contrary, is less defined, with no significant correlation along any of the MNI coordinates after correcting for multiple comparisons (ρX = 0.201, pX = 0.156; ρY = 0.219, pY = 0.154; ρZ = −0.159, pZ = 0.443, mixed-effects models and Bonferroni-corrected; Fig. 3). These findings support a fine-grained intrinsic organization of spontaneous neural dynamics in the extended auditory network that manifests across cortical and limbic regions and exhibits an anatomic gradient spanning the temporal cortex from posterior to anterior.
We next investigated whether intrinsic timescales at baseline could explain the timing of auditory processing. At a qualitative level, auditory intracranial event-related potentials (iERPs) show striking differences throughout the temporal lobe (Fig. 4A). iERPs in primary auditory regions (e.g., the transverse gyrus, Fig. 4A, top row) show early, short-lasting, and high-amplitude responses, while iERPs in the superior temporal gyrus have a later onset and duration (Fig. 4A, second row). By contrast, auditory responses in the hippocampus, amygdala, and entorhinal cortex are smoother and long-lasting (Fig. 4A, third to fifth rows), similar to previous reports (Halgren et al., 1980).
To quantify these response profiles, we restricted our analysis to electrodes that showed a significant 1-40 Hz iEEG response to the auditory stimuli compared with a prestimulus baseline (see Responsive electrodes, N = 67 of 270 total electrodes). For each responsive electrode, we computed its response onset and peak latencies (Fig. 4B). Cortical electrodes show generally faster responses than hippocampal and amygdalar ones both for onset (30 ms faster) and peak (50-60 ms faster). At the group level though, there is no significant effect of brain region on onset latency (F(3,55) = 1.867, p = 0.146) and just barely on peak latency (F(3,55) = 2.774, p = 0.0499, both mixed-effects models). In cortical subregions, the transverse gyrus shows the earliest responses, followed by the superior temporal gyrus/sulcus, inferior and middle temporal gyri (Table 3).
Interestingly, this variability in onset and peak latencies within and between brain regions can be partially explained when accounting for differences in intrinsic timescales (Fig. 4C). We computed a regression of response latencies on intrinsic timescales, which shows a highly significant main effect of timescale at baseline both on response onset (ρ = 0.353, p = 0.0009) and peak latency (ρ = 0.409, p = 0.0005, both mixed-effects models with random intercepts, to account for different patients). The strong regression of the onset of auditory responses on intrinsic timescales at baseline holds for electrodes within the temporal cortex (ρ = 0.457, pcorr = 0.0017) and hippocampus (ρ = 0.816, pcorr = 0.0013, Fig. 4D) (mixed-effect models and Bonferroni-corrected). The other within-region regressions do not reach significance, except for the peak latency in the hippocampus (ρ = 0.734, pcorr = 0.031). Moreover, a significant regression result persists when splitting each patient's trials into two experimental halves, suggesting that the observed results are robust across the experimental session (ρ1 = 0.429, p1 = 0.001; ρ2 = 0.364, p2 = 0.01 for the first and second half of the experiment, respectively). These results show that intrinsic timescales at baseline can explain both the onset and peak latencies of auditory responses throughout the temporal lobe: regions that are characterized by fast intrinsic timescales exhibit a fast reaction to incoming auditory stimuli, while the hippocampus, amygdala, and entorhinal cortex are mediated by slower ongoing dynamics and show slower auditory responses.
To further explore and confirm the observed hierarchy of intrinsic neural timescales, we additionally characterized their aperiodic neural activity via the spectral exponent (Fig. 5). The average PSD shows qualitative differences across the four ROIs (Fig. 5A). The cortex exhibits a characteristic oscillatory peak at ∼10 Hz, and a relatively fast decay, while the hippocampus displays the strongest power, which for low frequencies decays relatively gently, but after 70 Hz much faster (Fig. 5A). We quantified the nonoscillatory part of the power spectra for each electrode via the spectral exponent (i.e., the slope in log-log space) in a lower (20-35 Hz, as commonly reported in the literature) (Gao et al., 2017; Miskovic et al., 2019; Lendner et al., 2020) and upper range (80-150 Hz), corresponding to high γ-power (Lachaux et al., 2012). The lower range was chosen after considering typical ranges used in the literature, which vary across studies, and compromising between consistency with previous studies and recommended methodological considerations (for a detailed explanation of the choice of the frequency band and control analyses, see PSD and spectral exponent).
The spectral exponent in the 20-35 Hz range shows a strong ordering, with electrodes in the temporal cortex having the steepest exponent, followed by electrodes in the entorhinal cortex, hippocampus, and amygdala (Fig. 5B), with a significant effect of region (F(3,256) = 80.665, p = 1.11 × 10−16). This result confirms the ordering observed for intrinsic timescales (Fig. 2B), with a different and complementary measure. Moreover, all pairs of cortical-limbic areas have significant differences in their 20-35 Hz exponent (Table 4), while the difference between temporal and entorhinal cortex is slightly below significance threshold (Table 4, pCTX-ENT = 0.054). Exponents in cortical subregions do not show marked differences from each other (Table 3).
The spectral exponent in the 80-150 Hz range also shows a significant main effect of region (F(3,256) = 79.156, p = 1.11 × 10−16) (Fig. 5C). This effect is mainly driven by the difference between the hippocampus (with an exponent of 4.5 on average across electrodes) and all other regions (Table 4), which instead have very similar exponent values, ranging between 2.4 and 2.6 on average (Fig. 5C). The particularly steep hippocampal spectral exponent for high frequencies reflects the abrupt change of slope in the power spectrum, which forms a knee at ∼70 Hz (Fig. 5A).
As the lower range spectral exponent reflects the same ordering of brain regions as intrinsic timescales, we explored its spatial organization. Similar to the intrinsic timescales, we observe an anatomic modulation of spectral exponents along the temporal lobe, indicated by a significant, albeit weaker, correlation between spectral exponent and MNI X coordinates (ρX = −0.188, pX = 9.99 × 10−4, mixed-effects model and Bonferroni-corrected), but no significant correlation along other axes (p > 0.12). This information provides further support for a gradient organization of neural dynamics within the extended auditory cortical network.
Last, the spectral exponent within the hippocampus/amygdala only shows a weak correlation along the X axis (ρX = −0.252, pX = 0.029, mixed-effects model and Bonferroni-corrected). When correlating the lower spectral exponent with response onset or peak latencies, there is no significant correlation, neither in all responsive electrodes grouped together (ρonset = 0.066, ponset = 0.529; ρpeak = −0.096, ppeak = 0.430, mixed-effects models) nor within any of the individual brain regions (Table 5).
Discussion
We provide evidence for a hierarchy of spontaneous intrinsic neural dynamics in the extended human auditory network, which in turn explains a hierarchy in the processing of incoming auditory stimuli. At a macroscopic level, the temporal cortex assumes a “low” position along this hierarchy, highlighted by a steep spectral exponent and short intrinsic timescales, which likely mediate short temporal receptive windows (Honey et al., 2012; Norman-Haignere et al., 2022). On the contrary, the hippocampus and amygdala exhibit longer intrinsic timescales and have flatter spectral exponents. This suggests that the hippocampus and amygdala assume a “higher,” or integrative, function in a temporal lobe hierarchy, as longer receptive time windows, indicated by longer timescales, may be necessary for information integration (Murray et al., 2014; Golesorkhi et al., 2021b). By contrast, a flatter exponent may indicate a shift toward excitation (Gao et al., 2017), or increased neural noise (Alnes et al., 2021).
Intrinsic timescales and spectral exponent reveal a hierarchy in the temporal lobe
Our findings are in line with previous reports of a hierarchical organization in the visual and somatosensory modalities (Murray et al., 2014; Wang, 2020), where neural timescales progressively increase along the cortical hierarchy. Previous investigations of intrinsic timescales in humans have mainly relied on hemodynamic and magnetoencephalographic measures, and have shown fast spontaneous dynamics in the temporal lobe compared with higher-level areas, such as the Prefrontal cortex, albeit only at a macroscopic level (Raut et al., 2020; Golesorkhi et al., 2021a). Apart from timescales, oscillatory power and the spectral exponent also show an intrinsic organization (Frauscher et al., 2018; Mahjoory et al., 2020). iEEG oscillatory peaks transition from faster to slower frequencies along the posterior-to-anterior temporal cortex (Frauscher et al., 2018), while primary auditory regions show weaker α and stronger high-γ power in their baseline activity compared with secondary auditory areas (Billig et al., 2019). Here, we refine these observations by exploring the hierarchy of intrinsic timescales within the extended auditory network of the temporal lobe.
From a signal processing perspective, timescales quantify the autocorrelation decay of neural signals, while the spectral exponent measures the power decay of aperiodic neural activity (He et al., 2010; Hasson et al., 2015). A steeper exponent may reflect decreased higher-frequency activity, a rotation in the power spectra (Podvalny et al., 2015), or lower levels of neural noise (Voytek et al., 2015; Alnes et al., 2021). As several mechanisms can explain changes in the steepness of power spectra, associating those to neural timescales is neither trivial nor unambiguous. At a physiological level, timescales are considered an indicator of a neural system's memory capacity (Hasson et al., 2015), while the steepness of the spectral exponent around the lower range we studied here is considered a proxy of excitation-to-inhibition balance (Gao et al., 2017).
Importantly, similar to timescales, synaptic excitation and inhibition also manifest hierarchically: while advancing through the visual hierarchy, excitatory connections increase, myelin content decreases, and the expression of genes involved in synaptic transmission increases (Wang, 2020). In our data, the 20-35 Hz spectral exponent was steeper in the temporal cortex than in the hippocampus or amygdala, similar to previous reports (Frauscher et al., 2018), and possibly reflecting higher levels of inhibition, compatible with previous reports of increased inhibition in sensory regions (Wang, 2020).
In our results, the macroscopic ordering that we identify via timescales is mirrored by the spectral exponent and reflects the neurobiological proximity that one would expect between the temporal/entorhinal cortex and hippocampus, which are all characterized by a laminar organization of pyramidal neurons, as opposed to the amygdala whose basolateral nucleus consists primarily of pyramidal cells without preferential orientation, and with a much higher neural density (Dumas et al., 2011).
Overall, our findings support the notion that properties of neural dynamics are intrinsic (Wainio-Theberge et al., 2022); to this, we add that they are also local in nature. Taking advantage of the fine spatial resolution of intracranial recordings in humans, we show that a hierarchy of intrinsic neural dynamics of the extended auditory network manifests as a continuous gradient along the posterolateral to anteromedial axis, following the anatomy of the temporal lobe, both for intrinsic timescales and spectral exponent.
Extending the hierarchy of intrinsic timescales to hippocampus and amygdala
Importantly, contrary to the vast majority of existing studies (Honey et al., 2012; Murray et al., 2014; Gao et al., 2020; Norman-Haignere et al., 2022), we extend the characterization of intrinsic neural dynamics beyond cortical electrodes by including limbic structures, such as the hippocampus and the amygdala. Previous studies have shown that prefrontal or parietal regions assume the role of “higher-order” areas (Honey et al., 2012; Rocchi et al., 2021). Here, we expand these well-studied hierarchies by showing that the hippocampus and amygdala can also be positioned in a “higher” order compared with sensory areas, both in terms of intrinsic dynamics (slower timescales and flatter exponent) and auditory response latencies.
To date, only few studies have attempted to characterize intrinsic timescales in the human hippocampus and amygdala. These report intermixed results, with one MRI study showing gradients of timescales in the hippocampus in the range of few seconds (Raut et al., 2020), and a study of neural firing reporting no differences in timescales between the two structures (Hagemann et al., 2022). In our work, we did not find evidence for gradients of timescales within the hippocampus and amygdala. There are several possible explanations for these diverging results across studies, including the different overall temporal sensitivity of the recorded signals, ranging from seconds to milliseconds, the electrode coverage, or, in the case of hemodynamic responses, signal dropout (Raut et al., 2020), which all together make the comparison of timescales extracted from different recording modalities nontrivial (Manea et al., 2022).
Additionally, the extended auditory network includes the “what” and “where” pathways, comprising prefrontal and parietal areas (Rauschecker and Scott, 2009). The “what,” or rostral, pathway typically shows sustained responses and longer response latencies than the “where,” or caudal, pathway (Jasmin et al., 2019). This dissociation can be observed in nonhuman and human primates (Scott et al., 2011; Camalier et al., 2012; Hamilton et al., 2018). The lack of frontal or parietal electrode coverage in our patient cohort did not allow investigating how timescales are organized along the full extent of these pathways and how they would be positioned relative to hippocampus/amygdala in a putative hierarchy. Future investigations could expand beyond the temporal lobe, allowing a direct comparison of intrinsic timescales in limbic structures and frontal or parietal cortex.
Linking spontaneous intrinsic timescales and auditory processing
Although several studies have posited that short intrinsic timescales may mediate fast responses to incoming stimuli, we now provide direct evidence for the auditory system. Previous studies have either analyzed intrinsic timescales in the auditory system while assessing whole-brain dynamics (Gao et al., 2020; Raut et al., 2020; Golesorkhi et al., 2021a), without the specificity of our work for the auditory system, or have investigated intrinsic timescales during complex auditory stimuli such as speech, as they unfold over time (Honey et al., 2012; Norman-Haignere et al., 2022). Here, we show, for the first time, specifically for the extended auditory system, that a hierarchical organization in spontaneous neural activity is strongly related to the timing of processing short, evoked auditory stimuli.
Importantly, we show, in the same patients and recordings, that the diversity of intrinsic timescales partially explains the richness of auditory responses that are observed in temporal areas in terms of onset and peak latencies, at the single electrode level, with high spatial resolution. Although our analyses are correlational, we posit that this repertoire of spontaneous intrinsic timescales may support the auditory process itself, providing a variety of processing windows (Golesorkhi et al., 2021b) both at a macroscopic level across brain regions, and also at the millimeter level, following the anatomic organization of the temporal cortex. Here, we used pure tones as a simple experimental model of auditory processing. Future studies can examine how the characteristics of auditory stimuli, for example, their frequency or complexity, affect the interplay between spontaneous and evoked activity, and whether trial-by-trial changes in timescales may affect auditory processing and perception of individual sounds.
Last, although the spectral exponent mirrors the macroscopic hierarchy observed via intrinsic timescales, in our data there was no direct link to the timing of auditory responses. Although the spectral exponent is sensitive to auditory processing (Gyurkovics et al., 2022), or levels of attention (Waschke et al., 2021), it does not seem to directly relate to their timing. We speculate that the exponent may capture frequency-specific modulations in neural activity, rather than the response latency itself, which may be better explained by the temporal “memory” of a signal.
In conclusion, our results show a hierarchy of neural dynamics in the extended human auditory network that manifests across cortical and limbic structures, exhibits anatomic gradients with millimeter resolution, and can explain the temporal richness of neural responses to auditory stimuli.
Footnotes
This work was supported by Inselspital University Hospital Bern; the Interfaculty Research Cooperation “Decoding Sleep: From Neurons to Health & Mind” of the University of Bern to A.A., I.B., A.T., and S.L.A.; Swiss National Science Foundation to A.A. and 320030_188737 to A.T.; European Research Council CoG-725850 to A.A.; the Synapsis Foundation to A.A.; the University of Bern to A.A.; and Fondation Pierre Mercier pour la science to A.T. We thank Prof. Johannes Sarnthein for support with recordings and feedback on the manuscript.
M.O.B. holds shares with Epios SA, a medical device company based in Geneva. The remaining authors declare no competing financial interests.
- Correspondence should be addressed to Athina Tzovara at athina.tz{at}gmail.com