Abstract
Despite the prevalence of poverty worldwide, little is known about how early socioeconomic adversity affects auditory brain function. Socioeconomically disadvantaged children are underexposed to linguistically and cognitively stimulating environments and overexposed to environmental toxins, including noise pollution. This kind of sensory impoverishment, we theorize, has extensive repercussions on how the brain processes sound. To characterize how this impoverishment affects auditory brain function, we compared two groups of normal-hearing human adolescents who attended the same schools and who were matched in age, sex, and ethnicity, but differed in their maternal education level, a correlate of socioeconomic status (SES). In addition to lower literacy levels and cognitive abilities, adolescents from lower maternal education backgrounds were found to have noisier neural activity than their classmates, as reflected by greater activity in the absence of auditory stimulation. Additionally, in the lower maternal education group, the neural response to speech was more erratic over repeated stimulation, with lower fidelity to the input signal. These weaker, more variable, and noisier responses are suggestive of an inefficient auditory system. By studying SES within a neuroscientific framework, we have the potential to expand our understanding of how experience molds the brain, in addition to informing intervention research aimed at closing the achievement gap between high-SES and low-SES children.
Introduction
Auditory brain function is shaped by the environment (Blakemore and Cooper, 1970; Hyde and Knudsen, 2002). Environments that are less acoustically and/or socially stimulating can affect the neural systems that underlie auditory perception, leading to delayed auditory development (Eggermont et al., 1997; Chang and Merzenich, 2003), altered neural responses (Sanes and Constantine-Paton, 1983, 1985; Poon and Chen, 1992; Zhang et al., 2002; Hauber et al., 2013), and hyperactivity within the auditory cortex and midbrain (Noreña and Eggermont, 2003; Mulders and Robertson, 2013). The impact of sensory deprivation has been well studied in laboratory animals and humans (Roberts et al., 1998; Kral and Eggermont, 2007; Conway et al., 2009); yet, comparatively little is known about how normal-hearing human populations are affected by auditory impoverishment, conditions where acoustic stimulation is reduced but not absent. Here we examine the link between auditory function and environmental factors in adolescents with varying degrees of maternal education.
Maternal education, income, occupation, and social status combine to predict a child's socioeconomic status (SES), with maternal education being one of the strongest indicators of SES in studies of child development (Hoff et al., 2012). In addition to reduced access to nutrient-rich food and greater exposure to environmental toxins (Evans and Kantrowitz, 2002), socioeconomically disadvantaged children receive less linguistic and cognitive stimulation (e.g., fewer and simpler words) from their caregivers than their more privileged counterparts (Hart and Risley, 1995; Bradley and Corwyn, 2002; Rowe and Goldin-Meadow, 2009; Sheridan et al., 2012; Cartmill et al., 2013). This, combined with greater exposure to unstructured auditory stimulation, as the result of increased noise exposure (Evans and Kantrowitz, 2002), limits opportunities for high-quality learning experiences. Decades of research indicate that this reduced environmental stimulation has profound effects on academic outcomes and executive function (Brooks-Gunn and Duncan, 1997; Hackman and Farah, 2009; Raizada and Kishiyama, 2010). A recent wave of neuroscientific research has further endorsed the idea that sociocultural factors influence brain structure and function (Otero, 1997; Tomarken et al., 2004; Noble et al., 2007, 2012; D'Angiulli et al., 2008; Raizada et al., 2008; Kishiyama et al., 2009; Stevens et al., 2009; Hanson et al., 2011; Jednoróg et al., 2012; Sheridan et al., 2012; for review, see Lipina and Posner, 2012). This work has revealed that the neurobiological systems mediating higher-order functions like language, memory, and executive function—including the perisylvian region, hippocampus, and prefrontal cortex—are especially sensitive to disparities in SES (Noble et al., 2005, 2006, 2007, 2012; Raizada et al., 2008; Evans and Schamberg, 2009; Kishiyama et al., 2009; Hackman et al., 2010; Hanson et al., 2011; Sheridan et al., 2012). Specific to auditory processing, electrophysiological evidence suggests that low-SES children have increased difficulty suppressing irrelevant auditory input during an active-listening, selective-attention paradigm (D'Angiulli et al., 2008; Stevens et al., 2009; Neville et al., 2013). But the question remains as to whether impoverished auditory environments affect automatic sound processing in normal-hearing individuals. Drawing on studies of animals raised in altered environments, we hypothesized that limited environmental stimulation that accompanies socioeconomic disadvantages, in addition to affecting higher-order functions, compromises how the brain represents auditory signals, resulting in differences in auditory neural function even when attentional factors are removed.
To test our hypothesis, we assessed brain activity both in response to, and in the absence of, auditory input. We recorded auditory brainstem responses (ABRs) using a passive listening paradigm in a cohort of normal-hearing adolescents, grouping students based on the highest education level achieved by their mother. Passive electrophysiological recordings have been used extensively to study how the brainstem is fine-tuned by experience (Krishnan et al., 2005; Kraus and Chandrasekaran, 2010). This work has revealed that ABRs can, even under passive conditions, tap into how the nervous system was shaped by previous active engagement with sound. We predicted that adolescents with lower maternal education (no postsecondary education) would have lower literacy levels and working memory ability (Noble et al., 2007; Evans and Schamberg, 2009), coupled with weaker brainstem function than schoolmates with higher maternal education. This weakening, we posited, would translate into more inconsistent ABRs (Hornickel and Kraus, 2013), less robust responses to the input signal (Banai et al., 2009), and an altered resting brain state (Otero, 1997; Tomarken et al., 2004; Marshall et al., 2008).
Materials and Methods
Data were collected as part of a larger ongoing study on adolescent brain development currently underway in the Auditory Neuroscience Laboratory at Northwestern University. Informed written assent was obtained from each participant. Parents/guardians provided informed consent and also completed a questionnaire assessing health, education, language history, and ethnic/racial background. All experimental procedures were approved by Northwestern University's Institutional Review Board.
Participants
Inclusion criteria
All students were incoming or current ninth graders at the time of testing and met the following study inclusion criteria: (1) no history of poor maternal health during pregnancy (e.g., smoking, alcohol consumption, drugs, etc.), (2) normal birth weight, (3) no indication of a neurological disorder, (4) normal audiological status as confirmed by otoscopic examination, distortion product otoacoustic emissions, as well as air and bone conduction audiometry (pure tone averages thresholds >20 dB hearing level), and (5) an IQ >80 (Wechsler Abbreviated Scale of Intelligence Vocabulary and Matrix Reasoning subtests). In addition, participants were excluded if English was not their primary language, if either parent was educated outside of the United States, if the participant lived outside of the United States for >6 months, or if their primary caregivers were not biological relatives. The current analysis includes 66 participants (average age, 14.57 years; 30 females) enrolled in three high schools within the Chicago Public School system.
Participant groups
As part of the participant history questionnaire, the parent/guardian indicated the highest level of education achieved by the biological mother by selecting one of four possible categories: middle school, high school/GED (General Education Development) equivalent, college, graduate/professional. Participants were then divided into two groups that we refer to as the “low” and “high” maternal education groups (hereafter abbreviated low-MaternalEd and high-MaternalEd; n = 33 each, 15 females/group). The low-MaternalEd group included participants whose mothers had a high school education (n = 27) or less (n = 6). If the mother completed some postsecondary schooling, the participant was placed in the high-MaternalEd group. For 46 participants, additional information on the exact years and nature of the postsecondary education was obtained (low, 20; high, 26). For these 46 participants, the average years of maternal education was calculated to be 11.40 ± 2.04 years for the low group and 14.96 ± 1.82 years for the high group, with 12 years corresponding to the completion of senior year of high school (independent t test, t(44) = −6.26, p < 0.001; Table 1).
Distribution of maternal educational achievement within the low-MaternalEd and high-MaternalEd groups
In addition to differing in maternal education, the two groups also differed in paternal education (low, 11.75 ± 2.86 years; high, 13.57 ± 2.86, t(41) = −2.22, p = 0.03) and parental occupation (maternal: Kolmogorov–Smirnov Z = 1.39, p = 0.04; paternal: Z = 1.81, p = 0.003). Maternal and paternal occupations were coded into eight categories according to the General Social Survey (http://www3.norc.org/gss+website/). These categories were as follows: professional or technical, higher administration, clerical, sales, service, skilled, semiskilled, and unskilled work.
The groups were created to be matched on variables that have been previously shown to influence auditory function, including age (Salamy et al., 1975), sex (Krizman et al., 2012a), language background (Krizman et al., 2012b), and hearing thresholds (Hood, 1998; Anderson et al., 2013b; Tables 2, 3). In addition, the groups were created to have an equivalent sample of participants from the low and high groups from each of the three schools (Z = 0.246, p = 1.00). Another factor that was controlled was race/ethnicity, as assessed by parental report. The ethnic distribution was matched between groups (Z = 0.57, p = 0.89; two participants opted not to respond), with Hispanics representing a large majority of both the low and high groups (84 vs 70%, respectively). For the low group, the remaining 16% identified as African-American. In the high group, among the non-Hispanics, the distribution was 24, 3, and 3% for African-American, Caucasian, and Asian categories, respectively. Given that ethnicity and school environment are socioeconomic cofactors (Brooks-Gunn et al., 1996), we were careful to control for both. Through this careful matching, we were able to minimize heterogeneity within our sample, allowing us to more specifically address how early auditory experiences, as measured by maternal education, influence how the brain represents sound.
The low- and high-MaternalEd groups were matched with respect to age, hearing thresholds (pure tone average for right and left ear), and birth weight
Self-reported age of acquisition and extent of current exposure to English and Spanish for the low-MaternalEd and high-MaternalEd groupsa
Reading and executive function testing
In addition to obtaining IQ assessments, students were administered a standardized, age-normed test battery of reading ability and executive function (working memory).
Test of Silent Word Reading Fluency (TOSWRF)
A timed test (3 min) in which students must identify word boundaries in a row of letters. There are no spaces between words and every letter is part of a word (e.g., nospacesbetweenwords).
Test of Word Reading Efficiency (TOWRE)
A measure of real-word and nonword reading ability in a timed format. Students are given 45 s to accurately pronounce as many items in a list as possible. Real-word and nonword scores are combined into a final composite score.
Woodcock Johnson
The test battery included multiple subtests of the Woodcock Johnson (WJII) Achievement and the Cognitive Abilities tests (Woodcock et al., 2001).
Basic Reading Composite Score.
Measures real-word and nonword reading ability in a nontimed format. The subtest represents a composite of the Letter-Word Identification (real-word) and Word Attack (nonword) subtests.
Spelling.
Words are spelled from dictation.
Numbers Reversed.
A series of orally presented numbers (1–9) must be repeated in reverse order in this digit span test (e.g., the series 4, 7, 1 must be repeated 1, 7, 4).
Auditory Working Memory.
The student is asked to reorder an orally presented sequence containing digits and objects (e.g., cat, 4, turtle, 7, 9, pear) first repeating the objects and then the digits, each in the order presented (e.g., cat, turtle, pear; 4, 7, 9).
ABRs
The auditory brainstem is a communication hub within the CNS that serves as a bridge between the cochlea and the cerebral cortex. ABRs, which can be measured under passive listening conditions, reflect past and ongoing sensory experiences and relate to linguistic and cognitive function, as seen in auditory experts and populations with average-to-higher SES (for review, see Kraus and Chandrasekaran, 2010; Skoe and Kraus, 2010; Kraus et al., 2012).
Procedures for measuring ABRs followed published guidelines (Skoe and Kraus, 2010). The collection protocol lasted ∼20 min, during which participants sat comfortably watching a self-selected subtitled movie, while the ABR to 6000 stimulus presentations was passively collected. As is typical of most ABR paradigms, participants were instructed to attend to the movie and ignore the stimulus.
Stimulus
ABRs were recorded to a 40 ms synthesized [da] stimulus that was presented at a rapid rate (10.9/s) to the right ear at 80 dB SPL through an earphone placed in the ear canal. The left ear remained unblocked, making the movie sound track audible (<40 dB SPL) yet not intense enough to mask the stimulus.
This particular syllable was chosen because it is common to many languages of the world (Maddieson, 1984), its acoustic characteristics cause it to be perceptually challenging (i.e., a short syllable containing rapid changes; Tallal and Stark, 1981), and it has been used routinely to study auditory function in other populations (Banai et al., 2009; Krizman et al., 2012a). The stimulus begins with an initial noise burst that is followed by a formant transition period, during which the fundamental frequency (F0) and lower three formants (F1, F2, F3) ramp linearly (F0, 103–125 Hz; F1, 264–656 Hz; F2, 1700–1240 Hz; F3, 2580–2500 Hz).
Recording parameters
ABRs were recorded with a customized recording protocol within the AEP (auditory evoked potentials) module of the Navigator Pro, an evoked potential system (Natus Medical). Ag/AgCl-plated electrodes were placed at the vertex of the head (Cz, active electrode), right ear (reference electrode), and high forehead (ground electrode). Responses were online bandpass filtered from 100 to 2000 Hz, with an online artifact rejection threshold of ±23 μV. The stimulus was presented in alternating polarity. To maximize the response to the speech formants, responses to the alternating polarities were subtracted, following conventions described previously (Aiken and Picton, 2008; Skoe and Kraus, 2010). To gauge the consistency of the response over the course of the recording, two subaverages of 3000 trials were created.
Analyses
Three independent measures were extracted from the ABR using custom routines coded in Matlab (Mathworks): Response Consistency, Speech Encoding, and Spontaneous Neural Activity. Response Consistency reflects the stability of the ABR over the recording session and is measured by correlating the two subaverages (Hornickel and Kraus, 2013). The closer the r value is to 1, the more repeatable the response is over the 20 min session. r Values were Fisher transformed to z-scores for statistical analyses. Speech Encoding measures the strength with which the speech stimulus is represented in the ABR. The response was transformed into the frequency domain using fast Fourier analysis and the energy within the first formant frequency range (264–656 Hz) was averaged. The first formant carries important phonological information regarding the identity of the speech sound. Response Consistency and Speech Encoding were computed for the frequency-following region of the response, which for this stimulus falls between 11.4 and 40.6 ms (Krizman et al., 2012a). Spontaneous Neural Activity represents the magnitude of neural activity in the absence of a stimulus. It was calculated as the average root-mean-square magnitude of the response during the 15 ms prestimulus interval.
Operational definitions
Auditory impoverishment
We define “auditory impoverishment” as a reduction in auditory input relative to standard conditions. This reduction can arise either (1) because the sensory organs (inner ear) have been compromised by an insult, genetic mutation, disease, or normal aging and/or (2) because of limited acoustics or acoustic degradation within the environment. Environmental impoverishment comes in a variety of forms, ranging from complete sound isolation to a partial reduction in sound quality. Examples of auditory impoverishment include listening through a device that distorts the signal (e.g., cell phones), listening in reverberant or noisy conditions, hearing sounds that are less complex and/or socially less engaging, and hearing statistically fewer sounds overall (e.g., hearing fewer words per hour from caregivers). In this paper, we examine a specific type of auditory impoverishment, namely what happens when sensory organs are intact (and therefore hearing thresholds are normal) but the quality and quantity of a child's auditory input is limited as the result of reduced language-based social interactions and greater exposure to unstructured, and behaviorally irrelevant, sounds, such as ambient noise.
Auditory neural acuity
We use the term “auditory neural acuity” to collectively refer to our three measures of brainstem function. We define “auditory neural acuity” as the nervous system's ability to resolve and reliably transmit fine-grained information about acoustic signals within the environment. By controlling for hearing thresholds, we have minimized the role of the cochlea—the peripheral, sensory organ of hearing—to isolate the neural component of hearing.
Statistical analyses
To test the primary hypothesis that auditory impoverishment, as operationalized here by maternal education, affects how auditory signals are encoded in the brain, the low- and high-MaternalEd groups were compared using a one-way ANOVA. One-way ANOVAs were also used to compare the low and high groups on each of the behavioral measures of IQ, reading, and working memory. For each comparison, F and p statistics are reported, along with η2, the estimated effect size. Following Cohen's conventions (Cohen, 1988), an effect size between 0.01 and 0.059 is considered small, between 0.059 and 0.138 is medium, and ≥0.138 is large.
In a follow-up analysis, we used regression techniques to examine the relationship among the neural and behavioral measures. First, using multiple linear regression models, we tested which combination, if any, of the behavioral measures predict brainstem function. Second, using binary logistic regression models, we examined how the behavioral and neural measures uniquely and collectively predict group membership. Through this binary logistic analysis, we aimed to test whether neural acuity is a unique predictor of maternal education that cannot be accounted for by reading or cognitive ability. Binary logistic regression is a variant of regression, in which the dependent variable is categorical (i.e., low- vs high-MaternalEd). After initially examining colinearity among the independent variables via Pearson's correlations, we selected a subset of neural and cognitive variables to include in the linear and binary logistic regression models with the goal of reducing multicolinearity (redundancy) within the models and improving the estimate of predictability for each independent variable entered into the model. For both types of regression, predictors were entered simultaneously into the model.
For the binary logistic analysis, we tested three models: the first included all three brainstem measures (three independent variables); the second model included representative measures of brainstem function, reading, and working memory (three independent variables); and the third model also factored in current exposure to English and hearing thresholds. We included the third model because the amount of exposure to the target language can influence both cognitive and language function (Dixon et al., 2012), and also because speech-evoked ABRs are sensitive to language experience (Krishnan et al., 2005; Krizman et al., 2012b). Hearing thresholds can likewise affect cognitive, reading, and brainstem measures (Hood, 1998; Moeller et al., 2007). Although the groups were matched with respect to both current language exposure and hearing thresholds, given that they can affect the neural and behavioral measures under consideration, we included them in the binary logistic regression analysis to further rule out their role in explaining our findings. All statistical analyses were performed in SPSS v21 (IBM).
Results
Reading and cognitive measures
Group differences emerged for all literacy-related tests except the spelling test. In addition, the low-MaternalEd group performed more poorly than the high-MaternalEd group on working memory ability but only on the more linguistically engaging subtest. See Table 4 for summary statistics.
Group comparisons for the low- and high-MaternalEd groups on reading and cognitive measuresa
Neural measures
For the low-MaternalEd group, ABRs were less stable over time (Response Consistency: F(1,64) = 6.34, p = 0.01, η2 = 0.09) compared with the high-MaternalEd group (Fig. 1). This instability was accompanied by weaker encoding of speech-specific information (Speech Encoding: F(1,64) = 5.12, p = 0.03, η2 = 0.06; Fig. 2). In addition, for the low-MaternalEd group, responses were noisier overall, as evidenced by greater activity during the silent intervals when no stimulus was being played (Spontaneous Neural Activity: F(1,64) = 6.50, p = 0.01, η2 = 0.14; Fig. 3).
Adolescents with low maternal education (black) have less consistent auditory brain responses (ABRs) than adolescents with high maternal education (red). A, For the low-MaternalEd group (top) and high-MaternalEd group (bottom), the ABR during the first half of the recording (black/red, respectively) is compared with the response during the second half of the recording (gray). For the high group, the two waveforms align more closely in morphology than for the low group (p = 0.01). B, The average correlation (r value) between the first and second halves of the recording is 0.44 for the low group versus 0.57 for the high group (r = 1, when the two halves are identical). Mean + 1 SEM are plotted.
Adolescents with lower maternal education (black) have smaller auditory brainstem responses (ABRs) to the speech stimulus than adolescents with higher maternal education (red). A, The speech frequency spectrum is plotted. The band of energy in the 264–656 Hz range corresponds to the first formant of the speech syllable. The first formant carries important phonological information regarding the identity of the speech sound. B, Bottom, The frequency spectrum of the ABR is plotted for the two groups. For the response to the first formant, the average energy is reduced in the low-MaternalEd group compared with the high-MaternalEd group (p = 0.03), suggesting that the speech stimulus is encoded less robustly in the low group. Mean + 1 SEM are plotted for both groups.
Adolescents with lower maternal education (black) have more spontaneous neural activity than those with higher maternal education (red). A, Neural activity is plotted for each participant in the low-MaternalEd (top) and high-MaternalEd (bottom) groups during the 15 ms of silence preceding each stimulus presentation. Time 0 refers to when the stimulus starts playing. B, The average energy is greater in the low group compared with the high group (p = 0.01). Increased energy during the absence of auditory input is argued to be indicative of greater neural noise.
Correlations among measures
Table 5 presents the correlations among the brainstem, reading, and working memory variables. Response Consistency was correlated with the other two brainstem variables. Yet, the amount of Spontaneous Neural Activity and the strength of Speech Encoding did not relate, suggesting that they reflect distinct neural mechanisms. For the subset of subjects for whom the exact years of maternal education was known (n = 46), we also examined relationships between this variable and each of the neural and behavioral measures (Table 5). This subanalysis revealed that more years of maternal education was associated with higher Response Consistency and more robust Speech Encoding (Table 5).
Correlations among brainstem, cognitive, and years of maternal education variablesa
Multiple linear regression analysis
Using multiple linear regression analysis modeling, we tested whether reading ability and auditory working memory had a combined influence on brainstem function. Based on the outcomes of the one-way ANOVAs and the colinearity among variables (Table 5), we selected two independent variables (TOSWRF, Auditory Working Memory) and one dependent measure (Response Consistency) to include in the model. Response Consistency was selected as the representative brainstem variable because it was correlated with the two other measures of brainstem function but the other two measures were not interrelated, suggesting that Response Consistency may capture aspects of the other two neural measures. TOSWRF was chosen as the primary measure of reading because it was correlated with the other measures of reading, but unlike the other reading measures (TOWRE, Basic Reading) it did not correlate with either measure of working memory (Digits Reversed, Auditory Working Memory). Auditory Working Memory was selected over Digits Reversed because the low- and high-MaternalEd groups did not differ statistically on the Digits Reversed measure, and the effect size was comparatively low. In addition, compared with the Numbers Reversed subtest, this Auditory Working Memory test has a higher receptive language requirement due to the open set nature of the test (Flanagan, 2001). By requiring greater manipulation of the input, this test may also require greater demands from the frontal executive function system (St Clair-Thompson, 2009).
The results of the regression indicated that reading and working memory were not significant predictors of brainstem function (R2 = 0.001, F(2,65) = 0.038, p = 0.963). Adding in maternal education group, as a binary predictor variable, the combined model predicted 11% of the variance of brainstem function (R2 = 0.110, F(3,65) = 2.546, p = 0.064), with maternal education being the only significant individual predictor (standardized regression coefficient B = 0.159, t = 2.748, p = 0.008).
Binary logistic regression
To examine how the neural and behavioral measures combined to predict group membership, we used binary logistic regression modeling. The results of the three models are reported in Table 6. All three models produced statistically significant results. Model 1, which included all brainstem variables, predicted ∼65% of the cases. In Model 2, when factoring in the representative reading, working memory, and brainstem variables, the prediction increased to 75.8% correct, with Response Consistency and Auditory Working Memory making unique, independent contributions to the model (both p < 0.05) and with reading ability (TOSWRF) adding to the predictive power of the model without carrying significant predictive power on its own (p = 0.08). In Model 3, we confirmed that neither peripheral hearing levels nor current exposure to English contributed to the model.
Binary logistic regression
Discussion
We examined the link between SES and adolescent brain function in a normal-hearing population using maternal education as a proxy for SES (Bradley and Corwyn, 2002). In addition to finding that low maternal education is associated with poorer performance on tests of reading and working memory, we find that ABRs from the low-maternalEd group were noisier, more variable, and represented the input signal more weakly. Because low maternal education increases the risk for auditory impoverishment (Evans, 2004), poorer auditory neural acuity in the low maternal education group is argued to arise from impoverished early auditory experiences, which, as our findings suggest, become “embedded” in a person's biology (Hertzman and Boyce, 2010) to negatively shape the nervous system's automatic response to sound.
The impact of small differences in SES
We studied a cohort of adolescents who likely have had different auditory experiences, as operationalized by differences in maternal education (Bradley and Corwyn, 2002), but were matched in age, sex, hearing, ethnicity, and the schools they attended. This homogeneity of our sample permitted more strict evaluation of the link between maternal education, a component of SES, and brain function while controlling for many confounding variables (D'Angiulli et al., 2008; Helvik et al., 2009; Pungello et al., 2009). By recruiting from the Chicago Public School system, a “majority-minority” district that serves primarily low-income populations, most, if not all, of the participants in our study could be considered socioeconomically disadvantaged (Chicago Public Schools, 2013), with our lower and higher groups representing gradients on the lower end of the socioeconomic spectrum. As such, our study is capturing a relatively narrow slice of the SES spectrum. Nevertheless, we observe striking differences in auditory neural acuity, reading ability, and working memory based on differences in maternal education within this lower end.
It is unclear at this point whether our findings generalize to other populations, or what the full ramifications of these neural differences might be. But we believe that this study makes an important first step by establishing that auditory impoverishment and auditory neural acuity are linked. In addition, given that our two groups had similar demographics, our findings highlight the potential for small differences in early auditory experiences to cascade into larger neural and cognitive gaps over time (Brooks-Gunn and Duncan, 1997; Haskins and Rouse, 2005; Fryer and Levitt, 2006; Stevens et al., 2009). Based on the outcomes of this study, we plan to undertake a larger research program that will delve more deeply into understanding the transient and longstanding effects of early socioeconomic adversity on auditory neural activity, cognitive function, and social mobility across different ages, ethnicities, and geographies.
Normal auditory processing depends on the quality and quantity of environmental stimulation
We screened for conductive and sensorineural hearing loss to ensure that all participants had normal hearing thresholds at the time of testing and that audibility did not differ between groups. This is an important experimental control given that ear infections and hearing loss are more common in underprivileged populations (Paradise et al., 1997; Olatoke et al., 2008; Helvik et al., 2009; Siddartha et al., 2012). By controlling for hearing thresholds, we are in a stronger position to argue that the differences in auditory neural acuity between the low and high groups reflect differences in environmental stimulation throughout childhood. Children from lower-SES backgrounds are exposed to less complex and linguistically rich input in addition to hearing fewer words per hour from their caregivers (Hart and Risley, 1995; Sheridan et al., 2012; Cartmill et al., 2013). Low-SES children also watch more television than their higher-SES peers, which may increase their passive exposure to sound (Evans, 2004), but further limits opportunities to interact with a diverse, as well as an emotionally and socially engaging, soundscape (Kuhl et al., 2003; Sanes and Woolley, 2011). In addition to being underexposed to linguistically stimulating environments, socioeconomically disadvantaged children face greater exposure to noise pollution (Fidell, 1978; Evans and Kantrowitz, 2002; Kohlhuber et al., 2006), further compromising the quality of the auditory input by increasing the amount of unstructured, irrelevant input. Thus, while their aggregate auditory input might not be diminished, the complexity and clarity of the auditory input and how they interact with sound is atypical. This combination of environmental factors, we argue, leads to diminished neural encoding of auditory signals in children with low maternal education.
Noise in, noise out: noisy auditory environment increases neural noise
Within urban populations, income and amount of noise exposure are known to be correlated (Fidell, 1978), with the average noise exposure being ∼15 dB SPL higher in households earning ∼$35,000/year compared with those earning $70,000/year (70 dB vs 55 dB SPL, income adjusted for inflation; Fidell, 1978). Additionally, although ambient noise levels differ by neighborhood, socioeconomic gradients within a neighborhood enable wealthier residents to afford higher-quality housing that dampens ambient noise (Fyhri and Klæboe, 2006). Therefore, even within our population, there may be inequalities in noise exposure. While ambient noise levels may be too low to decrease audibility, chronic low-level noise exposure has been shown to decrease reading and cognitive performance (Cohen et al., 1973; Bronzaft and McCarthy, 1975; Hygge et al., 2002), change how children discriminate and attend to auditory stimuli (Cohen et al., 1973; Bronzaft and McCarthy, 1975; Evans et al., 1995), and leave the auditory system more vulnerable in the long term (Evans and Kantrowitz, 2002; Kujawa and Liberman, 2006). As has been shown in laboratory animals, chronic noise exposure can also alter neural function by reducing the brain's sensitivity to sound (Zhou and Merzenich, 2012), delaying auditory development (Chang and Merzenich, 2003), accelerating age-related hearing loss (Kujawa and Liberman, 2006), and increasing neural noise, the spontaneous firing of neurons in the absence of stimulation (Costalupes et al., 1984; Noreña and Eggermont, 2003; Seki and Eggermont, 2003; Pienkowski and Eggermont, 2012; Mulders and Robertson, 2013).
Consistent with the idea that noisy auditory environments increase neural noise, we found that adolescents from the low-MaternalEd group, compared with those from the high group, have increased neural activity in the absence of sound input. Neural models indicate that when the input to a neuron is noisier, firing rate becomes more variable, ultimately limiting the amount of sensory information that can be transmitted (Stein et al., 2005). Additionally, when input to the auditory system is not coherent from one instance of a sound to the next, signal transmission is less effective, interrupting the ability to form a stable representation of the auditory world (Ahissar, 2007; Centanni et al., 2013; Hornickel and Kraus, 2013). We have observed this increase in neural noise and decrease in response consistency by probing the auditory system; however, we believe it is reflective of increased spontaneous activity throughout the brain. We conclude that increased neural variability and noise observed in the low-maternalEd group reflect an inefficient sensory system in which the ability to translate an external stimulus into a meaningful electrical signal is compromised (Centanni et al., 2013; Hornickel and Kraus, 2013). This noise and instability in the encoding of sensory signals may in turn affect higher-level functions that depend on that signal (Fox et al., 2010).
Auditory impoverishment may lead to the decoupling of sensory and cognitive functions
Our data advance the theory that auditory experiences help to forge the “scaffolding” that binds sensory and cognitive information in the brain (Conway et al., 2009; Kraus et al., 2012; Krizman et al., 2012b). We posit that this biological scaffolding, once formed, can be traversed in both “directions,” allowing for stable transmission of both bottom-up and top-down signals and the interlocking of these signals. Yet, if auditory impoverishment occurs early in life, this scaffolding may never fully form, or its neuroanatomical infrastructure may be intact but weaker, altering the typical relationships between sensory and cognitive functions (Sak and Ruben, 1981; Sininger et al., 1999; Kral and Eggermont, 2007; Conway et al., 2009). Although both sensory and cognitive systems were found to be linked to maternal education in our study, sensory and cognitive variables were uncorrelated. This is in contrast to other work, primarily from populations with average to high SES backgrounds, in which correlations between sensory and cognitive functions have been observed (Hornickel et al., 2009; Kraus et al., 2012). Although we leave open the possibility that our behavioral tests lack the sensitivity to illuminate weak correlations between sensory and cognitive abilities if they did exist (Stevens et al., 2009) and/or that the spread of SES is too narrow to see such relationships, the absence of a relationship could alternatively indicate that sensory and cognitive systems have not synched up, and that the normal processes that lead to a symbiotic tethering of sensory and cognitive systems have been compromised. We theorize that when this scaffolding is absent or malformed it further weakens the already deficient sensory and/or cognitive systems, such that sensory systems cannot benefit from top-down cognitive processes that refine how incoming sounds are processed, and cognitive processes cannot benefit from that refined sensory input.
By suggesting that alterations to the prefrontal cortex and other higher-order systems may lead to and subsequently reinforce a breakdown in sensory–cognitive feedback, this sensory–cognitive scaffolding theory provides new mechanistic insight into the previously described differences in higher-order functions associated with SES (Noble et al., 2007). This breakdown, coupled with reduced auditory acuity, may then translate into disparities in how auditory information is processed in both active (D'Angiulli et al., 2008; Stevens et al., 2009) and passive (current study) listening contexts.
Auditory function and social mobility
Intergenerational social mobility in the United States is among the lowest of the developed economies (Beller and Hout, 2006): children born into poverty are unlikely to move out of poverty. While there are myriad reasons for this low social mobility, we argue that reductions in fundamental auditory neural function may play a part. Because auditory disorders run in families (e.g., Maziade et al., 2000), auditory neural acuity may be transmitted from one generation to the next by genetic mechanisms. Yet, our study emphasizes that auditory neural acuity may also reflect the auditory environment that is shared between mother and child. Moreover, given that the heritability of some traits is altered under suboptimal environmental conditions (Turkheimer et al., 2003; Jukes and Grigorenko, 2010), this supports the possibility that environmental factors play an especially prominent role when the auditory environment is impoverished.
What can be done to break this intergenerational negative feedback loop and reverse the effects of poor environmental conditions on auditory neural acuity? By documenting how socioeconomic factors impede sensory and cognitive function, it may be possible to use that information to develop appropriate intervention strategies that target specific neurobiological processes. Music education may be one promising remediation strategy (Fitzpatrick, 2006; Kraus and Chandrasekaran, 2010; Patel, 2011). Classroom-based auditory enrichment, in the form of assistive listening devices, may also be a successful option. In children with dyslexia, this type of remediation has produced more stable ABRs (Hornickel et al., 2012), a characteristic of the response that was compromised in our group with lower maternal education and which has been previously shown to be a factor for reading ability (Centanni et al., 2013; Hornickel and Kraus, 2013). There is accumulating evidence that the auditory brainstem remains malleable throughout life (Carcagno and Plack, 2011; Hornickel et al., 2012; Song et al., 2012; Anderson et al., 2013a; Strait et al., 2013), further supporting the prospect that acuity can be improved with appropriate intervention strategies. Interventions that foster quality parent–child interactions at home also offer great promise in bolstering the neurobiological systems compromised by socioeconomic conditions (Neville et al., 2013). Optimistically, it may just be a matter of finding the right kind of remediation, knowing when such training is most effective, and delivering it in a way that is cost-effective and appropriate for use in a variety of classroom and home settings (Moore, 2002; Als et al., 2004; Raizada and Kishiyama, 2010). However, this is not to downplay the necessity to improve the social and educational systems that led to disparities in maternal education in the first place. Our findings suggest that even small gains in maternal education—the difference between a high school degree and a few college courses—could make a difference in a child's sensory and cognitive development and break the intergenerational negative-feedback loop that affects social mobility (Brooks-Gunn and Duncan, 1997).
Footnotes
This work was supported by Supported by Northwestern University Knowles Hearing Center, NSF SMA-1015614, and the Mathers Foundation. We thank Vance Gao for his help with data analysis, as well as Douglas Savitsky, Trent Nicol, Ellie Thompson, and Travis White-Schwoch for their comments on an earlier draft of this manuscript. We also thank the students and their families for participating and the schools and teachers, especially Kate Johnston, for their support of the project. We also extend our appreciation to Rafael Escobedo-Quiroz and the other members of the Auditory Neuroscience Laboratory, past and present, for their help in data collection and participant recruitment.
The authors declare no competing financial interests.
- Correspondence should be addressed to Nina Kraus, PhD, Northwestern University, 2240 Campus Drive, Evanston IL, 60208. nkraus{at}northwestern.edu