Abstract
Dyslexics are diagnosed for their poor reading skills, yet they characteristically also suffer from poor verbal memory and often from poor auditory skills. To date, this combined profile has been accounted for in broad cognitive terms. Here we hypothesize that the perceptual deficits associated with dyslexia can be understood computationally as a deficit in integrating prior information with noisy observations. To test this hypothesis we analyzed the performance of human participants in an auditory discrimination task using a two-parameter computational model. One parameter captures the internal noise in representing the current event, and the other captures the impact of recently acquired prior information. Our findings show that dyslexics' perceptual deficit can be accounted for by inadequate adjustment of these components; namely, low weighting of their implicit memory of past trials relative to their internal noise. Underweighting the stimulus statistics decreased dyslexics' ability to compensate for noisy observations. ERP measurements (P2 component) while participants watched a silent movie indicated that dyslexics' perceptual deficiency may stem from poor automatic integration of stimulus statistics. This study provides the first description of a specific computational deficit associated with dyslexia.
SIGNIFICANCE STATEMENT This study presents the first attempt to specify the mechanisms underlying dyslexics' perceptual difficulties computationally by applying a specific model, inspired by the Bayesian framework. This model dissociates between the contribution of sensory noise and that of the prior statistics in an auditory perceptual decision task. We show that dyslexics cannot compensate for their perceptual noise by incorporating prior information. By contrast, adequately reading controls' usage of previous information is often close to optimal. We used ERP measurements to assess the neuronal stage of this deficit. We found that unlike their peers, dyslexics' ERP responses are not sensitive to the relations between the current observation and the prior observation, indicating that they cannot establish a reliable prior.
- anchoring deficit
- computational neuroscience
- event-related potentials
- predictive coding
- perception
- statistical learning
Introduction
The controversy surrounding the deficits underlying dyslexics' difficulties is still unresolved. The prevailing theory claims that dyslexics' phonological representations, whose adequacy is crucial for efficient usage of the alphabetical code, are impaired (Snowling, 2000). However, dyslexics perform well on some tasks that rely on adequate phonological representations (for review, see: Ramus and Ahissar, 2012). A computationally motivated imaging study found that dyslexics' task-specific top-down tuning to phonological processing, as revealed in their thalamic response, is impaired (Díaz et al., 2012), suggesting that it may be access to phonological representations which is deficient in dyslexia (Boets et al., 2013; Ramus, 2014). Nevertheless, other studies have found that dyslexics' automatic responses to basic sounds are noisier than those of good readers (Nagarajan et al., 1999; Hornickel and Kraus, 2013). The anchoring deficit hypothesis (Ahissar et al., 2006; Ahissar, 2007) states that dyslexics' deficits stem from poor utilization of stimulus repetitions. In this framework dyslexia does not result from a deficit in stimulus processing (sensation) or alternatively from a deficit in memory representation, but from a deficit in the match between them (similar to the concept of predictive coding; Díaz et al., 2012).
The current study was designed to generalize and quantify the impaired-anchoring hypothesis in a computational model, which specifies the experimental conditions that enhance dyslexics' difficulties. A two-tone discrimination task was used because it is simple, quantifiable, and provides a reliable predictor of performance in phonologically related tasks (adults: Amitay et al., 2002; Ben-Yehudah et al., 2004; Oganian and Ahissar, 2012; children: Mengler et al., 2005; Banai and Yifat, 2012; Banai and Ahissar, 2013). Previously, we showed that individuals with dyslexia have difficulties benefitting from simple stimulus repetitions on this task (Ahissar et al., 2006; Ahissar, 2007; Oganian and Ahissar, 2012). Here we used a version of this task that contains no repetitions. Instead, it has richer statistics, which allowed us to analyze the impact of stimulus history on single trials. Specifically, it allowed us to separate the trials in which performance is expected to “benefit” from stimulus history from those trials in which these statistics are expected to hamper performance.
We used an on-line computational model (Raviv et al., 2012), which quantitatively specifies the integration of context on the two-tone discrimination task, to disentangle dyslexics' disabilities. The model posits that rather than comparing the second tone in each trial to the first one, listeners compare the frequency of the second tone to a weighted average of the frequency of the first tone and a memory trace of the frequencies of the tones presented in previous trials. This may allow listeners to reduce the disruptive effect of noise in the internal representation of the first tone (e.g., noise added in the retention interval) by combining this noisy representation with that of previous trials in a way that qualitatively resembles Bayesian reasoning (Lu et al., 1992; Preuschhof et al., 2010; Ashourian and Loewenstein, 2011). Note, however, that a Bayesian inference specifies exactly how the prior distribution should be integrated in perception. By contrast, the prior distribution is not explicitly learned in our model, and the extent to which previous trials affect perception is a parameter. Using the model we conclude that dyslexics' poorer perceptual performance is associated with suboptimal incorporation of prior knowledge in perception.
To further characterize the processes underlying the integration of the stimulus statistics in perception, we measured ERP responses that are sensitive to stimulus statistics (the P2 component; Tremblay et al., 2001, 2010). Unlike controls' P2, dyslexics' P2 did not show such sensitivity.
These results pave the way for a better understanding of dyslexics' perceptual deficits as a computational impairment related to the learning and incorporation of prior sensory information.
Materials and Methods
Participants.
Twenty-eight native Hebrew speakers (14 dyslexics and 14 good readers), who are students at the Hebrew University (mean age = 25.4 years; STD = 2.2; 18 females), were considered in this study. Recruitment was based on ads at the Hebrew University. Monetary compensation for participation was according to standard student rates. The study was approved by the Hebrew University Committee for the Use of Human Subjects in Research. All dyslexic participants (except one, whose reading, phonological, and reasoning profile matched the profile of the other dyslexic participants) had been diagnosed before the study by clinicians using a standard didactic test approved by the Israeli Ministry of Education. Six participants, with >2 years of formal musical education were excluded, so that musical training would not be a major contributor to their pitch sensitivity (Micheyl et al., 2006; Parbery-Clark et al., 2011). Two participants with poor Block Design scores (lower than a normalized score of 7) were also excluded from the study. All participants filled out a questionnaire regarding any neurological or psychiatric disorders. None of the participants reported any such disorders.
Procedure.
All participants were administered three sessions on three different days. In session 1 participants were administered a series of cognitive assessments. In session 2, ERPs were recorded. Participants watched a silent movie while a series of stimuli was presented to them. In session 3 the same series of stimuli was presented, and participants actively engaged in the frequency discrimination task.
General assessments.
General cognitive abilities were assessed using two standard tasks. (1) Nonverbal reasoning ability was measured with the Block Design, a standard test for assessing visuospatial reasoning (WAIS-III; Wechsler, 1997). (2) Short-term verbal memory was evaluated with the standard Digit Span task (forward and backward; Hebrew version of WAIS-III; Wechsler, 1997).
Phonological skills.
Phonological skills were assessed using two types of phonological tasks. (1) Phonological decoding was tested when single pseudoword and real-word reading were assessed using two standard Hebrew lists designed by Deutsch and Bentin (1996). One list contains 24 punctuated Hebrew words and the other contains 24 punctuated pseudowords; i.e., words with Hebrew morphology but no meaning. Both accuracy and rate were scored. (2) Phonological awareness was assessed using the Spoonerism task (MacKay, 1970; Möller et al., 2007). Participants heard (Hebrew) word pairs and were asked to switch the first phonemes of the two words and respond vocally (e.g., /laila tov/, “good night” in Hebrew, should be switched to /taila lov/). Both accuracy and rate were scored. We merged accuracy and speed in all phonological tests into a combined phonological score by averaging the z-score (relative to the control group distribution) within each subject.
Stimulation procedure for two-tone frequency discrimination.
Participants were presented with sequences (blocks) of tone pairs (50 ms, 70 dB each tone; 620 ms intertone intervals and 1380 ms interpair intervals). They were asked to indicate which of the two tones in each pair (i.e., trial) had a higher pitch. No feedback was provided.
Participants were presented with a fixed easy-to-difficult sequence of stimuli based on the average sequence of frequency differences that were given to participants (students, adequate readers with no, or only minor, musical background) in an adaptive staircase procedure (three down, one up), which converges at ∼80% correct (Nahum et al., 2010). Consequently, the average performance was similar in the two protocols. This pseudoadaptive sequence of tone pairs was chosen in order to use the same easy-to-difficult sequence of stimuli for all subjects, under both the passive and active conditions.
Each participant performed 700 trials in two separate blocks. (1) In the No-Reference protocol, on each trial, a tone was chosen randomly from 800 to 1200 Hz. The other tone was determined according to the frequency difference that was preassigned to that trial and could be higher or lower than the randomly chosen tone. The order of the tones was random. This block consisted of 300 trials. The difference between the two tones began at 25% and decreased gradually to 18%. (2) In the Reference-Interleaved condition, one of the tones was 1000 Hz in all trials (i.e., Reference tone) and the other tone was determined according to the frequency difference that was preassigned to that trial. The order of the tones was predetermined: odd number trials, Reference first; even number trials, Reference second. This block consisted of 400 trials. The difference between the two tones began at 25% and decreased gradually to 5–7%. The order of the blocks was counterbalanced across participants.
Parameter estimation of the implicit memory model.
We modeled the responses made by the participants using the Implicit Memory Model (IMM; Raviv et al., 2012). According to this model, choices in each trial are determined by the difference between the frequency of the second tone and a weighted average (M1) of the noisy memory of present and past frequencies of the first tone. Formally, where a(t) denotes the choice of the participant on trial t, sgn is the sign function, f2(t) is the frequency of the second tone on trial t, and where f1(t) is the frequency of the first tone on the trial, z(t) is an independent Gaussian random variable with zero mean and variance σ2, and we assume resetting of the initial conditions M1(t = 1) = f1(t = 1) + z(t = 1) (Shteingart et al., 2013). This model is characterized by two parameters, σ and η. The first parameter σ denotes the level of internal noise in the process of “sensing and memorizing” the first tone on each trial. The second parameter η denotes the weight of previously stored stimuli in the current comparison.
Therefore, the probability to respond “first tone higher” is as follows: where Φ is the cumulative normal distribution function and σe(t) is the effective response variance at trial t, which depends on σ, η, and t: For each participant, we estimated the two parameters, η and σ, which minimize the squared distance between the predicted response probabilities of the model and the observed responses in the frequency discrimination task. To assess the reliability of this estimate, we bootstrapped the trials for each participant by 1000 samples of 300 trials with replacement and re-estimating parameters.
Calculating the optimal η.
The optimal η is defined as the value of η that minimizes the probability of an error in the stimulation schedule for a fixed value of σ.
We defined mental difference (D(t)) and correct difference (C(t)) on each trial as follows:
Thus the probability to make a correct decision on each trial is as follows:
According to Sheppard's Median Dichotomy Theorem (Sheppard, 1899; Kendall et al., 1987, their p. 482):
where ρm =
The probability of a correct response is monotonous over ρm; thus it is sufficient to maximize ρm over η:
where ρf =
The optimal η is a solution to the equation: Resulting in where Since this computation assumes an infinite number of trials, the solution is only an asymptotic approximation of the optimal η. To verify its proximity to the optimal η, with a finite number of trials, we ran numerical simulations. We estimated the η that yielded the highest accuracy rate in the actual sequence of trials used in the experiment for various values of σ and compared the two solutions. The numerical and analytical calculations nearly overlapped.
We then estimated the optimal impact of implicit memory given participants' estimated σ; i.e., for each participant we found η*—the η that would maximize success on the auditory task. We defined inadequacy, or suboptimality, of the implicit memory weighting of each participant as the difference between η* and the estimated η.
ERP recordings and analyses.
Electrophysiological activity was recorded in a sound-attenuated room while participants heard the tone sequences either passively (session 2) or while actively performing the task (session 3). Sounds were produced using MATLAB (The MathWorks) and were presented by E-Prime 1.1 software (Psychology Software Tools). The EEG was recorded from 32 active Ag-AgCl electrodes mounted on an elastic cap using the BioSemi ActiveTwo tools and recording software. Electrode sites were based on the 10–20 system (American Electroencephalographic Society, 1991). Two additional electrodes were placed over the left and right mastoids. Horizontal EOG was recorded from two electrodes placed at the outer canthi of both eyes. Vertical EOG was recorded from electrodes on the infraorbital and supraorbital regions of the right eye in line with the pupil.
EEG and EOG signals were sampled at 256 Hz, amplified and filtered with an analog bandpass filter of 0.16–100 Hz. Off-line analysis was performed using Brain Vision Analyzer 1.05 software (Brain Products). The EEG signal was digitally bandpass filtered between 1 and 30 Hz to remove large drifts in signal and high-frequency noise. Independent component analysis was trained on the entire length of each block to identify components that reflect eye blink or eye movement-evoked electrical activity. An eye-related component was identified by its time correlation with the occurrence of blinks or saccades. This relationship between the identified component and eye blink activity was verified by controlling the component's scalp distribution that was typical of eye-related electrical activity. Data were referenced to the nose channel to remove external electrical influence. Artifact rejection was applied to the nonsegmented data according to the following criteria: any data point with EOG or EEG > ±100 μV was rejected along with the data ± 300 ms around it. In addition, if the difference between the maximum and the minimum amplitudes of two data points within an interval of 50 ms exceeded 100 μV, data ± 200 ms around it were rejected. Finally, if the difference between two adjacent data points was >50 μV, the data ± 300 ms around it were rejected. Trials containing rejected data points were omitted from further analysis (average omission of 17 trials per subject).
For ERP averaging across trials, the EEG was parsed to 2000 ms epochs starting 500 ms before the onset of the first stimulus in each pair and averaged separately for each electrode. The baseline was adjusted by subtracting the mean amplitude of the prestimulus period (500–150 ms before the onset of the first stimulus in the trial) of each ERP from each data point in the epoch. The prestimulus baseline period was calculated from this time interval to exclude effects of anticipatory responses that precede informative anticipated stimuli, such as contingent negative variation (CNV;Walter et al., 1964).
ERP analysis was based on the epochs that were recorded with electrode Cz (at the vertex of the scalp) after they were processed as described above. This electrode measured the most prominent response to the auditory stimuli. Each participant's data (accumulated during the two blocks) were analyzed separately.
Results
Participants' cognitive profile
Dyslexics' performance did not differ from controls' performance on general reasoning skills as measured by the Block Design (Snowling, 2000). However, as expected, their performance was poorer than the controls' performance on the measures of verbal memory and phonological skills (see Table 1). All of the dyslexic participants were poorer decoders than all of the control participants. Specifically, the combined phonological score (average z-scores across all phonological tasks) of the best reader in the dyslexic group was 0.3 SDs lower than that of the poorest reader in the control group.
The effect of the stimulus statistics on performance in frequency discrimination
To evaluate the impact of the stimulus statistics on perception parametrically, we used the two-tone frequency discrimination task. On each trial, participants are presented with two sequentially presented pure tones and are instructed to indicate which had a higher pitch (Fig. 1A–C; see also Materials and Methods). The frequencies of the two stimuli were drawn from a broad distribution, a protocol we denoted as the No-Reference protocol. Although frequency discrimination tasks are traditionally used as an assessment of low-level sensory bottlenecks, we have shown that in this (No-Reference) protocol, which is devoid of a repeating standard tone, performance is substantially affected by the statistics of previous trials (Raviv et al., 2012). When the two tones are higher than the average, performance has been shown to be better when the second tone is higher than the first. Similarly, when the two tones are lower than the average, findings have indicated that performance is better when the second tone is also lower than the first (Romo et al., 2002; Hairston and Nagarajan, 2007; Alcalá-Quintana and García-Pérez, 2011; Raviv et al., 2012). This result can be understood in a Bayesian framework where participants use prior information about the distribution of frequencies to compensate for noise in the representation of the memorized tone. Loosely speaking, this computation results in a “contraction” of the representation of the first tone to the mean frequency in the experiment.
Based on these findings, we divided the trials according to the sign of the predicted contribution of the statistics of the experiment to the trial. Specifically, we divided the trials into three types, based on their relationship to the mean frequency. Bias+ trials were defined as those trials in which the experiment's statistics was predicted to improve performance. Namely, (1) the frequencies of the two tones were either both higher than the mean frequency or both lower than the mean frequency and (2) the frequency of the second tone was more extreme than that of the first tone (higher when the two tones were above average and lower when they were below average; Fig. 1A,D,E, yellow zones). In these trials the contraction of the first tone toward the mean increases its perceived difference from the second tone, and was expected to improve performance. In contrast, Bias− trials were trials in which the statistics was predicted to hamper performance. Namely, when the two tones were above the average, the second tone was lower than the first and when they were lower than the average, the second tone was higher than the first (Fig. 1B,D,E, gray zones). In these trials, contracting the first tone toward the mean frequency decreased its perceived difference from the second tone and was expected to be detrimental to performance. Bias0 trials (Fig. 1C,D,E, white zones) were trials in which the first and the second tone flanked the mean frequency.
In line with a previous study (Raviv et al., 2012), we found that trial type had a marked effect on the ability of controls to discriminate the two tones. The probability that controls would correctly discriminate the frequencies of the two stimuli in the Bias+, Bias0, and Bias− regions spanned a broad range of 92, 81.8, and 54.9%, respectively. This was found despite the fact that the experiment was designed in a way that there was almost no difference in the “objective” difficulty of the three regions (quantified as the absolute difference between the frequencies of the two tones on a logarithmic scale depicted as the distance of the points from the diagonal in Fig. 1D,E). When quantifying performance in terms of d′, we found that the median d′ was significantly better in the Bias+ regions than in the Bias0 regions [Median (interquartile range - IQR), Bias+ d′: 3.2 (2); Bias0 d′: 2 (1.9); Wilcoxon test, p < 0.005] and was significantly better in the Bias0 regions compared with the Bias− regions (Bias− d′: 0 (2.4); Wilcoxon test, p < 0.001). In fact, the control group's performance in the Bias− regions was not significantly different from chance (Wilcoxon test, p = 0.55).
In contrast, dyslexics' performance in the Bias− region was significantly above chance (58.2% Wilcoxon test, p < 0.05), whereas their accuracy in the Bias+ regions was lower than that of the controls (72.4%). Their performance in the Bias0 region was intermediate (70.7%). When quantifying their performance in terms of d′, their median d′ was not significantly better in the Bias+ regions than in the Bias0 regions [Median (IQR), Bias+ d′: 1 (1); Bias0 d′: 1 (1); Wilcoxon test, p = 0.36], although it was significantly better in the Bias0 regions than in the Bias− regions [Bias− d′: 0.5 (0.8); Wilcoxon test, p < 0.02].
To test whether dyslexics' sensitivity to trial type was significantly less than that of the controls, we assessed the performance of each participant individually. This is illustrated in Figure 2, where the value of d′ of each participant in each region (Bias+, Bias0, and Bias−) is plotted for all participants (Fig. 2A, Controls; B, Dyslexics). To quantify the sensitivity to the prior distribution we computed the difference between the values of d′ in the Bias+ and Bias− regions for each participant (Fig. 2C). Overall, dyslexics' sensitivity, as measured by this difference, was significantly smaller than controls' sensitivity [Mann–Whitney test for Condition (Bias+ vs Bias−) × Group (controls vs dyslexics) interaction: z = 3.5, p < 0.001]. These results further indicate that dyslexics were less influenced by the statistics of the experiment than controls. Note that the dyslexics' but not the controls' performance was significantly above chance level even in the Bias− region, indicating that their reduced sensitivity to the prior distribution did not stem from a “floor effect.” That is, dyslexics' overall poorer performance cannot account for the smaller difference between their performance on the Bias+ and Bias− trials, since they performed better than chance in the Bias− region.
Modeling the effect of statistics with the IMM
The division of the trials into Bias+, Bias−, and Bias0 drew on the Bayesian framework, which specifies how performance in a noisy system can benefit maximally from accumulating environmental statistics and incorporating it into its decision making. As shown above, human performers use such statistics even in simple two-tone discriminations. Nevertheless, this incorporation of the statistics deviates from the predictions of an optimal Bayesian performer (Raviv et al., 2012). To quantify listeners' performance, we used a simplified model (IMM) in which listeners do not know the full distribution of the stimulus statistics. Rather, they weigh past trials using an exponentially decaying function (Raviv et al., 2012).
The model (Eq. 2) is characterized by two parameters: η, the contribution of previous trials (η = 0, no contribution; the larger the value of η, the larger the contribution), and σ,the level of internal noise (on a scale of percentage difference between the two tones: the larger the value of σ the noisier the within-trial representation). We used the IMM to estimate the values of η and σ for each participant. We tested the model by simulating it on the same task using the estimated parameters. Qualitatively, the differential performance in the three trial types and the different performance level of dyslexics and controls was captured by the model (Fig. 3). Quantitatively, the difference in performance between the Bias+ and Bias− regions was slightly larger in the experiment than in the model, in particular for the control participants.
One prediction of the Bayesian framework is that the effect of the stimulus statistics on behavior should increase when the level of internal noise increases. Qualitatively, this prediction is intuitive. If the representation of the stimuli is noiseless, performance cannot be improved by incorporating prior information. However, if the representation of a stimulus is noisy, prior information would be useful, and the noisier the representation, the larger the weight that should be given to this earlier in the discrimination task. One study in fact showed that in the visual modality, increasing the level of internal noise enhanced the contribution of prior knowledge to perception and decision making (Ashourian and Loewenstein, 2011).
The IMM model makes no assumptions regarding the relationships between participants' η and σ. Nevertheless the model can be used to determine the extent to which participants' weighting of previous trials (their η) was close to optimal given their within-trial noise (σ) and the stimulus statistics (in the sense of maximizing their success rate; Fig. 4A, green).
Analysis of the parameters characterizing our participants indicated that on average, the estimated value of σ was higher among the dyslexic participants [Median (IQR), controls: σ = 35 (40)%; dyslexics: σ = 98 (213)%; Mann–Whitney test, z = 2.2, p < 0.05]. Given their higher σ, the optimal model solution predicted that the dyslexics' η should also be higher than the control's η. However, the dyslexics' η was similar to that of the controls [Median (IQR), controls: η = 0.52 (0.5); dyslexics: η = 0.41 (0.53); Mann–Whitney test, n.s.].
As shown in Figure 4A, controls' ηs (blue squares) were nearly optimal (green line) given their σs, indicating that their weighting of the history was similar to that of an “ideal listener” (in the framework of the IMM). In contrast, dyslexics' ηs were lower than expected given their σs (red squares). To quantify this group difference, we calculated the difference between the optimal and the actual η for each participant. We found that Dyslexics' underweighting of implicit memory was significantly larger than the Controls' underweighting (Fig. 4B; Mann–Whitney test, z = 2.6, p < 0.01). Hence, given their internal noise level, Dyslexics, in contrast to Controls, do not give sufficient weight to prior information.
Note that our simple, two-parameter IMM model only assumes within-trial noise. In this framework, the dyslexics' underweighting of the stimulus statistics was suboptimal. An alternative account would posit that dyslexics' memory is noisy. However, introducing this to the model would require the addition of a third parameter, whereas our model accounts well for controls' performance. On the other hand, we could assume optimal weighting of the history (i.e., the weighting of the history is determined by the stimulus statistics) and replace (η) with a parameter that estimates the noise of previous trials (memory noise). However, this would require an additional optimality assumption.
Together, the findings show that in the framework of IMM, controls' weighting of past events was nearly optimal. In contrast, dyslexics' weighting of these past events was too low, indicating a deficit in adequate incorporation of prior knowledge into perception.
ERP results
The IMM model posits that in the two-tone discrimination task, the listener compares the frequency of the second tone to a linear combination of the frequencies of the first tone in the current and previous trials. This suggests a process in which the comparison is preceded by the formation of a combined representation. We hypothesized that if dyslexics' reduced weighting of previous trial stems from an impaired formation of an integrated representation, reduced sensitivity to stimulus statistics may be apparent even before the second tone is presented. To test this hypothesis, we measured ERPs. We focused on the dynamics and magnitude of the P2 component, which is an automatic response evoked by the auditory cortex (Sheehan et al., 2005; Mayhew et al., 2010). Previous studies, using both oddball (mismatched negativity, MMN; Haenschel et al., 2005; Baldeweg, 2007; Tong et al., 2009) and discrimination paradigms (Tremblay et al., 2001, 2010; Ross and Tremblay, 2009) have shown that the magnitude of this component increases with stimulus repetitions, suggesting that this component is sensitive to the statistics of the experiment. We hypothesized that P2's sensitivity to stimulus repetitions is a special case of its analog sensitivity to the congruency between the current stimulus and the prior distribution. Therefore, we predicted that the magnitude of the control's P2 would be larger in Bias+ trials than in Bias− trials, since the average distance of the first tone from the mean frequency was smaller in the Bias+ trials than in the Bias− trials (as shown in Fig. 1A,B). Consequently, the first tone in the Bias+ trials was more congruent with the prior than the first tone in the Bias− trials. We further predicted that dyslexics' P2 would not be as sensitive to trial type.
We recorded ERPs under both passive and active conditions in separate sessions. In the passive condition, the same series of stimuli was presented to the participants while they were watching a silent movie and were not asked to perform any task. For each participant in each of the specified trial types, we calculated the area under the curve between 150 and 250 ms after the first tone's onset as his/her individual P2 area. As predicted, we found that the controls' evoked response (Fig. 5A) was sensitive to the trial type. This was visible in the P2 component of the response to the first tone. In the Bias+ trials, this response was, on average, significantly larger than the response to the first tone of the Bias− trials (Wilcoxon test, p < 0.005). However, the dyslexics' P2 was not sensitive to trial type (Fig. 5B; Wilcoxon tests, p = 0.46). The difference between controls' and dyslexics' P2 sensitivity to trial type was also significant (Group × Condition interaction; Mann–Whitney test, z = 2.5, p < 0.05). We repeated this analysis when participants were asked to actively perform the task with the same stimuli, and found similar results. controls' P2 was sensitive to the trial type (Fig. 5C; Wilcoxon test, p < 0.05), although the magnitude of this effect was smaller than in the passive condition (perhaps due to masking by additional active components). Dyslexics' P2 did not significantly differ between the two trial types (Fig. 5D; Wilcoxon tests, p = 0.67). In this condition, the interaction was not significant (Condition × Group; Mann–Whitney test, z = 1.2, p = 0.22). Together, these results support the hypothesis that dyslexics' computational deficit is associated with a failure to reliably integrate their on-line representation with the prior distribution.
In principle, we expected that individuals' P2 sensitivity to the stimulus statistics would be correlated with their weighting of previous trials (η). This correlation is expected both if P2 reflects the result of the process of incorporation of previous trials into current memory (IMM), thus directly reflecting the magnitude of the contribution of previous trials, and if it measures the reliability of the memory trace, which is not captured in the simple IMM model. However, as shown in Figure 5, top right insets, single subject distribution, while at the group level the difference in P2 sensitivity was very consistent, the magnitude of P2 sensitivity to stimulus statistics was too small (in terms of signal-to-noise ratio) to enable a reliable calculation of these correlations.
Dyslexics' insensitivity to stimulus statistics in a protocol with a reference
Dyslexics' reduced sensitivity to stimulus statistics leads to nonintuitive predictions on specific trials that should be more challenging for them in other protocols with more structured statistics. We examined the interleaved reference protocol (Nahum et al., 2010). In this protocol, a reference (1000 Hz) tone is presented on every trial. However, on odd trials it is presented first (Ref first), whereas on even trials it is presented second (Ref second). Intuitively (and consistent with the anchoring deficit hypothesis), dyslexics' performance is expected to be impaired in both, since both types of trials contain a repeated reference. However, the rationale behind the IMM model prompts a specific prediction with respect to different trial types, since previous trials are specifically integrated into the representation of the first tone on each trial. It therefore predicts that controls should benefit substantially from previous statistics in Ref first trials. In these trials, integrating previous statistics reduces the noisiness of the representation of the first tone without modifying its mean. In contrast, in Ref second trials, the second tone is (approximately) equal to the mean of the distribution. Therefore, the incorporation of the prior (approximately the mean) to the representation of the first tone in the same manner decreases the perceived difference and hampers performance. If dyslexics' sensitivity to the statistics of the experiment is indeed reduced, the difference between their performances on the two trial types should be significantly smaller than that of controls.
To test this prediction, all participants performed this interleaved protocol. In line with the IMM prediction, and consistent with previous observations (see also Romo and Salinas, 2003; Nahum et al., 2010), we found that the performance of controls in Ref first trials was substantially and significantly better than their performance in the Ref second trials (Fig. 6, blue; Ref first: mean 86 ± 2% SEM correct; Ref second: 69 ± 4% correct; Wilcoxon test, p < 0.001). In contrast, in line with the IMM's prediction, dyslexics' performance did not differ between these trial types (Fig. 6 red; Ref first: 64 ± 4% correct; Ref second: 66 ± 4% correct; Wilcoxon test, p = 0.71). The group difference in sensitivity to trial types was also significant (Condition × Group interaction; Mann–Whitney test for interaction between Reference position and group, z = 3, p < 0.005).
Discussion
Summary of results
We dissociated the factors underlying discrimination performance into factors that stem from internal noise within a trial and those that stem from context; i.e., the stimulus statistics. We found that the context effects were significantly and substantially larger in the population of adequate readers, compared with dyslexics. When calculating the optimal weighting of previous trials, given the within-trial noise and the stimulus statistics, we found that dyslexics' weighting but not controls' weighting was indeed significantly lower than optimal. Finally, to tap dyslexics' specific difficulty in adequately weighting previous trials, we recorded ERPs. These recordings suggested that dyslexics' perceptual difficulties are associated with inadequate automatic updating and integrating of the context of the experiment with their current stimulus response.
Dyslexics' specific pattern of reduced sensitivity to the first tone in the pair, as suggested by the IMM model and observed in the ERP measures, also yielded nonintuitive predictions for a novel protocol that motivated the last experiment. Here the reference tone was presented either as the first or as the second tone. The findings showed that dyslexics' difficulties were specific to Ref first trials.
We interpret all these results as an indication that dyslexics (but not controls) underweigh history in perceptual decisions. However, dyslexics' weighting of history could alternatively be close to optimal while their deficit could be associated with increased memory noise (a possibility that goes beyond the IMM, since it requires additional assumptions). In that case, increasing the weight of previous trials may not enhance their performance since it will also increase the memory noise.
Relation to previous studies of dyslexia
The IMM is an extended formalization of the anchoring hypothesis of dyslexia. This theory was motivated by the observation that controls' better performance in psychoacoustic tasks (McAnally and Stein, 1996; Witton et al., 1998; Ramus et al., 2003; Ahissar et al., 2006; Ahissar, 2007) and in speech discrimination (McArthur and Bishop, 2005; Ahissar et al., 2006; Boets et al., 2007) is restricted, to a large extent, to experiments in which a single reference stimulus or a small range of reference stimuli are used (Ahissar et al., 2006; Ahissar, 2007; Chandrasekaran et al., 2009; Banai and Ahissar, 2010; Oganian and Ahissar, 2012). The theory posits that repeated stimuli serve as an anchor, and thus boost performance when these stimuli are subsequently used. According to this theory, the ability to track such simple regularities is deficient in dyslexia.
The IMM presented here proposes a specific, well-defined computation that is impaired in dyslexics' incorporation of stimulus statistics. As such, the model specifies the conditions where incorporating previous trials is expected to improve perception and those where it is expected to hamper perception. Thus, unlike the intuitive anchoring hypothesis, stimulus repetition is not necessary (e.g., Bias+ trials in a protocol with no reference), and may even not be beneficial (e.g., Ref second trials in the reference interleaved protocol). The IMM predicts that dyslexics will only perform worse than controls on the trials that benefit from stimulus statistics.
The IMM assumes that the stimulus statistics are continuously learned and used. We found that this was indeed the case for controls and only to a lesser extent for dyslexics. The ERP recordings further illustrated the analog nature of the automatic tracking mechanisms. The traditional ERP measure of automatic tracking of sounds, the MMN component, is sensitive to the violation of repetition, and is studied in the context of oddball paradigms (Näätänen et al., 1978). The MMN component is often smaller among dyslexics (Baldeweg et al., 1999; Kujala et al., 2003; Bishop, 2007). The findings here show that controls' automatic ERP response is parametrically sensitive to the distance of the stimulus from the peak of the stimulus distribution in previous trials. No such sensitivity was found in the dyslexic population.
The anchoring hypothesis and its computational implementation are inconsistent with the hypothesis that poor phonological representations are the core deficit in dyslexia (Snowling, 2000) for at least two reasons. Not only did we find difficulties in the processing of nonphonological stimuli, but in the computational framework, dyslexics' main difficulty lay in poor usage of context. Specifically they used an overly low η given their somewhat higher σ.
However, our computational model is tightly related to hypotheses that associate dyslexics' difficulties with a failure to make effective predictions that can facilitate task performance (“predictive coding”; Díaz et al., 2012). However, it is also compatible with hypotheses that dyslexics are less resilient to external noise (Sperling et al., 2005, 2006; Beattie et al., 2011; Conlon et al., 2012; Partanen et al., 2012; the “noise exclusion hypothesis”). According to the Bayesian framework underpinning the IMM, the prior information is used to compensate for the noise in the representation of the stimuli. We found that dyslexics do not properly adjust the weight of previous trials to the level of internal noise (Fig. 4). Functionally, this results in reduced noise exclusion. However, putting this broad concept in a computational framework leads to a counterintuitive prediction: when the context is compromising (e.g., Bias− like conditions leading to disruptive predictions), dyslexics should not do worse and may even do better than controls since this mechanism for “noise exclusion” is biased by the prior statistics.
Other studies have suggested that dyslexics' implicit sequence learning is impaired (Vicari et al., 2003; Stoodley et al., 2006; Jiménez-Fernández et al., 2011). This phrasing is very broad. For example, it is not clear whether the representation of syllables should be intact, while the representation of words, which are perhaps formed by implicit sequential learning (conditional probabilities between adjacent syllables; Saffran et al., 1996), should be impaired. Our computational model assumes poor incorporation of basic attributes (zero-order statistics), as observed by the poor usage of the mean frequency of the experiment. It predicts that the same deficit may lead to reduced sensitivity to the prevalence of single syllables.
Other studies (Ramus and Szenkovits, 2008; Boets et al., 2013; Ramus, 2014) have proposed that phonological representations remain intact, but that dyslexics' access to these representations is inefficient, perhaps due to impaired connectivity between the superior temporal areas that encode auditory stimuli and the frontal areas (e.g., Broca) that use them. This hypothesis assumes a clear distinction between representations and access, although it does not specify what “access” is and which retrieval conditions should pose difficulties (e.g., whether implicit priming should be impaired). Our own perspective, as described above, is that perception is never devoid of context. Ease of retrieval is heavily affected by the availability of adequate predictions, which can substantially facilitate the process. Thus, difficulty in using priors may impair the efficiency of retrieval. Nevertheless, the nature of retrieval processes in different behavioral contexts is far from being understood.
Recently, it was shown that a hierarchical model of perception is able to reconstruct or decode the dynamics underlying generated series of stimuli (Yildiz et al., 2013). Our observations may perhaps be interpreted within this framework, and suggest that dyslexics' deficit resides in the higher and slower level of the hierarchy, which is characterized by larger time constants, or in their impaired communication with lower levels (Boets et al., 2013). The slower dynamics in this higher level are able to track cumulating evidence and form a prior based on the underlying statistics of the stimuli. In turn, these predictions are used to modulate the lower levels of the hierarchy to better match the upcoming stimuli and the requirements of the task. This coincides with data showing that among dyslexics, task-related top-down modulation is impaired (Díaz et al., 2012).
Implications, limits, and future directions
We posited that inadequate usage of priors in the context of spoken or written language underlies dyslexics' reading deficit. This generalization is based on the fact that proficient reading of single words and even pseudowords, which are impaired in dyslexia, heavily relies on priors related to sound sequences (e.g., phonological, morphological and orthographic; Norris, 2013). In fact, dyslexics' reading of familiar words is more serial (van der Leij and van Daal, 1999; Ziegler et al., 2003; Zoccolotti et al., 2005; Martens and de Jong, 2006), and their “visual word form area,” an area of reading expertise that develops in parallel to readers' gradual reliance on word priors, i.e., word familiarity, is not adequately developed (Paulesu et al., 2001; Shaywitz et al., 2002; McCandliss et al., 2003).
However, even within the framework of the simple IMM there are open questions. First, we also found a group difference for within-trial noise (σ), which we did not expect, and should be further investigated. Additionally, it is unclear which dimensions of the priors are impaired along with frequency, e.g., intervals (Banai and Ahissar, 2006), or frequency changes in time (Goswami et al., 2011), and at which timescales (the scale of phonemes, syllables, words, or phrases). Moreover, dyslexics' deficit could also include other modalities, such as the spatial distribution of visual stimuli (Moores et al., 2011; Franceschini et al., 2012). A better understanding of these features is particularly important when designing novel training procedures to improve dyslexics' reading skills.
Footnotes
This work was supported by the Israel Science Foundation Grant 616/11, the EPFL-HUJI Brain Collaboration, and the Gatsby Charitable Foundation.
The authors declare no competing financial interests.
- Correspondence should be addressed to Merav Ahissar, Psychology Department, Hebrew University of Jerusalem, Jerusalem 9190501, Israel. msmerava{at}gmail.com