Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Reduced Neural Integration of Letters and Speech Sounds in Dyslexic Children Scales with Individual Differences in Reading Fluency

  • Gojko Žarić ,

    gojko.zaric@maastrichtuniversity.nl

    Affiliations Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, University of Maastricht, Maastricht, Netherlands, Maastricht Brain Imaging Center (M-BIC), Maastricht, Netherlands

  • Gorka Fraga González,

    Affiliations Department of Psychology, University of Amsterdam, Amsterdam, Netherlands, Rudolf Berlin Center, Amsterdam, Netherlands

  • Jurgen Tijms,

    Affiliations Department of Psychology, University of Amsterdam, Amsterdam, Netherlands, IWAL Institute for Dyslexia, Amsterdam, Netherlands, Rudolf Berlin Center, Amsterdam, Netherlands

  • Maurits W. van der Molen,

    Affiliations Department of Psychology, University of Amsterdam, Amsterdam, Netherlands, Rudolf Berlin Center, Amsterdam, Netherlands, Amsterdam Brain and Cognition, University of Amsterdam, Amsterdam, Netherlands

  • Leo Blomert †,

    † Deceased.

    Affiliations Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, University of Maastricht, Maastricht, Netherlands, Maastricht Brain Imaging Center (M-BIC), Maastricht, Netherlands

  • Milene Bonte

    Affiliations Department of Cognitive Neuroscience, Faculty of Psychology and Neuroscience, University of Maastricht, Maastricht, Netherlands, Maastricht Brain Imaging Center (M-BIC), Maastricht, Netherlands

Abstract

The acquisition of letter-speech sound associations is one of the basic requirements for fluent reading acquisition and its failure may contribute to reading difficulties in developmental dyslexia. Here we investigated event-related potential (ERP) measures of letter-speech sound integration in 9-year-old typical and dyslexic readers and specifically test their relation to individual differences in reading fluency. We employed an audiovisual oddball paradigm in typical readers (n = 20), dysfluent (n = 18) and severely dysfluent (n = 18) dyslexic children. In one auditory and two audiovisual conditions the Dutch spoken vowels/a/and/o/were presented as standard and deviant stimuli. In audiovisual blocks, the letter ‘a’ was presented either simultaneously (AV0), or 200 ms before (AV200) vowel sound onset. Across the three children groups, vowel deviancy in auditory blocks elicited comparable mismatch negativity (MMN) and late negativity (LN) responses. In typical readers, both audiovisual conditions (AV0 and AV200) led to enhanced MMN and LN amplitudes. In both dyslexic groups, the audiovisual LN effects were mildly reduced. Most interestingly, individual differences in reading fluency were correlated with MMN latency in the AV0 condition. A further analysis revealed that this effect was driven by a short-lived MMN effect encompassing only the N1 window in severely dysfluent dyslexics versus a longer MMN effect encompassing both the N1 and P2 windows in the other two groups. Our results confirm and extend previous findings in dyslexic children by demonstrating a deficient pattern of letter-speech sound integration depending on the level of reading dysfluency. These findings underscore the importance of considering individual differences across the entire spectrum of reading skills in addition to group differences between typical and dyslexic readers.

Introduction

Though many children learn to read without any problems, 5% to 10% of children are affected by developmental dyslexia and never acquire proficient reading skills despite normal cognitive abilities and schooling opportunities [1], [2]. A lack of reading fluency has been pinpointed as the most persistent and impaired characteristic of developmental dyslexia [3], although studies also show a large inter-individual variability in the level of reading (dys)fluency [4], [5]. A first important step in obtaining reading fluency in alphabetic orthographies, the formation of solid letter-speech sound correspondences, poses an immediate hurdle for beginning dyslexic readers [6][9]. Accordingly, in addition to tackling well-known difficulties in phonological processing, i.e. recognizing and manipulating the sound structure of language [10], [11], many dyslexia interventions contain a condition focused on teaching letter-speech sound correspondences [12][14]. In recent years, the neurofunctional basis of the difficulties in building these correspondences has been elucidated by neuroimaging findings showing reduced neural integration of letters and speech sounds in dyslexic children [15][17] and adults [18]. How this reduced neural integration scales with individual differences in the level of reading (dys)fluency remains an interesting and open question.

In largely transparent languages, accurate letter-speech sound identification and discrimination is typically observed after one year of reading instruction [19]. However, evidence from neuroimaging studies shows a dissociation between knowing which letters belong to which speech sounds and the automatic neural integration of these associations, with the latter showing a much more protracted period of incremental learning throughout primary school in typical readers [20], [21]. One of the first studies in this domain recorded magnetoencephalography (MEG) in literate Finnish adults and indicated a dynamic process of letter-speech sound integration with audiovisual convergence starting around 225 ms in bilateral temporal cortices [22]. Subsequent functional magnetic resonance imaging (fMRI) evidence revealed the involvement of the superior temporal sulcus, as well as auditory cortex (superior temporal gyrus/planum temporale) [23]. More specifically, results indicated a modulating influence of letters on speech sounds as a consequence of feedback from the superior temporal sulcus to the superior temporal gyrus, but only if the letter and speech sound were presented simultaneously and in a passive listening paradigm [23][25].

Evidence for the protracted developmental time-course of letter-speech sound integration comes from a series of studies measuring electroencephalography (EEG) responses in a passive crossmodal “oddball” paradigm [16], [21], [26]. In an oddball paradigm, at a latency of 100–200 ms, a mismatch negativity (MMN) response is elicited by a rare sound (deviant, or oddball) that is presented in a sequence of frequent (standard) sounds [27]. The MMN represents an automatic change detection response sensitive to deviation from traces in auditory short-term memory [27], [28]. Multiple experiments have shown that the MMN is sensitive to language-specific speech sounds in adults and in children [27], [29][37], as well as to audiovisual integration [21], [26], [38][42]. Furthermore, in school-aged children the speech evoked MMN response may be followed by an additional late negativity (Late MMN or LN) in a broader time window from 300 to 700 ms, the amplitude of which decreases in adulthood [43][45], suggesting fast and automatic discrimination and representation of speech syllables in adults while the developing brain still needs additional processing resources to reach to the same result. In the paradigm of Froyen and colleagues [16], [21], [26], in an auditory and two audiovisual experimental blocks, the Dutch speech sound/a/was presented as the standard stimulus, and the Dutch speech sound/o/as the deviant stimulus. In the audiovisual blocks, the letter ‘a’ was presented together with the standard and deviant speech sounds, leading to a double violation: the deviant speech sound/o/violated the expectation built by both the standard speech sound/a/and the letter ‘a’. In order to investigate the temporal window of integration the letter ‘a’ was presented either simultaneously, or 200 ms before speech sound onset. Due to the double violation, literate adults showed a significant enhancement of the MMN response in the audiovisual as compared to auditory blocks, but, reminiscent of the fMRI findings [23][25], only when letters and speech sounds were presented simultaneously, indicating an early and automatic audiovisual integration [26]. Using the same paradigm, 8 year-old typically reading children after one year of reading instruction only showed a significant audiovisual enhancement of a late negativity (LN) evoked around 650 ms after speech sound onset and only when the letter preceded the speech sound, indicating a slow and immature neural integration of letters and speech sounds [21]. A more developed, bit still not adult-like pattern was apparent in 11 year-old typically reading children after four years of reading instruction. That is, these more advanced readers showed a cross-modal enhancement of the earlier MMN when the letter preceded the speech sound, and a cross-modal LN enhancement when the letter and speech sound were presented simultaneously [21]. Although adults did not show these late letter effects in this passive paradigm using simple speech stimuli [26], orthographic influences on spoken language processing around 400–700 ms have been reported in adults when using a more complex metaphonological task [46], [47]. Thus, whereas the observed cross-modal MMN enhancement may be a neurophysiological marker of initial and automatic letter-speech sound integration and/or representation [27], [28]), the crossmodal LN enhancement may reflect more elaborate, explicit associative processes that, in children but not in adults, are recruited for the integration of simple letter-speech pairs.

The same crossmodal MMN paradigm [16], [21], [26], was used in 11 year-old dyslexic children after 4 years of reading instruction, and indicated reduced integration of letters and speech sounds, with significant letter effects only in the late LN window [16]. Further evidence for a reduced sensitivity to letter-speech sound associations in dyslexia comes from fMRI studies showing an underactivation of superior temporal cortex during letter-speech sound integration in 9-year-old dyslexic children [15] and adults [18], as well as from the absence of a crossmodal MMN enhancement for spoken-written syllable pairs in dyslexic adults [41]. Similarly, recent studies suggest reduced orthographic-phonological integration in dyslexics on word reading tasks [48][51].

In the current study, we investigate EEG measures of letter-speech sound integration in 9-year-old typical and dyslexic readers and specifically test whether they scale with behavioral measures of reading fluency and reading-related skills. To this end we employ an adapted version of the audiovisual paradigm of Froyen and colleagues [16] in typical readers (n = 20) and age-matched dyslexic children (n = 36) with variable levels of reading (dys)fluency. By focusing on 9-year old children, we extend previous ERP findings to a different age group, 2.5 years after reading instruction, intermediate between the previously studied age groups, namely typical readers with 1 or 4 years of reading instruction [21] and dyslexic readers with 4 years of reading instruction [16]. This age range also coincides with the age of participants from a previous fMRI study reporting reduced neural letter-speech sound integration in dyslexics, as well as correlations between these neural measures and behavioral reading scores [15], allowing us to further investigate the same processes with a different technique at the same level of development. As the morphology, amplitude and timing of ERP components are known to be age dependent [52][56], we chose to use a chronologically age matched control group, rather than a reading age matched control group. Finally, including a larger group of dyslexic children with variable levels of reading fluency, allowed us to demonstrate brain-behavior relations not only across typical and dyslexic readers but also to show interindividual differences within the group of dyslexic readers.

Methods

Participants

In total, 61 children (41 dyslexic and 20 typical readers) participated in the EEG experiment. All children were 3rd graders, native Dutch speakers, having received 2.5 years of reading instruction. Data of 56 children were included in the analysis, including 36 dyslexic readers (9.0 years old, range: 8.2–9.9; 16 girls; 6 left-handed, as assessed by an adapted version of Annett’s handedness questionnaire [57], and 20 age-matched typical readers (8.8 years old range: 8.3–9.5; 12 girls; 2 left-handed). Data of five children with dyslexia were discarded, 4 children did not complete the EEG measurement and 1 child moved excessively during the EEG measurement.

Dyslexic children were recruited from a specialized institute for dyslexia and reading problems. Prior to the present study, they were diagnosed as dyslexic after an extensive cognitive psycho-diagnostic procedure by the institute. Each of the dyslexic children scored in the lowest 10 percentile of the age appropriate group on standard reading tests (see below). Other behavioral scores such as phonological skills or rapid automatized naming (RAN), were not included as selection criteria. Typical readers had reading score percentile of 25 or higher and were recruited from 3rd grades of primary schools in the same area and had similar social-demographic characteristics. All children had a normal hearing and normal or corrected to normal vision, as reported by parents. Comorbidity with behavioral and/or attention disorders was assessed with the Child Behavior Checklist (CBCL) from the Achenbach system of empirically based assessment (ASEBA) [58] which was completed by the parents, and used as an exclusion criterion. Children with below average IQ were also excluded from the experiment. Written informed consent was obtained from the parents of each child. Each child received a small present after the experiment and travel costs were reimbursed. The approval for the research was obtained from a local ethical committee of the Developmental Psychology department of the University of Amsterdam.

Behavioral tests

Prior to the EEG measurements, each of the children performed standard language tests including the word reading, spelling, letter-speech sound identification, letter-speech sound discrimination, rapid automized naming (RAN) and basic reaction time subtests of the 3DM battery (Dyslexia Differential Diagnosis; 3DM, [19], as well as a one-minute word reading test (EMT) [59], and reading of a short story ‘De kat’ (‘The cat’) [60]. In addition, we used a paper and pencil version of the RAVEN Coloured Progressive Matrices to assess non-verbal IQ scores (RAVEN CPM) [61]. We performed a median split of the 36 dyslexic children based on their fluency score to form groups of dysfluent dyslexic readers (n = 18) and severely dysfluent dyslexic readers (n = 18) (for similar approaches see [62][64]). Because our three individual word reading fluency measures were highly correlated (r >.919), reading fluency was quantified using a composite score of the 3DM word reading tests, the EMT, and “De Kat”. Subject characteristics and the results of the behavioral tests are shown in Table 1. Group differences were tested using one way ANOVAs for each of the pairwise group comparisons and are reported in the Results section.

thumbnail
Table 1. Behavioral reading scores: Descriptive data and statistical group comparisons for typical readers, dysfluent and severely dysfluent dyslexic readers.

https://doi.org/10.1371/journal.pone.0110337.t001

Stimuli

The stimuli were the same as in Froyen et al. [16], [21], [26] including auditory stimuli consisting of the Dutch vowels/a/and/o/spoken by a native Dutch female speaker. The vowels were digitally recorded (sampling rate 44.1 kHz, 16 bit quantization), band-pass filtered (180–10.000 Hz), resampled at 22.05 kHz and matched for loudness with Praat software [65]. Phoneme duration was 384 ms for vowel/a/and 348 ms for vowel/o/. Vowels were presented binaurally through headphones at a comfortable listening level of ∼65dB as measured with an analog loudness meter. The visual stimulus was a white, lower case letter “a”, presented in size 40 “Arial” font in the center of a computer screen with a black background. The stimuli were presented using Presentation 14.4 (Neurobehavioral Systems, Inc., Albany, CA).

Experimental Design

In one auditory and two audiovisual oddball paradigms, we presented the vowel/a/as the standard (83%) and the vowel/o/as the deviant (17%) stimulus. In the auditory condition participants listened to the vowels while watching silent movies. In the two audiovisual conditions, the presentation of the standard or deviant vowel was paired with a white letter “a”, presented for 500 ms at the center of a black computer screen. Between the successive letter presentations a white fixation cross appeared in the same location. Both auditory and audiovisual trials had a trial length of 1700 ms. The two audiovisual conditions differed in the stimulus onset asynchrony (SOA) between letter and vowel presentation. In the first audiovisual condition (AV0), the letter and vowel appeared simultaneously, and in the second audiovisual condition (AV200), the letter appeared 200 ms before the vowel. During the audiovisual conditions, participants performed a simple visual target detection task to ensure fixation on the screen. The visual target consisted of a color picture of a wrapped present, which was presented instead of the letter in 3,5% of the trials (10 trials per block) and required a button press. This paradigm was the same as the one used in Froyen et al., [16], [21], [26] except for the following two changes. First, the trial length was increased from 1250 ms to 1700 ms because the experiment was part of a larger study involving coherence analysis at lower frequencies. Second, in order to obtain sufficient deviant trials within 76 minutes of measurement time, the number of deviants was increased from 10% to 17%, a percentage that was previously shown to lead to reliable mismatch responses [29], [66]. Participants performed three blocks of each condition, each block consisting of 288 trials. Per condition a total of 150 deviant and 714 standard trials were presented. Standards and deviants were presented in pseudo-random order: two successive deviants were required to be separated by at least three standards, and deviants could not be one of the first two trials of a block. The blocks of the same condition were presented in consecutive fashion, the order of the conditions was pseudorandomized and counterbalanced, with one of the cross-modal conditions always being presented first.

EEG data recording and analysis

EEG data were recorded at a 1024 Hz sampling rate and a DC-104 Hz recording bandwidth using the Biosemi Active Two system (Biosemi, Amsterdam, Netherlands). EEG was measured from 64 active-channels, placed according to the 10–20 international system (Electro-cap International Inc., with CMS and DRL electrodes at the approximate locations of PO1 and PO2), and the CMS was used as a recording reference. Eye-movements were measured using 4 additional Flat-Type Active electrodes, 2 were placed below and above left eye and 2 at the outer canthi of each eye. In addition 2 electrodes were placed on the right and left mastoids and used for offline re-referencing. The offset range of the electrodes was kept between −20 mV and 20 mV.

For the analysis we discarded the first 2 trials of each block, as well as the first epoch after each deviant and, in the audiovisual conditions, the first 2 epochs after catch trials to account for possible movement artifacts due to button presses [67]. Per participant, we included data of the 150 deviant trials and of 150 standard trials chosen randomly from the pool of 480 standard trials. Data were analyzed using the v11.0.0b EEGLAB toolbox [68](http://www.sccn.ucsd.edu/eeglab) and custom Matlab scripts (MATLAB 2011b, The MathWorks, Natick, MA). EEG data pre-processing included offline rereferencing to the average of the left and right mastoids, applying a bandpass filter of 1–70 Hz and downsampling to 256 Hz. The EEG data was epoched from −500 ms to 1200 ms with respect to the auditory (Auditory and AV0 condition) or visual (AV200) stimulus and baseline corrected with respect to the mean signal in the 500 ms baseline period. Removal of artifacts was performed in two steps. First, the data were visually inspected and epochs containing non-stereotypical artifacts including high amplitude, high-frequency muscle noise, swallowing and electrode cable movements, were rejected. Secondly, stereotypical artifacts, including eye movements, eye blinks and heart beat artifacts were corrected with extended INFOMAX ICA [69] as implemented in EEGLAB. ICA was performed on 64 scalp channels resulting in 64 components per condition per participant which were classified as EEG or artificial activity based on scalp topography, time course and power-frequency spectra [70]. Single electrodes containing high amplitude noise were interpolated using spherical interpolation after the data was reconstructed. Component activations representing non-brain artifacts were removed, and EEG data were reconstructed from the remaining component activations representing brain activity. In typical readers, the reconstructed data was based on a mean (SD) number of 41 (4) components per subject in the auditory condition, 47 (4) in the Av0 condition and 44 (5) in the Av200 condition. The number of EEG components in the two subgroups of dyslexic children corresponded to - dysfluent dyslexics: 43 (4) components per subject in the auditory condition, 48 (4) in the Av0 condition and 46 (4) in the Av200 condition; severely dysfluent dyslexics: 42 (8) components per subject in the auditory condition, 45 (7) in the Av0 condition and 44 (7) in the Av200 condition. The reconstructed data was baseline corrected and low pass filtered at 30 Hz.

Event related potentials (ERPs) were obtained by averaging the epochs per stimulus per condition for each participant separately. In typical readers, the mean (SD) number of epochs included in the averages for the standards and deviants corresponded to 149 (1) and 149 (1) in the auditory condition, 148 (2) and 149 (2) in the AV0 condition and 148 (2) and 149 (1) in the AV200 condition. In dysfluent dyslexics these values were 149 (1) and 148 (2) in the auditory condition, 148 (2) and 148 (2) in the AV0 condition and 148 (2) and 148 (2) in the AV200 condition. Lastly, in severely dysfluent these values were 148 (5) and 148 (5) in the auditory condition, 148 (2) and 148 (2) in the AV0 condition and 148 (2) and 147 (3) in the AV200 condition.

We evaluated the presence of the mismatch negativity (MMN) and late negativity (LN) by analyzing ERPs evoked by standard and deviant stimuli in the auditory and both audiovisual conditions, in each of our subject groups. As maximal cross-modal effects were expected across fronto-central electrodes [16], [21], analysis was performed on four fronto-central sites (Fz, Cz, FC3, FC4). The timing of the MMN and LN responses was measured by determining individual peak latencies within their respective time windows: (1) 100 to 250 ms for the MMN and (2) 600 ms–750 ms for the LN [16], [21]. The amplitude of the MMN and LN responses was measured by determining the mean amplitude in the 50 ms around the individual peak latencies. We first applied mixed model ANOVAs with stimulus (2 levels: standard and deviant), condition (3 levels: Au, Av0, Av200) and electrode (4 levels) as within subjects factors and group (3 levels: typical, dysfluent dyslexic and severely dysfluent dyslexic readers) as between subjects factor. Because MMN and LN responses were detected in all conditions and all subject groups, we subsequently performed our analyses of crossmodal letter effects on the difference waves obtained by subtracting the ERPs evoked by standards from those of deviants. MMN/LN amplitude and latency measures were analyzed using mixed model ANOVAs on pairs of conditions (Au-Av0; Au-Av200) and electrode (4 levels: Fz, Cz, FC3, FC4) as within subject factors. We report Greenhouse-Geisser corrected p-values. In a further analysis, we separately analyzed the N1 and P2 responses to standard and deviant stimuli to investigate the relation between the crossmodal MMN effects and these event-related responses. N1 and P2 amplitudes were statistically assessed using repeated measures ANOVAs with Stimulus (2 levels: standard, deviant) and Electrode (4 levels: Fz, Cz, FC3, FC4) as within subject factors.

We used linear regression to investigate the relation between individual differences in ERP correlates of letter-sound associations and behavioral measures of reading fluency and related skills. ERP measures consisted of (1) MMN latency in the two audiovisual (AV0 and AV200) conditions, and (2) the MMN and LN letter-effect in the two audiovisual conditions (i.e. the MMN/LN amplitude enhancement in the audiovisual as compared to auditory conditions, [16], [21], [26]). For linear regression, ERP measures were quantified using composite scores across 4 fronto-central electrodes (Fz, Cz, FC3 and FC4). Our selection of behavioral measures of interest was based on (1) our research aim, i.e. to test the relation between reading fluency and cross-modal ERP measures, and (2) results of a large-scale study in Dutch, Hungarian and Portuguese school children in which factor analysis of test scores on 3DM subtests identified verbal working memory, phonemic awareness, RAN, letter-speech sound accuracy and letter-speech sound speed as separate factors [71]. Thus, reading fluency, was quantified using a composite score of the 3DM word reading tests, the EMT, and “De Kat”. Furthermore, the other behavioral measures included reading accuracy (3DM), letter-speech sound coupling accuracy (spelling, letter-speech sound identification and discrimination, r >.406) and speed (spelling, letter-speech sound identification and discrimination, r >.391), phonological awareness (3DM phoneme deletion accuracy) and rapid auditory naming (3DM RAN total score). We employed simple linear regressions with one of the ERP measures as a predictor and one of the behavioral composite scores as a dependent variable to probe relations of each ERP measure specifically to the behavior. As we conducted several linear regressions, i.e. 12 for latencies and 24 for amplitudes, once for all the children and once only for dyslexics, we accounted for multiple comparisons by using FDR corrections [72] and report p values relative to obtained q(FDR) thresholds.

Results

Behavioral tests: Typical readers, Dysfluent and Severely Dysfluent Dyslexic readers

The dyslexic readers were divided into a dysfluent and a severely dysfluent group, using a median split based on their reading fluency scores (see Methods). A comparison of behavioral test scores showed significantly lower scores for severely dysfluent as compared to dysfluent dyslexics on all word reading tests (fluency as well as accuracy), as well as on rapid automatized naming (Table 1). The dyslexic subgroups did not significantly differ on any of the other behavioral tests. Both dyslexic groups scored significantly lower than typical readers on all language tests with the exception of the accuracy of letter-speech sound discrimination and identification (comparable in dysfluent and typical readers) and the speed of letter-speech sound discrimination (comparable in all three groups). Non-verbal IQ and speed of processing (reaction times) were comparable across the three groups.

Auditory MMN and LN effects

Because these measures served as a baseline for assessing cross-modal enhancement effects, we first analyzed the MMN and LN responses evoked in the auditory condition. In each of the three subject groups, the deviant elicited a comparable auditory MMN, with the expected fronto-central topographical distribution (Figures 1 and 2). Comparable MMN responses lead to a main effect of Stimulus (F(1, 53) = 105.26, p = .000, η2 = .665), without significant group differences. We also did not observe significant group differences in the latency of the auditory MMN.

thumbnail
Figure 1. Grand average event-related potentials (ERP) and difference waves.

Grand average ERPs averaged over 4 frontocentral electrodes for standard (dotted line), deviant (dashed line) and their difference (solid line) in auditory (Au) and two audiovisual conditions (Av0 and Av200) with time intervals of interest shaded in light grey. N1 and P2 peaks of standard and deviant ERPs are also marked in synchronous audiovisual condition (Av0).

https://doi.org/10.1371/journal.pone.0110337.g001

thumbnail
Figure 2. Difference waves and topographical maps.

I. Difference waves (averaged over 4 frontocentral electrodes) with time intervals of interest shaded in gray. Significant differences between conditions are marked with asterisks (*p< = .05;**p< = .01): green asterisk – Av0 vs. Au difference, red asterisk – Av200 vs. Au difference. II. Topographical distribution of average amplitudes in difference waves over 64 scalp electrodes.

https://doi.org/10.1371/journal.pone.0110337.g002

In each group, the vowel deviant additionally elicited a weaker and more sustained auditory LN response around 670–690 ms after stimulus onset (Figures 1 and 2). Statistical analyses indicated a comparable auditory LN response across groups, with a significant main effect of Stimulus (F(1, 53) = 64.44, p = .000, η2 = .549) and no significant group differences for amplitudes or latencies. In sum, in each of the three groups the deviant vowel sound/o/elicited comparable MMN and LN responses with respect to the standard vowel sound/a/.

Letter-effects: MMN and LN amplitude enhancements

In both cross-modal conditions (Av0 and Av200) the deviant vowel sound/o/elicited MMN and LN responses with respect to the standard vowel sound/a/(Fig. 1 and Fig. 2). In these blocks, the letter “a” was presented together with the standard and deviant vowels, leading to a double violation: the deviant speech sound/o/violated the expectation built by both the standard speech sound/a/and the letter ‘a’. To assess the neural integration of letters and speech sounds we tested whether this double violation significantly enhanced the cross-modal as compared to the auditory MMN and LN responses [16], [21], [26]. To this end, we first performed a mixed model ANOVA with stimulus (standard, deviant), condition (auditory, AV0, AV200) and electrodes (4 fronto-central electrodes) as within-subject factors and subject group as between-subject factor. In the MMN window, this overall analysis indicated stronger responses to deviants as compared to standards (Stimulus F(1,53) = 294.61, p = .000, η2 = .848), and stronger cross-modal as compared to auditory (mismatch) responses (Condition F(2,99) = 50.61, p = .000, η2 = .488; Stimulus*Condition: F(2,105) = 12.29, p = .000, η2 = .188), as well as a tendency for groups to differ in their response amplitudes across conditions (Condition*Group (F(4,99) = 2.18, p = .081, η2 = .076). Pairwise group comparisons showed that this tendency was driven by significant differences between severely dysfluent dyslexics and typical readers (Condition*Group: F(2,69) = 3.87, p = .027, η2 = .097), while there were no differences in other cases.

Also in the later LN window, the overall mixed-model ANOVA indicated stronger responses to deviants as compared to standards (Stimulus F(1,53) = 267.28, p = .000, η2 = .835) and stronger cross-modal as compared to auditory (mismatch) responses (Condition F(1,73) = 9.47, p = .001, η2 = .152; Stimulus*Condition: F(2,99) = 8.78, p = .000, η2 = .142) together with significant group differences in response amplitudes across conditions (Condition*Group: F(3,73) = 3.87, p = .015, η2 = .127). Pairwise group comparisons showed that the significant group difference in response amplitudes across conditions was driven by significant differences between severely dysfluent dyslexics and typical readers (Condition*Group: F(1,51) = 4.49, p = .027, η2 = .111) as well as between dysfluent dyslexics and typical readers (Condition*Group: F(1,49) = 6.80, p = .006, η2 = .159), but not by a difference between the two dyslexic groups (F(1,46) = 0.48, p = .549, η2 = .014).

We further analyzed these condition and group effects by performing mixed model ANOVAs on the MMN difference waves with Condition (AV0 versus Auditory; AV200 versus Auditory) and electrodes (4 fronto-central electrodes) as within subjects factors. In the simultaneous cross-modal condition (AV0), significant MMN amplitude enhancements were observed in typical readers (F(1,19) = 10.92, p = .004, η2 = .365) as well as in dysfluent dyslexics (F(1,17) = 7.63, p = .013, η2 = .310), whereas in severely dysfluent dyslexics this enhancement was not significant (F(1,17) = 3.43, p = .082, η2 = .168). When letter presentation preceded speech sound onset (AV200) significant cross-modal MMN enhancements were found in all three subject groups (typical readers: F(1,19) = 7.21, p = .015, η2 = .275; dysfluent dyslexics: F(1,17) = 5.93, p = .026, η2 = .259; severely dysfluent dyslexics: F(1,17) = 6.09, p = .024, η2 = .264).

As for the LN window, in typical readers the presence of a letter significantly enhanced the LN amplitude in both cross-modal conditions (AV0 vs. Auditory: F(1,19) = 10.90, p = .004, η2 = .365; AV200 vs. Auditory: F(1,19) = 7.85, p = .011, η2 = .292). In contrast, in the dyslexic groups the LN letter effects only resulted in non-significant trends (AV0 vs. Auditory, dysfluent: F(1,17) = 3.93, p = .064, η2 = .188; severely dysfluent: F(1,17) = 3.98, p = .062, η2 = .190; Av200 vs. Auditory, dysfluent: (F(1,17) = 3.44, p = .081, η2 = .168; severely dysfluent: F(1,17) = 2.41, p = .139, η2 = .124).

In summary, in both cross-modal conditions, our results showed early (MMN) and late (LN) letter effects in typical readers. Although the group differences in MMN and LN letter effects were subtle, the dysfluent group resembled typical readers in the presence of an MMN letter effect, whereas they resembled the severely dysfluent group in the lack of a significant LN letter effect. The severely dysfluent group differed most from typical readers in both the lack of a significant LN letter effect and in only showing a significant MMN letter effect in the AV200 condition.

Letter effects: MMN and LN latency

We investigated the timing of the letter effects, by analyzing MMN and LN latencies in both crossmodal conditions. As already indicated by our regression analysis, groups tended to differ in MMN response latency in the simultaneous crossmodal (AV0) condition (Group (F(2,53) = 5.85, p = .062, η2 = .181). In particular, severely dysfluent dyslexics showed a significantly shorter MMN latency than both the typical readers (F(1,36) = 8.59, p = .006, η2 = .193), and the dysfluent dyslexics (F(1,34) = 7.65, p = .009, η2 = .184), whereas the latter groups showed similar latencies (F(1,36) = 0.07, p = .787, η2 = .002). No significant group differences were found for MMN latency in the asynchronous cross-modal condition (AV200), or for LN latency in either of the cross-modal conditions.

ERP - behavior relations

We used linear regression to investigate the relation between individual differences in MMN/LN responses in the crossmodal conditions and behavioral measures of reading fluency and related skills. Results showed that such a relation was present for the latency of the MMN response elicited in the simultaneous crossmodal condition (AV0). First, across both typical and dyslexic readers, MMN latency in the AV0 condition significantly correlated (significance threshold q(FDR) = .010) with individual differences in word reading fluency (r = .368; p = .005; Table 2, Figure 3), word reading accuracy (r = .398; p = .002) and phoneme deletion accuracy (r = .343; p = .001). When restricting the analysis to the dyslexic readers, a significant relation was only shown for reading fluency (r = .554; p = .000). Remarkably, in both cases, longer MMN peak latencies were found in the more fluent readers, a result that is further investigated in our analysis of the N1-P2 responses underlying the MMN effect. None of the other ERP-behavior regressions reached statistical significance.

thumbnail
Figure 3. Correlations of word reading fluency and accuracy with the MMN latency in Av0 condition.

Relation of composite fluency (3DM word reading fluency, EMT, ‘De Kat’) and 3DM Accuracy (HFW, LFW, PW) raw scores with MMN latency in Av0 condition with the strength of correlation represented by r.

https://doi.org/10.1371/journal.pone.0110337.g003

thumbnail
Table 2. Regression of word reading fluency and accuracy with the MMN latency in Av0 condition: Regression coefficients for behavioral measures and latency of MMN in Av0 condition averaged over 4 electrodes (significant p values for q(FDR) = .010 are denoted in bold).

https://doi.org/10.1371/journal.pone.0110337.t002

Relation between MMN and N1-P2 responses

In a final analysis we aimed to further investigate the origin of the shorter MMN latency in severely dysfluent dyslexics as compared to both other groups, as well as the corresponding regression results in the simultaneous cross-modal (AV0) condition (Figure 3) by analyzing the timing of the MMN with respect to the evoked N1 and P2 responses (Figure 1). Results showed that in severely dysfluent dyslexics, letter-speech sound pairs only elicited a deviancy effect in the N1 window (main effect of Stimulus F(1,17) = 28.32, p = .000, η2 = .625), but not in the P2 window (F(1,17) = .03, p = .860, η2 = .002), as can also be seen in Figure 1. In contrast, typical readers as well as dysfluent dyslexics showed a deviancy effect in both the N1 and the P2 windows (main effects of Stimulus typical readers N1: F(1,19) = 32.27, p = .000, η2 = .629; P2: F(1,19) = 6.49, p = .020, η2 = .255); dysfluent dyslexics N1: F(1,17) = 14.24, p = .002, η2 = .456 and P2: F(1,17) = 10.15, p = .005, η2 = .374). The absence of a P2 deviancy effect in the severely dysfluent dyslexics thus most likely explains their shorter MMN latency.

Discussion

The present study investigated ERP measures of letter-speech sound integration in typically reading and dyslexic children after 2,5 years of reading instruction and specifically investigated how these ERP measures relate to individual differences in reading fluency and related skills. For this purpose we employed a passive oddball paradigm and tested cross-modal enhancement of the MMN and LN due to an audiovisual violation (vowel sound/o/vs. vowel sound/a/and letter ‘a’) as compared to an auditory only violation (vowel sound/o/vs. vowel sound/a/; [21]). Our results show robust neural integration of letters and speech sounds in the typical readers, albeit at a different temporal integration window compared to adults [26]. Furthermore, they confirm and extend previous findings in dyslexic children by demonstrating reduced letter-speech sound integration, with different patterns of integration deficiency depending on the level of reading dysfluency (Table 3).

Typical readers: enhanced (and non-selective) neural sensitivity?

Typical readers after 2,5 years of reading instruction showed cross-modal enhancements of the MMN and LN responses independent of the precise timing of letter and speech sound presentation (AV0 and Av200), indicating a broad temporal window of integration. This contrasts with previous results of adults, who only showed a cross-modal enhancement in the MMN window, and only when letters and speech sounds were presented simultaneously [26]. Furthermore, the present results showed a different pattern of cross-modal effects as compared to previous findings in both typical beginning readers after 1-year of reading instruction and advanced typical readers after 4-years of reading instruction [21]. That is, in the younger readers cross-modal enhancement only reached significance in the LN window and only when the letter preceded the speech sound (AV200). In the advanced readers this pattern shifted with significant cross-modal enhancement in the earlier MMN window when the letter preceded the speech sound (AV200) and additionally in the LN window when letters and speech sounds were presented simultaneously (AV0). In comparison to the younger readers, the observed emergence of a cross-modal MMN enhancement may reflect the initial development of an early detection and integration of letter-sound pairs in 9-year-old children [27], [28]. In comparison to the more advanced readers, who only showed a cross-modal LN enhancement in the simultaneous AV0 condition, the unspecific cross-modal LN enhancement independent of the timing of letter presentation, may suggest immature and non-selective further processing of letter-sound congruency in 9-year-olds. On a more general level, the pattern of stronger and/or less selective cross-modal effects in our intermediate group of typical readers is consistent with a developing neural system for audiovisual speech and letter-sound processing that undergoes dynamic changes throughout primary school years [6], [73], [74]. In particular, they may reflect enhanced neural sensitivity for letter-sound congruency after 2.5 years of reading instruction, before developing into a more selectively tuned system with advanced reading practice. Similar nonlinear patterns of an initial sensitivity increase during early school years, followed by reduced and more selective sensitivity in adults, have been observed for print selective visual N1 responses [53] and ERP phonological priming effects during an auditory lexical decision task [52]. On the other hand, also methodological differences between the present study and the previous ERP studies by Froyen and colleagues [16], [21], [26] may have (at least partly) contributed to the relatively strong cross-modal ERP effects in our typical readers. For example, unlike these previous studies, we used active electrodes for EEG data acquisition and ICA-based EEG preprocessing, which both lead to improvement of ERP signal-to-noise ratio. Another factor that may have increased signal-to-noise ratio is an increase in trial length from 1250 ms to 1700 ms. In addition, we employed a full within-subjects design as compared to a partial within-subjects design [21]. To further understand the development of brain mechanisms for letter-speech sound integration and their relation to reading acquisition, it is important to apply similar cross-modal neuroimaging paradigms in a longitudinal set-up and follow the same children while they are learning to read.

Dyslexic children: crossmodal MMN effect scales with reading dysfluency

ERP data of dyslexic children showed normal auditory MMN and LN responses together with reduced effects of letter-speech sound congruency. Most interestingly results indicated different patterns of integration deficiency depending on the level of reading fluency. In particular, in our regression analysis behavioral measures of reading fluency significantly predicted the latency of the MMN response in the simultaneous cross-modal condition (AV0), not only across typical and dyslexic readers, but also within the group of dyslexic readers. Further analyses indicated that this relation could be explained by a short-lasting, reduced MMN response, encompassing only the N1 window in severely dysfluent dyslexics as compared to a longer lasting MMN response, encompassing both the N1 and P2 windows in dysfluent dyslexic and typical readers. In addition, in the AV0 condition, MMN latency correlated significantly with reading accuracy and phonological awareness (phoneme deletion) across typical readers and dyslexics, but not within the dyslexic group alone.

Although relatively little is known about the functional relevance of the P2, recent evidence suggests that this response is specifically sensitive to the audiovisual integration of orthographic and phonological units [47] as well as of visual articulatory gestures and speech [74][77]. For example, in an elegant study by Baart et al. [75], audiovisual integration of articulatory gestures (lipread speech) and sine-wave speech was compared between participants trained to perceive the sine-wave speech in either a ‘speech’ or ‘nonspeech’ mode. Whereas lipread speech modulated the N1 response in both processing modes, the P2 was only modulated in listeners who recognized the sine-wave speech as speech, thus suggesting more general audiovisual convergence in the N1 window followed by speech specific audiovisual integration in the P2 window. This speech specificity of the P2 response would also concur with its putative neural source in the posterior superior temporal sulcus [78], a region involved in the integration of audiovisual speech [79] and of letters and speech sounds [23], [80]. Moreover activity in this region has been found to scale with interindividual differences in audiovisual speech perception (McGurk effect) in typically reading children and adults [73], [81] and to show reduced letter-speech sound integration in dyslexic children [15]. By analogy, we would speculate that the N1 deviancy effect shown in the simultaneous audiovisual (AV0) condition indicates audiovisual convergence common across the three children groups, followed by a speech specific integration of letters and speech sounds in the P2 window in typical readers and dysfluent dyslexics, which is reduced or absent in severely dysfluent dyslexics. Recently, a diminished speech-specificity of audiovisual integration was also found for spoken-written syllable pairs in Finnish dyslexic adults [41].

A similar pattern of results or relation with behavior (reading fluency) was not observed in the asynchronous crossmodal (AV200) blocks, possibly indicating a diminished recruitment of speech specific integration processes in this condition. This may also explain why typical adult readers, with selectively tuned neural systems for letter-speech sound integration, only show congruency effects when letters and speech sounds are presented simultaneously (AV0 blocks, [23], [26]). These findings deserve further study however, especially in relation to the nature of possible audiovisual integration deficits in dyslexics, because in a previous study with 11 year old dyslexic readers, behavioral measures of word reading fluency (MMN), non-word reading fluency and letter-phoneme matching (LN) were found to correlate with MMN/LN letter effects in the AV200 condition [16].

Further evidence for brain-behavior relations with respect to letter-speech sound integration comes mainly from fMRI studies. [15], for example found a positive correlation between letter-speech sound congruency effects measured in the planum temporale/Heschl’s sulcus and both reading accuracy and the speed of letter-speech sound matching across 9-year old typical and dyslexic readers. Cross-modal fMRI congruency effects measured in the planum temporale during a rhyme judgment task were also found to positively correlate with literacy skills [82] as well as with phonological awareness [17] in 8–11 year old typical readers. However, in the latter study no correlation was found between cross-modal integration and phonological awareness in dyslexic children, which was suggested to indicate a decoupling of these processes in dyslexia [17]. Although most studies thus find a relation between reading and/or phonological skills and ERP/fMRI indices of letter-speech sound integration, the type of relation shows some variability as well as its presence or absence in dyslexic readers. These differences may relate to the dependency of neural effects of letter-speech sound congruency on task demands [25], [46] the type of speech/letter units, e.g. consonants, vowels, syllables [38], [83] and/or the depth of the orthography [84]. Importantly, the present results show that the severity of reading (dys)fluency may also lead to differences in the observed brain-behavior correlations across studies. It would thus be interesting to further investigate these brain-behavior relationships in a larger-scale cross-national study including children of varying reading (dys)fluency, and deep and shallow orthographies, using the same paradigm with stimuli of grain sizes with different relevance in the different orthographies. Furthermore, the observed relation between reading fluency/accuracy and crossmodal ERP modulations across typical as well as dyslexic readers is compatible with a continuity model of reading dysfunction [80], although our results do not exclude the possibility of dyslexia subtypes.

Given the present focus on letter-speech sound integration, it may seem surprising that behavioral measures of letter-speech sound coupling showed the weakest differences across reading groups and did not correlate with ERP measures of letter-speech sound coupling. However, these findings are in agreement with previous ERP and fMRI evidence revealing a dissociation between children’s knowledge of which letters belong to which speech sounds and the automatic neural integration of these associations [19][21]. The present results thus provide additional support for the notion that behavioral measures alone may not to be a sensitive indicator for letter–speech sound integration [6], [21].

Normal auditory responses and subtle reduction of LN letter effect in dyslexics

Results showed comparable auditory MMN and LN responses to vowels in all three groups. This absence of group differences was expected because vowels typically elicit normal MMN and/or LN responses in dyslexic children [16], [85] whereas deficient change detection responses are typically reported for more subtle speech changes involving e.g. stop consonants (e.g. [85], [86]). Correspondingly, we did not find any correlations between auditory MMN and LN responses and behavioral measures of reading or reading-related skills.

Our results also confirm weak neural integration of letters and speech sounds in 9 year old dyslexic children [15], with reduced or absent audiovisual LN effects in both groups and a short-lived MMN effect in the AV0 condition in severely dysfluent dyslexics. The timing of these effects was different than those previously reported in 11 year old dyslexics, for whom MMN letter effects did not, but the LN letter effect in the AV200 condition did, reach significance [16]. These differences could relate to developmental or strategy changes in dyslexic readers after 2.5 versus 4 years of reading instruction, but could also relate to methodological differences between the two studies and/or differences in the severity of reading dysfluency (see above).

Interestingly, an attenuated late negativity to spoken/ba/deviants as compared to/da/standards has been proposed as a potential endophenotype for dyslexia [36]. That is, this negativity in a broader time window (300–700 ms) overlapping with our crossmodal LN (600–750 ms) was found to be attenuated in both dyslexic children and their siblings without dyslexia [36], and to be associated with rare variants in a candidate gene region for dyslexia [43]. As compared to perceptual aspects of letter-speech sound congruency in the MMN window, the observed LN letter effect may reflect cognitive, explicit associative and/or attentional processes [16], [21], [36], [87], present depending on familiarity and complexity of the stimuli. Whereas in adults these type of late orthographic-phonological interactions may only occur during complex metaphonological tasks [46], [47] and pseudoword-word priming tasks [51], in typically reading children they seem to be recruited during the integration of simple letter-vowel pairs (present findings and [21]), letter strings [88], integration of audiovisual words [89] and a visual lexical decision task with phonological distractors [49], with disrupted recruitment in dyslexic children (present findings and [16], [49]). If this process is disrupted, as was the case in both dyslexic groups then the automaticity in adulthood may not be reached [18], as it may be prevented by an incapability to access and/or manipulate the representations [15], [51], [90], and/or reduced attentionally-mediated integration [36], [43], [50], [51]. The additional attenuation of the early MMN in severely dyslexic children may signal a more basic failure in forming a clear letter-speech sound representation [6], and concurs with the suggested timing of sublexical orthographic-phonological integration during online visual word recognition [91]. Similarly, Roeske and colleagues [92] observed an association between specific genetic markers and a speech evoked late negativity across dyslexic children independently of the severity of their reading problems, while a significant association with the earlier speech evoked MMN was present only in the most severe dyslexics. Thus, the crossmodal MMN enhancement would represent a successful representation of an audiovisual stimulus, and the crossmodal LN enhancement would characterize availability of the represented stimulus for further manipulation.

Finally, Blau and colleagues [15] showed that Dutch dyslexic children of the same age as our participants, exhibit reduced neural integration of letters and speech sounds in the planum temporale/Heschl sulcus and the superior temporal sulcus, thus showing both reduced integration in the STS and feedback to the auditory cortex [23]. As the severely dysfluent dyslexics in our study showed both early and late deficiencies, while dysfluent dyslexics showed only a late deficiency, it would be interesting to investigate whether the early MMN deficiency would stem from the reduced integration in the STS while the LN deficiency would be a product of a reduced feedback.

Conclusion

The present ERP study investigated letter-speech sound integration in 9-year-old typically reading and dyslexic children by using a well-studied oddball paradigm with simple vowel sounds and letters and demonstrated that early (speech specific) audiovisual integration processes scale with individual differences in reading (dys)fluency. Results further indicated enhanced neural sensitivity to letter-speech sound associations in 9-year-old typical readers, together with disrupted sensitivity in dyslexic readers. In future studies, this ERP paradigm could be used to investigate whether and how systematic training of phonological skills and/or letter speech-sound coupling changes this disrupted neural integration of letters and speech sounds in dyslexic children. Furthermore, an extension to longitudinal designs using units of different complexity and grain sizes (e.g. vowels, consonants, syllables, words), possibly in the context of more naturalistic on-line reading tasks, could lead to a more complete understanding of the dynamic neural changes contributing to successful and hampered reading acquisition.

Acknowledgments

We dedicate this paper to our co-author, professor Leo Blomert, who passed away on November 25, 2012. His contributions to the project prior to his passing were significant.

This project is part of the research program “Fluent reading acquisition neurocognitively decomposed: the case of dyslexia (HCMI 10–59)” funded by the Netherlands Initiative Brain and Cognition (NIHC), a part of the Netherlands Organization for Scientific Research (NWO) under grant number 056-14-015.

We thank Suzanne van Grieken, Marlena van Langevelde, Gert-Jan Munneke, Jitka Annen, Mandy Meijer, Helene Vos and Jorinde Wesseling for their help during data collection and Nienke van Atteveldt for useful comments on the manuscript. Last but not least, a special thanks goes to the children participating in the research and their parents.

Author Contributions

Conceived and designed the experiments: GŽ LB MB. Performed the experiments: GŽ GFG JT. Analyzed the data: GŽ MB. Contributed to the writing of the manuscript: GŽ MB MvdM.

References

  1. 1. Blomert L (2005) Dyslexie in Nederland. Amsterdam: Uitgeverij Nieuwezijds. Available: http://www.boomtestuitgevers.nl/upload/Dyslexie_in_Nederland_Leo_Blomert.pdf. Accessed 2014 April 16.
  2. 2. Lyon GR, Shaywitz SE, Shaywitz BA (2003) A definition of dyslexia. Ann Dyslexia 53: 1–14 Available: http://link.springer.com/10.1007/s11881-003-0001-9 Accessed 2014 February 14.
  3. 3. Shaywitz SE, Morris R, Shaywitz BA (2008) The education of dyslexic children from childhood to young adulthood. Annu Rev Psychol 59: 451–475 Available: http://www.ncbi.nlm.nih.gov/pubmed/18154503 Accessed 2013 February 10.
  4. 4. Katzir T, Kim , Wolf M, Morris R, Lovett MW (2008) The Varieties of Pathways to Dysfluent Reading: Comparing Subtypes of Children With Dyslexia at Letter, Word, and Connected Text Levels of Reading. J Learn Disabil 41: 47–66 Available: http://www.ncbi.nlm.nih.gov/pubmed/18274503 Accessed 2013 September 10.
  5. 5. Leinonen S, Müller K, Leppänen PHT, Aro M, Ahonen T, et al. (2001) Heterogeneity in adult dyslexic readers: Relating processing skills to the speed and accuracy of oral text reading *. Read Writ 14: 265–296
  6. 6. Blomert L (2011) The neural signature of orthographic-phonological binding in successful and failing reading development. Neuroimage 57: 695–703 Available: http://www.ncbi.nlm.nih.gov/pubmed/21056673 Accessed 2013 January 29.
  7. 7. Ehri LC (2005) Learning to Read Words: Theory, Findings, and Issues. Sci Stud Read 9: 167–188 Available: http://www.tandfonline.com/doi/abs/10.1207/s1532799xssr0902_4.
  8. 8. Share DL (1995) Phonological recoding and self-teaching: sine qua non of reading acquisition. Cognition 55: 151–218 discussion 219-26 Available: http://www.ncbi.nlm.nih.gov/pubmed/7789090.
  9. 9. Wimmer H, Schurz M (2010) Dyslexia in regular orthographies: manifestation and causation. Dyslexia 16: 283–299 Available: http://www.ncbi.nlm.nih.gov/pubmed/20957684 Accessed 2013 December 9.
  10. 10. Snowling M (1998) Dyslexia as a Phonological Deficit: Evidence and Implications. Child Psychol Psychiatry Rev 3: 4–11 Available: http://www.journals.cambridge.org/abstract_S1360641797001366.
  11. 11. Ramus F, Szenkovits G (2008) What phonological deficit? Q J Exp Psychol (Hove) 61: 129–141 Available: http://www.ncbi.nlm.nih.gov/pubmed/18038344 Accessed 2013 March 3.
  12. 12. Aravena S, Snellings P, Tijms J, van der Molen MW (2013) A lab-controlled simulation of a letter–speech sound binding deficit in dyslexia. J Exp Child Psychol 115: 691–707 Available: http://linkinghub.elsevier.com/retrieve/pii/S0022096513000659 Accessed 2013 May 27.
  13. 13. Bus AG, Ijzendoorn MH Van (1999) Phonological Awareness and Early Reading: A Meta-Analysis of Experimental Training Studies. J Educ Psychol 91: 403–414
  14. 14. Tijms J, Hoeks J (2005) A computerized treatment of dyslexia: benefits from treating lexico-phonological processing problems. Dyslexia 11: 22–40 Available: http://www.ncbi.nlm.nih.gov/pubmed/15747805 Accessed 2013 March 4.
  15. 15. Blau V, Reithler J, van Atteveldt N, Seitz J, Gerretsen P, et al. (2010) Deviant processing of letters and speech sounds as proximate cause of reading failure: a functional magnetic resonance imaging study of dyslexic children. Brain 133: 868–879 Available: http://www.ncbi.nlm.nih.gov/pubmed/20061325 Accessed 2013 May 22.
  16. 16. Froyen D, Willems G, Blomert L (2011) Evidence for a specific cross-modal association deficit in dyslexia: an electrophysiological study of letter-speech sound processing. Dev Sci 14: 635–648 Available: http://www.ncbi.nlm.nih.gov/pubmed/21676085 Accessed 2013 February 15.
  17. 17. McNorgan C, Randazzo-Wagner M, Booth JR (2013) Cross-modal integration in the brain is related to phonological awareness only in typical readers, not in those with reading difficulty. Front Hum Neurosci 7: 388 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3719029&tool=pmcentrez&rendertype=abstract Accessed 2013 August 7.
  18. 18. Blau V, van Atteveldt N, Ekkebus M, Goebel R, Blomert L (2009) Reduced neural integration of letters and speech sounds links phonological and reading deficits in adult dyslexia. Curr Biol 19: 503–508 Available: http://www.ncbi.nlm.nih.gov/pubmed/19285401 Accessed 2013 August 15.
  19. 19. Blomert L, Vaessen AA (2009) 3DM Differential diagnostics for dyslexia: Cognitive analysis of reading and spelling. Amsterdam: Boom Test Publishers.
  20. 20. Booth JR, Burman DD, Van Santen FW, Harasaki Y, Gitelman DR, et al. (2001) The development of specialized brain systems in reading and oral-language. Child Neuropsychol 7: 119–141 Available: http://www.tandfonline.com/doi/abs/10.1076/chin.7.3.119.8740 Accessed 2013 September 9.
  21. 21. Froyen DJW, Bonte ML, van Atteveldt N, Blomert L (2009) The long road to automation: neurocognitive development of letter-speech sound processing. J Cogn Neurosci 21: 567–580 Available: http://www.ncbi.nlm.nih.gov/pubmed/18593266 Accessed 2014 April 8.
  22. 22. Raij T, Uutela K, Hari R (2000) Audiovisual integration of letters in the human brain. Neuron 28: 617–625 Available: http://www.ncbi.nlm.nih.gov/pubmed/11144369.
  23. 23. Van Atteveldt N, Formisano E, Goebel R, Blomert L (2004) Integration of letters and speech sounds in the human brain. Neuron 43: 271–282 Available: http://www.ncbi.nlm.nih.gov/pubmed/15260962.
  24. 24. Van Atteveldt NM, Formisano E, Blomert L, Goebel R (2007) The effect of temporal asynchrony on the multisensory integration of letters and speech sounds. Cereb Cortex 17: 962–974 Available: http://www.ncbi.nlm.nih.gov/pubmed/16751298 Accessed 2013 August 28.
  25. 25. Van Atteveldt NM, Formisano E, Goebel R, Blomert L (2007) Top-down task effects overrule automatic multisensory responses to letter-sound pairs in auditory association cortex. Neuroimage 36: 1345–1360 Available: http://www.ncbi.nlm.nih.gov/pubmed/17513133 Accessed 2013 August 26.
  26. 26. Froyen D, Van Atteveldt N, Bonte M, Blomert L (2008) Cross-modal enhancement of the MMN to speech-sounds indicates early and automatic integration of letters and speech-sounds. Neurosci Lett 430: 23–28 Available: http://www.ncbi.nlm.nih.gov/pubmed/18023979 Accessed 2013 October 21.
  27. 27. Näätänen R (2001) The perception of speech sounds by the human brain as reflected by the mismatch negativity (MMN) and its magnetic equivalent (MMNm). Psychophysiology 38: 1–21.
  28. 28. Näätänen R, Paavilainen P, Rinne T, Alho K (2007) The mismatch negativity (MMN) in basic research of central auditory processing: a review. Clin Neurophysiol 118: 2544–2590 Available: http://www.ncbi.nlm.nih.gov/pubmed/17931964 Accessed 2013 August 6.
  29. 29. Bonte ML, Poelmans H, Blomert L (2007) Deviant neurophysiological responses to phonological regularities in speech in dyslexic children. Neuropsychologia 45: 1427–1437 Available: http://www.ncbi.nlm.nih.gov/pubmed/17187830 Accessed 2013 February 2.
  30. 30. Cheour-Luhtanen M, Alho K, Kujala T, Sainio K, Reinikainen K, et al. (1995) Mismatch negativity indicates vowel discrimination in newborns. Hear Res 82: 53–58 Available: http://www.sciencedirect.com/science/article/pii/037859559400164L Accessed 2013 March 18.
  31. 31. Cheour M, Shestakova A, Alku P, Ceponiene R, Näätänen R (2002) Mismatch negativity shows that 3–6-year-old children can learn to discriminate non-native speech sounds within two months. Neurosci Lett 325: 187–190.
  32. 32. Csépe V (1995) On the origin and development of the mismatch negativity. Ear Hear 16: 91–104 Available: http://scholar.google.com/scholar?hl=en&btnG=Search&q=intitle:On+the+origin+and+development+of+the+mismatch+negativity#0 Accessed 2013 December 3.
  33. 33. Eulitz C, Lahiri A (2004) Neurobiological Evidence for Abstract Phonological Representations in the Mental Lexicon during Speech Recognition. J Cogn Neurosci 16: 577–583.
  34. 34. Kasai K, Nakagome K, Iwanami A, Fukuda M, Itoh K, et al. (2002) No effect of gender on tonal and phonetic mismatch negativity in normal adults assessed by a high-resolution EEG recording. Brain Res Cogn Brain Res 13: 305–312 Available: http://www.ncbi.nlm.nih.gov/pubmed/11918996.
  35. 35. Lohvansuu K, Hämäläinen JA, Tanskanen A, Bartling J, Bruder J, et al. (2013) Separating mismatch negativity (MMN) response from auditory obligatory brain responses in school-aged children. Psychophysiology 50: 640–652 Available: http://www.ncbi.nlm.nih.gov/pubmed/23577828 Accessed 2014 January 23.
  36. 36. Neuhoff N, Bruder J, Bartling J, Warnke A, Remschmidt H, et al. (2012) Evidence for the late MMN as a neurophysiological endophenotype for dyslexia. PLoS One 7: e34909 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3351484&tool=pmcentrez&rendertype=abstract Accessed 2013 May 22.
  37. 37. Winkler I, Kujala T, Tiitinen H, Sivonen P, Alku P, et al. (1999) Brain responses reveal the learning of foreign language phonemes. Psychophysiology 36: 638–642 Available: http://www.ncbi.nlm.nih.gov/pubmed/10442032.
  38. 38. Andres AJD, Oram Cardy JE, Joanisse MF (2011) Congruency of auditory sounds and visual letters modulates mismatch negativity and P300 event-related potentials. Int J Psychophysiol 79: 137–146 Available: http://www.ncbi.nlm.nih.gov/pubmed/20888870 Accessed 2013 February 21.
  39. 39. Colin C, Radeau M, Soquet A, Demolin D, Colin F, et al. (2002) Mismatch negativity evoked by the McGurk-MacDonald effect: a phonetic representation within short-term memory. Clin Neurophysiol 113: 495–506 Available: http://www.ncbi.nlm.nih.gov/pubmed/11955994.
  40. 40. Mittag M, Takegata R, Kujala T (2011) The effects of visual material and temporal synchrony on the processing of letters and speech sounds. Exp Brain Res 211: 287–298 Available: http://www.ncbi.nlm.nih.gov/pubmed/21516331 Accessed 2013 October 16.
  41. 41. Mittag M, Alho K, Takegata R, Makkonen T, Kujala T (2013) Audiovisual attention boosts letter-speech sound integration. Psychophysiology 50: 1034–1044 Available: http://www.ncbi.nlm.nih.gov/pubmed/23848280 Accessed 2013 October 21.
  42. 42. Stekelenburg JJ, Vroomen J, de Gelder B (2004) Illusory sound shifts induced by the ventriloquist illusion evoke the mismatch negativity. Neurosci Lett 357: 163–166 Available: http://www.ncbi.nlm.nih.gov/pubmed/15003275 Accessed 2013 June 6.
  43. 43. Czamara D, Bruder J, Becker J, Bartling J, Hoffmann P, et al. (2011) Association of a rare variant with mismatch negativity in a region between KIAA0319 and DCDC2 in dyslexia. Behav Genet 41: 110–119 Available: http://www.ncbi.nlm.nih.gov/pubmed/21104116 Accessed 2013 August 20.
  44. 44. Hommet C, Vidal J, Roux S, Blanc R, Barthez MA, et al. (2009) Topography of syllable change-detection electrophysiological indices in children and adults with reading disabilities. Neuropsychologia 47: 761–770 Available: http://www.ncbi.nlm.nih.gov/pubmed/19126410 Accessed 2013 October 25.
  45. 45. Cheour M, Korpilahti P, Martynova O, Lang AH (2001) Mismatch negativity and late discriminative negativity in investigating speech perception and learning in children and infants. Audiol Neurootol 6: 2–11 Available: http://www.ncbi.nlm.nih.gov/pubmed/11173771.
  46. 46. Lafontaine H, Chetail F, Colin C, Kolinsky R, Pattamadilok C (2012) Role and activation time course of phonological and orthographic information during phoneme judgments. Neuropsychologia 50: 2897–2906 Available: http://www.ncbi.nlm.nih.gov/pubmed/22971812 Accessed 2013 October 31.
  47. 47. Pattamadilok C, Perre L, Ziegler JC (2011) Beyond rhyme or reason: ERPs reveal task-specific activation of orthography on spoken language. Brain Lang 116: 116–124 Available: http://www.ncbi.nlm.nih.gov/pubmed/21211833 Accessed 2013 October 31.
  48. 48. Hasko S, Bruder J, Bartling J, Schulte-Körne G (2012) N300 indexes deficient integration of orthographic and phonological representations in children with dyslexia. Neuropsychologia 50: 640–654 Available: http://www.ncbi.nlm.nih.gov/pubmed/22245008 Accessed 2014 July 29.
  49. 49. Hasko S, Groth K, Bruder J, Bartling J, Schulte-Körne G (2013) The time course of reading processes in children with and without dyslexia: an ERP study. Front Hum Neurosci 7: 570 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3791381&tool=pmcentrez&rendertype=abstract Accessed 2013 October 17.
  50. 50. Savill NJ, Thierry G (2012) Decoding ability makes waves in reading: deficient interactions between attention and phonological analysis in developmental dyslexia. Neuropsychologia 50: 1553–1564 Available: http://www.ncbi.nlm.nih.gov/pubmed/22426204 Accessed 2014 July 9.
  51. 51. Savill NJ, Thierry G (2011) Reading for sound with dyslexia: evidence for early orthographic and late phonological integration deficits. Brain Res 1385: 192–205 Available: http://www.ncbi.nlm.nih.gov/pubmed/21316349 Accessed 2014 July 9.
  52. 52. Bonte ML, Blomert L (2004) Developmental dyslexia: ERP correlates of anomalous phonological processing during spoken word recognition. Brain Res Cogn Brain Res 21: 360–376 Available: http://www.ncbi.nlm.nih.gov/pubmed/15511652 Accessed 2013 October 24.
  53. 53. Maurer U, Brem S, Kranz F, Bucher K, Benz R, et al. (2006) Coarse neural tuning for print peaks when children learn to read. Neuroimage 33: 749–758 Available: http://www.ncbi.nlm.nih.gov/pubmed/16920367 Accessed 2014 February 12.
  54. 54. Oades RD, Dmittmann-Balcar A, Zerbin D (1997) Development and topography of auditory event-related potentials (ERPs): Mismatch and processing negativity in individuals 8?22 years of age. Psychophysiology 34: 677–693 Available: http://doi.wiley.com/10.1111/j.1469-8986.1997.tb02143.x Accessed 2014 July 10.
  55. 55. Shafer VL, Morr ML, Kreuzer JA, Kurtzberg D (2000) Maturation of mismatch negativity in school-age children. Ear Hear 21: 242–251 Available: http://cat.inist.fr/?aModele=afficheN&cpsidt=1410810 Accessed 2014 July 10.
  56. 56. Ponton CW, Eggermont JJ, Kwong B, Don M (2000) Maturation of human central auditory system activity: evidence from multi-channel evoked potentials. Clin Neurophysiol 111: 220–236 Available: http://www.ncbi.nlm.nih.gov/pubmed/10680557.
  57. 57. Annett M (1979) Family handedness in three generations predicted by the right shift theory. Ann Hum Genet 42: 479–491 Available: http://doi.wiley.com/10.1111/j.1469-1809.1979.tb00681.x Accessed 2014 April 16.
  58. 58. Achenbach TM, McConaughy SH (2003) The Achenbach System of Empirically Based Assessment (ASEBA). In: Reynolds CR, Kamphaus RW, editors. Handbook of psychological and educational assessment of children. New York, NY: Guilford Press. 406–430.
  59. 59. Brus BT, Voeten MJM (1973) Eén-minuut test, vorm A en B; Verantwoording en Handleiding. Nimegen: Berkenhout Testmateriaal.
  60. 60. De Vos T (2007) Schoolvaardigheidstoets Technisch Lezen. Amsterdam: Boom test uitgevers.
  61. 61. Raven J, Raven JC, Court JH (1998) Coloured progressive matrices. Oxford, UK: Oxford Psychologists Press.
  62. 62. Faulkner HJ, Levy BA (1999) Fluent and nonfluent forms of transfer in reading: words and their message. Psychon Bull Rev 6: 111–116 Available: http://www.ncbi.nlm.nih.gov/pubmed/12199304.
  63. 63. Fuchs D, Fuchs LS, Compton DL (2004) Identifying Reading Disabilities by Responsiveness-to-Instruction: Specifying Measures and Criteria. Learn Disabil Q 27: 216–227 Available: http://www.jstor.org/stable/1593674?origin=crossref Accessed 2014 April 11.
  64. 64. Stanovich KE, Feeman DJ, Cunningham AE (1983) The development of the relation between letter-naming speed and reading ability. Bull Psychon Soc 21: 199–202 Available: http://link.springer.com/10.3758/BF03334686 Accessed 2014 April 11.
  65. 65. Boersma P, Weenink D (2002) Praat: doing phonetics by computer (Version. 4.0.26). Available: http://www.fon.hum.uva.nl/praat.
  66. 66. Bonte ML, Mitterer H, Zellagui N, Poelmans H, Blomert L (2005) Auditory cortical tuning to statistical regularities in phonology. Clin Neurophysiol 116: 2765–2774 Available: http://www.ncbi.nlm.nih.gov/pubmed/16256430 Accessed 2013 March 19.
  67. 67. Luck SJ (2005) An introduction to the event-related potential technique. Cambridge, MA: The MIT Press.
  68. 68. Delorme A, Makeig S (2004) EEGLAB: an open source toolbox for analysis of single-trial EEG dynamics including independent component analysis. J Neurosci Methods 134: 9–21 Available: http://www.sciencedirect.com/science/article/pii/S0165027003003479 Accessed 2013 June 23.
  69. 69. Lee TW, Girolami M, Sejnowski TJ (1999) Independent component analysis using an extended infomax algorithm for mixed subgaussian and supergaussian sources. Neural Comput 11: 417–441 Available: http://www.ncbi.nlm.nih.gov/pubmed/9950738.
  70. 70. Jung TP, Makeig S, Humphries C, Lee TW, McKeown MJ, et al. (2000) Removing electroencephalographic artifacts by blind source separation. Psychophysiology 37: 163–178 Available: http://www.ncbi.nlm.nih.gov/pubmed/10731767.
  71. 71. Vaessen A, Bertrand D, Tóth D, Csépe V, Faísca L, et al. (2010) Cognitive development of fluent word reading does not qualitatively differ between transparent and opaque orthographies. J Educ Psychol 102: 827–842 Available: http://doi.apa.org/getdoi.cfm?doi=10.1037/a0019465 Accessed 2013 July 9.
  72. 72. Benjamini Y, Hochberg Y (1995) Controlling the false discovery rate: a practical and powerful approach to multiple testing. J R Stat Soc Ser B 57: 289–300 Available: http://www.jstor.org/stable/10.2307/2346101 Accessed 2013 November 15.
  73. 73. Nath AR, Fava EE, Beauchamp MS (2011) Neural correlates of interindividual differences in children’s audiovisual speech perception. J Neurosci 31: 13963–13971 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3203203&tool=pmcentrez&rendertype=abstract Accessed 2014 April 4.
  74. 74. Knowland VCP, Mercure E, Karmiloff-Smith A, Dick F, Thomas MSC (2014) Audio-visual speech perception: a developmental ERP investigation. Dev Sci 17: 110–124 Available: http://www.ncbi.nlm.nih.gov/pubmed/24176002 Accessed 2014 February 25.
  75. 75. Baart M, Stekelenburg JJ, Vroomen J (2014) Electrophysiological evidence for speech-specific audiovisual integration. Neuropsychologia 53: 115–121 Available: http://www.ncbi.nlm.nih.gov/pubmed/24291340 Accessed 2014 February 24.
  76. 76. Pilling M (2009) Auditory event-related potentials (ERPs) in audiovisual speech perception. J Speech Lang Hear Res 52: 1073–1081 Available: http://www.ncbi.nlm.nih.gov/pubmed/19641083 Accessed 2014 March 4.
  77. 77. Van Wassenhove V, Grant KW, Poeppel D (2005) Visual speech speeds up the neural processing of auditory speech. Proc Natl Acad Sci U S A 102: 1181–1186 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=545853&tool=pmcentrez&rendertype=abstract Accessed 2013 June 24.
  78. 78. Liebenthal E, Desai R, Ellingson MM, Ramachandran B, Desai A, et al. (2010) Specialization along the left superior temporal sulcus for auditory categorization. Cereb Cortex 20: 2958–2970 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=2978244&tool=pmcentrez&rendertype=abstract Accessed 2014 April 3.
  79. 79. Beauchamp MS, Lee KE, Argall BD, Martin A (2004) Integration of auditory and visual information about objects in superior temporal sulcus. Neuron 41: 809–823 Available: http://www.ncbi.nlm.nih.gov/pubmed/15003179.
  80. 80. Van Atteveldt NM, Blau VC, Blomert L, Goebel R (2010) fMR-adaptation indicates selectivity to audiovisual content congruency in distributed clusters in human superior temporal cortex. BMC Neurosci 11: 11 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=2829029&tool=pmcentrez&rendertype=abstract.
  81. 81. Nath AR, Beauchamp MS (2012) A neural basis for interindividual differences in the McGurk effect, a multisensory speech illusion. Neuroimage 59: 781–787 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=3196040&tool=pmcentrez&rendertype=abstract Accessed 2014 March 19.
  82. 82. McNorgan C, Awati N, Desroches AS, Booth JR (2013) Multimodal Lexical Processing in Auditory Cortex Is Literacy Skill Dependent. Cereb Cortex. Available: http://www.ncbi.nlm.nih.gov/pubmed/23588185. Accessed 2014 March 5.
  83. 83. Mittag M, Thesleff P, Laasonen M, Kujala T (2013) The neurophysiological basis of the integration of written and heard syllables in dyslexic adults. Clin Neurophysiol 124: 315–326 Available: http://www.ncbi.nlm.nih.gov/pubmed/22939780 Accessed 2013 February 7.
  84. 84. Holloway ID, van Atteveldt N, Blomert L, Ansari D (2013) Orthographic Dependency in the Neural Correlates of Reading: Evidence from Audiovisual Integration in English Readers. Cereb Cortex. Available: http://www.ncbi.nlm.nih.gov/pubmed/24351976. Accessed 2014 March 4.
  85. 85. Csépe V (2003) Auditory event-related potentials in studying developmental dyslexia. In: Csépe V, editor. Dyslexia: Different brain, different behavior. New York, NY: Kluwer Academic Publishers – Plenum Publishers. 81–112.
  86. 86. Schulte-Körne G, Deimel W, Bartling J, Remschmidt H (1998) Auditory processing and dyslexia: evidence for a specific speech. Neuroreport 9: 337–340.
  87. 87. Shestakova A, Huotilainen M, Čeponienė R, Cheour M (2003) Event-related potentials associated with second language learning in children. Clin Neurophysiol 114: 1507–1512 Available: http://linkinghub.elsevier.com/retrieve/pii/S1388245703001342 Accessed 2014 April 4.
  88. 88. Maurer U, Brem S, Bucher K, Brandeis D (2005) Emerging neurophysiological specialization for letter strings. J Cogn Neurosci 17: 1532–1552 Available: http://www.ncbi.nlm.nih.gov/pubmed/16269095.
  89. 89. Jost LB, Eberhard-Moscicka AK, Frisch C, Dellwo V, Maurer U (2013) Integration of Spoken and Written Words in Beginning Readers: A Topographic ERP Study. Brain Topogr. Available: http://www.ncbi.nlm.nih.gov/pubmed/24271979. Accessed 2013 December 17.
  90. 90. Ramus F, Ahissar M (2012) Developmental dyslexia: the difficulties of interpreting poor performance, and the importance of normal performance. Cogn Neuropsychol 29: 104–122 Available: http://www.ncbi.nlm.nih.gov/pubmed/22559749 Accessed 2013 November 1.
  91. 91. Grainger J, Holcomb PJ (2009) Watching the Word Go by: On the Time-course of Component Processes in Visual Word Recognition. Lang Linguist Compass 3: 128–156 Available: http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=2740997&tool=pmcentrez&rendertype=abstract.
  92. 92. Roeske D, Ludwig KU, Neuhoff N, Becker J, Bartling J, et al. (2011) First genome-wide association scan on neurophysiological endophenotypes points to trans-regulation effects on SLC2A3 in dyslexic children. Mol Psychiatry 16: 97–107 Available: http://www.ncbi.nlm.nih.gov/pubmed/19786962 Accessed 2014 June 5.