Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

How to Capitalize on the Retest Effect in Future Trials on Huntington’s Disease

  • Catherine Schramm,

    Affiliations INSERM U955 E01, Neuropsychologie interventionnelle, Institut Mondor de Recherche Biomédicale, Créteil, France, INSERM UMRS1138 E22, Science de l'information au service de la médecine personnalisée, Centre de Recherche des Cordeliers, Université Paris 5, Université Paris 6, Paris, France, Université Paris Est, Faculté de Médecine, Créteil, France, Ecole Normale Supérieure, Institut d’Etude de la Cognition, Paris, France

  • Sandrine Katsahian,

    Affiliations INSERM UMRS1138 E22, Science de l'information au service de la médecine personnalisée, Centre de Recherche des Cordeliers, Université Paris 5, Université Paris 6, Paris, France, Assistance Publique-Hôpitaux de Paris, Service d’informatique et statistiques, Hôpital Européen Georges Pompidou, Paris, France

  • Katia Youssov,

    Affiliations INSERM U955 E01, Neuropsychologie interventionnelle, Institut Mondor de Recherche Biomédicale, Créteil, France, Université Paris Est, Faculté de Médecine, Créteil, France, Ecole Normale Supérieure, Institut d’Etude de la Cognition, Paris, France, Assistance Publique-Hôpitaux de Paris, Centre National de Référence pour la Maladie de Huntington, Hôpital Henri Mondor, Créteil, France

  • Jean-François Démonet,

    Affiliation Leenaards Memory Centre, Clinical Neurosciences Department, CHUV Lausanne, Lausanne, Switzerland

  • Pierre Krystkowiak,

    Affiliations Centre Hospitalier Universitaire d'Amiens, Service de neurologie, Amiens, France, EA 4559 - Laboratoire de Neurosciences Fonctionnelles et Pathologie (LNFP), Université de Picardie Jules Verne (UPJV), Amiens, France, SFR CAP-Santé (FED 4231), Amiens, France

  • Frédéric Supiot,

    Affiliation Hôpital Erasme ULB, Service de Neurologie, Bruxelles, Belgium

  • Christophe Verny,

    Affiliation CHU d'Angers, Centre de Référence des Maladies Neurogénétiques, Service de Neurologie, Angers, France

  • Laurent Cleret de Langavant,

    Affiliations INSERM U955 E01, Neuropsychologie interventionnelle, Institut Mondor de Recherche Biomédicale, Créteil, France, Université Paris Est, Faculté de Médecine, Créteil, France, Ecole Normale Supérieure, Institut d’Etude de la Cognition, Paris, France, Assistance Publique-Hôpitaux de Paris, Centre National de Référence pour la Maladie de Huntington, Hôpital Henri Mondor, Créteil, France

  • Anne-Catherine Bachoud-Lévi ,

    bachoud@gmail.com

    Affiliations INSERM U955 E01, Neuropsychologie interventionnelle, Institut Mondor de Recherche Biomédicale, Créteil, France, Université Paris Est, Faculté de Médecine, Créteil, France, Ecole Normale Supérieure, Institut d’Etude de la Cognition, Paris, France, Assistance Publique-Hôpitaux de Paris, Centre National de Référence pour la Maladie de Huntington, Hôpital Henri Mondor, Créteil, France

  • European Huntington's Disease Initiative Study Group and the Multicentre Intracerebral Grafting in Huntington's Disease Group

    Membership of the European Huntington's Disease Initiative Study Group and the Multicentre Intracerebral Grafting in Huntington’s Disease Group is provided in the Acknowledgments.

Abstract

The retest effect—improvement of performance on second exposure to a task—may impede the detection of cognitive decline in clinical trials for neurodegenerative diseases. We assessed the impact of the retest effect in Huntington’s disease trials, and investigated its possible neutralization. We enrolled 54 patients in the Multicentric Intracerebral Grafting in Huntington’s Disease (MIG-HD) trial and 39 in the placebo arm of the Riluzole trial in Huntington’s Disease (RIL-HD). All were assessed with the Unified Huntington’s Disease Rating Scale (UHDRS) plus additional cognitive tasks at baseline (A1), shortly after baseline (A2) and one year later (A3). We used paired t-tests to analyze the retest effect between A1 and A2. For each task of the MIG-HD study, we used a stepwise algorithm to design models predictive of patient performance at A3, which we applied to the RIL-HD trial for external validation. We observed a retest effect in most cognitive tasks. A decline in performance at one year was detected in 3 of the 15 cognitive tasks with A1 as the baseline, and 9 of the 15 cognitive tasks with A2 as the baseline. We also included the retest effect in performance modeling and showed that it facilitated performance prediction one year later for 14 of the 15 cognitive tasks. The retest effect may mask cognitive decline in patients with neurodegenerative diseases. The dual baseline can improve clinical trial design, and better prediction should homogenize patient groups, resulting in smaller numbers of participants being required.

Introduction

Huntington’s disease (HD) is an inherited neurodegenerative disorder involving motor, behavioral and cognitive impairments [1]. The cognitive disorders have a major impact on daily life, but most clinical trials focus on motor endpoints. This is because clinical trial endpoints must be able to capture both patient decline and treatment efficacy, and cognitive decline is much more difficult to capture within one year in patients at early disease stages [2] than motor decline. This difficulty of assessment results from the heterogeneity of cognitive changes (language, memory, etc.) and two opposing effects: the retest effect and patient decline due to disease progression. The retest effect is defined as an improvement in performance with repeated exposure to a task, with the greatest improvement occurring between the first two assessments [35]. This effect combines familiarity with the task and testing environment and the possible recall of responses [2]. The first assessment, during which everything is new to the patient, is always the most difficult.

The retest effect may have contributed to the failure of some neuroprotection trials, by adding noise to statistics comparing patients with different backgrounds at baseline, particularly in trials including small numbers of patients, such as those assessing biotherapy. One approach to neutralizing the retest effect is to carry out a second assessment (A2) shortly after the first (A1), and then discard the results obtained at A1 from the analysis, using performance at A2 as the baseline [2]. In addition, the retest effect (ΔA2-A1) can be used to improve the prediction of long-term patient performance. Indeed, in an observational longitudinal study in HD patients, the retest effect (ΔA2-A1 around 7 months) accounted for up to 36% of the variance of performance at A3 (ΔA3-A2 around 29 months) [6]. Likewise, in healthy elderly adults, performance at A3 (one year) is accurately predicted by the one week-interval retest effect (ΔA2-A1) [7].

However, the impact of the retest effect in clinical trials, which include additional variability (placebo effect, hope, anxiety about treatment and randomization), remains unknown. Two trials, the Multicentric Intracerebral Grafting in Huntington’s Disease (MIG-HD) [8] and Riluzole in Huntington’s Disease (RIL-HD) [9] trials, were designed with a short-term test-retest procedure. We used the MIG-HD trial (i) to assess whether the retest effect modified performance and whether our strategy of using the second assessment as a baseline was sensitive to cognitive decline in the long-term (A3) and (ii) to evaluate whether introducing the retest effect (ΔA2-A1) into the model of disease progression in patients improved the predictive value of the model in the long term (A3). Finally, we transferred the models obtained for the MIG-HD cohort to the RIL-HD cohort, to assess their predictive value in another population.

Materials and Methods

Participants and design

Patients were enrolled in two separate trials: the MIG-HD trial (N = 54, Ref NCT00190450, PI AC Bachoud-Lévi) [8], which is currently underway, and the placebo group of the cognitive ancillary study of the RIL-HD trial conducted only in France (N = 39, Ref NCT00277602, study coordinator Sanofi) [9]. Both trials were approved by the institutional review board (Comité Consultatif de Protection des Personnes dans la Recherche Biomédicale) of Henri-Mondor Hospital at Créteil (MIG-HD the September 25, 2001, and RIL-HD the December 18, 2002). Patients had signed an inform consent. The data were analyzed anonymously.

The MIG-HD trial is a phase II randomized trial assessing the efficacy of cell transplantation in HD patients at early stages of the disease. Patients were assessed at inclusion (A1), then 35 days (SD = 15) later (A2). They were randomized at one year (A3), to determine the timing of transplantation (M13-M14 for the early graft group or M33-M34 for the late graft group). Patients were followed up until 52 months.

The RIL-HD trial is a phase III multinational, randomized, placebo-controlled, double-blind study, for which a cognitive ancillary study was conducted in France from 1999 to 2004, on patients with moderately advanced HD. Patients were assessed at inclusion (A1), 15 days (SD = 8) later (A2) and at one year (A3), with randomization at A2.

The demographic features for patients at A1 are displayed in Table 1.

thumbnail
Table 1. Characteristics of patients at their inclusion (A1) in the MIG-HD and RIL-HD trials.

https://doi.org/10.1371/journal.pone.0145842.t001

Clinical assessments

The Unified Huntington’s Disease Rating Scale (UHDRS) [10] and additional cognitive tests were used in both studies. Motor score reflected both voluntary and involuntary capacity and ranged from 0 to 124 (highest severity). Functional disability was assessed with Total Functional Capacity (TFC, range: 13 to 0) and Independence Scale (IS, range 100 to 0) scores, with lower scores indicating greater functional impairment, and the Functional Assessment Scale (FAS, 25 to 50), with higher scores indicating greater functional impairment. The severity and frequency of behavioral dysfunctions were quantified with the behavioral part of the UHDRS (range: 0 to 88), with higher scores indicating greater impairment. Global cognitive efficiency was evaluated with the Mattis Dementia Rating Scale (MDRS) [11]. Several tasks were used to assess attention and executive functions: letter fluency (for P, R and V in French) determined for 1 minute, the Symbol Digit Modalities Test (SDMT), the three components of the Stroop test (color naming, word reading, and color-word interference), each assessed for 45 seconds [12], categorical fluency (for animals) assessed for 1 minute [13],[14], the Trail-Making Test forms A and B (TMT A and B) [15], scoring the time taken to link 25 points, with a maximal time of 240 seconds, and figure cancellation tasks [16], in which patients were asked to cross out one, two and then three figures from a panel of signs, in 90 seconds, with lower scores indicating greater cognitive impairment. Short-term and long-term memory were evaluated with the Hopkins Verbal Learning Task (HVLT) including immediate recall, delayed recall and recognition tasks [17],[18]. By contrast to the other tasks, the HVLT was assessed with alternating parallel forms.

Each patient performed one motor test, three functional tests, one behavioral test and 15 cognitive tests at each assessment point.

Statistical Analysis

Evaluation of the retest effect in the MIG-HD cohort.

For each task, we used Student’s t-tests for paired data to compare performances, first between A1 and A2, to measure the potential retest effect, then between A1 and A3, to assess the decline over a one-year period and between A2 and A3, to determine whether discarding the A1 data unmasked a decline that was otherwise undetectable.

Modeling of performance for the MIG-HD cohort.

For each task, we selected the multivariate linear model best predicting the data at one year, by stepwise selection [19] with the Akaike Information Criterion (AIC) [20]. We used an iterative algorithm (stepwise selection) to select, without prior assumptions, the best predictive factors from a set of 10 variables (performance at A1, retest, age, sex, education level expressed as the number of years spent studying, parental inheritance, age of parent at disease onset, CAG repeat length, time since disease onset and the nature of the first symptom appearing at disease onset (motor, cognitive or psychiatric), as determined by the clinician or, if no clinician’s assessment was available, by the family or the patient). Lower AIC values indicate a better fit of the model to the data. The first model selection step was carried out for patients with complete data sets only. Estimates of regression coefficients were refined, by recalculating each model, using all the available complete data for the selected variables. The retest is the difference: performance at A2 –performance at A1 and is denoted ΔA2-A1. For each task, performance at A3 (P) was predicted as follows: where age, education level and age of parent at onset are expressed in years; the first symptom could be motor, cognitive or psychiatric; β0 is the intercept and, for each variable, βvariable is its associated regression coefficient (0 for the variables not selected). For quantitative variables, βvariable was multiplied by the value of the variable. For qualitative variables (sex, inheritance and first symptom), “woman”, “maternal inheritance” and “motor symptom” constituted the reference factors, such that βwoman = βmaternal = βmotor = 0. Calculation of the associated 95% predictive interval (95% PI) is explained in the supplemental data (S1 Text).

External validation on the RIL-HD cohort.

We used models constructed from data for the MIG-HD cohort to predict performances at A3 for each patient in the RIL-HD cohort. Then, for each task, we measured the concordance between observed (O) and predicted (P) values, using the intraclass correlation coefficient (ICC) and the coefficient of determination (). The ICC was calculated with a two-way mixed effect model [21] and evaluates agreement between observed (O) and predicted (P) performances at A3 in the RIL-HD cohort. The coefficient of determination () is the percentage of the observed performance variance explained by the model constructed from MIG-HD data. It assesses the degree to which observed performance at A3 in the RIL-HD cohort is accurately predicted by the model, as follows: where i refers to a patient and m is the mean observed performance at A3. = 1 indicates a perfect predictive value of the model, whereas ≤ 0 indicates that the model is not informative.

Analyses were performed with R 2.13 software (http://www.r-project.org/). All tests were two-tailed and values of P < 0.05 were considered significant.

Results

Evaluating the retest effect in the MIG-HD cohort

We assessed the retest effect between A1 and A2 in the MIG-HD cohort. Performance improved in seven cognitive tasks, and remained stable in the other cognitive, motor and functional tasks, except for FAS score, which declined between A1 and A2 (Fig 1).

thumbnail
Fig 1. Impact of the retest effect in the MIG-HD cohort.

SDMT: Symbol Digit Modalities Test; Stroop C, W and C/W: Stroop color, word and color/word interference; MDRS: Mattis Dementia Rating Scale; TMT A, B: Trail-Making Test A and B; TFC: Total Functional Capacity; IS: Independence Scale; FAS: Functional Assessment Scale. The red curve represents the baseline (reference score A1) and the blue curve shows the mean relative score one month later (A2). The portion of the blue curve beyond the red curve indicates performance improvement between A1 and A2. Paired t-tests, significance: * P<0.05, ** P<0.01, *** P<0.001.

https://doi.org/10.1371/journal.pone.0145842.g001

We assessed decline between A1 and A3 and between A2 and A3 in the MIG-HD cohort (Fig 2). The use of A2 as the baseline increased the number of tasks for which a decline in performance was detected from three to nine, but FAS score was the only motor or functional performance affected. Indeed, FAS performance declined between A1 and A3 but not between A2 and A3. Behavioral performance improved between A2 and A3.

thumbnail
Fig 2. Observed performance at one year (A3), with A1 or A2 used as the baseline, in the MIG-HD cohort.

SDMT: Symbol Digit Modalities Test; Stroop C, W and C/W: Stroop color, word and color/word interference; MDRS: Mattis Dementia Rating Scale; TMT A, B: Trail-Making Test A and B; TFC: Total Functional Capacity; IS: Independence Scale; FAS: Functional Assessment Scale. The red curve represents the baseline (reference score). The blue (or green) curve corresponds to the mean relative score one year later (A3), with A1 (or A2 for the green curve) used as the baseline. A green curve within the blue curve indicates that the decline was easier to detect if A2 was used as the baseline, rather than A1. Paired t-tests, significance: * P<0.05, ** P<0.01, *** P<0.001.

https://doi.org/10.1371/journal.pone.0145842.g002

Modeling of performance in the MIG-HD cohort

Table 2 displays the regression coefficients of the predictive model for each task, for the MIG-HD cohort. Performance at A1 was predictive of performance at A3 in all tasks. Introducing the difference in performance between A1 and A2 (ΔA2-A1) into the models improved the prediction of performance at A3 for 14 of the 15 cognitive tasks, for behavioral and motor performance and TFC. Larger numbers of CAG repeats were associated with a poorer FAS and IS scale scores and poorer motor performance, but better behavioral performance. Women outperformed men in 7 of the 15 cognitive tasks. Sex had no effect on motor and functional performances, whereas behavioral performance was better in women than in men. Higher education levels were associated with better performance at A3 for all components of the HVLT.

The regression coefficients presented in Table 2 are those used in the predictive models. For example, the performance at A3 in letter Fluency 1’ is given by the following formula:

The equations associated with the predictive models for each task are detailed in S1 Table. Moreover, S2 Table gives additional parameters for calculation of the 95% PI.

External validation on the RIL-HD cohort

For each task, we determined the predictive value of models by calculating the ICC and (Fig 3). Performance in the RIL-HD trial was well predicted for 14 of 20 tasks by the models developed with data for the MIG-HD cohort ( ≥ 0.5 and ICC ≥ 0.6).

thumbnail
Fig 3. External validation of models in the RIL-HD cohort, based on Re2 and ICC.

SDMT: Symbol Digit Modalities Test; Stroop C, W and C/W: Stroop color, word and color/word interference; MDRS: Mattis Dementia Rating Scale; TMT A, B: Trail-Making Test A and B; HVLT: Hopkins Verbal Learning Task; TFC: Total Functional Capacity; IS: Independence Scale; FAS: Functional Assessment Scale. N: number of patients in the RIL-HD cohort for whom all the data required for the predictive model were available. Re2: coefficient of determination for external validation. ICC: intraclass correlation coefficient. 95% CI: 95% confidence interval. a: Re2 = -0.7. The red line represents the limit for a high-quality model (Re2 > 50% of the observed variance explained by the model).

https://doi.org/10.1371/journal.pone.0145842.g003

Discussion

The design of clinical trials for neurodegenerative diseases could be improved by methodological approaches based on our knowledge of the patient’s cognitive performances. However, cognitive knowledge is obtained mostly through longitudinal follow-up in observational studies, which may not include variability factors inherent to clinical trials. The retest effect may impede observations of cognitive decline in patients with Huntington’s disease. We therefore assessed its impact in two long-term clinical trials in HD patients, with a short interval between the first and second assessments (MIG-HD, RILH-HD). We first determined whether there was a detectable retest effect between the first two assessments (A1 and A2), and then evaluated the impact of this effect one year later (A3). We found that replacing A1 with A2 as the baseline unmasked a decline that would not otherwise have been detected. Indeed, the comparison between A2 and A3 showed declines that were not apparent in the comparison between A1 and A3. We also modeled patient performance and showed how the inclusion of the retest effect in patient performance models would improve trial design.

At one year, decline was observed in a few cognitive tasks (SDMT, MDRS and the HVLT immediate recall), the motor task and all functional tasks. However, consistent with previous findings [2], there was a pronounced retest effect in cognitive tasks (letter fluency, SDMT, Stroop color and color/word interference, TMT A and 2- and 3-figure cancellation tasks), but no such effect in motor and functional assessments. This retest effect may hamper the objective detection of cognitive decline, with a major impact in tasks with a high cognitive demand, obscuring performance decline over a one-year period [22]. Neutralization of the retest effect is particularly important in clinical trials, because some patients may already have been exposed to testing whereas others have not, adding background noise to the overall performance data. Assuming that the retest effect is maximal at the second assessment, the use of this assessment as the baseline can decrease the impact of the retest effect on subsequent assessments. By discarding performances at A1 and using the performance measured at A2 as the baseline, we unmasked a decline in six tasks (Stroop color and color/word interference, recognition part of HVLT, TMT A and 2- and 3-figure cancellation), demonstrating the efficacy of this strategy for small samples. However, the improvement in behavioral performance [23], contrasting with the decline in other task performances, may reflect the patients’ hopes and expectations of treatment.

The HVLT constitutes a specific case: we alternated parallel forms because of the strength of item recall in declarative memory tasks [24]. However, alternation was not used for other tasks, because parallel forms are of no interest for procedural tasks or tasks with a strong motor output (SDMT, TMT A and verbal fluency tasks) [25]. The use of parallel forms should also be limited because of their low intrasubject equivalence, potentially introducing noise into longitudinal performance analyses. Furthermore, the ceiling effect observed in patients with high scores in the HVLT, MDRS and TMT tasks limits the utility of neutralizing the retest effect.

However, the retest effect depends not only on the nature of the task, but also on the population assessed [26]. Indeed, Cooper et al. [27], [28] demonstrated the existence of a retest effect in categorical fluency assessment in healthy participants but not in patients with Alzheimer’s disease or mild cognitive impairments. Likewise, we found no retest effect for this task in HD patients.

In addition to masking decline, the retest effect may provide information about disease progression [7]. This suggests that combining a strategy based on the individual performance of patients and the nature of the tasks may be useful. Indeed, the modeling of patient performance at one year for each task showed that ΔA2-A1 performance, even in the absence of a significant retest effect, accurately predicted performance for most cognitive tasks in HD and for motor and behavior tasks and TFC. ΔA2-A1 performance appears to be more frequently selected by stepwise algorithms than sociodemographic and genetic variables. We also arbitrated between parameters to strengthen our models. For example, both the number of CAG repeats and age at onset are eligible variables [29], but they are correlated [3032], so only one of these factors should be included in the model [33]. We decided to include the number of CAG repeats, as age at onset is subject to some degree of subjectivity. Likewise, rather than using the performance in one task to explain performance in another task (e.g. using motor score to explain TFC [34]), we limited the set of eligible variables to demographic variables. Finally, we did not include handedness in our models, because 90% of the patients were right-handed.

This approach made it possible to include a larger number of covariates in our models than in those of previous studies and to prioritize them through the selection algorithm. For example, the number of CAG repeats has been reported to affect general verbal and spatial abilities [35], whereas our stepwise selection suggested that it was predictive of performance in the 3-figure cancellation task, which has a spatial nonverbal component. Indeed, the number of CAG repeats was found to have less impact than the sex of the patient in verbal tasks (letter and categorical fluencies) and sex was not included in the model described in the previous study. Furthermore, dichotomization of the number of CAG repeats variable (small and large numbers of repeats) may have resulted in greater importance being assigned to this variable than in models, such as ours, in which the number of CAG repeats was treated as a continuous variable. Like Ruocco [36], Kieburtz [37] and Feigin et al. [38], we showed that the number of CAG repeats improved in the prediction of motor performance, but not TFC. Finally, higher education levels were associated with a better performance, for all HVLT components.

The small number of patients enrolled in the MIG-HD study is a potential limitation in the search for predictive factors for future studies. However, external validation on the RIL-HD cohort, through calculation of the intraclass correlation coefficient and the determination coefficient (), demonstrated the reproducibility and robustness of our models, regardless of the differences between the two trials. Indeed, patients in the MIG-HD trial were not randomized until one year (A3), whereas those in the RIL-HD study were randomized at the second assessment (A2). Consequently, the patients in the MIG-HD study approached the intervention with greater hope, whereas those in the placebo group of the RIL-HD study may have been aware of a lack of improvement during the follow-up period. This difference may account for the poor prediction of behavioral performance in the RIL-HD study ( < 0). By contrast, the difference in time interval between A1 and A2 in the two studies had no impact on prediction quality, further demonstrating the validity of the models. The models were constructed with data from patients with relatively mild disease. They may, therefore, not be applicable to patients with more advanced HD. Indeed, retest effects would be expected to be smaller in patients with more severe disease.

Our findings indicate that the retest effect is a limitation in clinical trials, but that both its neutralization, through the use of a second assessment as a baseline, and its integration into task modeling would be beneficial in future trials. For example, our predictive models may facilitate the identification of rapid decliners [39], defined as individuals whose observed performance is worse than predicted. Indeed, in longitudinal clinical trials, treatment effects could be masked in such patients, as already shown for Alzheimer’s disease [40]. The identification of such patients is helpful for trial design, in two ways. First, the exclusion of such patients would probably decrease intersubject variability, making it possible to decrease sample size. Second, rapid decliners could be uniformly allocated to the different arms of the study by stratified randomization, to ensure the constitution of comparable groups, in terms of both baseline data and disease progression.

Our findings suggest that the retest effect is detrimental, if uncontrolled, in clinical trials for neurodegenerative diseases, such as Huntington’s disease. We show here that if two assessments are performed a short time apart, use of the second assessment as the baseline increases the chances of detecting an effect of treatment, if there is one. In addition, including the retest effect in models renders the resulting models more predictive, making it possible to refine the design of future trials. This constitutes a great stride forward in cognitive assessments in clinical trials.

Supporting Information

S2 Table. M matrix for calculating the 95% prediction interval for performance at A3 for each task.

https://doi.org/10.1371/journal.pone.0145842.s002

(DOCX)

S1 Text. Statistical explanation for calculation of the 95% prediction interval for performance at A3, for each task.

https://doi.org/10.1371/journal.pone.0145842.s003

(DOCX)

Acknowledgments

The authors thank Julie Sappa from Alex Edelman & Associates for her language corrections.

We thank the neurologists and the neuropsychologists from the MIGHD group trial who collected the data: A-C. Bachoud-Lévi (Henri Mondor hospital, Créteil, Principal investigator), M-F Boissé (Henri Mondor hospital, Créteil, Neuropsychologist), L. Lemoine (Henri Mondor hospital, Créteil, Neuropsychologist), C. Verny (Angers hospital, Site coordinator), G. Aubin (Angers hospital, Neuropsychologist), J-F Demonet (CHU Rangueil, Toulouse, Site coordinator), F. Calvas (CHU Rangueil, Toulouse, Investigator), P. Krystkowiak (Roger Salengro hospital, Lille and CHU d’Amiens, Amiens, Sites coordinator), C. Simonin (Roger Salengro hospital, Lille, Investigator), M. Delliaux (Roger Salengro hospital, Lille, Neuropsychologist), P. Damier (Hôpital Nord Laennec, Nantes, Site coordinator), P. Renou (Hôpital Nord Laennec, Nantes, Investigator), F. Supiot (Erasme hospital, Bruxelles, Site coordinator), H. Slama (Erasme hospital, Bruxelles, Neuropsychologist).

We thank the EHDI Study group: A-C. Bachoud-Lévi (Henri Mondor hospital Créteil, Principal investigator of the RIL ancillary study), J. S. Guillamo (Henri Mondor hospital Créteil), M-F Boissé (Henri Mondor hospital Créteil, Neuropsychologist), A. Dürr (Fédération de Neurologie, Pitié-Salpêtrière hospital, Paris), F. Bloch (Fédération de Neurologie, Pitié-Salpêtrière hospital, Paris), O. Messouak (Fédération de Neurologie, Pitié-Salpêtrière hospital, Paris), C. Tallaksen (Fédération de Neurologie, Pitié-Salpêtrière hospital, Paris), B. Dubois (Fédération de Neurologie, Pitié-Salpêtrière hospital, Paris), A. Engles (Hôpital Roger Salengro, Lille), P. Krystkowiak, (Hôpital Roger Salengro, Lille) A. Destee (Hôpital Roger Salengro, Lille), A. Memin (Hôpital Roger Salengro, Lille), S. Thibaut-Tanchou (Hôpital Roger Salengro, Lille), F. Pasquier (Hôpital Roger Salengro, Lille, Neurology), J-P. Azulay (CHU Purpan, Toulouse), M. Galitzky (CHU Purpan, Toulouse), O. Rascol (CHU Purpan, Toulouse), H. Mollion (Pierre Wertheimer hospital, Lyon), E. Broussolle (Pierre Wertheimer hospital, Lyon), M. Madigand (La Beauchée hospital, Saint-Brieuc), F. Lallement (La Beauchée hospital, Saint-Brieuc), C. Goizet (Haut-Lévêque hospital, Pessac), F. Tison (Haut-Lévêque hospital, Pessac) S. Arguillère (CHG du Pays d’Aix, Aixen-Provence), F. Viallet (CHG du Pays d’Aix, Aixen-Provence) S. Bakchine (Maison Blanche hospital, Reims), J. Khoris, (Gui de Chauliac hospital, Montpellier), M. Pages (Gui de Chauliac hospital, Montpellier), W. Camu (Gui de Chauliac hospital, Montpellier), F. Resch (Charles Nicolle hospital, Rouen), D. Hannequin (Charles Nicolle hospital, Rouen), F. Durif (Gabriel Montpied hospital, Clermont-Ferrand), D. Saudeau (CHRU Bretonneau,Tours), A. Autret (CHRU Bretonneau,Tours).

Author Contributions

Conceived and designed the experiments: ACBL CS SK. Analyzed the data: CS SK. Wrote the paper: ACBL CS SK. Collected the data: ACBL KY JFD PK FS CV LCL.

References

  1. 1. Bates G, Tabrizi S, Jones L. Huntington’s Disease. 3rd ed. Oxford: Oxford University Press; 2014.
  2. 2. Bachoud-Lévi A-C, Maison P, Bartolomeo P, Boissé M-F, Dalla-Barba G, Ergis A-M, et al. Retest effects and cognitive decline in longitudinal follow-up of patients with early HD. Neurology. 2001;56(8):1052–8. pmid:11320178
  3. 3. Salthouse TA, Schroeder DH, Ferrer E. Estimating retest effects in longitudinal assessments of cognitive functioning in adults between 18 and 60 years of age. Dev Psychol. 2004;40(5):813–22. pmid:15355168
  4. 4. Collie A, Maruff P, Darby DG, McStephen M. The effects of practice on the cognitive test performance of neurologically normal individuals assessed at brief test-retest intervals. J Int Neuropsychol Soc. 2003;9(3):419–28. pmid:12666766
  5. 5. Stout JC, Queller S, Baker KN, Cowlishaw S, Sampaio C, Fitzer-Attas C, et al. HD-CAB: a cognitive assessment battery for clinical trials in Huntington’s disease. Mov Disord. 2014;29(10):1281–8. pmid:25209258
  6. 6. Duff K, Beglinger LJ, Schultz SK, Moser D, McCaffrey R, Haase R, et al. Practice effects in the prediction of long-term cognitive outcome in three patient samples: a novel prognostic index. Arch Clin Neuropsychol. 2007;22(1):15–24. pmid:17142007
  7. 7. Duff K, Beglinger LJ, Moser DJ, Paulsen JS, Schultz SK, Arndt S. Predicting cognitive change in older adults: the relative contribution of practice effects. Arch Clin Neuropsychol. 2010;25(2):81–8. pmid:20064816
  8. 8. Bachoud-Lévi A-C, Hantraye P, Peschanski M. Fetal neural grafts for Huntington’s disease: a prospective view. Mov Disord. 2002;17(3):439–44. pmid:12112189
  9. 9. Landwehrmeyer GB, Dubois B, de Yébenes JG, Kremer B, Gaus W, Kraus P, et al. Riluzole in Huntington’s disease: a 3-year, randomized controlled study. Ann Neurol. 2007;62(3):262–72. pmid:17702031
  10. 10. Kremer HPH, Hungtington Study Group. Unified Huntington’s disease rating scale: reliability and consistency. Mov Disord. 1996;11:136–42. pmid:8684382
  11. 11. Mattis S. Mental status examination for organic mental syndrome in the elderly patient. In: Bellak L, Karasu TB, eds. Geriatric psychiatry: a handbook for psychiatrists and primary care physicians. New York: Grune & Stratton, 1976:p77–121.
  12. 12. Golden CJ. Stroop colour and word test. Age. 1978;15:90.
  13. 13. Butters N, Wolfe J, Granholm E, Martone M. An assessment of verbal recall, recognition and fluency abilities in patients with Huntington’s disease. Cortex. 1986;22(1):11–32. pmid:2940074
  14. 14. Cardebat D, Doyon B, Puel M, Goulet P, Joanette Y. Formal and semantic lexical evocation in normal subjects. Performance and dynamics of production as a function of sex, age and educational level. Acta Neurol Belg. 1990;90(4):207–17. pmid:2124031
  15. 15. Reitan RM. Validity of the trail making test as an indicator of organic brain damage. Percept Mot Skills. 1958;8(3):271–6.
  16. 16. Zazzo R, Stambak M. Le test des deux barrages: Une épreuve de pointillage. Neuchatel, Switzerland: Delachaux et Niestlé; 1960.
  17. 17. Brandt J. The Hopkins verbal learning test: Development of a new memory test with six equivalent forms. Clin Neuropsychol. 1991;5(2):125–42.
  18. 18. Rieu D, Bachoud-Lévi A-C, Laurent A, Jurion E, Dalla Barba G. Adaptation française du «Hopkins verbal learning test». Rev Neurol. 2006;162(6):721–8.
  19. 19. Hocking RR. A Biometrics invited paper. The analysis and selection of variables in linear regression. Biometrics. 1976;32(1):1–49.
  20. 20. Akaike H. Information Theory and an Extension of the Maximum Likelihood Principle. In Selected Papers of Hirotugu Akaike. New York: Springer; 1998.
  21. 21. Shrout PE, Fleiss JL. Intraclass correlations: uses in assessing rater reliability. Psychol Bull. 1979;86(2):420–8. pmid:18839484
  22. 22. Snowden J, Craufurd D, Griffiths H, Thompson J, Neary D. Longitudinal evaluation of cognitive disorder in Huntington’s disease. J Int Neuropsychol Soc. 2001;7(1):33–44. pmid:11253840
  23. 23. Tabrizi SJ, Reilmann R, Roos RAC, Durr A, Leavitt B, Owen G, et al. Potential endpoints for clinical trials in premanifest and early Huntington’s disease in the TRACK-HD study: analysis of 24 month observational data. Lancet Neurol. 2012;11(1):42–53. pmid:22137354
  24. 24. Benedict RH, Zgaljardic DJ. Practice effects during repeated administrations of memory tests with and without alternate forms. J Clin Exp Neuropsychol. 1998;20(3):339–52. pmid:9845161
  25. 25. Beglinger LJ, Gaydos B, Tangphao-Daniels O, Duff K, Kareken D, Crawford J, et al. Practice effects and the use of alternate forms in serial neuropsychological testing. Arch Clin Neuropsychol. 2005;20(4):517–29. pmid:15896564
  26. 26. McCaffrey RJ, Westervelt HJ. Issues associated with repeated neuropsychological assessments. Neuropsychol Rev. 1995;5(3):203–21. pmid:8653109
  27. 27. Cooper DB, Lacritz LH, Weiner MF, Rosenberg RN, Cullum CM. Category fluency in mild cognitive impairment: reduced effect of practice in test-retest conditions. Alzheimer Dis Assoc Disord. 2004;18(3):120–2. pmid:15494616
  28. 28. Cooper DB, Epker M, Lacritz L, Weiner M, Rosenberg RN, Honig L, et al. Effects of practice on category fluency in Alzheimer’s disease. Clin Neuropsychol. 2001;15(1):125–8. pmid:11778573
  29. 29. Rosenblatt A, Kumar BV, Mo A, Welsh CS, Margolis RL, Ross CA. Age, CAG repeat length, and clinical progression in Huntington’s disease. Mov Disord. 2012;27(2):272–6. pmid:22173986
  30. 30. Stine OC, Pleasant N, Franz ML, Abbott MH, Folstein SE, Ross CA. Correlation between the onset age of Huntington’s disease and length of the trinucleotide repeat in IT-15. Hum Mol Genet. 1993;2(10):1547–9. pmid:8268907
  31. 31. Lee JM, Ramos EM, Lee JH, Gillis T, Mysore JS, Hayden MR, et al. CAG repeat expansion in Huntington disease determines age at onset in a fully dominant fashion. Neurology. 2012;78(10):690–5. pmid:22323755
  32. 32. Langbehn DR, Hayden MR, Paulsen JS and the PREDICT-HD Investigators of the HuntingtonStudy Group. CAG-repeat length and the age of onset in Huntington disease (HD): a review and validation study of statistical approaches. Am J Med Genet B Neuropsychiatr Genet. 2010;153B(2):397–408. pmid:19548255
  33. 33. Næs T, Mevik B-H. Understanding the collinearity problem in regression and discriminant analysis. J Chemom. 2001;15(4):413–26.
  34. 34. Marder K, Zhao H, Myers RH, Cudkowicz M, Kayson E, Kieburtz K, et al. Rate of functional decline in Huntington’s disease. Neurology. 2000;54(2):452. pmid:10668713
  35. 35. Brandt J, Bylsma FW, Gross R, Stine OC, Ranen N, Ross CA. Trinucleotide repeat length and clinical progression in Huntington’s disease. Neurology. 1996;46(2):527–31. pmid:8614526
  36. 36. Ruocco HH, Bonilha L, Li LM, Lopes-Cendes I, Cendes F. Longitudinal analysis of regional grey matter loss in Huntington disease: effects of the length of the expanded CAG repeat. J Neurol Neurosurg Psychiatry. 2008;79(2):130–5. pmid:17615168
  37. 37. Kieburtz K, MacDonald M, Shih C, Feigin A, Steinberg K, Bordwell K, et al. Trinucleotide repeat length and progression of illness in Huntington’s disease. J Med Genet. 1994;31(11):872–4. pmid:7853373
  38. 38. Feigin A, Kieburtz K, Bordwell K, Como P, Steinberg K, Sotack J, et al. Functional decline in Huntington’s disease. Mov Disord. 1995;10(2):211–4. pmid:7753064
  39. 39. Carcaillon L, Berrut G, Sellalm F, Dartigues J-F, Gillette S, Péré J-J, et al. Diagnosis of Alzheimer’s disease patients with rapid cognitive decline in clinical practice: interest of the Deco questionnaire. J Nutr Health Aging. 2011;15(5):361–6. pmid:21528162
  40. 40. Noda A, Kraemer HC, Taylor JL, Schneider B, Ashford JW, Yesavage JA. Strategies to reduce site differences in multisite studies: a case study of Alzheimer disease progression. Am J Geriatr Psychiatry. 2006;14(11):931–8. pmid:17068315