Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Performance of the 2010 Classification Criteria for Rheumatoid Arthritis: A Systematic Literature Review and a Meta-Analysis

Abstract

Objectives

To evaluate the performance of the American College of Rheumatology (ACR)/European League Against Rheumatism (EULAR) 2010 classification criteria for rheumatoid arthritis (RA) with a systematic literature review and a meta-analysis.

Methods

PubMed, Embase, Cochrane Library and the abstracts of the ACR and EULAR meetings (2010–2012) were searched for original articles or abstracts with the following inclusion criteria: 1) recent onset arthritis, with at least one swollen joint and no alternative diagnosis; 2) the ACR/EULAR 2010 criteria as index test; 3) the prescription of methotrexate (MTX) or disease modifying antirheumatic drugs (DMARDs) at any time during follow-up as reference standard. Data were pooled using the bivariate model. Three meta-analyses were performed with MTX (primary analysis), DMARDs or their combination (secondary analyses) as reference standard. Heterogeneity was formally tested and explored performing an influence analysis.

Results

The search identified 1,277 references. Six full papers and 4 abstracts met the inclusion criteria. With MTX as reference standard, sensitivity (95% confidence interval, CI) was 0.80 (0.74,0.85), specificity 0.61 (0.56,0.67), positive likelihood ratio (LR) 2.11 (1.92,2.32), negative LR 0.31 (0.25,0.38) and the diagnostic odds ratio (DOR) was 6.74 (5.49,8.28). Using DMARDs as reference standard, sensitivity was 0.73 (0.64,0.80), specificity was 0.74 (0.68,0.79), LR+2.85 (2.53,3.22), LR− 0.35 (0.27,0.45) and DOR 8.03 (6.4,10.09). Using the combination of MTX and DMARDs as reference standard, intermediate results were obtained. The influence analysis detected one potentially influential study. However, its exclusion from the meta-analysis did not have a clinically relevant impact on the results.

Conclusions

The new classification criteria have good sensitivity, lower specificity and an overall moderate diagnostic accuracy. These results confirm that the criteria have classificative and not diagnostic function.

Introduction

In the last few decades, the recognition of the central role of an early diagnosis and the early administration of disease modifying antirheumatic drugs (DMARDs), particularly methotrexate (MTX), greatly improved the management of rheumatoid arthritis (RA) [1]. The more effective diagnostic and treatment strategies led to a better control of the disease with a deeper suppression of synovitis [2] and the prevention of radiological progression of bone erosions in the joints [3]. All these improvements have led to a great reduction of the risk of permanent disability [4], [5] which is the most significant long-term detrimental consequence of the disease.

For these reasons, the early recognition of RA has become a central issue in clinical practice, although the absence of a single and reliable test to identify the disease does not always allow an immediate diagnosis. The American College of Rheumatology (ACR) proposed in 1987 a set of classification criteria, developed in patients with longstanding disease with the aim to be specific rather than sensitive [6]. These criteria, that were initially meant for the enrolment of patients in clinical trials, but in some cases are used for diagnosis, have shown an unsatisfactory performance in the setting of early arthritis, especially due to a low sensitivity [7]. The inadequate performance of the 1987 criteria led in recent years to the development of prognostic algorithms, meant to discriminate, at the time of symptom onset, patients with higher probability of persistent disease amenable to treatment with DMARDs from those with self-limiting arthritis [8], [9].

In 2010 the ACR and the European League Against Rheumatism (EULAR) jointly developed new classification criteria, aiming to allow earlier patient classification, treatment and inclusion in clinical trials [10]. The new criteria were developed from inceptional cohorts of inflammatory arthritis data [11] which subsequently integrated with expert opinion [12] and finally validated in external early arthritis cohorts. For their development, the reference standard for diagnosis was the prescription of MTX or other DMARDs within the first year of observation. This was considered the best available reference standard, reflecting the risk of chronicity and erosive damage [10]. Since the criteria were developed using a reference standard that can be fulfilled after a follow-up, their value is not only classificative, but also prognostic. The opinion of an expert rheumatologist was not considered an acceptable reference standard because the widespread knowledge of the 1987 classification criteria would have likely lead to a circularity bias.

The 2010 criteria include tender and swollen joint count, acute phase reactants, anti-cyclic citrullinated peptide antibodies (ACPA) or rheumatoid factor (RF), and symptom duration (Table 1). These clinical and laboratory data are combined into a score ranging from 0 to 10. In the validation cohorts, the areas under the Receiving Operator Characteristics (ROC) curve (AUC) for the new criteria ranged from 0.66 to 0.82, and the score that determined an optimal discrimination was between 6 and 7. The cut-off of 6 was afterwards chosen in order to improve sensitivity. However, detailed information on the sensitivity and specificity of the new criteria, along with the corresponding likelihood ratios (LR), has not been reported in the original paper. Moreover, there was a certain variability in accuracy measures across different cohorts [13].

thumbnail
Table 1. 2010 ACR/EULAR classification criteria for rheumatoid arthritis.

https://doi.org/10.1371/journal.pone.0056528.t001

After their first presentation at the ACR congress in 2009, the new criteria have been tested in a number of external early arthritis cohorts with a wide variability in the overall performance [14][17]. These studies are slightly different in terms of population, but the main differences are in the assessment of the reference standard. In fact, many studies considered expert opinion as reference standard [15], as well as disease persistency, appearance of bone erosions or the 1987 classification criteria [17]. Only some of the studies considered the prescription of MTX or other DMARDs as reference standard for diagnosis as indicated in the newly developed criteria.

Because of the increasing knowledge on the 2010 RA classification criteria it seemed timely to summarise the results of the available literature. The present study aims to evaluate the performance of the 2010 ACR/EULAR classification criteria in populations of individuals with early arthritis. For this purpose, we performed a systematic literature review and subsequent meta-analyses in which we considered the 2010 classification criteria a s index test, the prescription of MTX or other DMARDs as reference standard for the classification of RA as done during the development of the criteria.

Methods

The Preferred Reporting Items for Systematic Reviews and Meta-analyses (PRISMA) guidelines for the reporting of systematic reviews and meta-analysis were followed to conduct this review [18]. A pre-specified protocol, including research question, search strategy, inclusion criteria for the articles and methods for the analysis, was developed before the beginning of the study.

Data Sources and Search

The search was performed by one of the authors (GS) and a control search by a second author (CAS). We searched Medline (PubMed), Embase and Cochrane databases from November 2009 (when the criteria were presented for the first time) to May 2012. The search strategies are shown in table 2. The search strategy was based on terms related to RA, classification and diagnosis. The references of the included studies were manually screened to search for further papers. The abstracts of the ACR (2010 and 2011) and EULAR congresses (2010, 2011, 2012) were examined to look for additional studies. No language or publication restrictions were applied and studies were not selected based on quality. Filters developed for the identification of diagnostic studies were not utilised since such filters may result in omission of relevant studies [19].

Study Selection

Studies should include subjects presenting with recent onset arthritis, with at least one swollen joint and no definite diagnosis that could explain symptoms (that is, the same population in which classification criteria should be applied) [10]. Finally, only data on patients with RA or undifferentiated polyarthritis (UPA) at baseline were included. The 2010 classification criteria were the index test, in the score format; the presence of bone erosions was not included since it was not tested in the data-driven phase of the development. Moreover, a definition of typical erosions had not yet been presented during the timespan that we examined. The prescription of MTX or DMARDs was considered as reference standard. In particular, we included studies that considered MTX (alone) as reference standard or overall DMARDs (including also MTX). Corticosteroids were not included.

Diagnostic cohort studies (prospective and retrospective) and case control studies were eligible for inclusion in the review. The presence of sufficient data to build a 2×2 table of diagnostic performance was required. Two reviewers independently screened titles and abstracts. The full-text of the potentially eligible articles was obtained; inclusion assessment was performed by one reviewer and checked by a second. Disagreements were resolved by consensus. Only the most recent and complete report was included in the case of studies reported in multiple publications or abstracts.

Data Extraction and Quality Assessment

We extracted data using a standardised form. Results were extracted as 2×2 tables. Studies were assessed for methodological quality by using the modified version of the Quality Assessment of Diagnostic Accuracy Studies (QUADAS) tool proposed by the Cochrane Collaboration [20]. Data extraction and quality assessment were done by one reviewer and checked by a second reviewer.

Data Synthesis and Analysis

Sensitivity and specificity were calculated for each 2×2 set of data. Heterogeneity among n included studies was visually evaluated plotting sensitivity and specificity on a ROC graph, and separately tested by the Chi-square test using n−1 degree of freedom [21].

In the presence of heterogeneity and negative correlation between sensitivity and specificity, as commonly seen in diagnostic studies, the bivariate model was used to estimate summary sensitivity and specificity with 95% confidence intervals (CI) and to derive a hierarchical summary receiving operator characteristic (HSROC) curve [22][24].Summary positive and negative likelihood ratios (LR+, LR−) and the diagnostic odds ratio (DOR) were derived for each analysis.

Three different analyses were performed. Primary analysis used MTX as reference standard for diagnosis. Secondary analysis used DMARDs as reference standard for diagnosis. Since MTX is a particular type of DMARDs, and in order to increase the number of available studies and hence the precision of the meta-analysis, a third analysis used MTX or DMARDs as reference standard was performed. For this analysis, if data on both MTX and DMARDs were reported separately, data on DMARDs were used.

To evaluate the robustness of our findings, we performed an influence analysis calculating Cook’s distance for each study. This index is a measure of the influence of a study on the model parameters and can be used to check for particularly influential study. Studies with Cook’s distance ≥1 were considered potentially influent. A further meta-analysis was then performed after excluding these potentially influential studies.

Publication bias was indirectly evaluating the symmetry on the funnel plot of logDOR. Forest plots were used to represent sensitivity and specificity of the primary studies.

Finally to verify the presence of publication bias, a funnel plot was implemented.

Review Manager (RevMan) version 5 (The Nordic Cochrane Centre, The Cochrane Collaboration, Copenhagen) was used to build the forest plot showing sensitivity and specificity of single studies and the risk of bias tables. Stata, version 11, (StataCorp, College Station, Texas) was used to perform all analyses. In particular the generalized linear mixed model approach to bivariate meta-analysis of sensitivity and specificity was obtained using the ‘metandi’ command of Stata.

Results

The search strategy identified 1277 studies. Figure 1 reports the flow-chart of the process of study selection. 6 articles were included after the evaluation of all titles and abstracts [14], [16], [25][28]. Moreover, 4 abstracts from the ACR and EULAR congresses were eligible for inclusion [29][32]. In total, 4 papers used MTX as unique reference standard, 3 used DMARDs as unique reference standard, while 3 studies reported separate data on both MTX and DMARDs. Two studies [15], [33] were excluded since they used a composite reference standard based on concomitant expert opinion and DMARDs and expert opinion was not considered an acceptable reference standard for this review. Table 3 and Figure 2 summarise the main features of the included studies. Figure 2 shows the point estimates with 95% CIs of sensitivity and specificity for the included studies. Sensitivity ranged from 0.68 to 0.88 when MTX was considered as reference standard, and from 0.59 to 0.87 when DMARDs were used. The specificity ranged from 0.50 to 0.72 with MTX as reference standard, and from 0.64 to 0.88 with DMARDs.

thumbnail
Figure 1. PRISMA flow-chart describing the selection process in the systematic literature review.

https://doi.org/10.1371/journal.pone.0056528.g001

thumbnail
Figure 2. Sensitivity and specificity of the included studies. A)

methotrexate as reference standard B) disease modifying antirheumatic drugs as reference standard C) methotrexate+ disease modifying antirheumatic drugs as reference standard.

https://doi.org/10.1371/journal.pone.0056528.g002

The evaluation of the methodological quality of the included studies is shown in Figure 3. All studies had a low risk of bias for most of the items except for blinding that was not explicitly mentioned in any of the studies. For this reason, the blinding of the results of the reference standard and the index test was considered unclear. All the included studies had an overall low risk of bias; for this reason, additional analyses, excluding those for low-quality studies, were not performed, since the results of the meta-analysis would not be affected by a low methodological quality.

thumbnail
Figure 3. Assessment of the risk of bias in the included studies.

3A) Methodological quality graph: authors’ judgment about each methodological quality item presented as percentages across all included studies. 3B) Methodological quality summary: authors’ judgment about each methodological quality item for each included study. +: low risk of bias; ?: unclear risk of bias; -: high risk of bias.

https://doi.org/10.1371/journal.pone.0056528.g003

Both sensitivity and specificity showed significant heterogeneity, with a Chi-square for differences across studies of 29.01 (p<0.0001) and 17.09 (p = 0.009) respectively.

Table 4 summarises the results of all the meta-analyses. Using MTX as reference standard, pooled sensitivity (95% CI) was 0.80 (0.74, 0.85) and pooled specificity (95% CI) was 0.61 (0.56, 0.67), a LR+ of 2.11 (1.92, 2.32) and a LR− of 0.31 (0.25, 0.38).

Using DMARDs as reference standard, sensitivity was 0.73 (0.64,0.80), specificity was 0.74 (0.68,0.79), LR+ was 2.85 (2.53,3.22) and LR− was 0.35 (0.27,0.45).

The third analysis, combining MTX and DMARDs, led to intermediate results (Figure 4).

thumbnail
Figure 4. Results of the meta-analyses. A)

methotrexate as reference standard; 7 studies included, 3845 participants. B) disease modifying antirheumatic drugs s as reference standard; 6 studies included, 3018 participants. C) methotrexate+disease modifying antirheumatic drugs as reference standard; 10 studies included, 4134 participants. The black square indicates the point estimate of sensitivity and specificity for the meta-analysis, the dotted line indicates the 95% confidence interval. The continuous line is the hierarchical summary receiving operator characteristics curve. The dots represent the primary studies.

https://doi.org/10.1371/journal.pone.0056528.g004

The influence analysis using the Cook’s distance identified Van Der Linden’s study as a potentially influential [27]. After removing this study the pooled sensitivity changed from 0.80 (0.74, 0.85) to 0.79 (0.72, 0.84), and specificity from 0.61 (0.56, 0.67) to 0.60 (0.53, 0.68). The influence of each single study is reported in Figure 5.

thumbnail
Figure 5. Results of meta-analyses excluding single studies.

Overall meta-analysis, based on the primary analysis using methotrexate as reference standard, and meta-analysis excluding a single study are reported. The table and the legend refer to the study that has been excluded.

https://doi.org/10.1371/journal.pone.0056528.g005

Funnel plot did not show any systematic asymmetry suggesting the absence of publication bias (Figure 6).

thumbnail
Figure 6. Funnel plot.

The distribution of the studies in the funnel plot does not suggest the presence of publication bias. In fact, studies are distributed by each side of the plot, with moreover a lower number of studies with positive results. In the case of publication bias, the opposite situation would have been expected.

https://doi.org/10.1371/journal.pone.0056528.g006

Discussion

Since the new classification criteria for RA were first presented, several studies have provided a wide range of results by evaluating their performance in the setting of early arthritis. These studies were different in terms of population (mainly due to the recruitment criteria and calendar year into the existing cohorts) and reference standard. Moreover, treatment patterns differed significantly among them. This variability did not allow the drawing of consistent conclusions from a single study. Therefore it seemed appropriate to review the available literature on this topic.

We adopted a search strategy meant to be sensitive rather than specific; the limited time period under exam allowed this approach. We did not use pre-specified filters for diagnostic accuracy studies [19]. Since the search was designed to be as broad as possible, it is unlikely that relevant studies were missed. Even in this way, the number of studies that were eligible for the inclusion was limited. This is the main limitation of this review, since it did not allow to investigate the causes of heterogeneity in detail and, for the same reason, subgroup analyses were not feasible.

The majority of the validation studies did not use the same reference standard highlighted in the development of the criteria and many of them used expert opinion or the 1987 ACR criteria. In particular, 13 studies that were selected for detailed review have afterwards been excluded because of a wrong reference standard. These studies were excluded because of the possibility of the introduction of considerable circularity bias. Same reference standard used to develop the new criteria were adopted, that is the prescription of MTX or other DMARDs [10]. The results obtained showed an overall moderate performance of the criteria with acceptable values of sensitivity through all reference standards. One the other hand, specificity was lower in the analyses based on MTX, and MTX with DMARDs. The analysis based on overall DMARDs as reference standard showed a better specificity of the 2010 classification criteria. In fact, MTX might not have been given to all patients that would require it due to contraindications or national differences in prescriptions. On the other hand sensitivity does not increase accordingly because a twofold higher increase of subjects treated with DMARDs not fulfilling the 2010 criteria is observed.

The values of LR+ were around 2.5, far from the values of LR+ >10 that ideally identify a test having the power to detect the disease. However, LR− were around 0.3, which is closer to the optimal values of LR− <0.1 [34]. This indicates the patients not fulfilling the 2010 classification criteria have high probability of not having RA (and will not develop it later, if we consider that the criteria identify patients that might fulfil an outcome after a follow-up), while patients fulfilling them will not certainly develop RA. In contrast with this consideration, the DORs suggest a good performance, though this is probably driven by low values of LR−.

The need to perform differential diagnosis before the application of the criteria might have determined the underestimation of their accuracy even if the majority of studies excluded patients that were afterwards diagnosed with disease other than RA or UPA. Differential diagnosis is not always feasible in the early stages; this might have increased the number of false positives resulting in the reduction of specificity.

In the original presentation paper, detailed data on sensitivity, specificity and performance were not reported. The values of the areas under the ROC curve (AUC) reported in the validation cohorts ranged from 0.66 to 0.82. The best performance was achieved in the Norwegian cohort of early arthritis, while in the remaining two the AUCs were <0.7; this indicates a test with moderate performance [10]. The differences between the populations could be due to differences in recruitment and treatment. In the external validation cohorts, only the proportion of patients correctly classified was reported, while overall estimates of accuracy were not presented. Since not all information on the performance of the 2010 criteria is provided in the original paper, results could not be directly compared our with data from the inception of the criteria.

As expected, significant heterogeneity was found among diagnostic studies. This finding did not permit the separate combination of sensitivity and specificity. For this reason we adopted a bivariate model that allows for the negative correlation between sensitivity and specificity. Given the low number of studies, investigating heterogeneity by stratified meta-analysis or meta-regression was not feasible. For this reason we evaluated the robustness of our results by performing an influence analysis [21]. However, though one study was demonstrated to be potentially influential, a subsequent meta-analysis after the exclusion of the influential study led to comparable results.

All the included studies had an overall good methodological quality. The items dealing with blinding were not explicitly addressed by the papers, therefore the risk of bias related to these points was considered unclear for all studies. Despite this, given to the retrospective design and the unavailability of the criteria at the time of the clinical assessment and therapeutic decision, this did not clearly biased the results. The last two items (on the report of uninterpretable results and withdrawals), even though not directly reported, were considered satisfactory, since, as suggested by the Cochrane collaboration, they probably had no influence on the results [20]. Since the methodological quality of all the included studies was good, the results of the meta-analysis have likely not been affected by the methodological quality of the primary studies.

A meta-analysis on the diagnostic performance of the 1987 ACR classification criteria demonstrated an overall sensitivity and specificity of 0.77 in the subgroup of patients with early arthritis and a better performance among cases of established arthritis [7]. These results are not substantially different from those of our meta-analyses: the new criteria, specifically designed for early arthritis, do not perform better than the old ones in the same setting. However, it has to be kept in mind that the 1987 criteria were developed based on expert opinion while the 2010 criteria used a more practical reference standard, such as the use of specific treatment regimes. Therefore, the 2010 criteria could be more useful in a clinical setting because they classify patients based on a relevant prognostic aspect of the disease with an impact on clinical management. Nevertheless, the function of the criteria is classification, and not diagnosis; based on current knowledge the 2010 criteria should not be used as a guide to start specific treatments for RA.

The external validation studies in the meta-analysis on 1987 criteria had a cross-sectional case-control design that overestimates the accuracy while the cohort study design (such as in the development of the criteria) tends to underestimate accuracy [35], and this might have influenced our results as well.

One of the merits of the 2010 classification criteria is the fact that it is not necessary to reach six weeks of symptom duration to classify a patient as having RA. Only one study [14] applied the new criteria in a population of very early arthritis, but their performance did not seem substantially different that in the remaining early arthritis population.

The creation of an optimal tool for the classification of rheumatic disease, with RA in particular, has always been challenging. This is mainly due to the absence of gold standards for diagnosis. The decision to prescribe either DMARDs or MTX is dependent from expert opinion and reflects a judgement on the severity of arthritis, for this reason their performance does not seem to be superior to the old ones. However developing the criteria avoiding to use expert opinion as reference standard gives an opportunity to innovate clinical practice, and we could expect that the presentation of the new classification criteria might significantly modify the approach to RA. In fact, the results of the present meta-analysis evaluate the 2010 criteria in the setting of the current management of RA by rheumatologists, but the final judgement on the criteria will probably be drawn after prospective studies, evaluating their long-term impact on disease outcomes.

Author Contributions

Conceived and designed the experiments: CAS CM RC GS. Performed the experiments: GS CAS. Analyzed the data: CAS AZ. Contributed reagents/materials/analysis tools: CAS AZ RC CM. Wrote the paper: GS CAS AZ RC CM.

References

  1. 1. McInnes IB, O’Dell JR (2010) State-of-the-art: rheumatoid arthritis. Ann Rheum Dis 69: 1898–1906.
  2. 2. Sakellariou G, Scirè CA, Montecucco C, Caporali R (2013) In patients with early rheumatoid arthritis, the new ACR/EULAR definition of in clinical remission identifies patients with persistent absence of functional disability and suppression of ultrasonographic synovitis. Ann Rheum Dis 72: 245–249.
  3. 3. Breedveld FC, Combe B (2011) Understanding emerging treatment paradigms in rheumatoid arthritis. Arthritis Res Ther 13: S3.
  4. 4. Scirè CA, Verstappen SM, Mirjafari H, Bunn DK, Lunt M, et al. (2011) Reduction of functional disability in inflammatory polyarthritis by early and persistent suppression of joint inflammation: results from the Norfolk Arthritis register. Arthritis Care Res (Hoboken) 63: 945–952.
  5. 5. Knevel R, Schoels M, Huizinga TW, Aletaha D, Burmester GR, et al. (2010) Current evidence for a strategic approach to the management of rheumatoid arthritis with disease-modifying antirheumatic drugs: a systematic literature review informing the EULAR recommendations for the management of rheumatoid arthritis. Ann Rheum Dis 69: 987–994.
  6. 6. Arnett FC, Edworthy SM, Bloch DA, McShane DJ, Fries JF, et al. (1988) The America Rheumatism Association 1987 revised criteria for the classification of rheumatoid arthritis. Arthritis Rheum 31: 315–324.
  7. 7. Banal F, Dougados M, Combescure C, Gossec L (2009) Sensitivity and specificity of the American College of Rheumatology 1987 criteria for the diagnosis of rheumatoid arthritis according to disease duration: a systematic literature review and meta-analysis. Ann Rheum Dis 68: 1184–1191.
  8. 8. Visser H, le Cessie S, Vos K, Breeedveld FC, Hazes JM (2004) How to diagnose rheumatoid arthritis early: a prediction model for persistent (erosive) arthritis. Arthritis Rheum 46: 357–365.
  9. 9. van der Helm-van Mil AH, le Cessie S, van Dongen H, Breedveld FC, Toes RE (2007) A Prediction Rule for Disease Outcome in Patients With Recent-Onset Undifferentiated Arthritis. How to Guide Individual Treatment Decisions. Arthritis Rheum 56: 433–440.
  10. 10. Aletaha D, Neogi T, SIlman A, Funovits J, Felson DT, et al. (2010) 2010 rheumatoid arthritis classification criteria: an America College of Rhuematology/European League Against Rheumatism collaborative initiative. Ann Rheum Dis 69: 1580–1588.
  11. 11. Funovits J, Aletaha D, Bykerk V, Combe B, Dougados M, et al. (2010) The 2010 American College of Rheumatology/European league Against Rheumatism classification criteria for rheumatoid arthritis: methodological report phase I. Ann Rheum Dis. 69: 1589–1595.
  12. 12. Neogi T, Aletaha D, Silman AJ, Naden RL, Felson DT, et al. (2010) The 2010 American College of Rheumatology/European League Against Rheumatism classification criteria for rheumatoid arthritis: Phase 2 methodological report. Arthritis Rheum 62: 2582–2591.
  13. 13. Corrao S, Calvo L, Licata G (2011) The new criteria for classification of rheumatoid arthritis: what we need to know for clinical practice. Eur J Intern Med 22: 217–219.
  14. 14. Cader MZ, Filer A, Hazlehurst J, de Pablo P, Buckley CD, et al. (2011) Performance of the 2010 ACR/EULAR criteria for rheumatoid arthritis: comparison with the 1987 ACR criteria in a very early synovitis cohort. Ann Rheum Dis 70: 949–955.
  15. 15. Varache S, Cornec D, Morvan J, Devauchelle-Pensec V, Berthelot J, et al. (2011) Diagnostic accuracy of ACR/EULAR 2010 criteria for rheumatoid arthritis in a 2-year cohort. J Rheumatol 2011 38: 1250–1257.
  16. 16. Kaneko y, Kuwana M, Kamed H, Takeuchi T (2011) Sensitivity and specificity of 2010 rheumatoid arthritis classification criteria. Rheumatology (Oxford) 50: 1268–1274.
  17. 17. Cornec D, Varache S, Morvan J, Devauchelle-Pensec V, Berthelot JM, et al.. (2012) Comparison of the 1987 ACR criteria for predicting a 10-year diagnosis of rheumatoid arthritis. Joint Bone Spine doi:10.1016/j.jbspin.2012.01.015.
  18. 18. Moher D, Liberati A, Tetzlaff J, Altman DG (2009) the PRISMA group (2009) Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. J Clin Epidemiol 62: 1006–1012.
  19. 19. Leeflang MM, Scholten RJ, Rutjes AW, Reitsma JB, Bossuyt PM (2006) Use of methodological search filters to identify diagnostic accuracy studies can lead to the omission of relevant studies. J Clin Epidemiol 59: 234–240.
  20. 20. Reitsma JB, Rutjes AWS, Whiting P, Vlassov VV, Leeflang MMG, et al.. (2009) Chapter 9: Assessing methodological quality. In: Deeks JJ, Bossuyt PM, Gatsonis C (editors), Cochrane Handbook for Systematic Reviews of Diagnostic Test Accuracy Version 1.0.0. The Cochrane Collaboration. Available from: http://srdta.cochrane.org/.
  21. 21. Dinnes J, Deeks J, Kirby J, Roderick P (2005) A methodological review of how heterogeneity has been examined in systematic reviews of diagnostic test accuracy. Health Technology Assess 9: 1–113.
  22. 22. Harbord RM, Deeks JJ, Egger M, Whiting P, Sterne JA (2007) A unification of models for meta-analysis of diagnostic accuracy studies. Biostatistics 8: 239–251.
  23. 23. Harbord RM, Whiting P (2009) metandi: Meta-analysis of diagnostic accuracy using hierarchical logistic regression. Stata J 9: 211–229.
  24. 24. Reitsma H (2011) Meta-analysis of diagnostic accuracy data. Diagnostic test accuracy author training, Amsterdam, 2011. http://srdta.cochrane.org/sites/srdta.cochrane.org/files/uploads/10_Stat_Methods_2011.pdf.
  25. 25. Alves C, Luime JJ, van Zeben D, Huisman A, Weel AEAM, et al. (2011) Diagnostic performance of the ACR/EULAR 2010 criteria for rheumatoid arthritis and two diagnostic algorithms in an early arthritis clinic (REACH). Ann Rheum Dis 70: 1645–1647.
  26. 26. Britsemmer K, Ursum J, Gerritsen M, van Tuyl L, van Schaardenburg D (2011) Validation of the 2010 ACR/EULAR classification criteria for rheumatoid arthritis: slight improvement over the 1987 ACR criteria. Ann Rheum Dis 70: 1468–1470.
  27. 27. Kawashiri SY, Suzuki T, Okada A, Yamasaki S, Tamai M, et al. (2013) Musculoskeletal ultrasonography assists the diagnostic performance of the 2010 classification criteria for rheumatoid arthritis. Mod Rheumatol 23: 36–43.
  28. 28. van der Linden MM, Knevel R, Huizinga TWJ, van der Helm-van Mil AHM (2011) Classification of rheumatoid arthritis. Comparison of the 1987 America College of Rheumatology criteria and the 2010 American College of Rheumatology/American League Against Rheumatism Criteria. Arthritis Rheum 63: 37–42.
  29. 29. Berglin EH, Rantapaa-Dahlqvist SM (2011) Comparison of the 1987 ACR and 2010 ACR/EULAR Classification Criteria for Rheumatoid Arthritis in Clinical Practice. Arthritis Rheum 63 (S10): S117.
  30. 30. Nakagomi D, Ikeda K, Okubo A, Iwamoto T, Suzuki Y, et al. (2012) Ultrasonographic assessment of synovitis improves the accuracy of 2010 America College of Rheumatology/American League against rheumatism classification criteria for rheumatoid arthritis to predict development of a methotrexate-requiring disease. Ann Rheum Dis 71 (S3): 98.
  31. 31. Sakellariou G, Scirè CA, Montecucco C, Caporali R (2012) Diagnostic performance of the 2010 ACR/EULAR classification criteria for rheumatoid arthritis in a cohort of early arthritis. Ann Rheum Dis 71 (S3): 334.
  32. 32. Tamai M, Arima K, Uetani M, Iwamoto N, Kita J, et al. (2011) Evaluation of the Patients with Early Arthritis by 2010 RA Criteria in Conjunction with MRI of Wrists and Finger Joints. Arthritis Rheum 63 (S10): 318.
  33. 33. Chitale S, Goodson NJ, Sharpley DW, Thompson RL, Moots RJ, et al. (2010) New 2010 ACR-EULAR rheumatoid arthritis classification criteria are highly sensitive in patients with early rheumatoid arthritis. Arthritis Rheum 62 (S10): S733.
  34. 34. Jaeschke R, Guyatt G, Lijmer J (2002) Diagnostic tests. In: Guyatt G, Rennie D editors. User’s guides to the medical literature. Chicago: AMA press, 121–140.
  35. 35. Lijmer JG, Mol BW, Heisterkamp S, Bonsel GJ, Prins MH, et al. (1999) Empirical evidence of design-related bias in studies of diagnostic tests. JAMA 282: 1061–1066.