Skip to main content

Main menu

  • HOME
  • ONLINE FIRST
  • CURRENT ISSUE
  • ALL ISSUES
  • AUTHORS & REVIEWERS
  • SUBSCRIBE
  • BJGP LIFE
  • MORE
    • About BJGP
    • Conference
    • Advertising
    • eLetters
    • Alerts
    • Video
    • Audio
    • Librarian information
    • Resilience
    • COVID-19 Clinical Solutions
  • RCGP
    • BJGP for RCGP members
    • BJGP Open
    • RCGP eLearning
    • InnovAIT Journal
    • Jobs and careers

User menu

  • Subscriptions
  • Alerts
  • Log in

Search

  • Advanced search
British Journal of General Practice
Intended for Healthcare Professionals
  • RCGP
    • BJGP for RCGP members
    • BJGP Open
    • RCGP eLearning
    • InnovAIT Journal
    • Jobs and careers
  • Subscriptions
  • Alerts
  • Log in
  • Follow bjgp on Twitter
  • Visit bjgp on Facebook
  • Blog
  • Listen to BJGP podcast
  • Subscribe BJGP on YouTube
British Journal of General Practice
Intended for Healthcare Professionals

Advanced Search

  • HOME
  • ONLINE FIRST
  • CURRENT ISSUE
  • ALL ISSUES
  • AUTHORS & REVIEWERS
  • SUBSCRIBE
  • BJGP LIFE
  • MORE
    • About BJGP
    • Conference
    • Advertising
    • eLetters
    • Alerts
    • Video
    • Audio
    • Librarian information
    • Resilience
    • COVID-19 Clinical Solutions
Research

Comparison of Centor and McIsaac scores in primary care: a meta-analysis over multiple thresholds

Brian H Willis, Dyuti Coomar and Mohammed Baragilly
British Journal of General Practice 2020; 70 (693): e245-e254. DOI: https://doi.org/10.3399/bjgp20X708833
Brian H Willis
Institute of Applied Health Research, University of Birmingham, Birmingham.
Roles: Medical Research Council clinician scientist
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Dyuti Coomar
Institute of Applied Health Research, University of Birmingham, Birmingham.
Roles: Research fellow
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
Mohammed Baragilly
Roles: Research fellow
  • Find this author on Google Scholar
  • Find this author on PubMed
  • Search for this author on this site
  • Article
  • Figures & Data
  • Info
  • eLetters
  • PDF
Loading

This article has a correction. Please see:

  • Correction - May 01, 2020

Abstract

Background Centor and McIsaac scores are both used to diagnose group A beta-haemolytic streptococcus (GABHS) infection, but have not been compared through meta-analysis.

Aim To compare the performance of Centor and McIsaac scores at diagnosing patients with GABHS presenting to primary care with pharyngitis.

Design and setting A meta-analysis of diagnostic test accuracy studies conducted in primary care was performed using a novel model that incorporates data at multiple thresholds.

Method MEDLINE, EMBASE, and PsycINFO were searched for studies published between January 1980 and February 2019. Included studies were: cross-sectional; recruited patients with sore throats from primary care; used the Centor or McIsaac score; had GABHS infection as the target diagnosis; used throat swab culture as the reference standard; and reported 2 × 2 tables across multiple thresholds. Selection and data extraction were conducted by two independent reviewers. QUADAS-2 was used to assess study quality. Summary receiver operating characteristic (SROC) curves were synthesised. Calibration curves were used to assess the transferability of results into practice.

Results Ten studies using the Centor score and eight using the McIsaac score were included. The prevalence of GABHS ranged between 4% and 44%. The areas under the SROC curves for McIsaac and Centor scores were 0.7052 and 0.6888, respectively. The P-value for the difference (0.0164) was 0.419, suggesting the SROC curves for the tests are equivalent. Both scores demonstrated poor calibration.

Conclusion Both Centor and McIsaac scores provide only fair discrimination of those with and without GABHS, and appear broadly equivalent in performance. The poor calibration for a positive test result suggests other point-of-care tests are required to rule in GABHS; however, with both Centor and McIsaac scores, a score of ≤0 may be sufficient to rule out infection.

  • Centor score
  • diagnosis
  • McIsaac score
  • meta-analysis
  • pharyngitis
  • primary health care

INTRODUCTION

Pharyngitis is one of the most common reasons for consulting a GP. Over the winter period, around 6% of GP consultations in the UK tend to be for patients presenting with a sore throat, which represents more than 3.5 million consultations.1 Although, in many cases, pharyngitis has a viral aetiology, 20%–35% of cases may be caused by bacteria — specifically, group A beta-haemolytic streptococcus (GABHS).2,3 Worldwide, infection with group A streptococci (GAS) places a significant burden on global health, and around 500 million people will die from GAS-related diseases each year.4

In order to stratify patients most at risk of GABHS, the Centor score was developed. Each of four clinical features — absence of cough, purulent pharyngeal exudate, anterior cervical lymphadenopathy, and temperature of >38°C — is scored with 1 or 0, depending on whether it is present;5 scores range from 0 (when none of the features are present) to 4 (when all are present). In the original study, conducted in an emergency department in the US, a score of 3 was associated with a 30.1%–34.1% probability of GABHS.5 McIsaac independently derived a prediction system based on a cohort of patients from primary care.6 In essence, it modifies the Centor system to include an extra variable — age. For those aged between 3 years and 14 years, 1 is added to the score, whereas, for those aged ≥45 years, 1 is subtracted from the score; hence, a patient presenting with a sore throat may have a McIsaac score of anything between −1 and 5.6

Many health systems have recommended the use of Centor or McIsaac scores in their guidelines to help manage patients with acute pharyngitis.7–10 In the UK, the Centor score is one of two prediction rules recommended by the National Institute for Health and Care Excellence (NICE).10 Although the extent to which these rules are used in UK general practice is unclear, a recent survey of 266 GPs in Denmark reported that approximately half used the Centor score and 15% used the McIsaac score — this was in spite of the fact that the McIsaac score is the recommended rule in Denmark for diagnosing GABHS.9

The question of which rule is likely to yield the most accurate diagnosis of GABHS for patients presenting to general practice is difficult to answer based on existing research. Only one primary study to date — reported in two articles by Fine et al 11,12 — provides the data to allow a direct comparison. Furthermore, comparisons at individual thresholds are meaningless, as those thresholds are not equivalent to each other — for example, a Centor score of 3 is not equivalent to a McIsaac score of 3, as the latter is calculated with an extra variable (that of age). To compare the tests, an overall assessment across all thresholds is required, such as may be provided by a receiver operating characteristic (ROC) curve.

Although meta-analysis allows the aggregation of multiple studies, either to produce a summary (sensitivity, false positive rate) point or a summary ROC curve,13–15 both of these methods are constrained by the inclusion of only one (sensitivity, false positive rate) data point per study, where the false positive rate = 1 − specificity. When a study reports data at multiple thresholds, an arbitrary choice has to be made on which threshold to use when extracting the data for meta-analysis. Recent developments in meta-analysis methods allow this constraint to be relaxed so, if individual studies provide data at multiple thresholds, all of the data may be included for analysis;16 as such, the unit of interest for each study becomes its ROC curve and not just an individual (sensitivity, false positive rate) pair. This provides the basis for generating a summary ROC (SROC) curve for the Centor and McIsaac scores based on all the data reported in the primary studies.

In many healthcare systems, the Centor score and McIsaac score are used by GPs and primary care professionals to diagnose group A beta-haemolytic streptococcus (GABHS); however, there is no previous meta-analysis that has compared their performances in primary care. This comparative meta-analysis demonstrates that the Centor score and McIsaac score have broadly similar performance characteristics in diagnosing GABHS infection in primary care. A score of ≤0 when using either system may have a role in ruling out GABHS infection in primary care; however, neither score is sufficiently accurate to rule in GABHS infection and, if applied as recommended, could lead to more than one in two patients being prescribed antibiotics inappropriately. Other point-of-care diagnostics that augment these scores are needed if rates of inappropriate antibiotic prescribing are to be reduced.

How this fits in

This study aimed to compare the performance of Centor and McIsaac scores in diagnosing patients with GABHS presenting to primary care with a sore throat.

METHOD

Data sources and searches

MEDLINE, EMBASE, and PsycINFO were searched for relevant studies; the search terms used are given in Supplementary Box S1. The data were supplemented by a manual review of the references from two published meta-analyses — one by Aalbers et al,17 the other by Willis and Hyde.18 The grey literature was not specifically searched because of a lack of evidence supporting its use in test accuracy reviews;19,20 however, for completeness, a Google Scholar search was also performed using the terms ‘McIsaac score’ and ‘Centor score’. The searches were limited to studies published between January 1980 and February 2019. No restrictions were placed on the language of publication. Duplicate references were discarded to get a cohesive set of studies ready to be reviewed for inclusion.

Study eligibility criteria

Studies were included if:

  • the study was a cross-sectional primary study;

  • the study population consisted of unselected patients presenting with a sore throat to primary care;

  • the study evaluated at least one of Centor or McIsaac scores;

  • the target diagnosis was GABHS;

  • the reference standard was culture from a throat swab; and

  • sufficient data were reported to complete the 2 × 2 table for as many thresholds as possible.

Two researchers independently screened the title and abstracts of all citations identified. Full texts were obtained for those articles not excluded at the screening stage, and the same two investigators independently assessed the studies for eligibility based on the above criteria. Disagreements were resolved through discussion and achieving consensus.

Data collection and quality assessment

Data were extracted on the following study characteristics:

  • aim;

  • test evaluated;

  • start and end date;

  • method of subject recruitment;

  • study location;

  • description of study population;

  • sample size;

  • reference standard;

  • conclusion of study authors;

  • 2 × 2 contingency table data (true positives, false positives, true negatives, and false negatives) for each reported threshold on a per-patient basis; and

  • any conflicts of interest.

The quality of each included study was assessed using the Quality Assessment of Diagnostic Accuracy Studies (QUADAS-2) tool,21 which assesses the risk of bias across a number of domains. The category ‘unclear’ was used when there was insufficient information reported in the study to come to a clear decision even after discussion.

The same two researchers who screened the initial abstracts independently extracted data, and performed the appraisal and quality assessment of each study. Disagreements were resolved through discussion and achieving consensus.

Synthesis and meta-analysis methods

The Different random Intercept Different random Slope (DIDS) model from the R package diagmeta (https://CRAN.R-project.org/package=diagmeta) was used to fit the data from the primary studies. This fits two linear mixed models — one for the false negative rate and one for the specificity — using the study as the grouping factor and allowing data from multiple thresholds for each study. Each linear mixed model has a random intercept and random gradient term, and the four random effects are assumed to have a four-dimensional multivariate normal distribution;16 these are used to generate an SROC curve.

An SROC curve and C-statistic (area under the curve [AUC]) was generated for Centor and McIsaac scores. Positive and negative likelihood ratios were derived for each of the thresholds with bootstrap confidence intervals (CIs). Assuming a null hypothesis that there is no difference between the C-statistics, the null distribution was derived empirically using a bootstrap sample of 1000. The level of significance was set to 0.05. For each test, the summary (sensitivity, false positive rate) pair corresponding to each threshold was also derived. Calibration plots of expected probabilities versus observed probabilities were derived for positive and negative test results after fitting an additive model to the logits of these probabilities using cubic splines.22 Each plot was corrected for optimism using a bootstrap sample of 1000 as recommended by Harrell.23

RESULTS

Study selection

The searches identified 80 citations. The full selection process (outlined in Figure 1) resulted in 18 studies2,6,11,12,24–37 being included in the review; 10 of these used the Centor score2,11,12,24–31 and eight used the McIsaac score.6,11,12,32–37 Only one study — reported by Fine et al 11,12 — provided sufficient data to allow a direct comparison between the two tests.

A flowchart of the primary studies’ selection decisions is given in Figure 1.

Figure 1.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 1.

Flow chart of studies selected.

Study characteristics

Full study characteristics are detailed in Table 1. Of those studies using the Centor score, eight were conducted in Europe2,24–27,29–31 and two in the US.11,12,28 Of those studies using the McIsaac score, three were conducted in Europe,32,33,36 four in North America,6,11,12,35,37 and one in Australia.34 Three studies were translated from Spanish.26,31,36 The only study to provide data on both the Centor and McIsaac scores (Fine et al 11,12) had a sample size that was more than 100 times larger than the next-largest study.

View this table:
  • View inline
  • View popup
Table 1.

Study characteristics

The median prevalence of GABHS for the studies using the Centor score was 26.4% (range: 4.7%–42.0%); for studies using the McIsaac score, it was 23.0% (range: 12.7%–44.8%). Exactly half of the studies using the Centor score provided data on all thresholds, and all studies provided data for two or more thresholds. A quarter of studies using the McIsaac score provided data on all thresholds, and all studies provided data for ≥4 thresholds. The ROC curves for each of the studies using the Centor score are shown in Figure 2; those for each of the studies using the McIsaac score are given in Figure 3.

Figure 2.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 2.

Centor score: ROC curves for each of the included studies. Each line corresponds to a single study and each dot corresponds to the (sensitivity, 1 − specificity) pair at a particular threshold for that study. ROC = receiver operating characteristic.

Figure 3.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 3.

McIsaac score: ROC curves for each of the included studies. Each line corresponds to a single study and each triangle corresponds to the (sensitivity, 1 − specificity) pair at a particular threshold for that study. ROC = receiver operating characteristic.

For two of the included studies using the McIsaac score, McIsaac was listed as the lead author.6,37

Risk of bias and applicability

There is no validated statistic for measuring between-study heterogeneity across ROC curves; however, Figure 2 and Figure 3 show that, for both tests, the ROC curves are widely distributed; this suggests there is heterogeneity between studies for both tests.

For many of the studies,2,6,25–28,30–35,37 the reporting was inadequate, which introduced uncertainty when assessing the risk of bias — for example, the method of patient selection was not always described and it was not always clear whether any subjects had been excluded. Often, it was not reported whether the reference standard was carried out blind to the test results, although it is unclear whether knowledge of the test results would have greatly influenced the results of a cultured throat swab. In general, the study populations were considered representative of those seen in the different forms of primary care.

In two studies31,37 there were discrepancies between the number of subjects recruited and the number used in analyses, thereby increasing risk of biased estimates for the statistics of interest. In addition, although in one study that used the Centor score the target condition was largely GABHS, it also included group C and group G streptococcal infection as the target condition.25 This could potentially affect the applicability of the findings of this study. QUADAS-2 results are given in Supplementary Figure S1 (McIsaac score) and Supplementary Figure S2 (Centor score).

Synthesis of results

The sensitivities, specificities, and positive and negative likelihood ratios for each threshold are given for both scores in Table 2. Figure 4 shows the SROC curves for the Centor and McIsaac scores, with points on each curve corresponding to particular thresholds; it is clear that the curves are very close to each other and this is confirmed by the C-statistic. For the Centor score, the C-statistic was 0.6888 (95% CI = 0.653 to 0.724) and for McIsaac’s score it was 0.7052 (95% CI = 0.624 to 0.778); the 95% CIs are for the sensitivity given the specificity. From the empirical distribution of the difference between C-statistics, a difference of 0.0164 has a corresponding P-value of 0.419; this suggests there is no statistically significant difference between the C-statistics for the two curves.

View this table:
  • View inline
  • View popup
Table 2.

Sensitivity and specificity of the Centor and McIsaac scores at different thresholds

Figure 4.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 4.

SROC curves for the McIsaac and Centor scores. SROC = summary receiver operating characteristic.

Two post-hoc sensitivity analyses were carried out. The first investigated the effect of excluding the largest study (that by Fine et al 11,12 and resulted in the C-statistics for the Centor and McIsaac scores being 0.6724 (95% CI = 0.610 to 0.731) and 0.7167 (95% CI = 0.632 to 0.788), respectively. As such, the effect is to decrease the C-statistic for the Centor score and to increase it for the McIsaac score. Again, the difference (0.0443) was not statistically significant (P = 0.188). In the second analysis, it was noted that two of the eight included studies that used the McIsaac score were led by the researcher who proposed it (namely, McIsaac);6,37 as such, only six studies evaluated the score independently. A sensitivity analysis was conducted in which the two studies led by McIsaac were excluded to evaluate the overall effects on the C-statistic. The C-statistic for the six independent studies was 0.6700 — lower than that when all studies were included in the analysis (0.7052) and that for the Centor score (0.6888).

The calibration plot for the post-test probabilities after a positive test result (positive predictive value [PPV]) for both scores, after correcting for optimism, is shown in Figure 4. The curves broadly coincide, with overfitting being particularly evident for expected PPVs above 0.5. Supplementary Figure S3 shows the calibration plot for the post-test probabilities for a negative test result after correcting for optimism. Here, the Centor score demonstrates better calibration than the McIssac score. For the derivation of both calibration plots, the prevalence of GABHS is assumed to be known.

Whether either test could be used to rule in, or rule out, infection is not fully addressed by the AUC. For a GABHS infection prevalence of 25%, using Bayes’ theorem the expected PPV for a McIsaac score of 5 is 59%; however, from the calibration curve this expected PPV is likely to translate into an observed PPV of around 49% (Figure 5). Thus, if a score of 5 is used as the threshold for prescribing antibiotics, a PPV of 49% translates into more than one in two patients receiving antibiotics unnecessarily. Although the expected PPV would increase with GABHS prevalence, the calibration curves show this would not substantially affect the observed PPV; as such, neither test is effective at ruling in GABHS.

Figure 5.
  • Download figure
  • Open in new tab
  • Download powerpoint
Figure 5.

Calibration plots (corrected for optimism) for the Centor and McIsaac scores for a positive test result when the prevalence of GABHS is assumed to be known. GABHS = group A beta-haemolytic streptococcus. PPV = positive predictive value.

These results lead to the question of whether these criteria can be used to rule out infection. For a McIsaac score threshold of 1, a negative test corresponds to a score of −1, or 0. Similarly, at a threshold of 0, a negative test is a score of −1. From Table 2, the negative likelihood ratios (LR−) for the McIsaac score at thresholds 0 and 1 are 0.15 and 0.23, respectively. Thus, at a prevalence of GABHS infection of 25%, from Bayes theorem, low McIsaac scores such as −1 and 0 give expected probabilities of infection of 4.8% and 7.1%. Equally, for the same prevalence, a Centor score of 0 gives an expected probability of infection of 8.1%. For low probabilities, the corresponding calibration curves are more erratic for the McIsaac score than the Centor score (Supplementary Figure S3), so it is not clear how well these expected probabilities translate into practice. Nonetheless, given a shared decision between GP and patient on what constitutes an acceptable risk of GABHS, a low score on either criteria may be considered as sufficient evidence not to pursue treatment or further investigation.

DISCUSSION

Summary

This is the first meta-analysis to compare the performances of the Centor and McIsaac scores in a primary care setting over multiple cut points. Although there were 10 studies that evaluated the Centor score and eight that evaluated the McIsaac score, only one primary study provided data that allowed a direct comparison of the two tests.

The meta-analysis demonstrated that the SROC curves were broadly aligned, with the curve for the McIsaac score lying slightly above that for the Centor score (see Figure 4); however, the difference was marginal and no statistically significant difference between the AUCs was found. Moreover, when those studies authored by McIsaac were excluded, a sensitivity analysis revealed that the AUC for the McIsaac score may be overstated. Nonetheless, this did not alter the conclusion that the two prediction scores have similar performance characteristics and that adding an age variable does not appear to improve the accuracy of the Centor score for diagnosing GABHS in primary care. When compared with the Centor system, the McIsaac rule changes the operating points on the SROC curve rather than improving on discrimination. In addition, with AUCs of approximately 0.7, both systems appear to be, at best, fair at differentiating those patients who have GABHS from those who do not.

The calibration of the models for both scores demonstrates over-confidence, with the expected PPVs diverging substantially from the observed PPVs for probabilities of >40%. The effect of this is that an expected PPV of 80% translates into an observed PPV of 55%. Furthermore, these plots are ‘best cases’ as they are based on the prevalence of GABHS being known for the setting. When the prevalence is unknown, the average across all studies may be used; however, in the studies that were included in this review, the prevalence of GABHS ranged between 4.7% and 44.8%, so using the average prevalence would likely lead to poorer calibration as a result.

Strengths and limitations

All of the studies provided data at ≥2 thresholds, justifying the approach of using a model that accommodates both multiple cut points and different numbers of cut points between studies. This allowed the two criteria to be compared across the whole of the ROC space. Furthermore, by using calibration plots, it was possible to provide evidence on each criteria’s likely performance in practice and when they are most likely to be useful to clinicians.

As a reference standard, the throat swab has limitations — its performance may depend on the operator and the conditions for incubation.38 Alternative reference standards, such as a rising titre of the antistreptolysin O (ASO) antibody, may be used, but these also vary with age, prevalence of streptococcus, and comorbidities.38 ASO testing is also rarely used by investigators; none of the included studies — or those excluded due to inadequacy of a reference test — used ASO testing.

The model used in this review benefits from being able to aggregate studies that provide data at multiple thresholds; however, this needs to be weighed against the necessity for continuity corrections when there are 0 cell entries in the 2 × 2 tables. Furthermore, at present, it is not clear how the DIDS model could include study-level covariates to investigate potential sources of heterogeneity.

Some authors have recommended the use of level-specific likelihood ratios.39 This requires defining test positives as test results that equal the threshold score only, not the threshold score and above, as is usual practice. This alternative definition of a test positive leads to an important property of a ROC curve (monotonicity) being violated;40 hence, with the approach used here, it is not possible to estimate level-specific likelihood ratios.

As part of internal validation, the authors used bootstrap methods to correct the calibration plots for optimism. Other methods have been proposed that use leave-one-out cross-validation to derive a validation statistic so the internal validity of the summary estimates may be assessed;41 it is also possible to use other information, such as the test positive rate, to derive an estimate that is tailored to the setting of interest.42,43 However, a shortcoming with all of these methods, including the method used here, is that they are rarely subject to external validation; without this, it is difficult to make assertions on the transferability of the results.

Comparison with existing literature

A recent review of guidelines for diagnosing acute pharyngitis44 revealed that both the Centor and McIsaac prediction scores are incorporated into guidance for Europe and North America. The Centor score is one of two prediction rules recommended for managing patients with a sore throat in the UK,10 while, in Denmark9 and Germany,8 the McIsaac score is recommended. This demonstrates that these scores are considered relevant to the diagnosis of acute pharyngitis in a number of countries. Therefore, it is perhaps surprising that only two reviews17,18 have evaluated the Centor score in primary care and no systematic review has evaluated the McIssac score in primary care. None of the reviews to date have used a model that was able to accommodate data from multiple thresholds per study in the analysis. Previous reviews17,18 have treated each threshold separately when aggregating studies, thereby ignoring potential correlations between thresholds at a primary-study level and at an SROC curve level. Furthermore, none of the reviews have sought to establish how well the prediction rules calibrate in practice.

As a comparison, the two previous reviews on the Centor score reported positive likelihood ratios for a threshold of 3 — 2.68 (95% CI = 1.92 to 3.75)17 and 2.35 (95% CI = 1.51 to 3.67)18 — and these were inflated compared with the ratios presented here. However, the negative likelihood ratios for a threshold of 1 were comparable: 0.27 (95% CI = 0.16 to 0.46)17 and 0.28 (95% CI = 0.23 to 0.45).18

NICE has recently recommended using either the Centor or the FeverPAIN score to assess the symptoms of patients with acute pharyngitis.10 Although the latter was derived from a UK population, to date this is the only study on FeverPAIN25 and it is yet to be replicated in other independent populations; however, it is unclear whether the FeverPAIN score would lead to a marked improvement in discrimination and calibration, particularly when it shares many of the covariates of the scores that were reviewed here.

Implications for practice

Although the Centor score showed better calibration than the McIsaac system for a negative result, perhaps of more relevance is that, for estimated probabilities of <20%, the observed probabilities of GABHS in practice, given a negative test result, are consistently lower than the corresponding estimates. On this basis, a Centor score of 0 or a McIsaac score of ≤0 is likely to correspond to an actual risk of GABHS that is lower than the expected risk of 8.5% — as such, it is likely to be sufficient to rule out infection.

For a Centor or a McIsaac score of ≥1, it is less clear how to proceed. In general, the probability of GABHS for these scores is likely to be too high (>10%) to rule out infection and too low to rule in infection. NICE’s current recommendation is that a Centor score of ≥3 is sufficient grounds to consider prescribing antibiotics either immediately or as a delayed script with advice;10 however, the evidence presented here suggests that neither score can realistically identify patients with an observed risk of GABHS of >50%, irrespective of the expected risk. There is the potential that these recommendations could lead to inappropriate prescribing of antibiotics in a large percentage of cases.

In all instances, the GP should weigh up the public-health need to reduce the number of inappropriate antibiotic prescriptions and the individual patient’s need to treat a potential infection. With this in mind, an honest discussion with the patient about the likely GABHS risk and the GP’s obligation not to prescribe antibiotics inappropriately before deciding on management seems the most reasonable way to proceed.

Any substantive improvement in the diagnosis of GABHS-related pharyngitis is likely to require either a new prediction system or the use of point-of-care technologies to augment the existing clinical prediction tools.45

Notes

Funding

Brian H Willis was supported by funding from a Medical Research Council Clinician Scientist award (ref: MR/N007999/1).

Ethical approval

No ethical approval was required as this is a secondary analysis of data derived from published primary studies.

Provenance

Freely submitted; externally peer reviewed.

Competing interests

The authors have declared no competing interests.

Discuss this article

Contribute and read comments about this article: bjgp.org/letters

  • Received July 4, 2019.
  • Revision requested August 30, 2019.
  • Accepted October 15, 2019.
  • ©The Authors
http://creativecommons.org/licenses/by/4.0/

This article is Open Access: CC BY 4.0 licence (http://creativecommons.org/licences/by/4.0/).

REFERENCES

  1. 1.↵
    1. Gulliford MC,
    2. Dregan A,
    3. Moore MV,
    4. et al.
    Continued high rates of antibiotic prescribing to adults with respiratory tract infection: survey of 568 UK general practicesBMJ Open20144e006245
    OpenUrlAbstract/FREE Full Text
  2. 2.↵
    1. Dagnelie CF,
    2. Bartelink ML,
    3. van der Graaf Y,
    4. et al.
    Towards better diagnosis of throat infections (with group A beta-haemolytic streptococcus) in general practiceBr J Gen Pract199848427959962
    OpenUrlAbstract/FREE Full Text
  3. 3.↵
    1. Little P,
    2. Hobbs FD,
    3. Mant D,
    4. et al.
    Incidence and clinical variables associated with streptococcal throat infections: a prospective diagnostic cohort studyBr J Gen Pract2012DOI: https://doi.org/10.3399/bjgp12X658322.
  4. 4.↵
    1. Carapetis JR,
    2. Steer AC,
    3. Mulholland EK,
    4. Weber M
    The global burden of group A streptococcal diseasesLancet Infect Dis2005511685694
    OpenUrlCrossRefPubMed
  5. 5.↵
    1. Centor RM,
    2. Whitherspoon JM,
    3. Dalton HP,
    4. et al.
    The diagnosis of strep throat in adults in the emergency roomMed Decis Making198113239246
    OpenUrlCrossRefPubMed
  6. 6.↵
    1. McIsaac WJ,
    2. White D,
    3. Tannenbaum D,
    4. Low DE
    A clinical score to reduce unnecessary antibiotic use in patients with sore throatCMAJ199815817583
    OpenUrlAbstract/FREE Full Text
  7. 7.↵
    1. Chiappini E,
    2. Regoli M,
    3. Bonsignori F,
    4. et al.
    Analysis of different recommendations from international guidelines for the management of acute pharyngitis in adults and childrenClin Ther20113314858
    OpenUrlCrossRefPubMed
  8. 8.↵
    1. Windfuhr JP,
    2. Toepfner N,
    3. Steffen G,
    4. et al.
    Clinical practice guideline: tonsillitis I. Diagnostics and nonsurgical managementEur Arch Otorhinolaryngol20162734973987
    OpenUrl
  9. 9.↵
    1. Reinholdt KB,
    2. Rusan M,
    3. Hansen PR,
    4. Klug TE
    Management of sore throat in Danish general practicesBMC Fam Pract201920175
    OpenUrl
  10. 10.↵
    1. National Institute for Health and Care Excellence
    Sore throat (acute): antimicrobial prescribing. NG842018http://www.nice.org.uk/ng84 (accessed 21 Feb 2020).
  11. 11.↵
    1. Fine AM,
    2. Nizet V,
    3. Mandl KD
    Large-scale validation of the Centor and McIsaac scores to predict group A streptococcal pharyngitisArch Intern Med201217211847852
    OpenUrlCrossRefPubMed
  12. 12.↵
    1. Fine AM,
    2. Nizet V,
    3. Mandl KD
    Participatory medicine: a home score for streptococcal pharyngitis enabled by real-time biosurveillance: a cohort studyAnn Intern Med20131599577583
    OpenUrlCrossRefPubMed
  13. 13.↵
    1. Reitsma JB,
    2. Glas AS,
    3. Rutjes AW,
    4. et al.
    Bivariate analysis of sensitivity and specificity produces informative summary measures in diagnostic reviewsJ Clin Epidemiol20055810982990
    OpenUrlCrossRefPubMed
  14. 14.
    1. Chu H,
    2. Cole SR
    Bivariate meta-analysis of sensitivity and specificity with sparse data: a generalized linear mixed model approachJ Clin Epidemiol2006591213311332
    OpenUrlCrossRefPubMed
  15. 15.↵
    1. Rutter CM,
    2. Gatsonis CA
    A hierarchical regression approach to meta-analysis of diagnostic test accuracy evaluationsStat Med2001201928652884
    OpenUrlCrossRefPubMed
  16. 16.↵
    1. Steinhauser S,
    2. Schumacher M,
    3. Rücker G
    Modelling multiple thresholds in meta-analysis of diagnostic test accuracy studiesBMC Med Res Methodol201616197
    OpenUrl
  17. 17.↵
    1. Aalbers J,
    2. O’Brien KK,
    3. Chan WS,
    4. et al.
    Predicting streptococcal pharyngitis in adults in primary care: a systematic review of the diagnostic accuracy of symptoms and signs and validation of the Centor scoreBMC Med2011967
    OpenUrlCrossRefPubMed
  18. 18.↵
    1. Willis BH,
    2. Hyde CJ
    What is the test’s accuracy in my practice population? Tailored meta-analysis provides a plausible estimateJ Clin Epidemiol2015688847854
    OpenUrl
  19. 19.↵
    1. de Vet HCW,
    2. Eisinga A,
    3. Riphagen II,
    4. et al.
    Chapter 7: Searching for studies. In: Cochrane handbook for systematic reviews of diagnostic test accuracy Version 0.4 [updated September 2008]The Cochrane Collaboration2008
  20. 20.↵
    1. Brazzelli M,
    2. Lewis SC,
    3. Deeks JJ,
    4. Sandercock PAG
    No evidence of bias in the process of publication of diagnostic accuracy studies in stroke submitted as abstractsJ Clin Epidemiol2009624425430
    OpenUrlCrossRefPubMed
  21. 21.↵
    1. Whiting PF,
    2. Rutjes AW,
    3. Westwood ME,
    4. et al.
    QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studiesAnn Intern Med20111558529536
    OpenUrlCrossRefPubMed
  22. 22.↵
    1. Snell KI,
    2. Hua H,
    3. Debray TP,
    4. et al.
    Multivariate meta-analysis of individual participant data helped externally validate the performance and implementation of a prediction modelJ Clin Epidemiol2016694050
    OpenUrlCrossRefPubMed
  23. 23.↵
    1. Harrell FE Jr.
    Regression modeling strategies: with applications to linear models, logistic and ordinal regression, and survival analysis2nd ednNew York, NYSpringer2015
  24. 24.↵
    1. Alper Z,
    2. Uncu Y,
    3. Akalin H,
    4. et al.
    Diagnosis of acute tonsillopharyngitis in primary care: a new approach for low-resource settingsJ Chemother2013253148155
    OpenUrl
  25. 25.↵
    1. Little P,
    2. Moore M,
    3. Hobbs FDR,
    4. et al.
    PRImary care Streptococcal Management (PRISM) study: identifying clinical variables associated with Lancefield group A β-haemolytic streptococci and Lancefield non-group A streptococcal throat infections from two cohorts of patients presenting with an acute sore throatBMJ Open20133e003943
    OpenUrlAbstract/FREE Full Text
  26. 26.↵
    1. Marin Cañada J,
    2. Cubillo Serna A,
    3. Gómez-Escalonilla Cruz N,
    4. et al.
    [Is streptococcal pharyngitis diagnosis possible?]. [Article in Spanish]Aten Primaria2007397361365
    OpenUrlPubMed
  27. 27.↵
    1. Lindbæk M,
    2. Høiby EA,
    3. Lermark G,
    4. et al.
    Clinical symptoms and signs in sore throat patients with large colony variant β-haemolytic streptococci groups C or G versus group ABr J Gen Pract200555517615619
    OpenUrlAbstract/FREE Full Text
  28. 28.↵
    1. Atlas SJ,
    2. McDermott SM,
    3. Mannone C,
    4. Barry MJ
    The role of point of care testing for patients with acute pharyngitisJ Gen Intern Med2005208759761
    OpenUrlCrossRefPubMed
  29. 29.↵
    1. Chazan B,
    2. Shaabi M,
    3. Bishara E,
    4. et al.
    Clinical predictors of streptococcal pharyngitis in adultsIsr Med Assoc J200356413415
    OpenUrlPubMed
  30. 30.↵
    1. Seppälä H,
    2. Lahtonen R,
    3. Ziegler T,
    4. et al.
    Clinical scoring system in the evaluation of adult pharyngitisArch Otolaryngol Head Neck Surg19931193288291
    OpenUrlCrossRefPubMed
  31. 31.↵
    1. Regueras De Lorenzo G,
    2. Santos Rodríguez PM,
    3. Villa Bajo L,
    4. et al.
    [Use of the rapid antigen technique in the diagnosis of Streptococcus pyogenes pharyngotonsillitis]. [Article in Spanish]An Pediatr (Barc)201277193199
    OpenUrl
  32. 32.↵
    1. Stefaniuk E,
    2. Bosacka K,
    3. Wanke-Rytt M,
    4. Hryniewicz W
    The use of rapid test QuikRead go® Strep A in bacterial pharyngotonsillitis diagnosing and therapeutic decisionsEur J Clin Microbiol Infect Dis2017361017331738
    OpenUrl
  33. 33.↵
    1. Mistik S,
    2. Gokahmetoglu S,
    3. Balci E,
    4. Onuk FA
    Sore throat in primary care project: a clinical score to diagnose viral sore throatFam Pract2015323263268
    OpenUrlCrossRefPubMed
  34. 34.↵
    1. Dunne EM,
    2. Marshall JL,
    3. Baker CA,
    4. et al.
    Detection of group a streptococcal pharyngitis by quantitative PCRBMC Infect Dis201313312
    OpenUrlCrossRefPubMed
  35. 35.↵
    1. Tanz RR,
    2. Gerber MA,
    3. Kabat W,
    4. et al.
    Performance of a rapid antigen-detection test and throat culture in community pediatric offices: implications for management of pharyngitisPediatrics20091232437444
    OpenUrlAbstract/FREE Full Text
  36. 36.↵
    1. Flores Mateo G,
    2. Conejero J,
    3. Grezner Martinel E,
    4. et al.
    [Early diagnosis of streptococcal pharyngitis in paediatric practice: validity of a rapid antigen detection test]. [Article in Spanish]Aten Primaria2010427356361
    OpenUrlPubMed
  37. 37.↵
    1. McIsaac WJ,
    2. Goel V,
    3. To T,
    4. Low DE
    The validity of a sore throat score in family practiceCMAJ20001637811815
    OpenUrlAbstract/FREE Full Text
  38. 38.↵
    1. Ferretti JJ,
    2. Stevens DL,
    3. Fischetti VA
    1. Spellerberg B,
    2. Brandt C
    Laboratory diagnosis of Streptococcus pyogenes (group A streptococci)Ferretti JJ, Stevens DL, Fischetti VAStreptococcus pyogenes: basic biology to vlinical manifestationsOklahoma City, OKUniversity of Oklahoma Health Sciences Center2016
  39. 39.↵
    1. Haynes RB,
    2. Sackett DL,
    3. Gordon H,
    4. et al.
    Clinical epidemiology: how to do clinical practice research3rd ednLondonLippincott Williams & Wilkins2006
  40. 40.↵
    1. Krzanowski WJ,
    2. Hand DJ
    ROC curves for continuous dataLondonChapman and Hall/CRC2009
  41. 41.↵
    1. Willis BH,
    2. Riley RD
    Measuring the statistical validity of summary meta-analysis and meta-regression results for use in clinical practiceStat Med2017362132833301
    OpenUrl
  42. 42.↵
    1. Willis BH,
    2. Hyde CJ
    Estimating a test’s accuracy using tailored meta-analysis: how setting-specific data may aid study selectionJ Clin Epidemiol2014675538546
    OpenUrlCrossRefPubMed
  43. 43.↵
    1. Willis BH,
    2. Coomar D,
    3. Baragilly M
    Tailored meta-analysis: an investigation of the correlation between the test positive rate and prevalenceJ Clin Epidemiol201910619
    OpenUrl
  44. 44.↵
    1. Banerjee S,
    2. Ford C
    Clinical decision rules and strategies for the diagnosis of group A streptococcal infection: a review of clinical utility and guidelinesOttawaCADTH2018
  45. 45.↵
    1. National Institute for Health and Care Excellence
    Point-of-care diagnostic testing in primary care for strep A infection in sore throat. MIB1452018http://www.nice.org.uk/guidance/mib145 (accessed 21 Feb 2020).
Back to top
Previous ArticleNext Article

In this issue

British Journal of General Practice: 70 (693)
British Journal of General Practice
Vol. 70, Issue 693
April 2020
  • Table of Contents
  • Index by author
Download PDF
Download PowerPoint
Email Article

Thank you for recommending British Journal of General Practice.

NOTE: We only request your email address so that the person to whom you are recommending the page knows that you wanted them to see it, and that it is not junk mail. We do not capture any email address.

Enter multiple addresses on separate lines or separate them with commas.
Comparison of Centor and McIsaac scores in primary care: a meta-analysis over multiple thresholds
(Your Name) has forwarded a page to you from British Journal of General Practice
(Your Name) thought you would like to see this page from British Journal of General Practice.
CAPTCHA
This question is for testing whether or not you are a human visitor and to prevent automated spam submissions.
Citation Tools
Comparison of Centor and McIsaac scores in primary care: a meta-analysis over multiple thresholds
Brian H Willis, Dyuti Coomar, Mohammed Baragilly
British Journal of General Practice 2020; 70 (693): e245-e254. DOI: 10.3399/bjgp20X708833

Citation Manager Formats

  • BibTeX
  • Bookends
  • EasyBib
  • EndNote (tagged)
  • EndNote 8 (xml)
  • Medlars
  • Mendeley
  • Papers
  • RefWorks Tagged
  • Ref Manager
  • RIS
  • Zotero
Share
Comparison of Centor and McIsaac scores in primary care: a meta-analysis over multiple thresholds
Brian H Willis, Dyuti Coomar, Mohammed Baragilly
British Journal of General Practice 2020; 70 (693): e245-e254. DOI: 10.3399/bjgp20X708833
del.icio.us logo Digg logo Reddit logo Twitter logo CiteULike logo Facebook logo Google logo Mendeley logo
  • Tweet Widget
  • Facebook Like
  • Google Plus One
  • Mendeley logo Mendeley

Jump to section

  • Top
  • Article
    • Abstract
    • INTRODUCTION
    • METHOD
    • RESULTS
    • DISCUSSION
    • Notes
    • REFERENCES
  • Figures & Data
  • Info
  • eLetters
  • PDF

Keywords

  • Centor score
  • diagnosis
  • McIsaac score
  • meta-analysis
  • pharyngitis
  • primary health care

More in this TOC Section

  • Performance of ethnic minority versus White doctors in the MRCGP assessment 2016–2021: a cross-sectional study
  • Trends in the registration of anxiety in Belgian primary care from 2000 to 2021: a registry-based study
  • Strengthening the integration of primary care in pandemic response plans: a qualitative interview study of Canadian family physicians
Show more Research

Related Articles

Cited By...

Intended for Healthcare Professionals

BJGP Life

BJGP Open

 

@BJGPjournal's Likes on Twitter

 
 

British Journal of General Practice

NAVIGATE

  • Home
  • Current Issue
  • All Issues
  • Online First
  • Authors & reviewers

RCGP

  • BJGP for RCGP members
  • BJGP Open
  • RCGP eLearning
  • InnovAiT Journal
  • Jobs and careers

MY ACCOUNT

  • RCGP members' login
  • Subscriber login
  • Activate subscription
  • Terms and conditions

NEWS AND UPDATES

  • About BJGP
  • Alerts
  • RSS feeds
  • Facebook
  • Twitter

AUTHORS & REVIEWERS

  • Submit an article
  • Writing for BJGP: research
  • Writing for BJGP: other sections
  • BJGP editorial process & policies
  • BJGP ethical guidelines
  • Peer review for BJGP

CUSTOMER SERVICES

  • Advertising
  • Contact subscription agent
  • Copyright
  • Librarian information

CONTRIBUTE

  • BJGP Life
  • eLetters
  • Feedback

CONTACT US

BJGP Journal Office
RCGP
30 Euston Square
London NW1 2FB
Tel: +44 (0)20 3188 7400
Email: journal@rcgp.org.uk

British Journal of General Practice is an editorially-independent publication of the Royal College of General Practitioners
© 2023 British Journal of General Practice

Print ISSN: 0960-1643
Online ISSN: 1478-5242