Analysis

The good and the bad: using C reactive protein to distinguish bacterial from non-bacterial infection among febrile patients in low-resource settings

Abstract

C reactive protein (CRP), a marker for the presence of an inflammatory process, is the most extensively studied marker for distinguishing bacterial from non-bacterial infections in febrile patients. A point-of-care test for bacterial infections would be of particular use in low-resource settings where other laboratory diagnostics are not always available, antimicrobial resistance rates are high and bacterial infections such as pneumonia are a leading cause of death. This document summarises evidence on CRP testing for bacterial infections in low-income and middle-income countries (LMICs). With a push for universal health coverage and prevention of antimicrobial resistance, it is important to understand if CRP might be able to do the job. The use of CRP polarised the global health community and the aim of this document is to summarise the ‘good and the bad’ of CRP in multiple settings in LMICs. In brief, the literature that was reviewed suggests that CRP testing may be beneficial in low-resource settings to improve rational antibiotic use for febrile patients, but the positive predictive value is insufficient to allow it to be used alone as a single tool. CRP testing may be best used as part of a panel of diagnostic tests and algorithms. Further studies in low-resource settings, particularly with regard to impact on antibiotic prescribing and cost-effectiveness of CRP testing, are warranted.

Summary box

  • C reactive protein (CRP) testing may be beneficial in low-resource settings to improve rational antibiotic use for febrile patients, but the positive predictive value is insufficient to allow it to be used alone as a single tool.

  • More extensive cost-effectiveness data across multiple geographies are needed.

  • The cut-off used for CRP across studies varies widely and makes it difficult to select a universal cut-off threshold for diagnostic use.

Background

The management of febrile patients is a major problem in lower resource areas where access to diagnostics is limited. Fever symptoms can result from a variety of different infections, including parasites like malaria, bacterial or viral pathogens, which are difficult to distinguish from one another based on clinical presentation alone. While the widespread use of rapid diagnostic tests (RDTs) for malaria has transformed the management of fevers in tropical settings, it has been accompanied by an increase in antibiotic prescriptions, since in the absence of further diagnostics, malaria-negative patients are generally treated for bacterial infection.1 2 Unnecessary use of antibiotics is considered to be a major driver of development of antimicrobial resistance, an increasingly serious threat to global public health.

A rapid point-of-care test (POCT) to detect bacterial infection would be of particular use in resource-constrained settings, where other laboratory diagnostics such as blood culture and radiology are not always available, antimicrobial resistance rates are high, and bacterial infections such as pneumonia are a leading cause of death. Various potential biomarkers have been evaluated, of which C reactive protein (CRP), a marker for the presence of an inflammatory process, is the most extensively studied.3 A Cochrane review from 2014 focusing on acute non-severe respiratory infections in primary care concluded that CRP was the only sufficiently accurate biomarker for which POCTs are available that could safely and effectively reduce the prescribing of antibiotics.4 However, the majority of studies evaluating the diagnostic performance of CRP have taken place in high-resource settings. This document summarises existing data in low-income and middle-income countries (LMICs), and discusses the potential utility of CRP testing in this setting. The practical aim of this work is to help support communication and informed discussions among the global health community by showing the ‘good and the bad’ in a relatively unbiased manner. This work does not aim to be a systematic review but a pragmatic document that can help a broad community to start off discussions with a similar knowledge base. Data were gathered based on an unstructured search of the PubMed database for studies on CRP in LMICs, published before June 2019.

Diagnostic performance of CRP for distinguishing bacterial infections

Studies assessing the correlation between elevated CRP levels with presence of bacterial infections and the diagnostic accuracy of CRP for distinguishing between bacterial and non-bacterial infections in LMICs are detailed in table 1. Of seven studies that assessed the correlation between CRP levels and presence of bacterial infection, all seven found that CRP levels were significantly higher in patients with confirmed bacterial infections versus those without.5–13

Table 1
|
Studies assessing correlation between CRP levels and bacterial infection and diagnostic performance of CRP in LMICs

The performance of CRP for distinguishing bacterial from non-bacterial infections varied considerably across studies with areas under the receiver operating curve (AUROC) ranging from 0.62 to 0.91. The high variability in performance across studies may be due to a number of factors, including differences in the clinical presentation of the population studied, the degree of patient severity, the definition used for bacterial infections (the gold standard used; table 1), geographical location of the study, the age of the patients, the specific bacterial pathogens causing infection, concomitant infections and presence of other conditions causing elevated CRP.

Most of the studies identified were performed in African countries (6/8) with children under 5 years of age (6/10), facilitating comparison between studies. Studies carried out with inpatients, independently of the gold standard used for bacterial infection definition, presented high performances (0.72–0.87) except for the study with severe acute malnourished children, which reported an AUROC of 0.66. However, when looking at outpatients, two studies carried out in Tanzania in children and one study in Cambodia in individual 7–49 years reported very different performances. One study which used positive blood culture as gold standard reported a high AUROC of 0.83 and 74.2/77.8 sensitivity/specificity. However, the two other studies, which used a wider definition of bacterial infection (ie, the gold standard for bacterial infection includes microbiological results and/or symptoms in addition to positive cultures), reported lower performances: AUROC of 0.62, 44.6/78.5 sensitivity/specificity and 52.5/84.3 sensitivity/specificity, respectively. In line with this, Lubell et al described that the fact of being admitted as an inpatient had an independent effect of elevated levels of CRP (p=0.006).9

When looking specifically at HIV-positive and HIV-negative patients, Higdon et al found that patients who were HIV positive were more likely to have CRP levels of ≥40 mg/L, and within this group, older children and those with more severe pneumonia were also more likely to have higher CRP levels.7 Among HIV-negative patients, those from the African study sites were more likely to have CRP levels of ≥40 mg/L than those from the Asian sites.7

Three studies evaluated a CRP cut-off level of approximately 20 mg/L.6 9 14 Three studies assessed a higher cut-off of approximately 40 mg/L.6 7 10 Five studies calculated the optimal CRP cut-off for distinguishing bacterial from non-bacterial infections. Of these, Hildenwall et al and Mueller et al proposed a cut-off of 19 and 21.3 mg/L, respectively,11 14 while Mahende et al, Higdon et al, and Wangrangsimakul et al proposed higher cut-offs of 37.3, 37.1 and 36 mg/L, respectively.7 10 13

The method used to detect CRP differed across studies, with the majority of studies using quantitative CRP tests. A study by Phommasone et al was the only study to assess commercially available lateral flow CRP tests (DTS233 (Creative Diagnostics, USA), WD-23 (Assure Tech, China) and bioNexia CRPplus (bioMerieux, France)), in comparison to the Nycocard CRP test and reader.15 At a cut-off of 10 mg/L, all three tests had high sensitivity (ranging from 87% to 98%) and specificity (91% to 98%) in patients with fever in rural Laos, suggesting that lateral flow tests are a viable option in LMIC settings.

CRP levels have been shown to be elevated in patients with malaria, as well as those with bacterial infections (consequently, some studies excluded patients with malaria from their analyses). Studies that assessed the correlation between elevated CRP and malaria infection are detailed in table 2. All five studies found that CRP levels were significantly higher in patients with malaria versus those without.9 10 16–18 In Tanzania, 80% of children aged 2–59 months presenting with malaria had CRP levels >40 mg/L, the higher of the commonly used CRP cut-offs.10 Consistent with this, Lubell et al found no significant difference in CRP levels between patients with bacterial infections and patients with malaria.9 The confounding effect of malaria seems to be limited to clinical malaria, as Peto et al found that in the general population, only 7.6% of people who tested positive for subclinical malaria had a CRP of >10 mg/L.17

Table 2
|
Studies assessing correlation between CRP levels and malaria

Of note, CRP has also been widely evaluated as a predictor for serious infections, and serious bacterial infections (SBI) in particular, however such studies were essentially conducted in high-income countries. For example, CRP in combination with vital signs and objective symptoms measurements achieved a sensitivity of 97.1% (95% CI 94.3% to 98.7%) for identifying children aged 1 month to 16 years with serious infections in clinics and emergency departments in Belgium, classifying them into groups of low, intermediate and high risk with CRP levels of <20, 20–75 and >75 mg/L, respectively.19 Other studies in the Netherlands and Iran report a receiver operator characteristic curve area of 0.77 and 0.74 for CRP predicting SBI in children aged 1 month to 16 years, and infants aged 3 months or less, respectively.20 21 A cluster randomised controlled trial in Belgium furthermore proposes the use of CRP POCTs as a tool to rule out the need of hospital referral for children aged 1 month to 16 years with CRP levels <5 mg/L at primary healthcare level.22

Reduction in antibiotic prescription due to CRP testing

Only two prospective studies assessed the impact of CRP testing as a standalone tool on antibiotic prescriptions in LMICs (table 3), in combination with clinical judgement; both took place in South-East Asia. Althaus et al assessed use of CRP cut-offs of 20 mg/L and CRP of 40 mg/L to guide antibiotic prescription in adults and children aged ≥1 year attending primary care and presenting with fever.23 While the proportion of patients receiving antibiotics by the fifth day after the initial visit was slightly higher in the group for whom no CRP testing was performed compared with the two CRP groups, only the difference between the 40 mg/L cut-off and the control group was statistically significant. Do et al found that a significantly lower percentage of patients with non-severe acute respiratory tract infection who were diagnosed using CRP testing were prescribed antibiotics within 14 days, compared with those in whom no CRP testing was performed.24 This study used a CRP cut-off of 20 mg/L for patients aged 6–65 years, and a cut-off of 10 mg/L for those aged 1–5 years.

Table 3
|
Studies assessing reduction in antibiotic prescriptions associated with CRP testing in LMICs

In the study by Althaus et al, a higher proportion of patients with elevated CRP concentration were prescribed an antibiotic in the CRP groups versus the control group, suggesting that treatment was targeted.23 The limited impact of CRP testing on antibiotic prescriptions in this study was therefore unlikely to be due to non-adherence to the test results. However, in the study by Do et al, the majority of patients in Vietnam who received immediate antibiotic prescriptions had CRP measurements of <10 mg/L, suggesting that healthcare professionals did not always adhere to test results.24

There was a considerable difference between the percentages of patients prescribed antibiotics in the control groups of the studies by Althaus et al and Do et al (39% vs 78%).23 24 It is possible that healthcare professionals in the study by Althaus et al were more cautious with prescribing of antibiotics due to their participation in the study. However, the Althaus et al data are consistent with a retrospective study assessing antibiotic use in 97 230 patients with fever in primary care health centres in Thailand, in which only 46.9% were prescribed antibiotics.2 Given the differences between Althaus et al and Do et al in terms of adherence to test results and control groups, further studies are required to better elucidate the impact of CRP testing on reduction in antibiotic prescriptions.

Interestingly, two studies in Tanzania have used CRP testing as part of electronical decision algorithms to guide the healthcare staff decisions on the use of antibiotics and hospital referral for management of children aged 2–59 months in a primary healthcare setting. These algorithms combined CRP blood levels at a cut-off of 80 mg/L with other POCTs and/or clinical features, and allowed to reduce antibiotics use by up to 83.4% while at the same time improving clinical outcomes, in comparison with standard recommended clinical algorithms.25 26

Cost-effectiveness of CRP testing

Two studies have evaluated cost-effectiveness of CRP diagnostics in LMICs (table 4). In a model using data from the study of CRP testing in febrile patients in rural Laos,9 the NycoCard analyser was shown to avert 0.017 disability-adjusted life-years, with a median incremental cost-effectiveness ratio (ICER) of US$94, at a cut-off of 20 mg/L.27 This model assumed that patients were tested for CRP, dengue fever and scrub typhus, and that in patients who tested negative, antibiotics were prescribed randomly at a rate of 38%. The analysis suggested that CRP testing is likely to be cost-effective even at low willingness-to-pay thresholds.

Table 4
|
Studies assessing cost-effectiveness of CRP testing in LMICs

A model using data from the study in patients with acute respiratory infection in Vietnam28 showed that CRP testing at a cut-off of 20 mg/L (10 mg/L in those aged 1–5 years) can be cost-beneficial providing that adherence to test results is high (>70% for a unit cost of US$0.5 and >80% for a unit cost of US$1).29 This assumed an economic cost of antimicrobial resistance of US$4.1, based on published modelling data from Thailand.30 A higher cost of AMR led to cost-benefits of CRP testing at lower adherence and higher unit cost. Notably, this study assumed no difference in clinical outcomes between CRP-tested and non-tested patients, thus it may represent a conservative estimate.

Findings from the two studies detailed above are generally consistent with studies in higher resource settings, although willingness-to-pay thresholds used in high-income setting models are considerably higher.31–33 In Norway and Sweden, CRP POCT was associated with a cost per quality-adjusted life-year (QALY) gained of €9391.31 At a willingness-to-pay threshold of €30 000, there was a 70% probability of CRP being cost-effective. In the UK, CRP POCT in adults with acute respiratory tract infection had ICERs of £19 705 per QALY gained and £16.07 per antibiotic prescription avoided. At a threshold of £20 000 per QALY, the probability of CRP POCT being cost-effective was 0.49 (0.84 in those with lower respiratory tract infection).32

More studies would be needed to determine the optimal cost of a CRP rapid test in LMICs, which would obviously vary by country. The two published studies in Laos and Vietnam assume costs ranging from US$0.5 to US$3 per test which seem to be reasonably low and still compatible with the manufacturers’ capacities, assuming that current costs of approximately US$3 per test could be driven down if demand increases or if manufacturers accept selling for lower prices in LMICs while making more profit with higher prices in high-income countries.

Conclusion

While a significant correlation between elevated CRP levels and the presence of bacterial infections in febrile patients in low-resource settings has been confirmed across several studies, reported sensitivity and specificity values have varied considerably. As such, the optimal cut-off point for diagnostic utility in these settings is difficult to determine. A conservative cut-off of ~20 mg/L would limit the number of false negatives, ensuring that patients with serious bacterial infections are appropriately treated, but the potential impact on antibiotic prescribing would be reduced, since more patients would receive antibiotics unnecessarily. In one of the two studies assessing the impact of CRP testing on antibiotic prescriptions in low-resource settings, only the 40 mg/L cut-off was sufficient to significantly reduce the number of prescriptions23; however, the second study demonstrated a significant reduction at a cut-off of 20 mg/L (10 mg/L for the youngest patients).24 Several studies have demonstrated that CRP levels are also elevated in patients with malaria. In malaria-endemic settings, CRP testing should be performed in conjunction with a malaria RDT, to compensate for the confounding effect of malaria and to limit overtreatment of both conditions. CRP testing has been shown to be cost-effective in high-resource settings, but data in LMICs are limited. Two modelling studies have suggested potential cost benefits, but notably, in one study, the level of adherence to CRP test results had considerable impact on the cost-effectiveness of CRP testing.29 CRP testing may therefore be most beneficial as part of a diagnostic algorithm that includes test results and clinical symptoms as part of an integrated approach, giving healthcare professionals more confidence in the recommended treatment.

In summary, the use of CRP has advantages and disadvantages that need to be evaluated carefully for each specific use case (table 5). The collation of this evidence, both good and bad, can hopefully facilitate the discussions between researchers and global health decision makers bringing a little objectivity to what can be a much polarised discussion.

Table 5
|
Overview of the good and the bad