Browse Subject Areas

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Evaluating Operational Specifications of Point-of-Care Diagnostic Tests: A Standardized Scorecard

Evaluating Operational Specifications of Point-of-Care Diagnostic Tests: A Standardized Scorecard

  • Jonathan D. Lehe, 
  • Nádia E. Sitoe, 
  • Ocean Tobaiwa, 
  • Osvaldo Loquiha, 
  • Jorge I. Quevedo, 
  • Trevor F. Peter, 
  • Ilesh V. Jani


The expansion of HIV antiretroviral therapy into decentralized rural settings will increasingly require simple point-of-care (POC) diagnostic tests that can be used without laboratory infrastructure and technical skills. New POC test devices are becoming available but decisions around which technologies to deploy may be biased without systematic assessment of their suitability for decentralized healthcare settings. To address this, we developed a standardized, quantitative scorecard tool to objectively evaluate the operational characteristics of POC diagnostic devices. The tool scores devices on a scale of 1–5 across 30 weighted characteristics such as ease of use, quality control, electrical requirements, shelf life, portability, cost and service, and provides a cumulative score that ranks products against a set of ideal POC characteristics. The scorecard was tested on 19 devices for POC CD4 T-lymphocyte cell counting, clinical chemistry or hematology testing. Single and multi-parameter devices were assessed in each of test categories. The scores across all devices ranged from 2.78 to 4.40 out of 5. The tool effectively ranked devices within each category (p<0.01) except the CD4 and multi-parameter hematology products. The tool also enabled comparison of different characteristics between products. Agreement across the four scorers for each product was high (intra-class correlation >0.80; p<0.001). Use of this tool enables the systematic evaluation of diagnostic tests to facilitate product selection and investment in appropriate technology. It is particularly relevant for countries and testing programs considering the adoption of new POC diagnostic tests.


Decentralization of HIV antiretroviral treatment (ART) in resource-limited settings will require increased access to simple diagnostic technologies such as point-of-care (POC) tests suitable for use in primary health care facilities with minimal infrastructure, resources and skills. UNAIDS has identified POC tests as a key pillar of its HIV Treatment 2.0 framework in resource-limited settings [1]. While few HIV POC tests other than rapid serological test strips are widely used, global HIV testing demands are likely to be increasingly met by POC technologies for CD4+ T-cell counting (CD4), viral load quantification (VL), early infant diagnosis (EID), and HIV drug toxicity monitoring and resistance testing [2][4]. Point-of-care tests have the potential to improve access to ART and treatment program efficiency; for example recent studies demonstrated that POC CD4 testing reduced patient loss-to-follow-up and increased enrollment in ART in primary healthcare settings [5]. Given the growing uncertainty in global health funding [6], [7], interventions to improve ART efficiency are needed.

In recent years, investment by development donors and industry in the development of POC technologies for CD4, viral load and EID has resulted in a broader pipeline of products expected to reach the market in the next 1–3 years [8], [9]. In addition, a number of products for POC drug toxicity monitoring (clinical chemistry and hematology) exist but are not used widely. As ART is decentralized further, countries and testing programs will increasingly make decisions on which technologies to implement. Deployment of new technology on a national scale requires significant investment and diagnostic devices remain in use for years, hence selection of appropriate technology is important. However, the identification of appropriate products can be hampered by the lack of adequate and impartial information on their operational characteristics and suitability for the settings in which they will be used. The selection process may be non-standardized, subjective and susceptible to bias. Critical features defining the utility of each device in the targeted setting, e.g. test throughput, heat-stability, quality control, availability of service support, portability, ease of use or ability to run on batteries may be overlooked or assessed inconsistently across different products.

Here we describe the design of an objective, quantitative scorecard to assess the operational specifications POC test devices and its application to products for POC HIV CD4 count and drug toxicity monitoring.


Pre-selection of products

To identify products for assessment, a survey of CD4, hematology, and clinical chemistry products labeled as POC and available on the market was conducted. The following definition of POC was used: a fully or partially automated table-top, portable, handheld or disposable device able to be operated in a non-laboratory environment by non-technical staff to deliver a same-day, on-site, clinically relevant, diagnostic test result. A survey to identify POC diagnostic technologies to assess with the tool was undertaken. Products were identified from lists of known technologies maintained by the Instituto Nacional de Saúde, internet searches, and company enquiries. Search terms included “Point-of-care, test, diagnostics, technology”. All products identified in this search were catalogued and then assessed if commercially available by contacting the suppliers. All products that were not commercially available were excluded from the study. Information on product specifications was provided by suppliers via brochures and formal communications. Products were then pre-screened for their ability to conduct one or more of the following target tests, CD4 T+ lymphocyte enumeration, liver function (ALT and AST), kidney function (creatinine), glucose, cholesterol, full or partial blood cell count, and hemoglobin.

The shortlisted products were separated into five categories: (i) CD4+ T-cell count analyzers; (ii) multiple-parameter clinical chemistry analyzers (conducts more than 3 different types of tests); (iii) limited-parameter clinical chemistry analyzers (conducts 1–3 different tests); (iv) multiple-parameter hematology analyzers and; (v) limited-parameter hematology analyzers.

Scorecard Design

The selected devices were subjected to a quantitative analysis using a scorecard that rated each product based on a pre-defined set of ideal POC specifications, including factors such as portability, power source, heat stability, daily throughput, access to service and maintenance, availability of quality controls, ease of use, and cost. Thirty assessment criteria were scored for each product across six major categories as follows:

POC Features of Equipment - the degree to which a product's equipment can be deployed in settings with limited infrastructure, e.g. portability, throughput, and power source.

POC Features of Consumables - the degree to which the product consumables were appropriate for use in settings with limited infrastructure, e.g. heat stability and shelf life of reagents and controls.

Ease of Use - how suitable a product is for operators with minimal technical training, and the simplicity of the testing procedure.

Quality Control - assesses whether a product provides sufficient internal quality control mechanisms and is compatible with external quality assurance schemes.

Cost - the relative cost the equipment and consumables.

Distribution and Service - assesses whether local distribution, technical service and after-sales support is available.

The scorecard allows for each device to be scored on a scale of 1–5 (with 5 representing ideal) to assess the device's utility for POC testing in low-resource, primary health care settings. This overall score was made up from individual scores of the 30 assessment criteria each of which had pre-assigned score levels between 1 and 5 for different specifications. For example, the scoring for electrical requirement was as follows: No external power required: 5 points; rechargeable battery operated: 3 points; requires continuous stable electricity: 0 points. In addition, each individual score was also weighted based on its relative importance compared with other criteria. The scorecard with the criteria, scoring thresholds and weightings is provided in Table 1.

Table 1. Diagnostic technology scorecard assessment criteria with scoring thresholds and weightings.

Eight of the 30 criteria were considered key POC characteristics. These included features such as simplicity of use (requiring minimal operator skills), low cost instruments and consumables, presence of internal quality control, battery power, heat stability, availability of service and maintenance (Table 1).

Technology Assessments

Each of the products was scored by four independent scorers that were blinded to each other's assessments. The scorers were laboratory technicians with training and experience in medical diagnostics in low-resource settings. The overall score for each device was determined as the average of the four scores, and an unbiased arbitrator collated and analyzed the score data, and resolved any clear errors or discrepancies. Most criteria were objective (Table 1) and therefore if discrepancies across scorers arose related to these criteria, the scores were reviewed with the scorers to determine whether an error may have occurred. Any scores provided in clear error were corrected. Certain other criteria were more subjective and open to scorer interpretation (Table 1). If discrepancies across scorers arose related to these subjective criteria, they were investigated to ensure that there were no obvious errors or inconsistencies, but otherwise these discrepancies were assumed to be valid.

Statistical Analysis

Statistical analyses were done to evaluate the rank of each product and the reliability of the tool for each class of technology. As all products were judged by the same scorers, the Friedman test was used for a one-way repeated measures analysis of variance for ranks evaluation. We performed a two-way mixed model assuming absolute agreement and used single measures inter-class correlation (ICC) to assess the inter-rater reliability.

These analyses were conducted using SAS version 9.2 (SAS Institute Inc. USA). Tests used 2-sided p-values with α = 0.05 for level of significance and 95% confidence intervals.

Ethics Statement

This study was reviewed and approved by the Mozambique National Health Bioethics Committee. No human subjects participated in this study and no patient samples or patient data were used.


Of 49 diagnostic devices identified by the survey that met our definition of POC, 19 performed one or more of the target test parameters (Table 2). The remaining products did not conduct any of the target parameters and were not evaluated using the scorecard, e.g. blood gas analyzers were not included in the study. The 19 selected devices consisted of 2 CD4 analyzers, 6 multi-parameter blood chemistry analyzers, 4 limited-parameter chemistry analyzers, 3 multi-parameter hematology analyzers, and 4 hemoglobin analyzers (limited parameter hematology instruments).

Table 2. Point-of-care medical diagnostic products evaluated with the scorecard.

The scores across all product categories ranged from 2.78 to 4.40 out of 5 (Table 3). The range of scores enabled the ranking the devices. The ranking was statistically significant (p<0.01) in all product categories except for the CD4 and multi-parameter hematology devices (Table 3). Amongst the set of products evaluated in this study, the lowest ranking products achieved scores within 15–30% of the highest score of each category, suggesting relative homogeneity of operational characteristics as assessed by this tool. The widest range of scores by technology was seen amongst the multiple-parameter chemistry instruments indicating that these products were more diverse in characteristics, while scores amongst the hemoglobin meters were closest indicating similar characteristics in this product category (Table 3).

Table 3. Score, rank and inter-class correlation of point-of-care technologies assessed with the scorecard.

Of the 30 scorecard criteria, eight were considered key POC features that might disqualify devices from use. When assessing the 19 products for these key POC features, mean scores were high (>4.0) for the following criteria: availability of battery power, ease of reagent preparation and simplicity of test procedure, and availability of internal quality control. Mean scores were low (<3.0) for the following criteria: cost of equipment and reagents, heat stability of reagents (primarily due to low stability of control reagents) and weak distribution and service (due to lack of in-country providers). The multi-parameter chemistry, hematology and CD4 platforms scored lower on many of the key POC parameters than the single parameter platforms.

There was a high degree of agreement in the scores given to each product across the four scorers as well as in the scores for each of the 30 assessment criteria, with intra-class correlations (ICC) greater than 0.80 (p<0.001) for each product (Table 3), demonstrating the internal consistency of the tool.

A simplified representation of the performance of the top 2–3 scoring devices in each product category for technologies is illustrated in Figure 1. These spider plots chart the combined scores within the six major categories of assessment: cost, POC characteristics of the instrument, POC characteristics of the consumables, distribution and service, quality control and ease of use. These plots allow rapid assessment of performance across major operational characteristics and demonstrate that no product scored high in all areas.


We developed and tested an objective and standardized non-laboratory methodology for evaluating POC diagnostic devices being considered for deployment in decentralized healthcare settings. This tool enables the selection of products using unbiased criteria based on a set of ideal POC specifications that take into account key operational features such as portability, battery use, heat stability and ease of use, which are required for tests conducted in rural and low-resource settings with minimal infrastructure, electricity and skills. When applied to 19 currently available diagnostic devices, the scorecard successfully ranked the products and systematically identified the strengths and weaknesses of each platform.

The assessment method described here may reduce opportunity for bias in product selection and help health care programs invest in appropriate technology. The scorecard provides a way to objectively navigate the range of information provided by vendors and other sources, a task complicated by the large number of commercially available test devices and inconsistent information available for different products. While this scorecard does not replace assessment of clinical performance on patient samples, it can be used to shortlist devices for clinical evaluation or can be combined with test performance data to guide adoption and deployment decisions.

Although the scorecard ranked products, many scores were close and performance of products in different operational areas cannot be understood from the overall scores. The spider plots enabled key characteristics of different products to be more easily compared, as illustrated in Figure 1. These plots allow quick assessment of the trade-offs between technologies in areas such as quality control, ease of use, cost and service availability, especially amongst closely ranked products with similar overall scores.

Of the products evaluated, the multi-parameter chemistry, hematology and CD4 platforms generally scored lower than the single parameter CD4 platforms, indicating that the latter devices had more ideal POC features. As single parameter devices have drawbacks associated with the need to maintain multiple different platforms, future technology development should focus on improving the POC features of multi-parameter platforms to enable their deployability in resource-limited settings. Furthermore, across the 19 products, certain key POC features generally scored high, such as quality control and ease of use, while other characteristics such as cost and availability of service and distribution tended to score low. This highlights areas for manufactures and distributors to focus on when considering ways to improve design and delivery of current and future products.

The weighting and scoring system of the tool was based on the set of pre-defined ideal POC characteristics. These weights and scores can be changed if needed to customize the tool for different settings. In addition, a simplified version of the tool may be applicable in certain settings, for example based on the eight key POC criteria. The tool does not provide an absolute cut-off to define high performing or low performing devices, as this may vary with each set of products analyzed. Instead, within each category the tool ranks devices relative to each other in order to prioritize them for further evaluation and/or implementation, or to identify tradeoffs that need to be considered. For example, if ease of use, cost and service availability or good quality control are priorities for a testing program, devices that score higher on these parameters can be short-listed over products that have other strengths.

An advantage of the scorecard approach is that both qualitative and quantitative technology characteristics can be assessed. It therefore complements laboratory or clinical evaluations, which typically assess quantitative characteristics such as sensitivity and specificity, predictive values, bias, precision and reproducibility [10], [11]. One drawback of the scorecard is the potential for inaccurate scores due to incorrect information on products. As some of the data used to populate the tool may not be directly verifiable without a sample instrument, inaccurate information can influence the results. It is therefore necessary to confirm the source and integrity of the information used. Another drawback is that some information may not be known for new products that have had limited use, such as whether the technology will be compatible with external quality assurance schemes. In addition, 10 of the 30 assessment criteria scored were subjective and potentially open to different interpretation. These 10 criteria accounted for 35% of the total score weightings for each device. However, the results of the inter-rater reliability analysis in this study suggest that the overall variation in product scores was predominantly due to actual differences in product characteristics, rather than the subjectivity of certain assessment criteria or random variation attributed to the scorers.

The scorecard described here for assessing the operational characteristics of POC products can help improve the systemic validation of diagnostic tests, an area which is currently deficient [12][14]. Laboratories and public health managers faced with the challenge of selecting appropriate POC diagnostic devices can use this tool to help guide investment in new technologies and their deployment. This is particularly relevant because a number of new POC technologies are becoming available at a time when funding to expand treatment programs for HIV and other diseases is less certain. Careful selection and deployment of technologies that have both good technical performance and the appropriate operational features will be required.


We are grateful to Wilson da Silva (Vanderbelt University), Solon Kida (Association of Public Health Laboratories and Clotilde Nhumalo (Instituto Nacional de Saúde) for scoring the technologies.

Author Contributions

Conceived and designed the experiments: JL OT JQ TP IJ. Performed the experiments: JL NS OT. Analyzed the data: JL OL TP IJ. Contributed reagents/materials/analysis tools: JL NS OL JQ IJ. Wrote the paper: JL TP IJ OL.


  1. 1. UNAIDS Outlook Report 2010. UNAIDS website. Available: Accessed 2012 Apr 2.
  2. 2. Yager P, Domingo GJ, Gerdes J (2008) Point-of-care diagnostics for global health. Annu Rev Biomed Eng 10: 107–144. doi: 10.1146/annurev.bioeng.10.061807.160524
  3. 3. Peeling RW, Mabey D (2010) Point-of-care tests for diagnosing infections in the developing world. Clin Microbiol Infect 16: 1062–1069. doi: 10.1111/j.1469-0691.2010.03279.x
  4. 4. Schito ML, Peter TF, Cavanaugh S. Piatek AS, Young GJ, et al. (2012) Opportunities and challenges for cost-efficient implementation of new point-of-care diagnostics for HIV and tuberculosis. J Infect Dis 205 Suppl: S169–180. doi: 10.1093/infdis/jis044
  5. 5. Jani IV, Sitoe NE, Alfai ER, Chongo PL, Quevedo JI, et al. (2011) Effect of point-of-care CD4 cell count tests on retention of patients and rates of antiretroviral therapy initiation in primary health clinics: an observational cohort study. Lancet 378: 1572–1579. doi: 10.1016/s0140-6736(11)61052-0
  6. 6. McCoy D, Chand S, Sridhar D (2009) Global health funding: how much, where it comes from and where it goes. Health Policy 24: 407–417. doi: 10.1093/heapol/czp026
  7. 7. Ravishankar N, Gubbins P, Cooley RJ, Leach-Kemon K, Michaud CM, et al. (2009) Financing of global health: tracking development assistance for health from 1990 to 2007. Lancet 373: 2113–2124. doi: 10.1016/s0140-6736(09)60881-3
  8. 8. Pai NP, Pai M (2012) Point-of-care diagnostics for HIV and tuberculosis: landscape, pipeline, and unmet needs. Discov Med 68: 35–45.
  9. 9. Murtagh MM (2011) UNITAID Technical Report: HIV/AIDS diagnostic landscape; 2nd edition. UNITAID website. Available: Accessed 2012 Apr 2.
  10. 10. TDR Diagnostics Evaluation Expert Panel (2008) Evaluation of diagnostic tests for infectious diseases: general principles. Nature Reviews: Microbiology S21–33.
  11. 11. Stevens W, Gelman R, Glencross DK, Scott LE, Crowe SM, et al. (2008) Evaluating new CD4 enumeration technologies for resource-constrained countries. Nature Reviews: Microbiology; Evaluating Diagnostics: The CD4 Guide S36. doi: 10.1038/nrmicro2000
  12. 12. Jani IV (2010) The need for systematic evaluations of diagnostic tests. Clin Infect Dis 51: 609–610. doi: 10.1086/655763
  13. 13. Reid MC, Lachs MS, Feinstein A (1995) Use of methodological standards in diagnostic test research. Getting better but still not good. J Amer Med Assoc 274: 645–651. doi: 10.1001/jama.1995.03530080061042
  14. 14. Small PM, Perkins MD (2000) More rigour needed in trials of new diagnostic agents for tuberculosis. Lancet 356: 1048–1049. doi: 10.1016/s0140-6736(00)02724-0