Inter Laboratorycomparabilityofclinicalchemistrytesting Anewperspective
Inter Laboratorycomparabilityofclinicalchemistrytesting Anewperspective
net/publication/337317865
Article in International Journal of Sciences Basic and Applied Research (IJSBAR) · November 2019
CITATIONS READS
0 1,661
3 authors, including:
Perez Quartey
Garden City University College
17 PUBLICATIONS 28 CITATIONS
SEE PROFILE
All content following this page was uploaded by Perez Quartey on 19 November 2019.
http://gssrr.org/index.php?journal=JournalOfBasicAndApplied
---------------------------------------------------------------------------------------------------------------------------
a,b,c
Department of Medical Laboratory Technology, Garden City University College, Kenyase, Ghana
a
Email: [email protected]
b
Email: [email protected]
c
Email: [email protected]
Abstract
Background/Aim: Existing external quality assurance proficiency testing may not reflect routine „real-world‟
daily testing proficiency. The overall aim of the study was to conduct an inter-laboratory comparability testing
of some routinely measured clinical chemistry analytes by selected private medical laboratories using the
mentor-adept approach. Methodology: Aliquots of freshly separated plasma from a single individual (1ml) each
were sent to each of the selected labs and a Teaching hospital for same-day testing. All samples were sent as
„„blinded-samples‟‟ (labelled with anonymous names, ages and attached lab request forms with clinical
diagnoses) so that they would be tested as real patient samples. Results: All the labs met the acceptability
criteria range for both the z-score and Precision Index. However, many of the labs had their total analytical
errors for the tests outside the allowable total error ranges with both European and CLIA recommendations.
Conclusion: There are relative similarities in the z-score (inter-laboratory bias) and Precision Index (inter-
laboratory precision) among the labs. However, many of the labs did not meet recommended analytical goals for
total analytical errors on individual samples run in a day. In the light of the findings, it is highly recommended
that though laboratories should be aware of conduction of periodic external quality assurance exercises, such
exercises should be done using “blinded-samples” as utilized in this current study without prior notification of
the day and time of testing.
Keywords: Inter-laboratory; Mentor-adept; External Quality Assurance; z-score; Precision Index; Total
Analytical Error.
------------------------------------------------------------------------
* Corresponding author.
196
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
1. Introduction
Quality of clinical chemistry testing is very critical in diagnostic and overall patient assessment by expressing
reliability in testing results. Clinical chemistry testing presents as a very critical aspect of general laboratory
testing and it takes highly automated measurement methods and systems, combined with advanced information
technologies to accomplish this all-important task [1]. The overall uncertainty of the high-volume measurement
methods in clinical chemistry has decreased substantially in recent decade tough bias still remains as a
significant challenge. A key component of laboratory quality management system is Quality Assurance (QA)
which includes both Internal Quality control (done on regular daily basis) and External Quality Assurance
(EQA). EQA is done periodically, in collaboration with a higher laboratory authority through Proficiency
Testing (PT) or Inter-lab comparability (ILC) testing. with PT being the ideal method of choice in evaluating
laboratory performance [2, 3, 4, 5]. In cases where for some reasons there is lack of proficiency testing schemes,
inter-laboratory comparisons (ILCs) are preferred by accreditation bodies [5]. Typical method of assessment of
ILC results have been described with standard statistical protocols [4, 6, 7] with the use of „reference‟ materials
prepared by the QA provider for measurements. When it is not possible to apply this, the mentor-adept approach
of inter-laboratory comparison may be used [1]. Measurement results of analytes from the expert (mentor-lab),
„„considered‟‟ to be devoid of bias [1] are compared with that of the participating laboratories. The inter-
laboratory analysis of the results is done using recommended statistical methods. Statistically stringent or robust
analysis „„consensus‟‟ values calculated from participant results are recommended for the estimation of an
assigned value and are typically used for calculating the z-score, precision index (PI) and Total analytical error
(TE) [4, 5, 6, 8]. Presently, we are not aware of any external quality assurance exercise conducted for clinical
chemistry testing by medical laboratories in Ghana.
The overall aim of the study was to conduct an inter-laboratory comparability testing of some routinely
measured clinical chemistry analytes by some selected private medical labs in Kumasi (the second largest city in
the country) using the mentor-adept approach of performance evaluation.
Approval for the research was received from the Kwame Nkrumah University of Science and Technology
Research Committee.
Using the Mentor-Adept approach, the cross-sectional study was conducted on six (6) selected private
laboratories.
197
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
2.4. Adept-labs
Six (6) selected private labs in the Kumasi metropolis. The locations of two of the labs are relatively farther
away from the mentor-lab whiles the other four are in the vicinity of the mentor-lab. All labs were assigned
alphabetical codes for ethical anonymity.
After thoroughly explaining the aim, objectives and procedures of the study, venous blood was collected from a
single healthy donor. Freshly separated plasma was aliquoted into clean eppendorf tubes. Seven (7) of the 1ml
plasma aliquots were sent to each of the selected labs for same-day testing and reporting. All samples were sent
as „„blinded-samples‟‟ (labeled with anonymous names, ages and attached lab request forms with clinical
diagnoses) so that they would be tested as real patient samples.
2.6. Measurements
Comparability of precision, accuracy and total analytical error were assessed using standard lab quality
assurance statistical methods. Inter-laboratory comparability of analytical bias was assessed by calculation of
„„Consensus-based‟‟ Standard Deviation Index (z-score), using robust calculation involving exclusion of outlier
results [4].
Classification criteria:
0 to ±1 = Satisfactory
>1 to 2 = Acceptable
± > 2 = Unacceptable
198
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Classification criteria:
0 to ±1 = Satisfactory
>1 to 2 = Acceptable
± > 2 = Unacceptable
Inter-laboratory comparability of overall analytical performance was assessed with total analytical error and
comparing it with two recommended specifications of allowable total error (TEa) [4, 9, 10].
a. Total analytical error calculated as 1.65 x CV of lab % bias, where % bias is;
This error is compared with European recommended allowable total error for each analyte by Ricos and his
colleagues 2014 [9].
b. Lab means compared with CLIA [10] recommended allowable total error calculated as mentor mean
(target value) ± recommended % bias.
3. Results
The general characteristics, Z-scores and PI-scores of the participant labs are shown on results of the z-score
distributions in the participating labs are illustrated on table 1, 2 and 3 respectively. The results showed that all
the mean results of tests were within the acceptability criteria range (± 2 standard deviation points from the
consensus mean). All the mean results of Lab A were in the satisfactory range. The mean results that moved
from the satisfactory range into the acceptable range were urea and creatinine for Lab B, sodium and creatinine
for Lab C, HDL for Labs D and F, and both urea and triglyceride for Lab E. Additionally, all the participant labs
met the acceptability criteria for precision index (± 2). A detailed analysis however showed that the PI moved
from the satisfactory range to the acceptable range for sodium (Lab B) potassium (Lab C) and both LDL and
HDL for Lab D.
199
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Scores with no superscript means values are in satisfactory range (less than ± 1.0); lab means with superscript (α) means values
200
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Scores with no superscript means PI values are in satisfactory range (less than ± 1.0); Scores with superscript (α) means PI values are in acceptable range
(± 1.1 – 2.0).
201
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Table 4: Total analytical errors of the study labs compared by the European recommendations.
TEa: Allowable total error; β: Total analytical error beyond the allowable total error range for the analyte.
202
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Table 5: Total analytical errors (TE) of the study labs compared by CLIA 2019 recommendations.
Analyte (mmol/l) Lab A Lab B Lab C Lab D Lab E Lab F Mentor- lab mean TEa limits
β β β β
Sodium 146.7 121.5 105.1 145.5 137.9 143.3 149.2 145 - 153
β β
Potassium 4.1 3.9 3.5 4.1 3.8 4.3 4.2 3.9 - 4.5
β β β β β
Urea 3.8 2.5 3.8 3.6 5.3 4.7 2.3 2.09 - 2.51
β β β β
Creatinine 75.5 137.6 71.4 87.9 80.8 60.9 70.0 63.0 – 77.0
β β β β β
Cholesterol 3.8 3.2 3.9 4.9 3.8 3.8 3.21 2.89 – 3.53
β β β β
HDL 1.04 0.95 1.13 1.83 1.37 0.69 0.92 0.74 – 1.10
β β β β
Triglycerides 1.61 1.3 1.8 1.70 0.96 1.27 1.14 0.97 – 1.31
TEa: Allowable total error (CLIA, 2019), β: Total analytical error beyond the allowable total error range for the analyte.
203
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Table 6: Comparison of overall performance of the study labs based on European and CLIA total error
specifications.
Results are expressed as number (%) of analytes that passed allowable total errors by the recommended
specifications.
The overall performance was calculated in percentage points of the number of analyte measurements that passed
the recommended specifications out of the total number of measured analytes. The overall performance as
illustrated in tables 4, 5 and 6 showed that many of the labs had their total analytical errors for the tests outside
the allowable total error ranges with only Lab B scoring 4 correct analytical points (57.1%). Labs C, D and F
scored 1(14.3%) correct analytical point whilst Labs A and E did not score any point; with the use of the
European classification. In comparison to CLIA recommendation however, Lab A scored 4 (57.1%), Lab B
scored 5 (71.4%), Lab C scored 1 (14.3%), Labs D and Lab F scored 2 (28.6%), whilst lab E did not score any
point.
4. Discussion
Laboratory external quality assessment describes a method or process that allows testing conducted by a
laboratory to be compared to that of a source outside the laboratory which may be a peer group of laboratories
or a reference laboratory. For accrediting bodies and regulatory agencies, participation in EQA program
provides objective evidence of the quality of testing of patients‟ specimens [9,10]. In general, there is limited
data on EQA of clinical chemistry testing in Ghana. Presently, there is no oversight body that organizes External
Quality assurance tests of clinical chemistry in Ghana. This is the first study to report on the inter-laboratory
comparability of clinical chemistry testing in Ghana, using the mentor-adept approach. Without participation in
high-quality inter-laboratory comparability program, a laboratory may be unaware of periodic gradual or sudden
changes in performance of the testing system that may be caused by factors such as change in reagents or
calibrators, standardization changes, or instrument software changes. Participation in an inter-laboratory
comparability program can therefore offer awareness of shifts and trends and help verify the reliability and
quality of testing. An inter-laboratory program can also increase the confidence of laboratories in participation
in proficiency testing surveys.
204
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
This study typically assessed, the consensus-based z-score, precision index, and total analytical error of the
selected laboratories in some routinely measured clinical chemistry tests [4, 8, 11, 12]. This study presents the
likely events of the „„real-life‟‟ daily results presented to patients and requesting physicians alike when samples
are handled as routine patient samples for laboratory analysis and not as samples for an external quality
assurance work. In the current study, the results showed that all the mean test results (z-scores) and the Precision
indexes of the labs were within the acceptability criteria range (± 2). The exclusion of individual laboratory
results outliers for calculation of the „„consensus or peer-group‟‟ mean and standard deviation implies that
interpretation of the z-score and PIs typically relies on the statistical assumption that the results are normally
distributed and therefore introduces similarities. These parameters do not give an indication of the laboratory‟s
analytical performance but rather an idealized performance of the study group, though they are useful in
identifying very questionable analytical deficiencies in the event of a participant lab having a score for an
analyte outside the acceptable range. Thus, it has been advocated that in a single round of proficiency testing,
providers, participants and end-users should avoid classification and ranking of laboratories based on these
scores [13, 14,15]. The comparison of total analytical error for the analytes by the study labs using both
European and CLIA recommendations showed that many of the labs had their total analytical errors for the tests
outside the allowable total error ranges. A detailed comparison showed that most of the study labs failed with
the use of European recommendations as compared to CLIA recommendations. These discrepancies need
further investigations for reconciliation and harmonization in the light of differences in specifications and
recommendations.
Anytime a sample is sent to the lab, the requesting physician or patient does not care about the measurement
system, testing method or competency of the medical lab technologist doing the testing. All they expect is a
reliable result, irrespective of whatever was primarily used to produce the results. For example, a client who
walks into a lab with a plasma potassium level of 3.4 mmol/l, irrespective of whatever system, method, time of
testing or lab personnel expects a report result within an acceptable range (as close enough to 3.4 mmol/l).
External quality assurance typically involves an accredited scheme provider preparing reference materials and
sending them to the participating laboratories for measurements [16,17]. The results from the participating labs
are then returned to the scheme provider for analysis and feedback reported to the participating labs. The whole
process of the proficiency testing is done with the conscious awareness of the participating labs of their
involvement in an EQA exercise. External quality assurance programs like proficiency testing and inter-
laboratory comparison provide substantial information to measure overall testing performance. However, they
do not give a measure of the daily accuracy, reproducibility and overall reliability of individual testing. In our
view, EQA surveys may merely provide a snapshot in time and should not be used as a substitute for daily
quality. When EQA is conducted with the conscious awareness of the participating labs that the testing is being
carried out as part of a formal EQA program, there is a strong likelihood of bias in adhering to strict internal
quality control standards with the sole aim of achieving good results. “Acceptable” performance in the EQA
testing may therefore not reflect or guarantee testing reliability on individual samples run as real daily work.
205
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
Our main issue was to find out the comparability of test results as run by the labs in their real daily testing.
Therefore, in the present study no contact was made with the testing personnel to identify any instrumental,
methodological or internal quality assurance differences between the labs included in the study. The results of
the labs were therefore considered as true results that they would have reported, and dispatched to clients on a
„„real-life‟‟ basis.
1. All samples were sent to the selected labs as “blinded-patient samples”. All samples were labelled with
anonymous patient names, ages, gender and attached with corresponding request forms with different
clinical diagnoses;
2. All the selected labs performed testing without a conscious awareness of any inter-laboratory
comparability exercise and all results were collected as „„normal patient‟‟ results as routinely
dispatched by the labs. All these show that the results give the most likely test reporting that would
have been produced by the labs as part of their routine daily runs.
5. Conclusion
This study reports the first assessment of inter-laboratory comparability of clinical chemistry testing using the
mentor-adept approach in Ghana. The study identified that there are relative similarities in the z-score (inter-
laboratory bias) and Precision Index (inter-laboratory precision) among the selected labs. However, there are
considerable variations in the total analytical errors of testing. The differences in total analytical error as
compared to recommended allowable total error criteria may be a better indication of overall analytical
performance. In general, many of the laboratories did not meet the recommended analytical goals on individual
samples run in a day. These observed analytical deficiencies require critical investigation into the internal
quality assurance practices of laboratories in the country.
5.1. Recommendations
It is highly recommended that though laboratories should be aware of conduction of periodic external quality
assurance exercises, such exercises should be done using “blinded-samples” without prior notification of the day
and time of testing as utilized in this current study.
References
[1]. Theodorsson E Magnusson B, Leito I. Bias in clinical chemistry. Bioanalysis 2014; 6(21), 2855–2875.
[2]. ISO, IEC 17025 (2005) General requirements for the competence of testing and calibration
laboratories. International Organization for Standardization/International Electrotechnical Commission,
Geneva. https://www.iso.org/standard/39883.html. Accessed Feb 2019.
[3]. EA-4/18 (2010) Guidance on the level and frequency of proficiency testing participation.
http://www.european-accreditation.org/publication/ea-4-18-inf. EA-4/18: 2010. Accessed Feb 2019.
206
International Journal of Sciences: Basic and Applied Research (IJSBAR) (2019) Volume 48, No 5, pp 196-207
[4]. Karkalousos P, Evangelopoulos A (2011). Quality Control in Clinical Laboratories, Applications and
Experiences of Quality Control, Prof. Ognyan Ivanov (Ed.), ISBN: 978-953-307-236-4.
http://www.intechopen.com/books/applications-and-experiences-of-quality-control/quality-control-
inclinical-laboratories. Accessed Feb 2019.
[5]. Szewczak E, Bondarzewski A. Is the assessment of interlaboratory comparison results for a small
number of tests and limited number of participants reliable and rational? Accred Qual Assur 2016;
21:91–100.
[6]. ISO 13528 (2005) Statistical methods for use In proficiency testing by interlaboratory comparisons.
International Organization for Standardization, Geneva. https://www.iso.org/standard/35664.html.
Accessed Feb 2019.
[7]. Hyltoft Peterson P, Ricos C, Stockl D, Libeer JC, Baadenhuijsen H, Fraser C, Thienpont L. „„Proposed
guidelines for the internal quality control of analytical results in the medical laboratory.‟‟ Eur J Clin
Chem Clin Biochem 1996; 34:983-999.
[8]. Ricos C, Alvarez V, Cava F, Garcia-Lario JV, Hernandez A, Jimenez CV, Minchinela J, Perich C,
Simon M (2014) Desirable specifications for total error, imprecision, and bias, derived from intra- and
inter-individual biologic variation. http://www.westgard.com/biodatabase1.htm. Accessed Feb 2019.
[9]. Sciacovelli L, Secchiero S, Zardo L, Plebani M. The role of the External Quality Assessment.
Biochemia Medica. 2010; 20(2):160–4.
[10]. Kettelhut MM, Chiodini PL, Edwards H, Moody A. External quality assessment programmes raise
standards: evidence from the UKNEQAS parasitology subschemes. J Clin Pathol. 2003; 56(12):927–
32.
[11]. ISO 13528 (2015) Statistical methods for use in proficiency testing by interlaboratory comparisons.
International Organization for Standardization, Geneva. https://www.iso.org/standard/56125.html.
Accessed Feb 2019.
[12]. 2019 CLIA proposed changes to PT acceptable limits. https://www. westgard.com/2019-clia-
changes.htm. Accessed April 2019.
[13]. Fearn T, Fisher S, Thompson M, Ellison SLR. “A decision theory approach to fitness-for purpose in
analytical measurement”, Analyst 2002; 127:818-824.
[14]. Kristensen GB, Meijer P. Interpretation of EQA results and EQA-based trouble shooting. Biochem
Med 2017 27: 49-62.
[15]. Sareen R. Illuminating Z-Score in External Quality Assessment for Medical Laboratory. Health Care
Current Review 2018, 6:3.
[16]. ISO/IEC 17043:2010. Conformity assessment – General requirements for proficiency testing. Geneva:
International Organization for Standardization; 2010.
[17]. ISO 15189:2013. Medical laboratories – Requirements for quality and competence. International
Organization for Standardization; 2013.
207
View publication stats