20 February 2023

Types of Proficiency Testing Scheme

 

TYPES OF PROFICIENCY TESTING SCHEME

A.1

GENERAL

 

Proficiency testing has become an essential aspect of laboratory practice in all areas of testing, calibration, and inspection. Proficiency testing schemes vary according to the needs of the sector in which they are used, the nature of the proficiency test items, the methods in use and the number of participants. However, in their simplest form, most proficiency testing schemes possess the common feature of comparison of results obtained by one laboratory with those obtained by one or more different laboratories.

The nature of the test or measurement performed in Proficiency Testing Schemes governs the method of comparing performance. There are three basic types of laboratory examinations: quantitative, qualitative, and interpretive.

â The results of a quantitative measurement are numerical and are reported on an interval or a ratio scale. Test for quantitative measurement may vary in their precision, trueness, analytical sensitivity, and specificity. In quantitative proficiency testing schemes, numerical results are usually analysed statistically.

â The results of qualitative tests are descriptive and reported on a categorical or ordinal scale, e.g. identity of microorganisms, or by identification of the presence of a specific measurement (such as a drug or a grading of a characteristic). Assessment of performance by statistical analysis may not be appropriate for qualitative examinations.

â In interpretive tests, the “proficiency testing items” is a test result (e.g. descriptive morphology statement), a set of data (e.g. to determine a calibration line) or other set of information (e.g. a case study), concerning an interpretative feature of the participant’s competence.

Other Proficiency Testing scheme have additional features depending on their objective, as outlines in definition 3.7, Note 1, (a) to (h). Some common applications of those types of Proficiency Testing are discussed below and illustrated in Figure A.1. These schemes may be “single occasion” and performed once, or “continuous” and performed at regular intervals.

A.2

SEQUENTIAL PARTICIPATION SCHEMES

 

Sequential Participation Schemes (sometimes known as measurement comparison schemes) involve the Proficiency Test item being circulated successively from one participant to the next (i.e., sequential participation), or occasionally circulated back to the Proficiency Testing provider for rechecking. Model 1 in Figure A.1 provides a brief summary of this type of design, and the key features are typically those described below.

a.     A reference laboratory that is capable of providing a metrologically traceable assigned value with sufficiently small measurement uncertainty and reliability for the Proficiency Test item is used. For categorical or ordinal properties, the assigned value should be determined by consensus of experts or other authoritative source. It may be necessary for the Proficiency Test item to be checked at specific stages during the conduct of the Proficiency Testing scheme, in order to ensure that there is no significant change in the assigned value. 

b.     The individual measurement results are compared with the assigned value established by the reference laboratory. The coordinator should take into account the claimed measurement uncertainty of each participant, or the claimed level of expertise. It may be difficult to compare results on a group basis as there may be relatively few participants having measurement capabilities that closely match each other.

c.      Schemes involving sequential participation take time (in some cases, years) to complete. This causes a number of difficulties, such as

â Ensuring the stability of the item;

â The strict monitoring of the circulation among participants and the time allowed for measurement by individual participants, and

â The need to supply feedback on individual performance during the scheme’s implementation, rather than waiting until it finishes.

d.     Proficiency test items (measurement artifacts) used in this type of Proficiency test can include, for example, measurement reference standards (e.g. resistors, micrometers, and frequency counters) or, in medical programmes, histology slides with confirmed diagnosis.

e.     Schemes that follow this design but that are limited to situations where a single participant is tested independently are often called “measurement audits.”

f.      In some situations, the assigned value for a Proficiency Test item may be determined by consensus, after all participants (or in some situations, a subset of participants) have completed the measurement comparison

A.3

SIMULTANEOUS PARTICIPATION SCHEMES

 

A.3.1

General

 

Simultaneous participation Proficiency Testing Scheme usually involve randomly selected sub–samples from a source of material being distributed simultaneously to participants for concurrent testing. In some schemes, participants are required to take samples, which are then considered to be the Proficiency Test items for analysis. After completion of the testing, the results are returned to the Proficiency Testing Provider and compared with the assigned value(s) to give an indication of the performance of the individual participants and the group as a whole. Examples of Proficiency Test items used in this type of scheme include food, body fluids, agricultural products, water, soils, minerals and other environmental materials. In some cases, separate portions of previously established reference materials are circulated. Advice or educational comments are typically part of the report returned to participants by the Proficiency Testing Provider with the aim of promoting improvement in performance. Model 2 in Figure A.1 represents typical Proficiency Testing schemes of this type, usually for testing laboratories. Model 2 in Figure A.1 represents typical Proficiency Testing schemes of this type, usually for testing laboratories. Model 3 presents a type of scheme that is frequently used in conjunction with simultaneous Proficiency Testing Schemes, for oversight or educational purposes.

As discussed in Annex B, assigned values for these Proficiency Testing schemes may be determined in a variety of ways. However, either evaluations of performance are based on consensus values form participants (all participants, or a subset of “experts”) or evaluations can be on the basis of independently determined assigned values.

Known value schemes use assigned values that are determined independently of the participants and involve preparations of Proficiency test items with a number of known measurements or characteristics. Certified reference materials can also be used in these schemes, as their certified value and measurement uncertainty can used directly. A direct comparison can also be made between a proficiency test item and a certified reference material under repeatability conditions. However, care should be taken to ensure that the certified reference material is closely comparable with a Proficiency test item. Proficiency test items from previous proficiency testing rounds may be used in this type of scheme if the item has demonstrated stability. One special application of Proficiency Testing, often called “blind” Proficiency Testing is where the Proficiency test item is indistinguishable from normal customer items or samples received by the laboratory. This type of Proficiency Testing can be difficult since it requires coordination with a normal laboratory customer. In addition because, of unique packaging and shipping needs, bulk processing is usually not feasible and homogeneity testing is difficult.

A.3.2

Split–level Designs

 

A common design for Proficiency Testing is the “split–level” design, where similar (but not identical) levels of measurement are included in two separate Proficiency Test items. This design is used to estimate the participant’s precision at a specific level of a measurement. It avoids problems associated with replicate measurements on the same Proficiency Test item, or with the inclusion of two identical Proficiency test items in the same Proficiency Testing round.

A.3.3

Split–sample Testing Schemes

 

One special type of Proficiency Testing design that is often used by participants customers and some regulatory bodies is the “split–sample design”

Typically, split–sample proficiency testing involves comparisons of the data produced by small groups of participants (often only two). In these Proficiency Testing schemes, samples of a product of a material are divided into two or more parts, with each participant testing one part of the sample (see Figure A.1, model 5). Uses for this type of scheme include identifying poor accuracy, describing consistent bias and verifying the effectiveness of corrective actions. This design may be used to evaluate one or both participants as suppliers of testing services, or in cases where there are too few participants for appropriate evaluation of results. Under such schemes, one of the participants may be considered to operate at a higher metrological level (i.e. lower measurement uncertainty), due to the use of reference methodology and more advance equipment, etc. or through confirmation of its own performance through satisfactory participation in a recognized interlaboratory comparison scheme. Its results are considered to be the assigned values in such comparisons and it may act as an advisory or mentor laboratory to the other participants comparing split–image data with it.

 

A.3.4

Partial–process schemes

 

Special types of Proficiency Testing involve the evaluation of participants abilities to perform parts of the overall testing or measurement process. For example, some existing Proficiency Testing Schemes evaluate participants abilities to transform and report a given set of data (rather than conduct the actual test or measurement), to make interpretations based on a given set of data or Proficiency Testing items such as stained blood films for diagnosis, or to take and prepare samples or specimens in accordance with a specification.

A.4

EXTERNAL QUALITY ASSESSMENT PROGRAMMES (EQA)

 

EQA Programmes (such as those provided for laboratory medicine examinations) offer a variety of interlaboratory comparison schemes based on this traditional Proficiency Testing model, but with broader application to the schemes described in A.2 and A.3 and illustrated in Figure A.1. Many EQA Programmes are designed to provide insight into the complete path of workflow of the laboratory, and not just the testing processes. Most EQA programmes are continuous schemes that include long term follow up of laboratory performance. A typical feature of EQA Programmes is to provide education to participants and promote quality improvement. Advisory and educational comments comprise part of the report returned to participants to achieve this aim.

 

Some EQA Programmes assess performance of pre–analytical and post–analytical phases of testing, as well as the analytical phase. In such EQA Programmes, the nature of the Proficiency Test item may differ significantly from that used in traditional Proficiency Testing schemes. The “proficiency test item, requiring the participant to select an appropriate approach to testing or interpretation of results, and not just to perform the test. In “sample view” schemes, participants may be required to provide the “’proficiency test items” to the EQA provider (see Figure A.1, model 4). This may take the form of a processed specimen or sample (e.g. stained slide or fixed tissue), laboratory data (e.g. test results, laboratory reports or quality assurance/control records) or documentation (e.g. procedures or method verification criteria).

Model 1

Sequential

Model 2

Simultaneous

Model 3

Interpretive

Model 4

Sample Review

Model 5

Split Sample

Produce/procure test items artefacts

 

Produce/procure test items

 

 

Produce test items, develop questionnaire, or case study

Determine test items to be received from participants

Participants agree on analytes and sample types for comparison

Determine assigned value and its uncertainty

Determine assigned value and acceptable range of results

Distribute questionnaire, case sturdy or test items to participants

Distribute specifications to participants

Participant(s) split appropriate samples and send to others

Distribute to first participant

Distribute test items to participants

Receive results and interpretations from participants

Receive test items from participants

Participants share results or send to a coordinator

Have participants return item or send to next participant

Receive results and method information from participants

Determine acceptable criteria for responses and interpretations

Determine acceptable criteria for responses

Graph or otherwise compare results in this and previous studies

Review participant results and uncertainty for acceptability

Compare participants results and method information with acceptable range

Compare participant results and interpretation with criteria

Compare participants test items with criteria

Compare with pre– established criteria or discuss needs for action

Produce reports and issue advisory /educational comments

Produce reports and issue advisory/educational comments

Produce reports and issue advisory/educational comments

Produce reports and issue advisory/educational comments

Produce reports and records with any agreed conclusions, including data and graphs

No comments: