Emailing: _0125.htm

I think we need a definition of accuracy and precision - I think it is the
last one in the precision definitons that we are looking for.

I think that standard error is what we mean by accuracy?? if it is - what is
the correct expression - I do not think it has units.

Cheers,Sam

Sam

(attachments)

_0125.htm (636 Bytes)

I think that's a reasonable suggestion - especially since different people interpret these (and related) terms differently. I've picked out a couple of 'definitions' in the form that pathologists use (and in my experience this is a group who are precise and pedantic in a very professional way! - and they are the generators of much of the quantitative data in an EHR). In fact, they have come from a draft European standard from CEN/TC and are consistent with ISO and other international bodies. (Taken from R. Haeckel (Ed.) "Evaluation Methods in Laboratory Medicine")

"ACCURACY is the closeness of the agreement between the result of a measurement and a true value of the measurand". I.e it depends on knowledge of the TRUE value of the thing being measured, as during analyser calibration, when a consistent BIAS might be noted. It 'is usually expressed numerically by statistical measures of the inverse concept, INACCURACY of measurement', which is defined as "discrepancy between the result of a measurement and the true value of a measurand" - and which 'is usually expressed numerically as the error of measurement'. 'Inaccuracy, when applied to sets of results, describes a combination of random error of measurement and the systematic error of measurement.'

"PRECISION of a measurement is the closeness of agreement between independent results of measurement obtained by a measurement procedure under prescribed conditions". I.e. the variation obtained with repeated measurements on a single specimen. Precision thus 'depends only on the distribution of random errors of measurement. It is usually expressed numerically by statistical measures of imprecision of measurements'. "IMPRECISION is the dispersion of independent results of measurements obtained by a measurement procedure under specified conditions". 'It is usually expressed numerically as the repeatability standard deviation or reproducibility standard deviation of the results of measurement.' When applied to sets of results, imprecision 'depends solely on the dispersion of random error of measurement and does not relate to the true value of the measureable quantity'.

Cheers,
Tony

Tony Grivell wrote:

I think that's a reasonable suggestion - especially since different people interpret these (and related) terms differently. I've picked out a couple of 'definitions' in the form that pathologists use (and in my experience this is a group who are precise and pedantic in a very professional way! - and they are the generators of much of the quantitative data in an EHR). In fact, they have come from a draft European standard from CEN/TC and are consistent with ISO and other international bodies. (Taken from R. Haeckel (Ed.) "Evaluation Methods in Laboratory Medicine")

Tony, can you give me the full reference to this? I'll quote it in the Reference Model

"ACCURACY is the closeness of the agreement between the result of a measurement and a true value of the measurand". I.e it depends on knowledge of the TRUE value of the thing being measured, as during analyser calibration, when a consistent BIAS might be noted. It 'is usually expressed numerically by statistical measures of the inverse concept, INACCURACY of measurement', which is defined as "discrepancy between the result of a measurement and the true value of a measurand" - and which 'is usually expressed numerically as the error of measurement'. 'Inaccuracy, when applied to sets of results, describes a combination of random error of measurement and the systematic error of measurement.'

So this means that "+/-5%" is a statistically-based measure of the inaccuracy of the instrument or method being used. It does not make any statement about the individual measured value with respect to the "true" value - only that statistically, it is known to be within the 10% wide band centred on the true value.

"PRECISION of a measurement is the closeness of agreement between independent results of measurement obtained by a measurement procedure under prescribed conditions". I.e. the variation obtained with repeated measurements on a single specimen. Precision thus 'depends only on the distribution of random errors of measurement. It is usually expressed numerically by statistical measures of imprecision of measurements'. "IMPRECISION is the dispersion of independent results of measurements obtained by a measurement procedure under specified conditions". 'It is usually expressed numerically as the repeatability standard deviation or reproducibility standard deviation of the results of measurement.' When applied to sets of results, imprecision 'depends solely on the dispersion of random error of measurement and does not relate to the true value of the measureable quantity'.

a) so what about the "definition" of precision as significant places in a number, i.e. the level of preciseness to which a numerical result is reported - which is logically related to the definition above, since there is no point reporting to a higher degree of precision than actually available in the real-world measuring process...

b) the above definition would imply that we should report a standard-deviation of a notional population of meansurements of the same actual value....

c) should there be a merged definition of these concepts, as per this suggestion in HL7:

(quoting Gunther Schadow from CQ/MnM lists in HL7)
The NIST guide for uncertainty in measurements says that the
traditional notions of accuracy vs. precision should be superceded
by the one concept of uncertainty. So, any given measurement you
take is really a probability distribution over the measurement
domain. The probability distribution is typically described
parametrically. The NIST guide goes into quite specifics about
that and I have to say that it went a little bit past my memory.
But one of the ways they do specify their uncertainty is by
giving the mean and a standard deviation. That's often assuming
that your distribution is normal, which it often is due to the
central limit theorem.

But if it isn't, you need to know what your distribution type
and its parameters are.

In HL7 v3 we have a data type called Parametric Probability
Distribution, which is a generic type extension that works with
any base data type. In most cases we will have a PPD<PQ>.
The PPD<PQ> ends up having the properties:

   mean value
   unit
   distribution type code
   standard deviation

The distribution type code can distinguish normal, gamma, beta,
X^2, etc. The table of distribution types also summarizes how the
parameters mu and sigma relate to the specific parameters that
are usually used for each distribution type.

During the design of this we thought we would better use the
specific parameters for each distribution type, but those turned
out to be all deriveable from mu and sigma. The advantage of sending
mean and standard deviation consistently is that even if a
receiver does not understand the distribution type, he will get
a pretty good idea about the measurement from just looking at
mu and sigma.

I would encourage anyone with an interest in these matters to
review the V3DT semantics spec and particularly the table of
distribution types. This part has not received the same amount
of review as other parts, so errors are possible.

Recently in work we are doing here, I came to appreciate the
advantage of using moments as parameters instead of mean and
standard deviation. Well, first moment is the same as mean
but second moment has advantages over standard deviation when
combining population statistics. However, second moment and
standard deviation are also easily derivable. But I understand
that one could specify higher order moments to describe a
distribution.

regards
-Gunther

Tony Grivell wrote:

I think that’s a reasonable suggestion - especially since different people interpret these (and related) terms differently. I’ve picked out a couple of ‘definitions’ in the form that pathologists use (and in my experience this is a group who are precise and pedantic in a very professional way! - and they are the generators of much of the quantitative data in an EHR). In fact, they have come from a draft European standard from CEN/TC and are consistent with ISO and other international bodies. (Taken from R. Haeckel (Ed.) “Evaluation Methods in Laboratory Medicine”)

Tony, can you give me the full reference to this? I’ll quote it in the Reference Model

The quoted material is from a Glossary of Terms and Definitions in Rainer Haeckel (Ed.) “Evaluation Methods in Laboratory Medicine”, 1993, published by VCH Publishers, New York (ISBN 1-56081-274-5 (New York)). Because this is focused on laboratory technology, and because the Glossary acknowledges the sources it used (summarised above), you would probably prefer to use these primary sources - which are listed as:

  1. Comite Europeen de Normalisation (CEN)/TC 140 “In-vitro diagnostic systems”/WG 4 (1991). Measurement of quantities in samples of biological origin. Presentation of reference measurement procedures. Draft, Annex B.

  2. BIPM, IEL, IFCC, ISO, IUPAC, OIML (1991) International vocabulary of basic and general terms in metrology (VIM). Draft revision, Geneva, ISO 1989 (with some amendments from 1991).

“ACCURACY is the closeness of the agreement between the result of a measurement and a true value of the measurand”. I.e it depends on knowledge of the TRUE value of the thing being measured, as during analyser calibration, when a consistent BIAS might be noted. It ‘is usually expressed numerically by statistical measures of the inverse concept, INACCURACY of measurement’, which is defined as “discrepancy between the result of a measurement and the true value of a measurand” - and which ‘is usually expressed numerically as the error of measurement’. ‘Inaccuracy, when applied to sets of results, describes a combination of random error of measurement and the systematic error of measurement.’

So this means that “+/-5%” is a statistically-based measure of the inaccuracy of the instrument or method being used. It does not make any statement about the individual measured value with respect to the “true” value - only that statistically, it is known to be within the 10% wide band centred on the true value.

I would believe that to be a reasonable interpretation, if the +/- 5% is referred to as accuracy.

“PRECISION of a measurement is the closeness of agreement between independent results of measurement obtained by a measurement procedure under prescribed conditions”. I.e. the variation obtained with repeated measurements on a single specimen. Precision thus ‘depends only on the distribution of random errors of measurement. It is usually expressed numerically by statistical measures of imprecision of measurements’. “IMPRECISION is the dispersion of independent results of measurements obtained by a measurement procedure under specified conditions”. ‘It is usually expressed numerically as the repeatability standard deviation or reproducibility standard deviation of the results of measurement.’ When applied to sets of results, imprecision ‘depends solely on the dispersion of random error of measurement and does not relate to the true value of the measureable quantity’.

a) so what about the “definition” of precision as significant places in a number, i.e. the level of preciseness to which a numerical result is reported - which is logically related to the definition above, since there is no point reporting to a higher degree of precision than actually available in the real-world measuring process…

Although this seems to be a consistent sub-set of the above, I’m not sure that it’s not simply an older (and ‘imprecisely-defined’ concept). It would necessarily be a coarser and less informative quantity than a SD)

b) the above definition would imply that we should report a standard-deviation of a notional population of meansurements of the same actual value…

That’s my understanding of how the precision of a particular instrument/technique would be determined (formally-accredited pathology labs would record such a precision determination (incidentally of a known ‘standard’ material) for each automated instrument at least once each day. In practice (although this is a complex but important area), pathologists aim to have instrument imprecision for every analyte much smaller in magnitude than either of (a) normal intra-human biological variability for that analyte or (b) the range of short-term change that would be interpreted by a clinician as having physiological significance to the health of a patient. By doing this, instrument imprecision is implied to be irrelevant in making real-world clinical decisions when reported results are either out of reference-range or have changed dramatically over time for an individual

c) should there be a merged definition of these concepts, as per this suggestion in HL7:

(quoting Gunther Schadow from CQ/MnM lists in HL7)
The NIST guide for uncertainty in measurements says that the
traditional notions of accuracy vs. precision should be superceded
by the one concept of uncertainty. So, any given measurement you
take is really a probability distribution over the measurement
domain. The probability distribution is typically described
parametrically. The NIST guide goes into quite specifics about
that and I have to say that it went a little bit past my memory.
But one of the ways they do specify their uncertainty is by
giving the mean and a standard deviation. That’s often assuming
that your distribution is normal, which it often is due to the
central limit theorem.

My feeling is that we are not really comparing apples with apples. NIST tends to deal with ‘absolute’ quantities (maybe even fundamental constants), each of which has a single fixed value the knowledge of which we asymptote towards with better technology, and with probably vast numbers of measurements averaged. In contrast, a blood glucose, for example will vary between individuals and over short time-frames.

But if it isn’t, you need to know what your distribution type
and its parameters are.

In HL7 v3 we have a data type called Parametric Probability
Distribution, which is a generic type extension that works with
any base data type. In most cases we will have a PPD.
The PPD ends up having the properties:

mean value
unit
distribution type code
standard deviation

The distribution type code can distinguish normal, gamma, beta,
X^2, etc. The table of distribution types also summarizes how the
parameters mu and sigma relate to the specific parameters that
are usually used for each distribution type.

During the design of this we thought we would better use the
specific parameters for each distribution type, but those turned
out to be all deriveable from mu and sigma. The advantage of sending
mean and standard deviation consistently is that even if a
receiver does not understand the distribution type, he will get
a pretty good idea about the measurement from just looking at
mu and sigma.

This is the most general, I guess - but it is said that virtually all biological variation is normally-distributed (the central-limit theorem working on the situation that most biological parameters are the result of many independent influences during development.