Chapter 8
Data Analysis and
Presentation*
STUART
c.
BLACK
Nuclear Radiation Assessment Division,
Environmental Monitoring System Laboratory,
PO Box 93478, Las Vegas, Nevada 89193-3478, USA
8.1 INTRODUCTION
Analysis of data begins with a procedure that is commonly termed
validation. For the purposes of this chapter, validation means a thorough
check of the analytical method to ensure that an adequate quality control
process has been used. For analysis of data from a research or a method
development project, quality control must have been an integral part of the
method used for analysis, as is required by good laboratory practice. If, on
the other hand, the data to be analyzed originate from a monitoring or
surveillance program used to estimate environmental contamination, then
a comprehensive quality assurance plan covering both the sampling
performed and the analytical procedure should exist, and validation then
consists of verifying that the data quality objectives set forth at the
beginning of the program have been met.
Since this handbook is devoted to sampling methods as well as
instrumental methods of analysis, the following sections will address those
quality control procedures that relate to the bias and precision of the data
and to the adequacy of the sample that was analyzed. Obviously, the fact
*
Notice: Although the work described in this chapter has been supported by the US
Environmental Protection Agency, it has not been subjected to Agency review and
therefore does not necessarily reflect the views of the Agency and no official
endorsement should be inferred.
335
C. N. Hewitt (ed.), Instrumental Analysis of Pollutants
© Elsevier Science Publishers Ltd 1991
336
INSTRUMENTAL ANALYSIS OF POLLUTANTS
that a positive result has been obtained from an analysis ofa sample has little
significance unless the result can be related to something of interest and
unless the uncertainty associated with that result is known. The
relationship may be to appropriate blank samples or to results obtained in
control areas or in control samples. The uncertainty of a result is a
combination of the systematic and random errors that are an integral part
of the sampling and analytical procedures used to obtain that result.
Common practice has been to use the terms precision and accuracy,
implying that systematic errors (bias) and accuracy are synonymous.
However, accuracy is some combination of precision and bias where the
precision is high and the bias is low. Any other combination is inaccurate.
The final section of this chapter addresses the various means of
presenting the data so that noninvolved persons can readily understand the
results that were obtained from a project. However, for documenting the
results of an experiment, a true data base is preferable. The data base may
be in tabular form or in some form of electronic storage. It should include
all results from the project together with the factors that were used to
obtain the results, such as the analytical error, extraction efficiency,
calibration data, blank values, volume corrections, and so forth. This will
permit a complete reconstruction of the final results if that should ever
become necessary or in case a different method of presentation were to be
desired.
8.2 QUALITY CONTROL
Analysis of the output signal of an instrument requires comparison of the
magnitude of that output signal to a reference signal that may be the
instrumental background signal, the signal obtained when a reagent blank
is analyzed, or an appropriate contro1. 1 Also required is an estimate of the
error associated with the signal of interest. These two factors, signal
comparison and associated error,can be derived specifically if an effective
quality control (Qq program has been followed during the sample
collection and analysis procedures. An effective QC program will include
blanks, duplicates, calibration and check standards, splits, and other such
procedural requirements. 2 ,3 Each of these is discussed in the following
subsections.
8.2.1 Instrument calibration
Proper calibration of the analytical instrument will yield two significant
benefits, namely: it ensures that the analyte of interest will give a signal that
DATA ANALYSIS AND PRESENTATION
337
is detectable if of sufficient magnitude, and it allows definition of the
correlation between the signal from the instrument and the amount of
analyte present in the sample. To begin this procedure, the matrix in which
the analyte of interest is to appear on injection into the instrument should
be determined since it could be either aqueous or organic depending on the
properties of the analyte. The matrix blank is then determined, i.e. the
response of the instrument on injection of the matrix. An aqueous matrix
may be distilled and deionized (001) water or 001 water with dilute acid
or base while an organic matrix may be methylene chloride or methanol or
some other organic solvent. The matrix chosen should be the one that
interferes least with the analyte signal and produces little or no signal of its
own.
With an appropriate matrix chosen, a certified pure variety of an analyte
is added to it. The amount added can be estimated from prior knowledge
but should be enough to give a definite signal when injected into the
instrument. Depending on the magnitude of the resultant signal, higher
and/or lower amounts are analyzed so that there are at least three results
which span the expected range of the analyte in the samples to be analyzed.
The instrument response is then plotted versus either the concentration or
the amount of added analyte. 4 The desired outcome of this plot is a straight
line which can be described by an equation of the form C = aR + b, where C
is concentration or amount, R is the instrument response, a is the slope of
the line, and b is the intercept on the C-axis. For those cases where the
instrument response is nonlinear, intermediate concentrations of the
certified material may be analyzed to determine the shape of the response
more accurately.
8.2.2 Calibration check standard
Since certified pure materials are expensive, a calibration check standard
(CCS) commonly is prepared. This can be made of reagent grade materials
in some arbitrary concentration since the exact value is not necessary.5 Five
to seven portions of the CCS are analyzed immediately after the instrument
is calibrated to determine the mean value. A preliminary control chart is
prepared that is updated each time the CCS is analyzed. This standard is
analyzed each operating day or each shift to ensure that the instrument is
still in calibration. Once the response of the instrument to the CCS exceeds
the control limits, and remains so, then recalibration of the instrument
becomes necessary.
The CCS is useful because it can be made with less expensive chemicals,
because there is no real need to standardize it against national standards,
338
INSTRUMENTAL ANALYSIS OF POLLUTANTS
and because it is not necessary to expend the effort to make a precise
concentration. However, if the CCS is not stable over time, precise
concentration information may be required so that replacement CCSs can
be prepared until such time as recalibration of the analytical instrument
becomes necessary.
8.2.3 Laboratory control standard
The laboratory control standard (LCS) normally consists of a standard
reference material (SRM) in the same matrix as used for samples in the
analytical procedure. In the USA the National Bureau of Standards
produces a variety of SRMs that are useful for many types of analyses. 3 If
an SRM is not available, the alternative is a purified analyte produced by a
commercial concern or by a government laboratory that specializes in such
materials. The function of the LCS is to estimate the bias in the analytical
procedure and provide data on which the comparability of the results
obtained by the analytical method can be assessed. Comparability is an
important attribute of an analytical method since it implies that the
measurements made can be reproduced by laboratories that may be using
other methods.
The LCS is measured just after the calibration procedure is completed
and then with each 20 or so samples until 10 or more aliquots have been
measured, then the frequency can be reduced. A running average of the
LCS results is calculated and represents the bias (also called systematic
error) in the method.
8.2.4 Blanks
Once the analytical procedure has been calibrated and the calibration
controls have been identified, the result obtained from the procedure when
the analyte of interest is thought not to be present should be assessed. This
is done by analyzing blank samples. As a general rule, several types of
blanks can be devised for any sampling and analysis project. The type of
blank to be analyzed depends on which part of the procedure is to be tested
for its contribution to the final result. The important blanks and a brief
description of their use are listed below. 6
8.2.4.1 Field blanks
This blank is the most comprehensive. It normally consists of DDI water
placed in 5-10% of the sample containers that are sent out for sample
collection. At the sampling site these containers are opened during sample
collection. The DDI water is used to rinse the sampling instrument(s) after
DATA ANALYSIS AND PRESENTATION
339
cleaning but, before collecting the next sample, the DDI water is placed
back in the container, which is then sealed. If any sample preparation, such
as mixing, sieving, ball-milling, etc., is performed on the samples, the field
blank is then used to rinse the equipment after cleaning. This blank is then
analyzed along with the samples. Therefore this blank will contain any
contamination introduced into the sample by the collection apparatus, the
container, the sample preparation equipment, the reagents or the matrix.
8.2.4.2 Transit blank
This blank also consists of DDI water placed in the sample containers.
Normally about half of the field blanks can be used for transit blanks.
These blanks are used to assess any contamination due to leaching from the
sample container during shipment and storage. They are sent to the field
and returned to the laboratory unopened, stored during sample
preparation, and analyzed with the samples.
8.2.4.3 Sample preparation blank
This is frequently called the sample bank blank because samples generally
are sent to a central area (sample bank) for processing and numbering.
Every 20 samples or so, a sample container ofDDI water is used to rinse the
sample preparation equipment and the rinse collected and analyzed with
the regular samples.
8.2.4.4 Reagent blank
This blank consists of DDI water of the same volume as the samples to be
analyzed and is used to check for contamination in the reagents. It is
analyzed identically as the samples are analyzed, with the same amount of
reagents, at a rate of about one blank for each 20 samples. Additional
reagent blanks will be required whenever the reagents are changed.
8.2.4.5 Matrix blank
This blank is merely an aliquot of whatever final solvent is used in the
analytical procedure. It is injected into the analytical instrument without
further treatment and any result obtained is an indication ofcontamination
of the matrix material.
Although all of the blanks listed above should be collected during the
course of a project, only the field blanks, or the field and reagent blanks,
need to be analyzed at the start. However, if the blanks show any detectable
amounts of the analyte in question, then the other types of blanks may need
to be analyzed to locate the source of extraneous analyte signal. Even
340
INSTRUMENTAL ANALYSIS OF POLLUTANTS
though the various kinds of blanks may make up 30% of the total sample
load, it is possible to reduce this to 5% as experience is gained with the
samples and procedures, and especially if the field and reagent blanks show
negligible contamination.
8.2.5 Controls
The three most important parts of an analytical procedure are the
standards, the blanks and the controls. Assuming the first two parts have
been covered adequately, then a positive result may be obtained when
samples are analyzed. However, any such result has no significance unless it
can be related to a standard, a regulation, a normal baseline amount, or
something similar to these. If the result of analysis is to be related to a
standard, then the control is the reagent blank. Ifit is to be compared to a
regulation, it generally would be the amount above a local background and
the control would be the local baseline. Finally, if the result is to be related
to a normal baseline, then the control would be regional or national
average values.
8.2.5.1 Local control
The local control is composed ofsamples collected from areas that are close
in time and space to the area being studied. The sample collection sites
should be upwind and upstream (as regards surface and ground water flow)
of the area in question and the samples should be distributed so they are
affected by all the sources that impact that area except the source that is
suspected of causing the impacts being measured.
The local control is very useful because it is more directly relatable to the
area of study than regional or national controls. However, in many cases an
appropriate local control is impossible to obtain so the other types of
control are needed.
8.2.5.2 Regional control
The same criteria apply to a regional control, as stated above for the local
control. Because of the larger area available for choice of samples, this
control may be more readily available. An additional factor to be
considered in selecting a regional control is the demographic characteristic
of the two areas: study versus control. As an example, a farming area would
not be a suitable control for an industrial area.
8.2.5.3 National control
If resources are scarce, a national control may be derived from a literature
DATA ANALYSIS AND PRESENTATION
341
search as there exists a surprising amount of data that can be uncovered by
a diligent search. In fact, there may even be routine monitoring or
surveillance programs that can provide data useful for developing a
suitable control value. A good example of a national control is the set of
data on cancer rates published by the government that are used to
determine whether or not these rates are high in a given area.
The principal problem will be, in general, the lack of adequate data,
either in the type of analyte reported or in the quality of the data. If
collected several years in the past, then the QA may be poor or nonexistent
so interpretation would be difficult or impossible.
8.2.6 Detection limits
There have been many names for detection limits and many definitions for
them, but present practice is to base the detection limit, by whatever name,
on some multiplier of the standard deviation of sample analyses near zero
analyte concentration. The value of the multiplier is related, in turn, to the
acceptability of the errors that are possible in any analytical scheme. The
two principal errors are Type I (also called alpha error) that is the
probability of stating that a substance is present when it is not, and Type II
(also called beta error) that is the probability of stating a substance is not
present (not found) when it actually is present.
The American Society for Testing and Materials has issued a standard
that includes a procedure for determining detection limits? In that
procedure several samples are prepared by adding near zero concentrations
of the analyte to the matrix for which the analytical method was developed.
After analysis the standard deviation of the results is calculated. A
detection limit called the criterion of detection is then determined by
choosing to accept a Type I error of 5% and using a value of 1·645 from a
table of normal cumulative probabilities as a multiplier for the standard
deviation (s). Thus the criterion of detection becomes 1·645s. However, if
many samples with an analyte concentration of 1·645s are analyzed, only
half of them will give results exceeding the criterion of detection. This is
because the probability of a Type II error is 50%. To adjust the probability
of a Type II error to 5% to match the Type I error, the same multiplier is
used. The result is termed the limit of detection when the risk of making a
Type I and Type II error is 5% and is equal to (2)(l'645)s = 3·29s. The power
of a test is determined by 1- p, so when the risk ofType II error is 5%, then
the power of the test (probability of finding a substance when it is present) is
100% - 5% = 95%.
Although the preceding was based on inorganic analyses of water
342
INSTRUMENTAL ANALYSIS OF POLLUTANTS
samples, almost equivalent procedures apply for other types of analysis.
For analysis of samples containing radioactivity, the detection limit is
based on background counting rate, as well as the counting rate from the
sample, so the multiplier for the lower limit of detection (LLO) is increased
by the square root of 2. This arises from the normal propagation of errors
where the total error (St) is the square root of the background variance (s~)
plus the sample variance (s;):
St
=J
(8.1)
S b2 +S2
s
Since the LLO is based on analyte concentrations near background, then Sb
and Ss will be equal and the equation reduces to
(8.2)
Therefore the LLO becomes 4·65s where
background counts:
S
is the standard deviation of
LLO = j2(2)(1'645s) = 4·65s
(8.3)
This factor is also based on Type I = Type II = 5% false detection and false
nondetection probabilities. 8
Also, for analysis of organic contaminants in samples, a similar
procedure is used for setting detection limits. In this case it is called a
method detection limit and is based on the 99% confidence Iimit. 9 Since the
procedure specifies seven aliquots of a sample matrix containing the
analyte at near zero concentration, the value of t for a one-sided test at 0·99
confidence level with six degrees of freedom (7 - I = 6) is 3'14. In this case,
then, the MOL becomes 3·14s where s is the standard deviation of the seven
aliquots.
The three detection limits described above (limit of detection, lower limit
of detection, and method detection limit) are all computed by multiplying
the standard deviation of a background or low-level sample by a factor of 3
or 4. Some analysts use various multiples of method detection limit (MOL)
to describe method capability, such as quantitation limit (= 3 x MOL) and
practical quantitation limit (= 10 x MOL).lO Many analysts use the term
'trace' to describe any detectable amounts that are greater than the MOL
but less than the quantitation limit. Figure 8.1 shows the basic relationships
discussed in the preceding. Here J.ll is the mean of the background (or lowlevel) distribution and J.l2 is the mean of the sample distribution. Ifthe MOL
(C) was set at the 95% level, then all measured concentrations ~ C would be
detectable and the Type I error (a) would be 5% but the Type II error (p)
DATA ANALYSIS AND PRESENTATION
",
343
c
Fig. 8.1. Sample and background distributions.
would be greater, perhaps 30%. Such facts lead to the use of 4s in analyses
of radioactivity and 3s for organic analysis of samples.
8.2.7 Uncertainty of results
The uncertainty in the result generated by any analytical procedure is due
to some combination of two principal errors, namely random and
systematic. Systematic error is determined by examining the results from
analysis of standards, such as the LCS described above. The average of all
results from analysis of the LCS, when compared to the true value, is a
measure of the bias to be expected when using the chosen method. The bias
is due to the cumulation of systematic errors as they occur in the laboratory
that is using the method. Systematic errors are not normally statistical in
nature but are due to factors that consistently bias any result in one
direction. Such errors may include incorrect calibration of balances,
thermometers or pipettes as well as calculations of concentration, dilution,
etc. Random errors, on the contrary, are those errors that can be treated
statistically.ll To eliminate the systematic errors in a procedure requires
careful cross-checking of all the steps, with particular attention to any steps
where small changes may have a large effect.
8.2.7.1 Random errors
These statistical errors in an analytical method can be detected by
performing various duplicate or replicate analyses. For instance, to
determine the random error for all the steps in a procedure from collection
of the sample to the output of the analytical instrument, a duplicate
sampling program is used. A duplicate sample is one that is collected
adjacent, in time and space, to the principal sample. For example, if two air
sampling instruments are placed next to each other and both are then
operated for the same length of time, the collected air samples are
duplicates. If one large sample is divided into two parts, the result is termed
a split sample. This differs from the duplicate sample because any sampling
344
INSTRUMENTAL ANALYSIS OF POLLUTANTS
Table 8.1
Methods of error assessment
Replicate analyzed
Error(s) assessed
Duplicate sample
Split sample
Split extract
Sampling, homogeneity, processing and measurement
Homogeneity, processing and measurement
Measurement
error has been eliminated, and analysis of the splits yields an estimate of the
errors due to sample inhomogeneity, extraction, preparation and the
instrument response. Instrumental errors can be evaluated by analyses of
splits ofsample extracts. Errors in extraction of the analyte from the sample
can be determined if homogeneous aliquots of the sample can be obtained,
for example, if the sample is in solution. Table 8.1 puts these methods of
error assessment into perspective.
The total random error is an estimate of the precision ofa process and its
value is determined by the standard deviation (s) of the measurements. The
square of the standard deviation (S2) is the variance of the measurements
and variances can be added so the variance of the total procedure is just the
sum of all the component variances. In mathematical terms
s~
= s~
+ s~ + s~ + s~ + ...
(8.4)
where the subscripts of the variances are T for total, S for sampling, H for
homogeneity, E for extraction and M for measurement. The square root of
any of these values is the precision of that component of the method and the
square root of the sum is the precision of the total method. As an example,
suppose many duplicate samples were to be analyzed and a mean and
standard deviation calculated from the results. If that calculation yielded
80 ± 3 units, it could then be stated that the precision of the method was 3
units for samples containing about 80 units of analyte. This would be the
extent of the statement that could be made until data for samples with
different amounts of analyte were obtained since it would not be known
whether or not the relation between precision and concentration was linear.
Ifit was desired to put some confidence limits on the above data, then a
factor of2 could be used to approximate the 95% probability interval, that
is 2 x ±3 = ±6 units is the 95% probability interval for samples containing
about 80 units.
Since these calculations are based on pairs of analyses, either duplicates
or splits, the range is almost as efficient for expressing precision as the
DATA ANALYSIS AND PRESENTATION
345
standard deviation. The range is found by subtracting the lower result in a
pair from the higher one. If the range of paired observations is divided by
1-128, the result is an estimate of the standard deviation. To use this
technique, the range is calculated for all pairs, the ranges are summed and
the sum is divided by the number of pairs that were summed to get the mean
range, and this result divided by 1-128 to get the standard deviation
estimate.
For some methods, or for some analytes, the standard deviation is not
constant with changes in concentration but is a constant fraction of the
concentration. In this case it is preferable to use the coefficient of variation
(CV), which is defined as the standard deviation divided by the mean value
and is usually expressed in percentage (lOOs/x = CV). This CV can be used
as an estimate of precision in the same manner as the standard deviation or
the range can be used.
A third relationship may arise, that is the precision may be neither a
constant nor a constant proportion of the analyte as may be revealed when
precision is plotted against concentration. If this is so, then the formula for
a straight line (E = aC + b) must be developed and used to describe the
relationship. In this formula, E is the error, C is the concentration, a is the
slope of the line, and b is the intercept on the E axis. Linear correlation
analysis is used to obtain the parameters for the straight line (also called the
least-squares line).
8.2.7.2 Total uncertainty
As stated at the beginning of this section, the total uncertainty in an
analytical procedure is some combination of the bias and precision
(systematic and random errors). There have been some theoretical studies
of this subject but no particular method of combining these two errors has
become predominant. The one suggested method that appears reasonable
is to combine the various components in quadrature but adjusting the value
of the bias by 1/3. 8 The mathematical statement for this is
(8.5)
where UT is total uncertainty, ST is total random error, and B is the bias
(systematic error). For this equation, the s used should be one with a high
confidence level. The 95% level is recommended because the bias is known
with high confidence. To illustrate the use of this combination, suppose 10
samples were spiked with low levels of an analyte and then analyzed. When
the results were compared with the known spikes, suppose the outcome was
92 ± 3% then the 95% probability interval for precision is ±6% and the
346
INSTRUMENTAL ANALYSIS OF POLLUTANTS
bias is 92% -100% = -8%. The total uncertainty then is calculated as
follows:
(8.6)
This uncertainty value is used as a multiplier for the result of an analysis
and sets the 95% probability interval in which the true value lies. For
example, suppose a method for which the total uncertainty was 7·6% was
used to analyze a sample and the result was X units, then the uncertainty in
that result would be ±0·076X units. Therefore a result of 250 ppm would
have an uncertainty of 0'076(250) = 19 ppm and the true result would lie in
the interval from 231 to 269 ppm.
8.2.8 Comparability
Comparability is an important, and in some cases an essential,
characteristic of the data generated by an analytical method. Some analysts
have thought that this consisted only of stating the results in consistent
units such as the MKS or SI systems. Of course comparisons are easier
when the units used are identical, but comparability implies much more
than just use of any system of units. It implies that if other laboratories
analyze the same sample but use different methods the results will be
comparable. To achieve this a laboratory must participate in an
intercomparison program or produce comparable results when analyzing
standard reference materials or similar materials as produced by a national
standards office.
In 1983 a symposium entitled Quality Assurance for Environmental
Measurements was held in Boulder, Colorado. The proceedings of that
symposium includes many papers that discuss practical applications of the
principles outlined in this section. 12
8.3 DATA ANALYSIS
Preliminary to analyzing the data, some treatment of the results is required,
such as grouping the data, eliminating outliers, and handling less than
detectable values. Grouping the data is an obvious step. All the data should
be categorized as to the area from which collected, the sample type (soil, air,
water, etc.), and the constituent for which analyzed. This permits
calculation of the mean and standard deviation, or putting the data in some
kind of order for ease of analysis. If the quality control procedures of
Section 8.2 have been followed, then the accuracy and comparability of the
347
DATA ANALYSIS AND PRESENTATION
results produced by the analytical method will be known and only the
significance of the results remains to be demonstrated.
8.3.1 Rejection of outliers
Several methods have been, and presently are, used to reject data that
appear not to belong to a given data set. One of the better methods, at least
for data that can be described by a mean and standard deviation, for
deciding whether or not to reject a result has been described. 13
Theoretically, no result should be discarded unless an obvious error has
occurred and, in some cases, deviant results may indicate flaws in the
analytical procedure or may indicate 'hot spots' of environmental
contamination.
If it has been decided to reject possible outliers, then proceed as in the
following. Arrange the set of results in order from low to high:
(8.7)
and calculate the mean and standard deviation for the whole set. Next,
calculate the following statistic:
or
T= (XH - X)/s if a value appears high
T= (X - Xd/s if a value appears low
(8.8)
Finally, compare the calculated value of T with the value under either the
5% or I% column in the tabulation for tests ofdiscordancy shown in Table
8.2. If the calculated value of T is greater than the value in the tabulation for
the number of measurements made, then either the low or high result is an
outlier with that level of significance.
Table 8.2
Tabulation for tests of discordancy
Number of
measurements
9
(Example) ...
10
12
14
16
18
20
30
5%
2·11
2·18
2·29
2·37
2-44
2·50
2·56
2·74
1%
2-32
2-41
2·55
2-66
2·75
2·82
2·88
3-10
348
INSTRUMENTAL ANALYSIS OF POLLUTANTS
As an example, suppose 10 samples were analyzed and the results ranged
from 45 to 70 ppm. The mean and standard deviation are calculated as 56 ±
6 ppm. To determine whether or not the high value is an outlier, calculate T
from (70 - 56)/6 = 2·33. That value is between the 2·18 and 2·41 values for
10 measurements, as indicated in the tabulation. Therefore 70 ppm is an
outlier at the 5% level of significance but not at the 1% level.
8.3.2 Treatment of less-than values
It has been customary to treat less than detectable (e.g. <MDL) values as
zeroes when calculating statistics or when comparing data sets, and other
conventions have also been used. However, any of these tend to introduce a
bias into any calculations. One such convention has been to use the value of
the MDL for any value that is less than MDL(or LLD, or etc.). Using either
zero or the detection limit for all values in a data set that are less than the
limit yields a censored data set with an extremely biased average and
standard deviation. Although using half the distance between zero and a
detection limit, or the geometric mean in case of non-Gaussian
distributions, results in less bias, these adjustments are also inappropriate.
A better method for handling nondetectable data is to use probability
plotting on either normal or log-normal probability paper. The data are
ranked so that percentages are readily obtained. The data are then plotted
on the two kinds of probability paper. The kind on which the plotted points
most nearly approach a straight line determines the type of distribution of
the data. The value represented by the 50% point on the plot is the mean of
the data while the standard deviation is found by dividing the value at the
50% point by the value at the 84·15% point.
This discussion on handling less than detectable values applies to existing
data or reports, where the values cannot be changed. Presently, it is
recommended that no data be reported as 'nondetectable', or 'less than
MDL', or similar such expressions. It is always possible to calculate a value
for any measurement that is made, even if it is a negative number as may
occur when a sample value is subtracted from a background or blank value.
As an example of the points discussed above, consider the data displayed
in Table 8.3. These data are from a study of a contaminated area where
heavy metals had infiltrated the ground water system. The first column is a
common method of displaying the results of 20 samples taken from a
control area that is upstream of the contaminated area. Ignoring the
<MDL or equating them to 0 or to the MDL results in a censored
distribution with an inaccurate mean and standard deviation. If some
historical data exist, then it may be possible to approach the true mean by
349
DATA ANALYSIS AND PRESENTATION
91i
I
90
I
80
70
04~
X
f:
--·1
X
30
20
1.0
2.0
10
Concen1n1ion • ppm
Fig. 8.2. Probability plot of water concentration values.
plotting the data on probability paper, as shown in Fig. 8.2. Here column 1
data are ordered in column 3 and the sum of the cumulative number of
values is expressed as a percentage of the total as in column 4, that is the
seven values less than MOL represent 35% of the data, etc. The eyeballed
line drawn on the points is shown on the figure. The 50% intersection is the
mean of the data (2,0 mg/liter), which is very close to the actual mean shown
for column 2.
8.3.3 Data comparisons
The most common procedure used to test a set of data to determine whether
or not it is significantly different from other data is to use Student's t-test.
However, care must be exercised in using this test. The data in the two sets
to be compared must be approximately normal or modified to fit a normal
distribution. Much environmental data are distributed log-normally, which
means that the data are converted to logarithms before either plotting or
calculating any statistics. 14 If there are 10 or fewer values to be tested in the
data sets, then the type of distribution is not crucial.
To perform the t-test, assume that the background, reagent blank or
control sample data are known with high precision, and that the data from
the collected samples is to be compared with those data. Since it is expected
350
INSTRU:MENTAL ANALYSIS OF POLLUTANTS
that the sampled area will have a higher concentration of the analyte than
the background, blank or control samples, then a one-tailed i-test is used.
At this point access to a table of the t distribution is necessary. Based on the
confidence level chosen for the test and the degrees of freedom (actually one
less than the number of samples) a t value is selected from the table for use
as shown in eqn (8.9) to generate a test statistic. The procedure will be
clearer if a numerical example is presented as follows.
Assume the mean concentration of lead in control samples is 2·04 ppm in
water. Assume 20 water samples are taken from a suspect area and that the
mean is 2·60 ± 0·28 ppm (Table 8.3). If the desired confidence level is 95%,
Table 8.3
Concentration in water (mg/liter) (MOL = 1·93 mg/liter)
Control samples·
Number
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
.X'
s
I
2
2-3
2-3
3·1
2·0
<MOL
2·1
<MOL
<MOL
2·2
2·2
2·0
2·0
<MOL
2·8
2·0
<MOL
2-2
<MOL
<MOL
2·0
2·2
2'13 b
0·31
3·1
2·0
1-8
2·1
1·8
1·7
2·2
2·2
2·0
2·0
1·9
2·8
2·0
1·9
2·2
1·7
1·0
2·0
2·2
2·04
0·42
3
<MOL
<MOL
<MOL
<MOL
<MOL
<MOL
<MOL
2·0
2·0
2·0
2·0
2·0
2·1
2-2
2·2
2-2
2·2
2-3
2·8
3·1
Study samples
E%
2·5
3·2
2·7
2·8
2·7
2·0
35
2-3
2·9
2-3
2·7
3·0
60
65
85
90
2-4
2·5
2·6
2·8
2·7
2-3
2-4
2·8
2·5
2·60
0·28
• Column 1 is common listing of control results; column 2 is all results (preferred);
column 3 is column 1 ordered low to high; and ~ % is a cumulative percentage of
column 3.
b Assumes all < MOL are equal to the MOL; if < MOL were deleted, then.X' = 2·62.
351
DATA ANALYSIS AND PRESENTATION
then from a table of ( at 95% level and 19 degrees of freedom (20 - 1 = 19)
find the factor 1·729. The test statistic is then calculated as
u = (s/~
= 1'729(0'28)/fi = 0·11
(8.9)
where s = ±0·28. If the difference between the means of the control and
tested samples is greater than 0'11, then the concentration of lead in the
water in the suspect area is higher than in the control area at the 95%
confidence level. Since the difference between the means is 2·60 - 2·04 =
0·56 ppm and since this is greater than the 0·11 calculated above, then the
lead concentration in the water in the sampled area is significantly higher
than in the water from the control area.
The (-test is one of the simpler tests that can be used for comparing two
data sets, but there are others which vary with the type ofdistribution of the
data sets as well as some nonparametric tests which are independent of
distribution. ls One of the most generally used of the latter is the WilcoxonMann-Whitney test. For comparing two sets of data, assign rank to each
number in the two sets, starting from the lowest number (assign a 1) and
proceeding to the highest. Use an average rank for all numbers that are
equal. For example, use the first nine numbers of the study and control
samples in Table 8.3 and rank as shown in Table 8.4. Then choose the
significance level desired for the test, e.g. a = 0,05, and obtain the test
number from a table of'critical values of smaller rank sum for the W-M-W
test' (Ref. 15 contains such a table). In this example the value for samples
where n 1 = 10 and n 2 = 10 is 66. If the ranks assigned to the control sample
results are summed, the result is 58'5, as shown below. Since 58·5 is less than
Table 8.4
Ranking of selected data from Table 8.3
Control
2·3
3·1
2·0
1·8
2·1
J-8
1·7
2·2
2·2
Rank
10
17
4·5
2·5
6
2·5
1
7·5
7·5
L58·5
Study
2·5
3·2
2·7
2·8
2·7
2·0
2-3
2·9
2-3
Rank
12
18
115
15
115
4·5
10
16
10
352
INSTRUMENTAL ANALYSIS OF POLLUTANTS
the 66 obtained from the table, one can conclude that the concentration in
the study sample is greater than that in the control sample with only a 5%
probability of being in error.
For precise tests that compare results from various sampling programs,
the services of a statistician are indispensable and, if possible, the
statistician should be consulted prior to the initiation of any sampling
program.
8.3.4 Graphical presentation
A picture is worth a thousand words, to coin an expression, and this is at
least true for graphical presentation of scientific data. There is no fixed rule
for determining the form of the presentation, whether it should be a line,
bar or pie chart, etc. Therefore the empirical approach of testIng several
methods for the presentation of the pertinent data should be used as well as
40
(AI
..-. . ..
..
.. .-..- ..
-'
"(j
Do 20
I .. .. .
e
..... .
.5
.. • •
o ••
.- • •e
...
.
~
:~
·10
~
·20
1982
-'40
"(j
1983
1984
1985
1986
1988
...... e.-... .
(B)
Do
....
..
... .
..:
..
..........
1982
1987
1983
1984
1985
1986
.
1987
1988
Fig. 8.3. Tritium in water, time-series and running average plots.
353
DATA ANALYSIS AND PRESENTATION
attempting several types of data summing. For example, a pure time-series
plot ofdata as shown in Fig. 8.3(A) may present a confusing picture which is
no better than the normal tabular accumulation when wide variations in
the data occur. To clarify the apparent trend shown in Fig. 8.3(A), a running
six-month average of the data was calculated and then plotted as shown in
Fig. 8.3(8). This latter presentation is much more useful as it shows the
gradual upward trend in tritium concentration as well as two excursions
that suggest a pulse of contamination passing the sampling point. The use
of a six-month running average was purely arbitrary in this case; any
convenient averaging period can be used. A disadvantage of using running
averages is that the actual time of peak activity is shifted, in which case the
sequential plot or actual tabulation of the data should be consulted.
The graphs shown in Fig. 8.3 are termed rectilinear coordinate graphs
and are the most common type for displaying research data. Such graphs
can be modified to display much significant information about a set ofdata.
One common modification is to add the standard deviation (in some cases
the standard error) and range of data sets. The data displayed in Fig. 8.3 are
based on one analysis per month so range and standard deviation are
inappropriate. In Fig. 8.4 the data displayed are the mean, standard
deviation and range for 15 air sampling stations in Nevada. An additional
C
E
o
50
.e
~
e
c
0
45
40
y
E
l
Y
E
l
Y
30
25
0
15
:ci
10
l!!
I
T
35
..
';1
A
R
C
c 20
8c
(J
5
0
Jen
Feb
Mer
Apr
Mey
Jun
Fig. 8.4. Rectilinear plot with range. standard deviation. mean and location of
maximum concentration.
354
INSTRUMENTAL ANALYSIS OF POLLUTANTS
30
~
.
«
Ii
CD
20
"
~
U
Q.
.
.S
l....
.g.
12
6
•0
~
14
...
4
0
19!;6
~8
60
62
64
66
68
70
72
74
76
78
80
Fig. 8.5. Bar chart of strontium-90 concentration in bone.•. Bighorn sheep;
O. deer; ,~ cattle. Number of bone samples for each animal is indicated.
useful datum has been added to the figure; the city where the maximum air
concentration was detected.
The next most common form of graph for displaying research or
monitoring data is the bar chart. This type of graph is most useful for
comparing data from different sources. In our work of monitoring for
radioactivity, data on the concentration of radioactive strontium in the
bone of various species of animal have been collected for many years. The
most compact method of displaying all the accumulated data is by means of
a bar chart, as shown in Fig. 8.5. Such a graph shows not only the relative
concentration by species but also the change in concentration as the
amount of strontium in the environment changes over time.
There is a publication, now in its second edition,16 that includes an
extensive discussion of the different types of graphical display as well as
methods for designing them. A large variety of each of the types of graphs
are shown and explained, but the rectilinear coordinate and bar charts as
described above will suffice for most data display purposes.
REFERENCES
1. Sharaj, M., IIIman, D. & Kowalski, B., Chemometrics. John Wiley, New York,
Chapter 3, 1986.
2. Brown, K. W. & Black, S. C, Quality assurance and quality control data
validation procedures used for the Love Canal and Dallas lead soil monitoring
programs. Environ. Mon. Assmt, 3 (1983) 113-22.
3. Taylor, J. K., Quality assurance of chemical measurements. Anal. Chem., 53
(1981) 1588A-93A.
4. USEPA, Guidelines establishing test procedures for the analysis of pollutants
under the clean water act. Title 40, Code of Federal Regulations, Part 136 (40
CFR 136), Washington, DC, 1986.
DATA ANALYSIS AND PRESENTATION
355
5. Goldin, A. S., Evaluation of internal control measurements in radioassay.
Health Phys., 47 (1984) 361-74.
6. Black, S. c., Defining control sites and blank sample needs. In Principles of
Environmental Sampling, ed. L. H. Keith. American Chemical Society,
Washington, DC, 1987.
7. ASTM, Standard Practice for Intralaboratory Quality Control Procedures and a
Discussion on Reporting Low-Level Data. Designation: D4210-83, American
Society for Testing and Materials, Philadelphia, PA, 1983.
8. USEPA, Upgrading Environmental Radiation Data. EPA 520/1-80-012,
Washington, DC, Chapter 6, 1980.
9. Glaser, 1. A., Foerst, D. L., McKee, G. D., Quave, S. A. & Budde, W. L., Trace
analysis for wastewaters. Environ. Sci. Technol., 15 (1981) 1426-35.
10. USEPA, National Drinking Water Regulations for Synthetic Organic and
Inorganics. Title 40, Code of Federal Regulations, Part 141, Washington, DC,
1985.
II. ASTM, Standard Practicefor Determination ofPrecision and Bias ofMethods of
Committee D-J9 on Water. Designation: D2777-77, American Society for
Testing and Materials, Philadelphia, PA, 1977.
12. Taylor, 1. K. & Stanley, T. W. (eds), Quality Assurance for Environmental
Measurements. Special Technical Publication 867, American Society for
Testing and Materials, Philadelphia, PA, 1985.
13. Barnett, V. & Lewis, T., Outliers in Statistical Data. John Wiley, New York,
1978.
14. Gale, H. J., The lognormal distribution and some examples of its application in
the field of radiation protection. Report AERE-R-4736, AERE, Harwell, 1965.
15. Natrella, M. G., Experimental Statistics. National Bureau of Standards
Handbook 91, Washington, DC, 1966.
16. Schmid, C. F. & Schmid, S. E., Handbook ofGraphic Presentation, 2nd edn. John
Wiley, New York, 1979.