Analytical Methods Committee, AMCTB No 58
First published on 18th November 2013
A knowledge of the measurement uncertainty arising from sampling is crucial for the rational design of sampling programmes and the interpretation of their outputs (AMC Technical Briefs No 16, 2008). Usually the principal source of uncertainty is the lack of homogeneity (either spatial or temporal) in the object of sampling – the sampling target. Other sources of variation—the analytical and sampling processes—are also important.
Next consider the level of agreement between duplicate samples from the same target. Case A indicates target heterogeneity on a fine scale – in this situation the differences between duplicate composite samples are (on average) small, because the increments tend towards an average contamination. In Case B—heterogeneity on a coarser scale—there is a greater potential for duplicate samples to encompass areas of either high or low contamination, tending towards larger within-sampling uncertainty. Differences between duplicate samples would here range from small to potentially large, but greater on average than in Case A. Note also that the sample size—indicated on the diagram by the size of increments—influences the perception of local heterogeneity. As the size of sampling footprint increases in relation to the true scale of variation, the effect of local heterogeneity will tend to decrease.
In our example, the measurand is the average zinc concentration of the sampling target (in this example, the ‘target’ is the total quantity of effluent considered during a particular sampling event). The best approach to the evaluation of UfS—in the absence of prior knowledge about the underlying heterogeneity of the effluent—is to take a number (N) of duplicate effluent samples, say one pair of samples on each of N days spread out though the year (note that it is assumed for the sake of this example that there are no systematic changes in the measurand or the quality of sampling and measurement over the period of the investigation). Duplicate effluent samples are taken on the same day: practicality in visiting the sampling site means that the sampling procedure is effectively assigned this timescale. The aim then would be to analyse each sample in duplicate.1 This number N of sampling events, in this example visits to the effluent discharge point, is critical in determining the confidence interval around the uncertainty estimates and the overall cost of the operation; the higher the value of N, the better the estimate of UfS, but also the greater the effort and the higher the cost.
Once the data are assembled, nested analysis of variance can then be used to separate the effects of the three variable factors—analytical, sampling, and day-to-day changes in the effluent—expressed as standard deviations sa, ss, sd respectively. These estimates can then be combined to give the uncertainty of the chosen sampling programme.
Example 1 | RSD (percent) | ||
---|---|---|---|
True RSD | Analytical | Sampling | Between-target |
% | 5% | 10% | 15% |
Number of samples N | Observed range | Observed range | Observed range |
2 | 2.1–7.6 | 0.0–18.0 | 0.0–31.6 |
4 | 2.9–6.9 | 2.9–15.9 | 0.0–26.3 |
8 | 3.5–6.4 | 5.1–14.3 | 4.9–23.0 |
12 | 3.8–6.2 | 6.0–13.7 | 7.2–21.5 |
24 | 4.1–5.8 | 7.3–12.6 | 7.2–21.5 |
48 | 4.4–5.6 | 8.1–11.8 | 7.3–21.5 |
Example 2 | RSD (percent) | ||
---|---|---|---|
True RSD | Analytical | Sampling | Between-target |
% | 1% | 10% | 10% |
Number of samples N | Observed range | Observed range | Observed range |
2 | 0.4–1.5 | 2.2–17.6 | 0.0–23.0 |
4 | 0.6–1.4 | 4.2–15.5 | 0.0–18.7 |
8 | 0.7–1.3 | 5.8–14.0 | 0.0–16.1 |
12 | 0.8–1.2 | 6.6–13.4 | 1.2–14.9 |
24 | 0.8–1.2 | 7.6–12.3 | 2.7–14.8 |
• the confidence interval associated with estimates of uncertainty can be wide, even for a relatively large number of targets (N).
• The three sources of variation interact to affect the range of estimated values. The smaller the variance low in the hierarchy (from analytical upwards), the more precise the estimate of the next higher variance. Specifically, excellent analytical precision gives a smaller confidence interval in ss. Good precision in analysis and relatively low ss are both needed for precise determination of sd. This makes it clear that to determine the components of sampling uncertainty it is important to use as precise a measurement technique as might be available—even if it is not intended to use this technique for routine analysis. Also, the estimation of sampling uncertainty using a measurement technique that has a high relative standard deviation (say >10%, a value which is not uncommon for some trace analytical techniques), may produce measured values of uncertainty from sampling that themselves have an unacceptably large confidence interval.
• A value of N less than 8 gives fairly imprecise estimates of uncertainty. Given the fact that 4N analyses are required, a value of N larger than about 12 not only leads to escalating costs of testing, but also yields diminishing returns with respect to the precision of the estimates of uncertainty that are obtained. This suggests that N in the range 8 to 12 is likely to be a reasonable choice, in the absence of data to the contrary.2
Caution : Valid interpretation of this type of experiment depends on assuming a reasonably uniform analytical precision and between-sampling precision, from target to target. If successive targets vary widely in composition this assumption may not be justifiable, in which case a more complex statistical approach would be required, and expert advice should be sought.
M J Gardner [Atkins Limited]
This Technical Brief was written on behalf of the Subcommittee for Uncertainty from Sampling (Chair M H Ramsey) and approved by the Analytical Methods Committee on 23.10.13.
This journal is © The Royal Society of Chemistry 2014 |