Why We Need the Theory of Sampling

Kim H. Esbensen & Claas Wagner

Without representative sampling, measurement uncertainty is compromised. Here, we present the current Theory of Sampling versus Measurement Uncertainty debate. The verdict? Nolo contendere!

The purpose of sampling is to extract a representative amount of material from a ‘lot’ – the ‘sampling target’. It is clear that sampling must and can only be optimized before analysis. In a recent paper, we show how non-representative sampling processes will always result in an invalid aliquot for measurement uncertainty (MU) characterization [1].

A specific sampling process can either be representative – or not. If sampling is not representative, we have only undefined, mass-reduced lumps of material without provenance (called ‘specimens’ in the theory of sampling) that are not actually worth analyzing. Only representative aliquots reduce the MU of the full sampling-and-analysis process to its desired minimum; and it is only such MU estimates that are valid. Sampling ‘correctness’ (which we define later) and representativity are essential elements of the sampling process.

TOS Versus MU in Brief

Current Measurement Uncertainty (MU) approaches do not take sufficient account of all sources affecting the measurement process, in particular the impact of sampling errors. All pre-analysis sampling steps (from primary sample extraction to laboratory mass reduction and handling – sub-sampling, splitting and sample preparation procedures – to the final analytical test portion extraction) play an important, often dominating role in the total uncertainty budget, which, if not included, critically affects the validity of measurement uncertainty estimates.

Most sampling errors are not included in the current MU framework, including incorrect sampling errors (ISEs), which are only defined in the theory of sampling (TOS). If ISEs are not appropriately reduced, or fully eliminated, all measurement uncertainty estimates are subject to uncontrollable and inestimable sampling bias,which is not a similar to the statistical bias because it is not constant. The sampling bias cannot, therefore, be subjected to conventional bias-correction. TOS describes why all sources of sampling bias must be competently eliminated – or sufficiently reduced (in a fully documentable way) – to make MU estimates valid. TOS provides all the theoretical and practical countermeasures required for the task.

TOS must be involved before traditional MU to provide a representative analytical aliquot; otherwise, a given MU estimate does not comply with its own metrological intentions.

The Theory of Sampling – TOS – has been established over the last 60 years as the only theoretical framework that:

i) deals comprehensively with sampling
ii) defines representativity
iii) defines material heterogeneity
iv) furthers all practical approaches needed in achieving the required representative test portion.

The starting point of every measurement process is the primary lot. All lots are characterized by significant material heterogeneity – a concept only fully acknowledged and defined by TOS – where it is crucially subdivided into constitutional heterogeneity and distributional heterogeneity (spatial heterogeneity). The heterogeneity concept (and its many manifestations) are introduced and discussed in complete detail in the pertinent literature [4-12]. The full pathway from ‘lot-to-analytical-aliquot’ is complex (and in some aspects counter-intuitive because of specific manifestations of heterogeneity) and is subject to many types of uncertainty contributions in addition to analysis.

Unfortunately, the GUM [2] and EURACHEM/CITAC [3] guides focus only on estimating the total MU in a completely passive mode. TOS, on the other hand, focuses on the conceptual and practical active steps needed to minimize all sampling contributions to MU. The main thrust of our argument is that if the test portion is not representative – in other words, if all sampling error effects have not been reduced or eliminated where possible – all MU estimates are compromised. But this does not have to be the case – TOS to the fore!

The impact of heterogeneity

Proper understanding of the phenomenon of heterogeneity, its influence on sampling correctness and, most importantly, how heterogeneity can be counteracted in the sampling process requires a certain level of understanding. Here, we present a bare minimum of TOS tenets to give an appreciation of the deficiencies inherent in current MU approaches.

Before defining these concepts theoretically, Figure 1 illustrates selected examples of the many appearances of heterogeneity. In reality, heterogeneity exhibits almost infinite manifestations – a point that appears rather hopeless at first.

Figure 1. The almost infinite manifestations of heterogeneity in science, technology and industry. Materials and lots sometimes appear ‘uniform’ to the naked eye (E), but visual homogeneity does not guarantee that chemical characteristics are also distributed randomly. Heterogeneity can be structured (A, G, H), but is often highly irregular (B, C, D, I).

Heterogeneity plays out its role on all scales from constituent particles to full lot scale, and can only be fully understood by considering both constitutional and distributional aspects and characteristics. Irregular and segregated heterogeneity, for example stratification, layering, or significant individual grain irregularity, cannot be expected to follow any standard statistical distribution. Figure 1D shows a strongly heterogeneous spatial distribution – all grains with diameters above the average have been dyed blue, while all spherules with diameters below-average remain white. In this example, heterogeneity is partly physical and partly compositional because the spherules actually carry different grades/contents of the analyte in question. Compositional heterogeneity is also clear in Figure 1F: which grapes are ripe and ready for harvesting and which are not? These examples are only meant to illustrate the serious issues surrounding securing representative primary samples; there are many other examples in the literature.

Already, it should seem obvious that the notion of modeling every manifestation of heterogeneity within the fixed concepts of systematic and random variability is too simple to cover the almost infinite variations of the real world’s lot and material heterogeneity. This fact is argued and illustrated in full detail in the TOS literature. Previously, it has been argued that it is obvious that grab sampling will (always) fail in this context. Composite sampling is the only way forward [18].

For well-mixed materials (those that appear ‘homogenous’ to the naked eye, for example, Figure 1E), notions of simple random sampling have often been thought to lend support to the statistical assumption of systematic and random variability components; however, these comprise only a very minor proportion of materials with special characteristics, so clearly cannot be used to justify the same approach for significantly heterogeneous materials.

Theoretical analysis of the phenomenon of heterogeneity shows that the total heterogeneity of all types of lot material must be discriminated as two complements, the constitutional heterogeneity (CH) and distributional heterogeneity (DH). CH depends on the chemical and/or physical differences between individual ‘constituent units’ in the lot (particles, grains, kernels), generically termed ‘fragments’. Each fragment can exhibit any analyte concentration between 0 and 100 percent. When a lot (L) is sampled using single-increment procedures – grab sampling – CHL manifests itself in the form of a fundamental sampling non-representativity. The effect of this fundamental sampling error (FSE) – which is unavoidable using grab sampling – is the most fundamental tenet of TOS. CHL increases when the compositional difference between fragments increases; CHL can only be reduced by comminution, typically crushing.

DHL, on the other hand, reflects the irregular spatial distribution of the constituents  at all scales between the sampling tool volume and the entire lot. DHL is caused by the inherent tendency of particles to cluster and segregate locally (grouping) as well as more pervasively throughout the lot (segregation, layering) – or a combination of the two, as exemplified in the bewildering diversity of material heterogeneity in science, nature, technology and industry. DHL can only be reduced by mixing and/or by informed use of composite sampling with a tool that allows a high number of increments [4, 8, 11-13].

It is rarely possible to carry out forceful mixing of an entire primary lot. Therefore, if sampling lots have a high DHL, there is a very high likelihood for significant primary sampling errors: FSE + a grouping and segregation error (GSE). That is, of course, unless you pay close attention to the full complement of TOS principles.

The good news is that once you have acknowledged the principles of TOS, they are applicable to all stages and operations from the lot scale to the laboratory – representative sampling is scale-invariant. Lots come in all forms, shapes and sizes spanning the whole gamut of at least eight orders of magnitude, from microgram aliquots to million ton industrial or natural system lots.

It is critical to note that DHL is not a permanent, fixed property of the lot; GSE effects cannot be reliably estimated, as the spatial heterogeneity varies in both space and time as lots are manipulated, transported, on- and off-loaded, and so on. DHL can be changed intentionally (reduced) by forceful mixing, but can also be altered unintentionally, for example, by transportation, material handling or even by laboratory bench agitation.

An essential insight from TOS is that it is futile to estimate DHL based on assumptions of constancy. TOS instead focuses on the necessary practical counteracting measures that will reduce GSE as much as possible (the goal is full elimination) as an integral part of the sampling and sub-sampling process. In reality, it is rarely possible to completely eliminate GSE effects but they can always be brought under sufficient quantitative control [15].

Most materials in science, technology and industry are demonstrably not composed of many identical units. Instead the DHL irregularity is overwhelming (the illustrations in Figure 1 only show the tip of the iceberg). Lot heterogeneity (CHL + DHL), especially DHL, is simply too irregular and erratic to be accounted for by traditional statistical approaches that rely on systematic/random variability. In fact, this issue constitutes the primary difference between TOS and MU.

Representative sampling in practice

The focus of TOS is not on ‘the sample’ but exclusively on the sampling process that produces the sample – a subtle distinction with very important consequences. Without specific qualification of the sampling process, it is not possible to determine whether a particular sample is representative or not. Loosely referring to ‘representative samples’ without fully described, understandable and documented lot provenance and sampling processes is an exercise in futility. We repeat: a sampling process is either representative or it is not representative; there is no declination possible of this adjective.

The primary requirement in this context is sampling correctness, which means elimination of all bias-generating errors, termed ‘incorrect sampling errors’ (ISE) in TOS. After meeting this requirement (using only correct sampling procedures and equipment), the main thrust in TOS is to ensure an equal likelihood for all increments of the lot to be selected and extracted as part of an increment. This is called the ‘Fundamental Sampling Principle’ (FSP), without which all chances of representativity are lost. FSP underlies all other matters in TOS.

A unified approach for valid estimation of the sum of all sampling errors (TSE) and all analytical errors (TAE) was recently presented in the form of a new international standard, ‘DS 3077 Representative Sampling – HORIZONTAL standard’ [15]. This standard analyzes the general sampling scenario comprehensively, especially how heterogeneity interacts with the sampling process and what to do about it.

Pierre Gy and the History of TOS

Pierre Gy was born in Paris in 1924. He obtained a Masters degree in chemical engineering (1956), a PhD in physics (1960) and another PhD in mathematical statistics in 1975. Pierre has been awarded four major international scientific organization awards, including two gold medals from the Société de l‘industrie minérale, and the Lavoisier Medal.

A special issue of “Chemometrics and Intelligent Laboratory Systems” was dedicated to the scholar’s distinguished achievements in science and technology: “50 years of the Theory of Sampling (TOS)” [7], which includes a tribute to Pierre Gy that summarizes his scientific career; a three-paper series: “Sampling of particulate materials I-III” ; as well as “50 years of sampling theory – a personal history”; and a listing of his complete scientific oeuvre – the latter five comprise the last professionally published papers from the pen of Pierre himself. In it, he outlines the complex history and the reason behind the development of TOS in a fascinating web of personal, scientific and industrial stories. The special issue also forms the proceedings of the First World Conference on Sampling and Blending (WCSB1), which was dedicated to the 50-year anniversary.

A call for integration

We have shown that MU is not a fully comprehensive, universal, nor guaranteed approach to estimate a valid total measurement uncertainty if it does not include all relevant sampling effects. Around 60 years of theoretical development and application of TOS practice have shown that sampling, sample handling and sample preparation processes are associated with significantly larger uncertainty components, TSE, than the analysis (measurement) itself, TAE, typically multiplying total analytical error (TAE) by 10-50 times, dependent on the specific lot heterogeneity and sampling process in question. Occasional, very special deviations from this scenario cannot be generalized.

While GUM focuses on TAE only, the EURACHEM guide does point out some of the potential sampling uncertainty sources, but does not provide sampling operators with the necessary means to take appropriate actions. Only TOS specifies which types of errors can – and should – be eliminated (incorrect sampling errors, ISE) and which cannot, but which must instead be minimized (‘correct sampling errors’, CSE); our critique [1] crucially shows how!

Strikingly, the MU literature does not allow for sufficient understanding of the heterogeneity concept (CHL and DHL) or acknowledges the necessary practical sampling competence, both of which are essential for representative sampling. Incorrect sampling errors are non-existent in the MU framework, and the grouping and segregation error is only considered to an incomplete extent, leaving the TAE and the fundamental sampling error as the only main sources of measurement uncertainty here.

Also, the critically important sampling bias is only considered to a limited extent and only based on assumptions of statistical constancy. However, the main feature of the sampling bias is its very violation of constancy, which follows directly from a realistic understanding of heterogeneity. The only scientifically acceptable way to deal with the sampling bias is to eliminate it, as has been the main tenet of TOS since its inception in 1950 by its founder Pierre Gy (see sidebar). Here lies the major distinction between TOS and MU: TOS states that the sampling bias is a reflection of the incorrect sampling error effects interacting with a specific heterogeneity, whereas MU limits itself to acknowledging a statistical (constant) bias resulting from systematic effects attributable to protocols or people only [1].

MU is a top-down approach that is dependent upon an assumed framework of random and constant systematic effects (wrong), so individual uncertainty sources, such as GSE and ISE, are not subject to separate identification, concern, estimation, nor appropriate action (elimination/reduction). Indeed, the full measure of uncertainty sources connected to sampling, TSE, are almost completely disregarded in the MU approach. It is simply assumed that the analytical sample, which ends up as the test portion, has been extracted and mass reduced in a representative fashion. But if this assumption does not hold up, the uncertainty estimate of the analyte concentration is invalid; it will unavoidably be too small by an unknown, but significant (and variable) degree. The very different perspectives offered by MU and TOS are in serious need of clarification and reconciliation.

To that end, we call for an integration of TOS with the MU approach, illustrated by Figure 2, which shows three interconnected fishbone diagrams representing sampling, sample preparation and the standard complement of measurement uncertainty sources. Note that the intermediary complement represents the specific sample preparation error types, vividly illustrated and commented upon in a recent article in The Analytical Scientist [16].

Figure 2: Proposed TOS/MU integration using an augmented fishbone flow path diagram with three components: sampling, sample preparation and analytical measurement errors. The figure shows integration of TOS principal sampling uncertainty sources in an augmented MU framework. The standard MUanalysis fishbone diagram is shown on the right. TOS is charged with delivering a representative analytical aliquot, including all sampling aspects of sample preparation (crushing, mass-reduction, weighing a.o.) with a view to eliminating incorrect sampling errors [1].

To prevent underestimation of active uncertainty sources, we must integrate the effects of all three components (sample extraction stage, sample preparation stage and analytical stage), which can actually be done in a seamless fashion; there is no need to change the current framework for MUanalysis only to acknowledge the critical role played by TOS, which is to say, the framework for MUsampling.. These sampling and sample preparation branches should be implemented in every MU framework, as sampling uncertainty contributions logically must be dealt with ahead of the traditional measurement uncertainties. In this scheme, TOS delivers a valid, representative analytical aliquot (the horizontal arrow in Figure 2) as a basis for the now valid MUanalysis estimation.

An end to the debate?

Hopefully, we have explained the critical deficiencies in MU and have shown that TOS should be introduced as an essential first part in the complete measurement process framework, taking charge and responsibility of all sampling issues at all scales – along the entire lot-to-aliquot process. We want to see a much-needed reconciliation between two frameworks that have, for too long, been the subject of quite some antagonism. Indeed, the ‘debate’ between the TOS and MU communities has at times been unnecessarily hard, but we may add that such hostilities have been unilateral (always directed towards TOS). A ‘representative (!) example of this quite unnecessary confrontation can be appreciated in the appendix of a recent doctoral thesis [17].

Squabbling aside, we hope that our efforts are seen as a call for constructive integration of TOS and MU, and look forward to comments from the wider community.

Kim H. Esbensen is a research professor at the Geological Survey of Denmark and Greenland, Copenhagen, Denmark, and Claas Wagner is an energy and environmental consultant and specialist in feed and fuel QA/QC.

References

[1] K. H.  Esbensen and C. Wagner, “Theory of Sampling (TOS) vs. Measurement Uncertainty (MU) – a call for integration”, Trends in Analytical Chemistry (2014). doi:10.1016/j.trac.2014.02.007.

[2] GUM, “Evaluation of Measurement Data – Guide to the Expression of Uncertainty in Measurement”, JCGM 100:2008 (2008).

[3] EURACHEM, CITAC, “Measurement uncertainty arising from sampling: a guide to methods and approaches”, Eurachem, 1st edition, EUROLAB, Nordtest, UK RSC Analytical Methods Committee. Editors: M. H. Ramsey and S. L. R. Ellison. (2007).

[4] P. M. Gy, “Sampling for Analytical Purposes” (John Wiley & Sons Ltd, Chichester, 1998).

[5] P. M. Gy, “Proceedings of First World Conference on Sampling and Blending”, Special Issue of Chemometr. Intell. Lab. Syst., 74, 7–70 (2004).

[6] K. H. Esbensen et al., “ Representative process sampling – in practice: variographic analysis and estimation of total sampling errors (TSE). Proceedings 5th Winter Symposium of Chemometrics (WSC-5), Samara 2006”, Chemometr. Intell. Lab. Syst. 88 (1), 41–49 (2007).

[7] K. H. Esbensen and P. Minkkinen (Eds.)., “Special issue: 50 Years of Pierre Gy’s theory of sampling: proceedings: first world conference on sampling and blending (WCSB1). Tutorials on sampling: theory and practice”, Chemometr. Intell. Lab. Syst. 74 (1), 1-236 (2004).

[8] K. H. Esbensen and L. J. Petersen, “Representative Sampling, Data Quality, Validation – A Necessary Trinity in Chemometrics”, In: S. Brown, R. Tauler, and R. Walczak (Eds.) Comprehensive Chemometrics, volume 4, 1–20 (Oxford: Elsevier, 2010).

[9] L. Petersen, C. K. Dahl, and K. H. Esbensen, “Representative mass reduction: a critical survey of techniques and hardware”, Chemometr. Intell. Lab. Syst. 74, 95–114 (2004).

[10] L. Petersen et al., “Representative sampling for reliable data analysis. Theory of sampling”, Chemometr. Intell. Lab. Syst. 77 (1–2), 261–277 (2005).

[11] F. F. Pitard, “Pierre Gy’s Sampling Theory and Sampling Practice” (CRC Press, Boca Raton, 2nd edition, 1993).

[12] F. F. Pitard, “Pierre Gy’s Theory of Sampling and C. O. Ingamell’s Poisson Process Approach. Pathways to Representative Sampling and Appropriate Industrial Standards”, Dr. Techn. Thesis, Aalborg University, Denmark. ISBN: 978-87-7606-032-9.

[13] P. L. Smith, “A Primer for Sampling Solids, Liquids and Gases – Based on the Seven Sampling Errors of Pierre Gy” (ASA SIAM, USA, 2001).

[14] F. F. Pitard, D. Francois-Bongarcon, “Demystifying the Fundamental Sampling Error and the Grouping and Segregation Error for practitioners”, Proceedings World Conference on Sampling and Blending 5 (WCSB), 39-56, Santiago, Chile (2011).

[15] DS-3077, “Representative Sampling – HORIZONTAL Standard (draft CEN proposal). www.ds.dk. DS 3077:2013. ICS: 03. 120. 30; 13.080.05 (2013).

[16] H. Janssen, D. Benanou, and F. David,  “Three Wizards of Sample Preparation”, The Analytical Scientist, 18, 20–26 (2014).

[17] C. Wagner, “Non-representative sampling versus data reliability – Improving monitoring reliability of fuel combustion processes in large-scale coal and biomass-fired power plants”, PhD Thesis, Aalborg University, Denmark. ISBN: 978-87-93100-49-7 (2013).

[18] K. H. Esbensen, “The Critical Role of Sampling,” The Analytical Scientist, 12, 21-22 (2014).

Glossary

A B C D E F G H I J K L M N O P Q R S T U V W X Y Z

A

Aliquot

An aliquot is the ultimate sub-sample extracted in a 'Lot-to-Aliquot' pathway for analysis. By analogy, process analytical technology involves the extraction of virtual samples, which are defined volumes of matter interacting with a process analytical instrument.

Analysis

Analysis is the systematic examination and evaluation of the ultimate sub-sample of chemical, biological, or physical substance (Aliquot) to determine its composition, structure, properties, or presence of specific components.

Analytical Bias

Analytical bias is a systematic deviation of measured values from true values.  An analytical bias can arise from multiple sources, including instrument calibration errors, sample preparation techniques, operator method, or inherent methodological limitations. Unlike random errors, which fluctuate unpredictably, analytical bias consistently skews results in a particular direction. Identifying and correcting this bias is crucial to ensure the accuracy and reliability of analytical data (bias correction).

Analytical Precision

Analytical precision refers to the degree of agreement among repeated analyses of the same aliquot under identical conditions. It reflects the consistency and reproducibility of the results obtained by a given analytical method. High precision indicates minimal random analytical error and close clustering of analytical results around an average. Precision does not necessarily imply accuracy, as a method can be precise yet still yield systematically biased results. 

C

Composite Sampling

Composite sampling extracts a number (Q) of  Increments, established to capture the Lot Heterogeneity. Composite sampling is the only way to represent heterogeneous material. A composite sample is made by aggregating the Q increments subject to the Fundamental Sampling Principle (FSP). The required amount of increments for the requested Representativity Q can be carefully established to make sampling fit-for-purpose.

Compositional Heterogeneity (CH)

Compositional heterogeneity is the variation between individual fundamental units of a target material (particles, fragments, cells, ...). CH is an intrinsic characteristic of the target material to be sampled.

Correct Sampling Errors (CSE)
CSE are the errors that cannot be eliminated even when sampling correctly (unbiased) according to the Theory of Sampling (TOS). CSE are caused by Lot Heterogeneity and can only be minimised.
There are two Correct Sampling Errors (CSE):
  1. Fundamental Sampling Error (FSE)
  2. Grouping and Segregation Error (GSE)
Crushing
Crushing is the term used for the process of reducing particle size. Other terms are grinding, milling, maceration, comminution. Particle size reduction changes the Compositional Heterogeneity (CH) of a material. Composite Sampling and crushing are the only agents with which to reduce the Fundamental Sampling Error (FSE).

D

Data Format

Data must be reported as the measurement results and the Measurement Uncertainties stemming from sampling and analysis. Note that MUAnalysis and MUSampling are expressed as variances.

Data =            Measurement +/- (MUSampling ; MUAnalysis)

Example:       375 ppm +/- (85 ppm ; 18 ppm)

Note that the Uncertainties 85 ppm and 18 ppm are the square roots of MUSampling and MUAnalysis.

Data Uncertainty
Distributional Heterogeneity (DH)

Distributional heterogeneity is the variation between groups of fundamental units of a target material. Groups of units manifest themselves as Increments used in sampling. DH is an expression of the spatial heterogeneity of a material to be sampled (Lot).

DS3077:2024

This standard is a matrix-independent standard for representative sampling, published by the Danish Standards Foundation. This standard sets out a minimum competence basis for reliable planning, performance and assessment of existing or new sampling procedures with respect to representativity. This standard invalidates grab sampling and other incorrect sampling operations, by requiring conformance with a universal set of six Governing Principles and five Sampling Unit Operations. This standard is based on the Theory of Sampling (TOS).

webshop.ds.dk/en/standard/M374267/ds-3077-2024

Dynamic Lot

A dynamic lot is a moving material stream where sampling is carried out at a fixed location. For both Stationary Lots and Dynamic Lots, sampling procedures must be able to represent the entire lot volume guided by the Fundamental Sampling Principle.

F

Fractionation

Fractionation is a way of processing a Lot or Sample before sampling (or subsampling). Fractionation separates materials/lots into fractions according to particle properties, e.g. size, density, shape, magnetic susceptibility, wettability, conductivity, intrinsic, or introduced moisture ...

Fundamental Sampling Error (FSE)

FSE results from the impossibility to fully compensate for inherent Compositional Heterogeneity (CH) when sampling. FSE is always present in all sampling operations but can be reduced by adherence to TOS' principles. Even a fully representative, non-biased sampling process will be unable to materialise two samples with identical composition due to Lot Heterogeneity. FSE can only be reduced by Crushing (followed by Mixing / Blending) i.e. by transforming into a different material system with smaller particle sizes.

Fundamental Sampling Principle (FSP)

The Fundamental Sampling Principle (FSP) stipulates that all potential Lot Increments must have the same probability of being extracted to be aggregated as a Composite Sample. Sampling processes in which certain areas, volumes, parts of a Lot are not physically accessible cannot ensure Representativity.

G

Global Estimation Error (GEE)

The GEE is the total data estimation error, the sum of the Total Sampling Error (TSE) and the Total Analytical Error (TAE).

Governing Principles

Six Governing Principles (GP) describe how to conduct representative sampling of heterogeneous materials:

1) Fundamental Sampling Principle (FSP)

2) Sampling Scale Invariance (SCI)

3) Principle of Sampling Correctness (PSC)

4) Principle of Sampling Simplicity (PSS)

5) Lot Dimensionality Transformation (LDT), and

6) Lot Heterogeneity Characterisation (LHC).

Grab Sampling

Process of extracting a singular portion of the Lot. Grab sampling cannot ensure Representativity for heterogeneous materials. Grab sampling results in a sample designated a Specimen.

Grouping and Segregation Error (GSE)

The GSE originates from the inherent tendency of Lot particles, or fragments hereof, to segregate and/or to group together locally to varying degrees within the full lot volume. This spatial irregularity is called the Distributional Heterogeneity (DH). There will always be segregation and grouping of Lot particles at different scales. GSE plays a significant role in addition to the Fundamental Sampling Error FSE. Unlike FSE however, the effects from GSE can be reduced in a given system state by Composite Sampling and/or Mixing / Blending. GSE can in practice be reduced significantly but is seldomly fully eliminated.

H

Heterogeneity

Heterogeneity refers to the state of being varied in composition. It is often contrasted with homogeneity, which implies complete similarity among components, which is a rare case. For materials in science, technology and industry heterogeneity is the norm. Heterogeneity applies to various contexts, such as populations of non-identical units, bulk materials, powders, slurries, biological swhere multiple distinct components coexist.

Heterogeneity in context of the Theory of Sampling, is described using three distinct characteristics, Compositional Heterogeneity CH, Distributional Heterogeneity DH and Particle-Size Heterogeneity

 

Heterogeneity Testing (HT)

Heterogeneity tests are used for optimizing sampling protocols for a variable of interest (analyte, feature) with regards to minimising the Fundamental Sampling Error (FSE).

Experimental approaches available are the 50-particle method, the heterogeneity test (HT), the sampling tree experiment (STE) or the duplicate series/sample analysis (DSA), and the segregation free analysis (SFA).

Recently, sensor-based heterogeneity tests have been introduced which bring the advantage of cost-effective analysis of large numbers of single particles.

Homogeneity

An assemblage of material units with identical unit size, composition and  characteristics. There are practically no homogenous materials in the realm of technology, industry and commerce (mineral resources, biology, pharmaceuticals, food, feed, environment, manufacturing and more) of interest for sampling. With respect to sampling, it is advantageous to consider that all materials are in practice  heterogeneous.

I

Incorrect Delimitation Error (IDE)

The principle for extracting correct Increments from processes is to delineate a full planar-parallel slice across the full width and depth of a stream of matter (Dynamic Lot. IDE results from delineating any other volume shape. When a sampling system or procedure is not correct relative to the appropriate Increment delineation, a Sampling Bias will result. The resulting error is defined as the Increment Delimitation Error (IDE). Similar IDE definitions apply to delineation and extraction of increments from Stationary Lots.

Incorrect Extraction Error (IEE)

Increments must not only be correctly delimitated but must also be extracted in full. The error incurred by not extracting all particles and fragments within the delimitated increment is the Increment Extraction Error (IEE). IDE and IEE are very often committed simultaneously because of inferior design, manufacturing, implementation or maintenance of sampling equipment and systems.

Incorrect Preparation Error (IPE)

Adverse sampling bias effects may occur for example during sample transport and storage (e.g. mix-up, damage, spillage), preparation (contamination and/or losses), intentional (fraud, sabotage) or unintentional human error (careless actions; deliberate or ill-informed non-adherence to protocols). All such non-compliances with the criteria for representative sampling and good laboratory practices (GLP) are grouped under the umbrella term IPE. The IPE is part of the bias-generating errors ISE that must always be avoided.

Incorrect Sampling Errors (ISE)

There are four ISE, which result from an inferior sampling process. These ISE can and must be eliminated.

  1. Incorrect Delimitation Error (IDE) aka Increment Delimitation Error
  2. Incorrect Extraction Error (IEE) aka Increment Extraction Error
  3. Incorrect Preparation Error (IPE) aka Increment Preparation Error
  4. Incorrect Weighing Error (IWE) aka Increment Weighing Error
Incorrect Weighing Error (IWE)

IWE reflects specific weighing errors associated with collecting Increments. For process sampling, IWE is incurred when extracted increments are not proportional to the contemporary flow rate (dynamic 1-dimensional lots), at the time or place of extraction. IWE is often a relatively easily dealt with appropriate engineering attention. Increments, and Samples, should preferentially represent a consistent mass (or volume).

Increment

Fundamental unit of sampling, defined by a specific mass or correctly delineated volume extracted by a specified sampling tool.

L

Lot

a) A Lot is made up of a specific target material to be subjected to a specified sampling procedure.

b) A Lot is the totality of the volume for which inferences are going to be made based on the final analytical results (for decision-making). Lot size can range from being extremely large (e.g. an ore body, a ship) to very small (e.g. a blood sample).

c) The term Lot refers both to the material as well as to lot size (volume/mass) and physical characteristics. Lots are distinguished as stationary or dynamic lots. A stationary lot is a non-moving volume of material, a dynamic lot is a material stream (Lot Dimensionality). For both stationary and dynamic lots, sampling procedures must address the entire lot volume guided by the Fundamental Sampling Principle (FSP).

Lot Definition

Lot Definition describes the process of defining the target volume, which will be subjected to Sampling.

Lot Dimensionality

TOS distinguishes Lot volume  according to the dimensions that must be covered by correct Increment extraction. This defines the concept of 'lot dimensionality', an attribute which is independent of the lot scale. Lot dimensionality is a characterisation to help understand and optimise sample extraction from any lot at any sampling stage. There are four main lot types: 0-, 1-, 2- and 3-dimensional lots (0-D, 1-D, 2-D and 3-D lots).

Lots are classified by subtracting the dimensions of the lot that are fully 'covered' be the salient sampling extraction tool in question. The higher the number of dimensions fully covered in the resulting sampling operation, the easier it is to reduce the Total Sampling Error TSE.

Lot Dimensionality Transformation (LDT)

By the Governing Principle Lot Dimensionality Transformation LDT, stationary 0-D, 2-D and 3-D lots can in many cases advantageously be transformed into dynamic 1-D lots, enabling optimal sampling. However, the application of LDT has practical limits as some lots cannot be transformed (e.g. a body of soil, or a mine resource, biological cells). The optimal approach for such cases is penetrating one dimension with complete increment extraction (usually height) turning a 3-D lot into a 2-D lot.

Lot Heterogeneity

The lot heterogeneity is the combination of Compositional Heterogeneity, Distributional Heterogeneity and Particle-Size-Heterogeneity.

CH + DH + PH

Lot Heterogeneity Characterisation
Lot Heterogeneity Characterisation is the process of assessing Lot Heterogeneity magnitude. Logically, it is impossible to design a sampling procedure without knowledge of the Heterogeneity of target material. Lot Heterogeneity Characterisation is the process of determining Lot Heterogeneity when approaching a new sampling project. There are two principal procedures of determining Lot Heterogeneity, Replication Experiment (RE) for Stationary Lots, and Variographic Characterisation (VAR) for Dynamic Lots. Heterogeneity Tests determine Constitutional Heterogeneity as the irreducible minimum obtainable of Sampling Variance, excluding all other Sampling Error effects.

M

Mass-Reduction

Mass-reduction is a physical process that divides a given quantity into manageable sub-samples. Mass-reduction must ensure that these sub-samples are representative of the original quantity (Representative Mass Reduction – Subsampling

Measurement

The total process of producing numerical data about a Lot, including sampling and analysis is called Measurement. Simultaneously, sensor-based analytical technology combines virtual sampling and signal processing. For both types of measurements the principles and rules of the  Theory of Sampling apply.

Measurement Uncertainty (metrological term) (MU)

MU expresses the variability interval of values attributed to a quantity measured. MU is the effect of a particular error, e.g. a sampling error, or an analytical error  or of combined effects (see MUTotal).

MUsampling reflects the variability stemming from sampling errors

MUanalysis reflects the variability stemming from analytical errors

MUtotal is the effective variability stemming from both sampling and analysis

MUtotal= MUsampling+ MUanalysis

Mixing / Blending

Mixing and blending reduces Distributional Heterogeneity (DH) before sampling/sub-sampling. N.B. Forceful mixing is a much less effective process than commonly assumed.

P

Particle-Size-Heterogeneity (PH)

PH is the compositional difference due to assemblages of units with different particle sizes (or particle-size classes).

Pierre Gy

The founder of the Theory of Sampling (TOS), Pierre Gy (1924--2015) single-handedly developed the TOS from 1950 to 1975 and spent the following 25 years applying it in key industrial sectors (mining, minerals, cement and metals processing). In the course of his career he wrote nine books and gave more than 250 international lectures on all subjects of sampling. In addition to developing TOS, he also carried out a significant amount of practical R&D. But he never worked at a university; he was an independent researcher and a consultant for nearly his entire career - a remarkable scientific life and achievement.

Precision

Precision is a measure of the variability of quantitative results. The larger the variability, the smaller the precision. In practice, precision is measured as the statistical variance s2 of the quantitative results (square of the standard deviation).

Primary Sample

The initial mass extracted from the lot. The Primary Sample is the product of Composite Sampling and consists of Q Increments. Both the mass of the Primary Sample as well as the number of increments extracted influence the sampling variability. As the primary sampling stage often has by far the largest impact on MUTotal, optimisation always starts at this stage.

Principle of Sampling Correctness (PSC)

The Principle of Sampling Correctness (PSC) states that all TOS' Incorrect Sampling Errors (ISE) shall be eliminated, or a detrimental Sampling Bias will have been introduced.

Principle of Sampling Simplicity (PSS)

PSS states that sampling along the Lot-to-Aliquot can be optimised separately for each (primary, secondary, tertiary ....) sampling stage. Since the Primary Sampling stage is often the dominant source of sampling error, optimization logically shall always begin at this stage.

Process Periodicity Error (PPE)

PPE is incurred if short-, mid- or long-term periodic process behaviour is not corrected for, in which case it may contribute to a sampling bias.

A process sampling strategy must make use of a high enough sampling frequency to uncover such behaviours; the sampling frequency must as a minimum always be higher than twice the most frequent periodicity encountered.

Process Sampling Errors (PSE)

PSE come into effect when Dynamic Lots are being sampled without compensating for process trends or periodicities (Process Trend Error and Process Periodicity Error).

Process Trend Error (PTE)

PTE occurs if mid- to long-term process trends are not corrected for, in which case they may contribute to a Sampling Bias. PTE and Process Periodicity Error PPE may, or may not, occur simultaneously depending on the specific nature of the process to be sampled.

Q

Q

Number of Increments composited to a Sample.

R

R

R is the number of replications of a series of independent complete ‘Lot-to-AliquotMeasurements, made under identical conditions applied in a Replication Experiment.

Replication Experiment (RE)

The replication experiment RE consists of a series of independent complete ‘Lot-to-Aliquot’ analytical determinations, made under identical conditions. The number of replications is termed R. RE provides MUSampling + MUAnalysis.

Representative Mass Reduction – Subsampling

Representative Mass Reduction (RMR) aka sub-sampling. TOS argues why Riffle-Splitting and Vezin-sampling are the only options leading to Representative Mass Reduction.

Representativity

A sampling process is representative if it captures all intrinsic material features, e.g., composition, particle size distribution, physical properties (e.g. intrinsic moisture) of a Lot.Representativity is a characteristic of a sampling process in which the Total Sampling Error and Total Analytical Error have been reduced below a predefined threshold level, the acceptable Total Measurement Uncertainty.
Representativity is the prime objective of all sampling processes. The representativity status of an individual sample cannot be ascertained in isolation, if removed from the context of its full sampling-and-analysis pathway. The characteristic Representative can only be accorded a sampling process that complies with all demands specified by TOS (DS3077:2024).

S

Sample

Extracted portion of a Lot that can be documented to be a result of a representative sampling procedure (non-representatively extracted portions of a Lot are termed Specimens).

Sampling

Sampling is the process of collecting units from a Lot (sampling procedure; sampling process): Grab Sampling or Composite SamplingThere are only two principal types of sampling procedures: Grab Sampling or Composite Sampling.

Sampling Accuracy

Closeness of the analytical result of an Aliquot with regards to the true concentration of the Lot]/glossary]. NB. “sampling accuracy” = “sampling + analytical accuracy”

Sampling Bias

The Sampling Bias is the difference between the true Lot concentration and the average concentration from replicated sampling. Such a difference is a direct function of the Lot Heterogeneity and as such inconstant; it changes with each additional sampling and can therefore not be corrected for. This is the opposite to the Analytical Bias for which correction is often carried out.

Sampling Error Management (SEM)

SEM determines the priorities and tools for all sampling procedures in the following order:

  1. Elimination of Incorrect Sampling Errors (ISE) (unbiased sampling)
  2. Minimisation of the remaining Correct Sampling Errors (CSE)
  3. Estimation and use of s2(FSE) is only meaningful after complete elimination of ISE
  4. Minimisation of Process Sampling Errors
Sampling Manager

The Sampling Manager is the Legal Person accountable for ensuring that all sampling activities are conducted in accordance with scientifically valid principles to achieve representative results. They are responsible for managing the design, implementation, and evaluation of sampling protocols while balancing constraints such as material variability, logistics, and resource limitations. This role requires expertise in the Theory of Sampling (TOS), leadership, project management and stakeholder communication skills.

Sampling Precision

The Sampling Precision is the variance of the series of analytical determinations, for example from a Replication Experiment (RE). Sampling precision always includes the Analytical Precision, since all analysis is always based on an analytical Aliquot, which is the result of a complete 'Lot-to-Aliquot' sampling pathway. Therefore sampling precision = sampling + analysis precision.

Sampling Protocol

Document explaining the undertakings necessary for the sampling process. It contains the tools and procedures from Lot-to-Aliquot[/glossary].

Sampling Scale Invariance (SCI)

The Principle of SSI states that all Sampling Unit Operations (SUO) can be applied identically to all sampling stages, only the scale of sampling tools differs.

Sampling Uncertainty

Sampling Uncertainty is the difficulty of collecting a representative sample due to Lot Heterogeneity; the more heterogeneous the material, the higher the uncertainty associated with any sample attempting to represent the whole Lot.

Sampling Unit Operations (SUO)
A Sampling Unit Operation is a basic step in the 'Lot-to-Aliquot' pathway. Five practical SUOs cover all necessary practical aspects of representative sampling: Composite Sampling, Crushing, Mixing/ Blending, Fractionation, and Representative Mass Reduction - Subsampling.
Secondary Sample

A secondary sample is the product of Representative Mass Reduction - Subsampling from a Primary Sample. Identical nomenclature applies for further Representative Mass Reduction steps (Tertiary...).

Specimen

A specimen is a portion of a larger mass/volume (Lot) extracted by a non-representative sampling process. Grab Sampling results in a specimen.

Stakeholder

A Stakeholder is any entity interested in the result coming from sampling and analysis. Data representing stationary or flowing heterogeneous materials are requested by different parties with a multitude of differing objectives. Stakeholders can be internal, from commercial organisations, public authorities, research and academia or non-governmental organisations.

Stationary Lot

A Stationary Lot is a non-moving volume of material where sampling is carried at from multiple locations, each resulting in an Increment. For both Stationary Lots and Dynamic Lots, sampling procedures must address the entire Lot volume guided by the Fundamental Sampling Principle (FSP).

T

Theory of Sampling (TOS)

TOS Theory and Practice of Sampling: necessary-and-sufficient framework of Governing Principles (GP), Sampling Unit Operations (SUO), Sampling Error Management rules (SEM) together with normative practices and skills needed to ensure representative sampling procedures. TOS is codified in the universal standard DS3077:2024.

Total Analytical Error

TAE is manifested as the Measurement Uncertainty resulting only from analysis (MUAnalysis). TAE includes all errors occurring during assaying and analysis (e.g. related to matrix effects, analytical instrument uncertainty, maintenance, calibration, other), as well as human error.

Total Measurement Uncertainty

Whereas Measurement Uncertainty (MU) is traditionally only addressing analytical determination, e.g. concentration := 375 ppm +/- 18 ppm (MUanalysis), Theory of Sampling (TOS) stipulates reporting analytical results with uncertainty estimates from both sampling and analysis.  This gives users of analytical data the possibility to evaluate the relative magnitudes of MUsampling vs. MUanalysis, enabling fully informed assessment of the true, effective data quality involved. A complete data uncertainty must have this format:

MUTotal = MUSampling + MUAnalysis

The attribute Total Measurement Uncertainty (MUTotal) is the most important factor determining the attribute data quality.

Total Sampling Error (TSE)

The Incorrect Sampling Errors (ISE) and Correct Sampling Errors (CSE) add up to the effective Total Sampling Error (TSE). TSE is causing the Total Uncertainty resulting from material extraction along the sampling pathway from-lot-to-aliquot (MUSampling).

Total Uncertainty Threshold

The acceptable Total Measurement Uncertainty, which must include the Sampling Measurement Uncertainty (MUSampling) and Analytical Measurement Uncertainty (MUAnalysis).

U

V

Variographic Characterisation (VAR)

Variography is a variability characterisation of a dynamic 1-dimensional dynamic lot. A variogram describes variability as a function of Increment pair spacing (in time). Variography is also applied in geostatisctics in describing the variability as a function of spacing/distance between analyses.