Subgroup Analysis
Subgroup analysis divides process data into rational subgroups — small groups of measurements collected under similar conditions (same machine, operator, material lot, time window). Variation within subgroups estimates short-term process noise, while variation between subgroups reveals shifts and trends.
Why It Matters
Rational subgrouping is the foundation of meaningful SPC. The choice of subgroup size and composition determines whether control charts detect real process changes or just reflect noise. A poorly chosen subgroup strategy can make a capable process look unstable or an unstable process look capable.
The classic example: subgrouping consecutive parts from one machine captures within-machine variation but may miss between-machine differences. Subgrouping across machines captures total variation but masks machine-specific shifts. The "right" subgroup strategy depends on what sources of variation you need to detect versus those you want to include in baseline variability.
Subgroup size also affects statistical performance. Larger subgroups (n = 5–10) provide tighter control limits on X-bar charts (better shift detection) but may mask within-subgroup non-normality. Individual charts (n = 1) preserve distributional information but have wider limits and lower sensitivity to small shifts.
The EntropyStat Perspective
EntropyStat enhances subgroup analysis by applying the EGDF separately to each subgroup or to within-subgroup residuals, without requiring the normality assumption that traditional X-bar/R charts rely on. For individual measurements (subgroup size = 1), this is especially valuable because the Central Limit Theorem does not apply and normality cannot be invoked through averaging.
The ELDF provides a data-driven approach to subgroup validation. If you believe your subgroups are rational (homogeneous), the ELDF can test this by checking whether measurements within each subgroup form a single cluster. If the ELDF detects multiple clusters within a subgroup, it suggests the subgroup definition is mixing distinct process conditions — the subgroup is not truly rational.
This automated subgroup validation replaces the manual, judgment-based process of determining rational subgroups. Engineers typically rely on process knowledge to define subgroups, then hope the choice was correct. EntropyStat's homogeneity testing provides statistical confirmation or rejection of the subgrouping strategy, strengthening the foundation on which all subsequent SPC analysis is built.
Related Terms
Control Charts
Control charts are time-ordered plots of a process measurement with statistically derived upper and lower control limits. They visually separate normal process variation from signals that indicate the process has shifted or become unstable.
Statistical Process Control (SPC)
Statistical Process Control is a methodology that uses statistical methods to monitor and control a manufacturing process. SPC distinguishes between common-cause variation (inherent to the process) and special-cause variation (assignable to specific events).
Homogeneity Testing
Homogeneity testing determines whether a dataset comes from a single statistical population or contains multiple subpopulations. In manufacturing, non-homogeneous data indicates that the process was not operating in a single stable mode during data collection.
ANOVA (Analysis of Variance)
ANOVA is a statistical method that tests whether the means of three or more groups differ significantly. It partitions total variation into between-group and within-group components, determining if observed group differences exceed what random variation alone would produce.
Process Stability
Process stability means a process is operating in statistical control — only common-cause variation is present, and the process distribution is consistent over time. A stable process is predictable: its mean, spread, and shape do not change from sample to sample.
Related Articles
Process Drift Detection Without False Alarms
Process drift hides under false alarms. Shewhart charts catch sudden shifts but miss gradual process drift — while Nelson rules fire on stable data. Entropy-based homogeneity testing separates real drift from noise without chart configuration.
Mar 12, 2026
Hidden Clusters in Your Process Data — and Why Cpk Hides Them
Hidden clusters from multi-cavity molds, shift changes, and material lots produce aggregate Cpk that looks capable — while one subpopulation ships defects. ELDF detects what Cpk can’t see.
Mar 11, 2026
Small Sample Capability: How to Trust Cpk With Only 10 Parts
With a small sample of 10 parts, traditional Cpk has a confidence interval 0.6 units wide — your 1.38 could be anywhere from 1.05 to 1.71. Entropy-based methods extract more from limited data without the normality assumption.
Mar 7, 2026
See Entropy-Powered Analysis in Action
Upload your data and compare traditional SPC with entropy-based methods. Free demo — no credit card required.