Anderson-Darling Test
The Anderson-Darling test is a statistical goodness-of-fit test that measures how well data follows a specified distribution. It gives extra weight to the tails of the distribution, making it more sensitive than the Kolmogorov-Smirnov test for detecting departures from normality.
Why It Matters
The Anderson-Darling test is the most commonly used normality test in quality engineering software. Minitab, JMP, and most SPC packages default to Anderson-Darling when testing whether data follows a normal distribution. Its tail sensitivity makes it well-suited for quality applications where tail behavior directly determines defect rates.
In practice, the test is a prerequisite step before computing capability indices. The workflow goes: collect data → run Anderson-Darling → if normal, compute Cpk with standard formula; if not normal, transform the data or use non-parametric methods. This binary decision point is a weak link in traditional quality analysis because the test itself has known limitations.
With small samples (n < 20), the test lacks statistical power — it cannot reliably detect non-normality even when it exists. With large samples (n > 200), it becomes oversensitive — rejecting normality for trivial departures that have no practical impact on capability calculations. The "right" sample size for reliable Anderson-Darling results falls in a narrow range, and real manufacturing data does not always cooperate.
The EntropyStat Perspective
EntropyStat renders the Anderson-Darling test unnecessary for capability analysis. Since the EGDF learns the actual distribution shape directly from data, there is no normality gate to pass through. The traditional workflow of "test → decide → compute" becomes simply "compute."
This eliminates a significant source of error in quality analysis. Engineers no longer need to interpret p-values, choose significance levels, or decide whether a borderline Anderson-Darling result (say, p = 0.06) means the data is "normal enough." The EGDF produces accurate results regardless of the underlying distribution — normal, skewed, or multimodal.
That said, EntropyStat does use the Kolmogorov-Smirnov statistic as a fit validation tool — but for a fundamentally different purpose. Instead of testing whether data fits a pre-assumed distribution, the K-S test validates how well the EGDF itself fits the data. This is a quality check on the model, not a gatekeeping step that determines which formula to use.
Related Terms
Kolmogorov-Smirnov Test
The Kolmogorov-Smirnov (K-S) test is a nonparametric goodness-of-fit test that measures the maximum distance between an empirical cumulative distribution function and a reference distribution. It determines whether a sample plausibly comes from a specified distribution.
Normal Distribution
The normal (Gaussian) distribution is a symmetric, bell-shaped probability distribution fully described by its mean and standard deviation. It is the foundational assumption behind most classical statistical quality methods, including Cpk, Shewhart charts, and Six Sigma calculations.
Non-Normal Data
Non-normal data is process data whose distribution does not follow the Gaussian (bell curve) pattern. Common non-normal patterns in manufacturing include skewed distributions, bimodal distributions, truncated distributions, and heavy-tailed distributions.
Shapiro-Wilk Test
The Shapiro-Wilk test is a statistical test for normality that compares ordered sample values against their expected values under a normal distribution. It is widely considered the most powerful normality test for small to moderate sample sizes (n < 50).
Chi-Square Test
The chi-square test is a statistical test used for two purposes in quality engineering: testing goodness-of-fit (does observed data match an expected distribution?) and testing independence (are two categorical variables related?). It compares observed frequencies to expected frequencies across categories.
Related Articles
The Distribution Fitting Trap: Weibull, Lognormal, or None of the Above?
Distribution fitting replaces the normality assumption with a different guess. With typical sample sizes, Weibull, lognormal, and gamma all pass goodness-of-fit tests — giving different Cpk values. The distribution fitting step that should fix your analysis becomes its own error source.
Mar 13, 2026
Small Sample Capability: How to Trust Cpk With Only 10 Parts
With a small sample of 10 parts, traditional Cpk has a confidence interval 0.6 units wide — your 1.38 could be anywhere from 1.05 to 1.71. Entropy-based methods extract more from limited data without the normality assumption.
Mar 7, 2026
Why Your SPC Software Lies About Non-Normal Data
Your SPC software computes Cpk assuming data follows a bell curve — but 60–80% of manufacturing data doesn’t. That silent assumption produces capability numbers that are confidently wrong, costing real money in both directions.
Mar 6, 2026
See Entropy-Powered Analysis in Action
Upload your data and compare traditional SPC with entropy-based methods. Free demo — no credit card required.