Lognormal Distribution
The lognormal distribution describes data whose logarithm follows a normal distribution. It is right-skewed, bounded below by zero, and commonly arises in manufacturing processes involving multiplicative effects — such as particle sizes, surface roughness, and chemical concentrations.
Why It Matters
Many quality characteristics that engineers assume are normal are actually lognormal. Surface roughness (Ra), particle contamination counts, assembly clearances, and chemical purity measurements often exhibit the characteristic right skew of a lognormal distribution. The telltale sign: values cannot go below zero, the distribution tails off gradually to the right, and the data looks normal on a log scale.
The practical problem is that computing Cpk assuming normality on lognormal data produces misleading results. For right-skewed data, the normal model underestimates the upper tail probability (risk of exceeding the upper spec) and overestimates the lower tail probability (risk of going below the lower spec). A process that appears capable (Cpk = 1.4 under normality) might actually produce significant upper-spec-limit exceedances.
The traditional workaround is to log-transform the data and compute Cpk on the transformed scale. But this changes the interpretation — Cpk on log-transformed data does not directly correspond to the fraction nonconforming on the original measurement scale. Engineers must back-transform to get meaningful defect rate estimates, and this step is often skipped or done incorrectly.
The EntropyStat Perspective
EntropyStat handles lognormal data without transformation. The EGDF fits the distribution in the original measurement space, capturing the right skew directly. Capability indices are computed from the actual distribution shape, so the Cpk value reflects the true fraction nonconforming without requiring log transformation or back-transformation.
This eliminates a common source of confusion and error. When engineers log-transform data for capability analysis, specification limits must also be transformed — and asymmetric specifications on the original scale become asymmetric in a different way on the log scale. The EGDF works directly on the raw measurements with the original specification limits, producing results that are immediately interpretable.
The EGDF is also more robust than lognormal fitting for data that is "approximately lognormal." Real manufacturing data rarely follows any theoretical distribution exactly. A process might be mostly lognormal but with a few measurements suggesting a second mode (perhaps from a different material lot). The EGDF captures this deviation naturally, while forcing a lognormal fit would smooth over it — potentially hiding actionable process information.
Related Terms
Normal Distribution
The normal (Gaussian) distribution is a symmetric, bell-shaped probability distribution fully described by its mean and standard deviation. It is the foundational assumption behind most classical statistical quality methods, including Cpk, Shewhart charts, and Six Sigma calculations.
Non-Normal Data
Non-normal data is process data whose distribution does not follow the Gaussian (bell curve) pattern. Common non-normal patterns in manufacturing include skewed distributions, bimodal distributions, truncated distributions, and heavy-tailed distributions.
Distribution Fitting
Distribution fitting is the process of finding a probability distribution that best describes a dataset. Traditional methods involve selecting a parametric family (normal, Weibull, lognormal) and estimating its parameters, then validating the fit with a goodness-of-fit test.
Weibull Distribution
The Weibull distribution is a versatile probability distribution widely used in reliability engineering and failure analysis. Its shape parameter allows it to model increasing failure rates (wear-out), constant failure rates (random failures), or decreasing failure rates (early mortality).
Process Capability (Cpk/Ppk)
Process capability indices (Cpk and Ppk) quantify how well a manufacturing process can produce parts within specification limits. Cpk measures short-term capability using within-subgroup variation, while Ppk measures long-term performance using overall variation.
Related Articles
First Pass Yield vs. Cpk: Which Metric Tells the Real Story?
First pass yield says 98.2%. Cpk says 0.94. One measures what happened. The other predicts what will happen next. When they disagree, something important is hiding — and knowing which to trust prevents costly mistakes.
Mar 17, 2026
PPAP Submissions: Capability Evidence That Survives Customer Audits
Your PPAP got rejected — not for bad parts, but for bad statistics. OEM auditors now scrutinize whether your Cpk method matches your data. Build a PPAP capability evidence chain that withstands the toughest audits.
Mar 14, 2026
The Distribution Fitting Trap: Weibull, Lognormal, or None of the Above?
Distribution fitting replaces the normality assumption with a different guess. With typical sample sizes, Weibull, lognormal, and gamma all pass goodness-of-fit tests — giving different Cpk values. The distribution fitting step that should fix your analysis becomes its own error source.
Mar 13, 2026
See Entropy-Powered Analysis in Action
Upload your data and compare traditional SPC with entropy-based methods. Free demo — no credit card required.