1. Introduction to Design of Experiments

Design of Experiments (DOE) is a systematic approach to planning experiments to efficiently explore the relationship between input variables (factors) and output variables (responses). Pioneered by Sir Ronald A. Fisher in the 1920s and later developed by George Box, DOE provides a rigorous framework for scientific investigation and process optimization.

DOE offers several advantages over one-factor-at-a-time (OFAT) experimentation:

2. Fundamental Principles

Randomization

Random assignment of experimental units to treatments ensures that the effects of unknown or uncontrollable factors are distributed randomly across treatment groups, preventing systematic bias and validating statistical inference.

Replication

Repeating experimental runs provides an estimate of experimental error (pure error), enables assessment of statistical significance, and improves precision of effect estimates.

Blocking

Grouping experimental units into homogeneous blocks reduces variability by accounting for known sources of variation (e.g., day-to-day variation, batch effects), increasing the precision of factor effect estimates.

Fisher's Key Insight

"Block what you can, randomize what you cannot." This principle guides the design of efficient experiments that maximize information while controlling for extraneous sources of variation.

3. DOE Terminology

4. Full Factorial Designs

A full factorial design includes all possible combinations of factor levels. For k factors each at n levels, the design requires nk runs.

Full Factorial Size
N = n1 × n2 × ... × nk
For balanced designs with each factor at the same number of levels: N = nk
Example: 32 Full Factorial

Two factors (Temperature, Pressure) each at 3 levels:

  • Temperature: Low (100°C), Medium (150°C), High (200°C)
  • Pressure: Low (1 atm), Medium (2 atm), High (3 atm)

Total runs: 32 = 9 unique treatment combinations

5. Two-Level Factorial Designs (2k)

Two-level factorials are particularly useful for screening experiments and investigating first-order effects. Factors are coded to -1 (low) and +1 (high).

Coded Factor Values
x = (X - X̄) / (Xhigh - Xlow)/2
Where X is the natural value, X̄ is the center point, converting factors to [-1, +1] scale.

22 Factorial Design Matrix

Run A B AB
1-1-1+1
2+1-1-1
3-1+1-1
4+1+1+1

Effect Calculations for 2k Designs

Main Effect of Factor A
EffectA = ȲA+ - ȲA-
The average response at high level minus average response at low level.
Interaction Effect AB
EffectAB = (1/2)[(ȲA+B+ - ȲA-B+) - (ȲA+B- - ȲA-B-)]
Measures whether the effect of A depends on the level of B.

6. Fractional Factorial Designs

When full factorials become too large, fractional factorial designs provide a subset that estimates main effects and lower-order interactions with fewer runs.

Fractional Factorial Notation
2k-p design
k factors in 2k-p runs. The fraction is 1/2p of the full factorial.

Resolution

Design resolution indicates the degree of aliasing:

Example: 23-1 Design (Half-Fraction)

Three factors in 4 runs instead of 8. Generator: C = AB

Defining relation: I = ABC

Alias structure: A = BC, B = AC, C = AB

This is a Resolution III design.

7. ANOVA for Factorial Designs

Analysis of Variance (ANOVA) partitions total variation into components attributable to factors, interactions, and error.

Total Sum of Squares Decomposition (Two-Factor)
SSTotal = SSA + SSB + SSAB + SSError
Sum of Squares for Factor A
SSA = bn ∑i (Ȳi.. - Ȳ...)2
Where b = number of levels of B, n = replicates, Ȳi.. = mean at level i of A.

F-Tests for Significance

F = MSEffect / MSError
Compare to Fα, dfEffect, dfError. Reject H0 if F > Fcritical.
Source SS df MS F
Factor ASSAa-1SSA/(a-1)MSA/MSE
Factor BSSBb-1SSB/(b-1)MSB/MSE
A × BSSAB(a-1)(b-1)SSAB/dfMSAB/MSE
ErrorSSEab(n-1)SSE/df-
TotalSSTabn-1--

8. Effect Estimation and Significance

Contrast Method

For 2k designs, effects are estimated using contrasts:

Effect Estimate
Effect = (2/N) × Contrast = (2/N) ∑ (sign) × yi

Standard Error of Effects

With replication:

SEEffect = √(4MSE / nN)
Where n = replicates per treatment, N = total unique treatments.

Normal Probability Plot

For unreplicated designs, plot ordered effects against normal quantiles. Significant effects will deviate from the line formed by negligible (noise) effects.

9. Response Surface Methodology (RSM)

RSM is used to optimize processes by fitting polynomial models to experimental data and finding factor settings that maximize or minimize the response.

First-Order Model

y = β0 + ∑ βixi + ε

Second-Order Model

Full Quadratic Model
y = β0 + ∑ βixi + ∑ βiixi2 + ∑∑ βijxixj + ε
Includes linear, quadratic, and interaction terms to model curvature.

Central Composite Design (CCD)

CCD is the most popular RSM design, consisting of:

Rotatability Condition
α = (2k)1/4
For rotatable designs, prediction variance is constant at all points equidistant from the center.

Box-Behnken Design

An alternative RSM design that does not include corner points, useful when extreme factor combinations should be avoided. Requires fewer runs than CCD but does not estimate all model terms as efficiently.

10. Taguchi Methods

Developed by Genichi Taguchi, these methods focus on robust design - making processes insensitive to noise factors.

Signal-to-Noise Ratios

Smaller-is-Better
S/N = -10 log10[(1/n) ∑ yi2]
Larger-is-Better
S/N = -10 log10[(1/n) ∑ (1/yi2)]
Nominal-is-Best
S/N = 10 log10(Ȳ2 / s2)
Maximizes the ratio of mean to variance.

Orthogonal Arrays

Taguchi uses standardized orthogonal arrays (L8, L9, L16, etc.) as fractional factorial designs for efficient experimentation.

11. Optimal Design

Optimal designs use mathematical criteria to select the best experimental points.

D-Optimality

Maximizes |XTX|, minimizing the volume of the confidence ellipsoid for parameter estimates.

I-Optimality

Minimizes average prediction variance over the design region.

A-Optimality

Minimizes the trace of (XTX)-1, the average variance of parameter estimates.

When to Use Optimal Designs
  • Irregular experimental regions
  • Constraints on factor combinations
  • Mixture experiments
  • Limited resources for experimentation

References and Further Reading

  1. Montgomery, D.C. (2017). Design and Analysis of Experiments, 9th Edition. Wiley.
  2. Box, G.E.P., Hunter, J.S., & Hunter, W.G. (2005). Statistics for Experimenters, 2nd Edition. Wiley.
  3. Myers, R.H., Montgomery, D.C., & Anderson-Cook, C.M. (2016). Response Surface Methodology, 4th Edition. Wiley.
  4. Taguchi, G. (1986). Introduction to Quality Engineering. Asian Productivity Organization.
  5. Wu, C.F.J. & Hamada, M.S. (2021). Experiments: Planning, Analysis, and Optimization, 3rd Edition. Wiley.