Usage Notes

The functions in this chapter cover a wide variety of commonly used experimental designs. They can be categorized, not only based upon the underlying experimental design that generated the user’s data, but also on whether they provide support for missing values, factorial treatment structure, blocking and replication of the entire experiment, or multiple locations.

Typically, responses are stored in the input vector y. For a few functions, such as anovaOneway and anovaFactorial the full set of model subscripts is not needed to identify each response. They assume the usual pattern, which requires that the last model subscript change most rapidly, followed by the model subscript next in line, and so forth, with the first subscript changing at the slowest rate. This pattern is referred to as lexicographical ordering.

However, for most of the functions in this chapter, one or more arrays are used to describe the experimental conditions associated with each value in the response input vector y. The function splitPlot, for example, requires three additional input arrays: split, whole and rep. They are used to identify the split-plot, whole-plot and replicate number associated with each value in y.

Many of the functions described in this chapter permit users to enter missing data values using NaN (Not a Number) as the missing value code. Use function machine (or function machine with the double-precision) to retrieve NaN. Any element of y that is missing must be set to machine(6). See machine in Chapter 15, Utilities for a description. Functions anovaFactorial, anovaNested and anovaBalanced require complete, balanced data, and do not accept missing values.

As a diagnostic tool for validating model assumptions, some functions in this chapter perform a test for lack of fit when replicates are available in each cell of the experimental design.

Completely Randomized Experiments

Completely randomized experiments are analyzed using some variation of the one-way analysis of variance (Anova). A completely randomized design (CRD) is the simplest and most common example of a statistically designed experiment. Researchers using a CRD are interested in comparing the average effect of two or more treatments. In agriculture, treatments might be different plant varieties or fertilizers. In industry, treatments might be different product designs, different manufacturing plants, different methods for delivering the product, etc. In business, different business processes, such as different shipping methods or alternate approaches to a product repair process, might be considered treatments. Regardless of the area, the one thing they have in common is that random errors in the observations cause variations in differences between treatment observations, making it difficult to confirm the effectiveness of one treatment to another.

If observations on these treatments are completely independent then the design is referred to as a completely randomized design or CRD. The PyIMSL Numerical Libraries has two functions for analysis of data from CRD: anovaOneway and crdFactorial.

Both functions allow users to specify observations with missing values, have unequal group sizes, and output treatment means and standard deviations. The primary difference between the functions is that:

  1. anovaOneway conducts multiple comparisons of treatment functions; whereas crdFactorial requires users to make a call to multipleComparisons to compare treatment means.
  2. crdFactorial can analyze treatments with a factorial treatment structure; whereas anovaOneway does not analyze factorial structures.
  3. crdFactorial can analyze data from CRD experiments that are replicated across several blocks or locations. This can happen when the same experiment is repeated at different times or different locations.

Factorial Experiments

In some cases, treatments are identified by a combination of experimental factors. For example, in an octane study comparing several different gasolines, each gasoline could be developed using a combination of two additives, denoted below in Table 4.1, as Additive A and Additive B.

Table 4.1 — 2x2 Factorial Experiment
Treatment Additive A Additive B
1 No No
2 Yes No
3 No Yes
4 Yes Yes

This is referred to as a 2x2 or \(2^2\) factorial experiment. There are 4 treatments involved in this study. One contains no additives, i.e., Treatment 1. Treatment 2 and 3 contain only one of the additives and treatment 4 contains both. A one-way anova, such as found in anovaOneway can analyze these data as four different treatments. Three functions, crdFactorial, rcbdFactorial and anovaFactorial will analyze these data exploiting the factorial treatment structure. These functions allow users to answer structural questions about the treatments such as:

  1. Are the average effects of the additives statistically significant? This is referred to as the factor main effects.
  2. Is there an interaction effect between the additives? That is, is the effectiveness of an additive independent of the other?

Both crdFactorial and rcbdFactorial support analysis of a factorial experiment with missing values and multiple locations. The function anovaFactorial does not support analysis of experiments with missing values or experiments replicated over multiple locations. The main difference, as the names imply, between crdFactorial and rcbdFactorial is that crdFactorial assumes that treatments were completely randomized to experimental units. Function rcbdFactorial assumes that treatments are blocked.

Blocking

Blocking is an important technique for reducing the impact of experimental error on the ability of the researcher to evaluate treatment differences. Usually this experimental error is caused by differences in location (spatial differences), differences in time (temporal differences) or differences in experimental units. Researchers refer to these as blocking factors. They are identifiable causes known to cause variation in observations between experimental units.

There are several functions that specifically support blocking in an experiment: rcbdFactorial, lattice, and latinSquare. The first two functions, rcbdFactorial and lattice, support blocking on one factor.

A requirement of RCBD experiments is that every block must contain observations on every treatment. However, when the number of treatments ( t ) is greater than the block size ( b ), it is impossible to have every block contain observations on every treatment.

In this case, when \(t>b\), an incomplete block design must be used instead of a RCBD. Lattice designs are a type of incomplete block design in which the number of treatments is equal to the square of an integer such as t = 9, 16, 25, etc. Lattice designs were originally described by Yates (1936). The function lattice supports analysis of data from lattice experiments.

Besides the requirement that \(t=k^2\), another characteristic of lattice experiments is that blocks be grouped into replicates, where each replicate contains one observation for every treatment. This forces the number of blocks in each replicate to be equal to the number of observations per block. That is, the number of blocks per replicate and the number of observations per block are both equal to \(k=\sqrt{t}\).

In addition, the number of replicate groups in Lattice experiments is always less than or equal to \(k+1\). If it is equal to \(k+1\) then the design is referred to as a Balanced Lattice. If it is less than \(k+1\) then the design is referred to as a Partially Balanced Lattice. Tables of these experiments and their analysis are tabulated in Cochran & Cox (1950).

Consider, for example, a 3 × 3 balanced-lattice, i.e., \(k=3\) and \(t=9\). Notice that the number of replicates is \(r=k+1=4\). And the number of blocks per replicate and block size are both \(k=3\). The total number of blocks is equal to \(b=\mathrm{n\_locations}\cdot r\cdot(k-1)+1\). For a balanced-lattice, \(b=r\cdot k=(k+1) \cdot k=\left( \sqrt{t}+1 \right)\cdot\sqrt{t}=4\cdot 3=12\).

Table 4.2 — A 3x3 Balanced-Lattice for Nine Treatments in Four Replicates
Replicate I Replicate II
Block 1 (T1, T2, T3) Block 4 (T1, T4, T7)
Block 2 (T4, T5, T6) Block 5 (T2, T5, T8)
Block 3 (T7, T8, T9) Block 6 (T3, T6, T9)
Replicate III Replicate IV
Block 1 (T1, T2, T3) Block 4 (T1, T4, T7)
Block 2 (T4, T5, T6) Block 5 (T2, T5, T8)
Block 3 (T7, T8, T9) Block 6 (T3, T6, T9)

The Anova table for a balanced-lattice experiment, takes the form shared with other balanced incomplete block experiments. In these experiments, the error term is divided into two components: the Inter-Block Error and the Intra-Block Error. For single and multiple locations, the general format of the Anova tables for Lattice experiments is illustrated in Table 4.3 and Table 4.4.

Table 4.3 — The Anova Table for a Lattice Experiment at One Location
Source DF Sum of Squares Mean Squares
REPLICATES \(t-1\) SSR MSR
TREATMENTS(unadj) \(t-1\) SST MST
TREATMENTS(adj) \(t-1\) SSTa MSTa
BLOCKS(adj) \(r\cdot(k-1)\) SSBa MSBa
INTRA-BLOCK ERROR \((k-1)(r\cdot k-k-1)\) SSE MSE
TOTAL \(r\cdot t-1\) SSR
Table 4.4 — The Anova Table for a Lattice Experiment at Multiple Locations
Source DF Sum of Squares Mean Squares
LOCATIONS \(p-1\) SSL MSL
REPLICATES WITHIN LOCATIONS \(p(r-1)\) SSR MSR
TREATMENTS(unadj) \(t-1\) SST MST
TREATMENTS(adj) \(t-1\) SSTa MSTa
BLOCKS(adj) \(p\cdot r(k-1)\) SSB MSB
TOTAL \(p(k-1)(r\cdot k-k-1)\) SSTot

Latin Square designs are very popular in cases where:

  1. two blocking factors are involved
  2. the two blocking factors do not interact with treatments, and
  3. the number of blocks for each factor is equal to the number of treatments.

Consider an octane study involving 4 test vehicles tested in 4 bays with 4 test gasolines. This is a natural arrangement for a Latin square experiment. In this case there are 4 treatments, and two blocking factors, test vehicle and bay, each with 4 levels. The Latin Square for this example would look like the following arrangement.

Table 4.5 — A Latin Square Design for t=4 Treatments
  Test Vehicle
1 2 3 4
Test 1 A C B D
  2 D B A C
Bay 3 C A D B
  4 B D C A

As illustrated above in Table 4.5, the letters A-D are used to denote the four test gasolines, or treatments. The assignment of each treatment to a particular test vehicle and test bay is described in Table 4.5. Gasoline A, for example, is tested in the following four vehicle/bay combinations: (1/1), (2/3), (3/2), and (4/4).

Notice that each treatment appears exactly once in every row and column. This balance, together with the assumed absence of interactions between treatments and the two blocking factors is characteristic of a Latin Square.

The corresponding Anova table for these data contains information on the blocking factors as well as treatment differences. Notice that the F-test for one of the two blocking factors, test vehicle, is statistically significant (\(p=0.048\)); whereas the other, test bay, is not statistically significant (\(p=0.321\)).

Some researchers might use this as a basis to remove test bay as a blocking factor. In that case, the design can then be analyzed as a RCBD experiment since every treatment is repeated once and only once in every block, i.e., test vehicle.

Table 4.6 — Latin Square Anova Table for Octane Experiment
Source Degrees of Freedom Sum of Squares Mean Squares F-Test p-Value
Test Vehicle 3 1.5825 0.5275 4.83 0.048
Test Bay 3 0.0472 0.157 1.44 0.321
Gasoline 3 4.247 1.416 12.97 0.005
Error 6 0.655 0.109  
TOTAL 15 6.9575  

Multiple Locations

It is common for a researcher to repeat an experiment and then conduct an analysis of the data. In agricultural experiments, for example, it is common to repeat an experiment at several different farms. In other cases, a researcher may want to repeat an experiment at a specified frequency, such as week, month or year. If these repeated experiments are independent of one another then we can treat them as multiple locations.

Several of the functions in this chapter allow for multiple locations: crdFactorial, rcbdFactorial, lattice, latinSquare, splitPlot, splitSplitPlot, stripPlot, stripSplitPlot. All of these functions allow for analysis of experiments replicated at multiple locations. By default they all treat locations as a random factor. Function splitPlot also allows users to declare locations as a fixed effect.

Split-Plot Designs – Nesting and Restricted Randomization

Originally, split-plot designs were developed for testing agricultural treatments, such as varieties of wheat, different fertilizers or different insecticides. In these original experiments, growing areas were divided into plots. The major treatment factor, such as wheat variety, was randomly assigned to these plots. However, in addition to testing wheat varieties, they wanted to test another treatment factor such as fertilizer. This could have been done using a CRD or RCBD design. If a CRD design was used then treatment combinations would need to be randomly assigned to plots, such as shown below in Table 4.7.

Table 4.7 — Completely Randomized Experiments – Both Factors Randomized
CRD
W3F2 W1F3 W4F1 W2F1
W2F3 W1F1 W1F3 W1F2
W2F2 W3F1 W2F1 W4F2
W3F2 W1F1 W2F3 W1F2
W4F1 W3F2 W3F2 W4F3
W4F3 W3F1 W2F2 W4F2

In the CRD illustration above, any plot could have any combination of wheat variety (W1, W2, W3 or W4) and fertilizer (F1, F2 or F3). There is no restriction on randomization in a CRD. Any of the \(t=4\times 3=12\) treatments can appear in any of the 24 plots.

If a RCBD were used, all \(t=12\) treatment combinations would need to be arranged in blocks similar to what is described in Table 4.8, which places one restriction on randomization.

Table 4.8 — Randomized Complete Block Experiments – Both Factors Randomized Within a Block
RCBD
Block 1 W3F3 W1F3 W4F1 W4F3
W2F3 W1F1 W3F2 W1F2
W2F2 W3F1 W2F1 W4F2
Block 2 W3F2 W1F1 W2F3 W1F2
W4F1 W1F3 W3F2 W4F3

The RCBD arrangement is basically a replicated CRD design with a randomization restriction that treatments are divided into two groups of replicates which are assigned to a block of land. Randomization of treatments only occurs within each block.

At first glance, a split-plot experiment could be mistaken for a RCBD experiment since it is also blocked. The split-plot arrangement with only one replicate for this experiment is illustrated below in Table 4.9. Notice that it appears as if levels of the fertilizer factor (F1, F2, and F3) are nested within wheat variety (W1, W2, W3 and W4), however that is not the case. Varieties were actually randomly assigned to one of four rows in the field. After randomizing wheat varieties, fertilizer was randomized within wheat variety.

Table 4.9 — A Split-Plot Experiment for Wheat (W) and Fertilizer (F)
Split-Plot Design
Block 1 W2 W2F1 W2F3 W2F2
W1 W1F3 W1F1 W1F2
W4 W4F1 W4F3 W4F2
W3 W3F2 W3F1 W3F3
Block 2 W3 W3F2 W3F1 W3F3
W1 W1F3 W1F1 W1F2
W4 W4F1 W4F3 W4F2
W2 W2F1 W2F3 W2F2

The essential distinction between split-plot experiments and completely randomized or randomized complete block experiments is the presence of a second factor that is blocked, or nested, within each level of the first factor. This second factor is referred to as the split-plot factor, and the first is referred to as the whole-plot factor.

Both factors are randomized, but with a restriction on randomization of the second factor, the split-plot factor. Whole plots (wheat variety) are randomly assigned, without restriction to plots, or rows in this example. However, the randomization of split-plots (fertilizer) is restricted. It is restricted to random assignment within whole-plots.

Strip-Plot Designs

Strip-plot experiments look similar to split-plot experiments. In fact they are easily confused, resulting in incorrect statistical analyses. The essential distinction between strip-plot and split-plot experiments is the application of the second factor. In a split-plot experiment, levels of the second factor are nested within the whole-plot factor (see Table 4.11). In strip-plot experiments, the whole-plot factor is completely crossed with the second factor (see Table 4.10).

This occurs, for example, when an agricultural field is used as a block and the levels of the whole-plot factor are applied in vertical strips across the entire field. Levels of the second factor are assigned to horizontal strips across the same block.

Table 4.10 — Strip-Plot Experiments – Strip-Plots Completely Crossed
  Whole-Plot Factor
A2 A1 A4 A3
Strip Plot B3 A2B3 A1B3 A4B3 A3B3
B1 A2B1 A1B1 A4B1 A3B1
B2 A2B2 A1B2 A4B2 A3B2
Table 4.11 — Split-Plot Experiments – Split-Plots Nested within Strip-Plots
Whole-Plot Factor
A2 A1 A4 A3
A2B1 A1B3 A4B1 A3B3
A2B3 A1B1 A4B3 A3B1
A2B2 A1B2 A4B2 A3B2

As described in the previous section, in a split-plot experiment the second experimental factor, referred to as the split-plot factor, is nested within the first factor, referred to as the whole-plot factor.

Consider, for example, the semiconductor experiment described in Figure 4.1, “Split-Plot Randomization.” The wafers from each plater, the whole-plot factor, are divided into equal size groups and then randomly assigned to an etcher, the split-plot factor. Wafers from different platers are etched separately from those that went through another plating machine. Randomization occurred within each level of the whole-plot factor, i.e., plater.

Graphically, as shown below, this arrangement appears similar to a tree or hierarchical structure.

../../_images/csch4-figure1.png

Figure 4.1 — Split-Plot Randomization

Notice that although there are only three etchers, 12 different runs are made using these etchers. The wafers randomly assigned to the first plater and first etcher are processed separately from the wafers assigned to other plating machines.

In a strip-plot experiment, the second randomization of the wafers to etchers occurs differently, see Figure 4.2, “Strip-Plot Semiconductor Experiment.” Instead of randomizing the wafers from each plater to the three etchers and then running them separately from the wafers from another plater, the wafers from each plater are divided into three groups and then each randomly assigned to one of the three etchers. However, the wafers from all four plating machines assigned to the same etcher are run together.

../../_images/csch4-figure2.png

Figure 4.2 — Strip-Plot Semiconductor Experiment

Strip-plot experiments can be analyzed using stripPlot. Function stripPlot returns a strip-plot Anova table with the following general structure:

Table 4.12 — Strip-Plot Anova Table for Semiconductor Experiment
Source DF SS MS F-Test p-Value
Blocks 1 0.0005 0.0005 0.955 0.431
Whole-Plots: Plating Machines 2 0.0139 0.0070 64.39 0.015
Whole-Plot Error 2 0.0002 0.0001 0.194 0.838
Strip-Plots: Etchers 1 0.0033 0.0033 100.0 0.060
Strip-Plot Error 1 <0.0001 <0.0001 0.060 0.830
Whole-Plot x Strip-Plot 2 0.0033 0.0017 2.970 0.251
Whole-Plot x Strip-Plot Error 2 0.0011 0.0006  
TOTAL 11 0.0225  

Split-Split Plot and Strip-Split Plot Experiments

The essential distinction between split-plot and split-split-plot experiments is the presence of a third factor that is blocked, or nested, within each level of the whole-plot and split-plot factors. This third factor is referred to as the sub-plot factor. A split-plot experiment, see Table 4.13, has only two factors, denoted by A and B. The second factor is nested within the first factor. Randomization of the second factor, the split-plot factor, occurs within each level of the first factor.

Table 4.13 — Split-Plot Experiment – Split-Plot B Nested within Whole-Plot A
Whole-Plot Factor
A2 A1 A4 A3
A2B1 A1B3 A4B1 A3B2
A2B3 A1B1 A4B3 A3B1
A2B2 A1B2 A4B2 A3B3

On the other hand, a split-split plot experiment has three factors, illustrated in Table 4.14 by A, B and C. The second factor is nested within the first factor, and the third factor is nested within the second.

Table 4.14 — Split-Split Plot Experiment
Whole-Plot Factor
A2 A1 A4 A3

A2B3C2

A2B3C1

A1B2C1

A1B2C2

A4B1C2

A4B1C1

A3B3C2

A3B3C1

A2B1C1

A2B1C2

A1B1C1

A1B1C2

A4B3C2

A4B3C1

A3B2C2

A3B2C1

A2B2C2

A2B2C1

A1B3C1

A1B3C2

A4B2C1

A4B2C2

A3B1C2

A3B1C1

Contrast the split-split plot experiment to the same experiment run using a strip-split plot design (see Table 4.15). In a strip-split plot experiment factor B is applied in a strip across factor A; whereas, in a split-split plot experiment, factor B is randomly assigned to each level of factor A. In a strip-split plot experiment, the level of factor B is constant across a row; whereas in a split-split plot experiment, the levels of factor B change as you go across a row, reflecting the fact that for split-plot experiments, factor B is randomized within each level of factor A.

Table 4.15 — Strip-Split Plot Experiment, Split-Plots Nested Within Strip-Plot Factors A and B
    Factor A Strip Plots
    A2 A1 A4 A3
Factor B StripPlot B3

A2B3C2

A2B3C1

A1B3C1

A1B3C2

A4B3C2

A4B3C1

A3B3C2

A3B3C1

B1

A2B1C1

A2B1C2

A1B1C1

A1B1C2

A4B1C2

A4B1C1

A3B1C2

A3B1C1

B2

A2B2C2

A2B2C1

A1B2C1

A1B2C2

A4B2C1

A4B2C2

A3B2C2

A3B2C1

In some studies, split-split-plot or strip-split-plot experiments are replicated at several locations. Functions splitSplitPlot and stripSplitPlot can analyze these, even when the number of blocks or replicates at each location is different.

Validating Key Assumptions in Anova

The key output in the analysis of designed experiments is the F-tests in the Anova table for that experiment. The validity of these tests relies upon several key assumptions:

  1. observational errors are independent of one another,
  2. observational errors are Normally distributed, and
  3. the variance of observational errors is homogeneous across treatments.

These are referred to as the independence, Normality and homogeneity of variance assumptions. All of these assumptions are evaluated by examining the properties of the residuals, which are estimates of the observational error for each observation. Residuals are calculated by taking the difference between each observed value in the series and its corresponding estimate. In most cases, the residual is the difference between the observed value and the mean for that treatment.

The independence assumption can be examined by evaluating the magnitude of the correlations among the residuals sorted in the order they were collected. The PyIMSL function autocorrelation (see Chapter 8, Time Series and Forecasting) can be used to obtain these correlations. The autocorrelations, to a maximum lag of about 20, can be examined to identify any that are statistically significant.

Residuals should be independent of one another, which implies that all autocorrelations with a lag of 1 or higher should be statistically equivalent to zero. If a statistically significant autocorrelation is found, leading a researcher to conclude that an autocorrelation is not equal to zero, then this would provide sufficient evidence to conclude that the observational errors are not independent of one another.

The second major assumption for analysis of variance is the Normality assumption. In the PyIMSL Numerical Libraries, functions shapiroWilkNormalityTest, lillieforsNormalityTest , and chiSquaredNormalityTest (see Chapter 7,:doc:/stat/fit/index) can be used to determine whether the residuals are not Normally distributed. A small p‑value from this test provides sufficient evidence to conclude that the observational errors are not Normally distributed.

The last assumption, homogeneity of variance, is evaluated by comparing treatment standard errors. This is equivalent to testing whether \(\sigma_1=\sigma_2=\cdots=\sigma_t\), where \(\sigma_i\) is the standard deviation of the observational error for the i‑th treatment. This test can be conducted using homogeneity. To conduct this test, the residuals, and their corresponding treatment identifiers are passed into homogeneity. It calculates the p‑values for both Bartlett’s and Levene’s tests for equal variance. If a p‑value is below the stated significance level, a researcher would conclude that the within treatment variances are not homogeneous.

Multiple Testing

It is well known that the risk of committing a Type I error (i.e., a false positive or false discovery) increases with the number of tests being performed, even if all the tests use the same cut‑off value for significance. The purpose of a multiple testing correction is to control the risk of false positives when performing multiple tests of significance. Two functions are included in this chapter to correct for multiple tests: multipleComparisons performs multiple comparisons of treatment means using one of four methods to control the overall Type I error rate, and falseDiscoveryRates calculates false discovery rates and associated q‑values, given a set of p‑values resulting from multiple independent tests.

Missing Observations

Missing observations create problems with the interpretation and calculation of F-tests for designed experiments. The approach taken in the functions described in this chapter is to estimate missing values using the Yates method and then to compute the Anova table using these estimates.

Essentially the Yates method, implemented in yates, replaces missing observations with the values that minimize the error sum of squares in the Anova table. The Anova table is calculated using these estimates, with one modification. The total degrees of freedom and the error degrees of freedom are both reduced by the number of missing observations.

For simple cases, in which only one observation is missing, formulas have been developed for most designs. See Steel and Torrie (1960) and Cochran and Cox (1957) for a description of these formulas. However for more than one missing observation, a multivariate optimization is conducted to simultaneously estimate the missing values. For the simple case with only one missing value, this approach produces estimates identical to the published formulas for a single missing value.

A potential issue arises when the Anova table contains more than one form of error, such as split-plot and strip‑plot designs. In every case, missing values are estimated by minimizing the last error term in the table.