Contents
Example Using Standard Least Squares
You have data that contains the responses of 30 subjects after treatment by three drugs (Snedecor and Cochran 1967). You want to use the Standard Least Squares personality of the Fit Model platform to test whether there is a difference in the mean response among the three drugs.
1. Open the Drug.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select y and click Y.
When you select the column y to be the Y response, the Fitting Personality becomes Standard Least Squares and the Emphasis is Effect Leverage. You can change these options in other situations.
4. Select Drug and x. Click Add.
Figure 3.2 Completed Fit Model Launch Window
5. Click Run.
Figure 3.3 Standard Least Squares Report Window
After accounting for the effect of the x variable, since the p-value for Drug is large (0.1384), you cannot conclude that there a difference in the mean response among the three drugs.
The Standard Least Squares Report and Options
You can show or hide each report and plot in the Standard Least Squares report window using the options in the red triangle menus. Each response and each effect that you specify in the Fit Model launch window has an associated red triangle menu.
For a description of the red triangle options available for an effect, see
Effect Options.
Regression Reports
Regression reports provide summary information about model fit, effect significance, and model parameters.
Summary of Fit
The Summary of Fit report provides a summary of model fit.
Analysis of Variance
The Analysis of Variance report provides the calculations for comparing the fitted model to a simple mean model.
Parameter Estimates
The Parameter Estimates report shows the estimates of the parameters and a t-test for the hypothesis that each parameter is zero.
Note: For reports that do not appear by default, you can set them to always appear. Select File > Preferences > Platforms > Fit Least Squares.
Effect Tests
The effect tests are joint tests in which all parameters for an individual effect are zero. If an effect has only one parameter, as with simple regressors, then the tests are no different from the t-tests in the Parameter Estimates report.
Note: Parameterization and handling of singularities are different from the SAS GLM procedure. For details about parameterization and handling of singularities, see the
Statistical Details appendix.
Lack of Fit
The Lack of Fit report shows or hides a test assessing if the model has the appropriate effects. In the following situations, the Lack of Fit report does not appear:
• There are no exactly replicated points with respect to the X data, and therefore there are no degrees of freedom for pure error.
• The model is saturated, meaning that the model itself has a degree of freedom for each different x value. Therefore, there are no degrees of freedom for lack of fit.
When observations are exact replicates of each other in terms of the X variables, you can estimate the error variance whether you have the right form of the model. The error that you can measure for these exact replicates is called pure error. This is the portion of the sample error that cannot be explained or predicted by the form that the model uses for the X variables. However, a lack of fit test is not very useful if it has only a few degrees of freedom (not many replicated x values).
The difference between the residual error from the model and the pure error is called the lack of fit error. The lack of fit error can be significantly greater than pure error if you have the wrong functional form of a regressor, or if you do not have enough interaction effects in an analysis of variance model. In that case, you should consider adding interaction terms, if appropriate, or try to better capture the functional form of a regressor.
Table 3.7 shows information about the calculations in the Lack of Fit report.
Estimates
Estimates options provide additional reports and tests for the model parameters.
Show Prediction Expression
The Show Prediction Expression shows the equation used to predict the response.
Example of a Prediction Expression
1. Open the Drug.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select y and click Y.
4. Select Drug and x. Click Add.
5. Click Run.
6. From the red triangle menu next to Response, select Estimates > Show Prediction Expression.
Figure 3.4 Prediction Expression
Sorted Estimates
The Sorted Estimates option produces a different version of the Parameter Estimates report that is more useful in screening situations. This version of the report is especially useful if the design is saturated, since typical reports are less informative.
Example of a Sorted Estimates Report
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select F, Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
Figure 3.5 Sorted Parameter Estimates
The Sorted Parameter Estimates report also appears automatically if the Emphasis is set to Effect Screening and all of the effects have only one parameter.
Note the following differences between this report and the Parameter Estimates report:
• This report does not show the intercept.
• The effects are sorted by the absolute value of the t-ratio, showing the most significant effects at the top.
• A bar chart shows the t-ratio, with lines showing the 0.05 significance level.
• If JMP cannot obtain standard errors for the estimates, relative standard errors appear.
• If there are no degrees of freedom for residual error, JMP constructs t-ratios and p-values using Lenth’s Pseudo-Standard Error (PSE). These quantities are labeled with Pseudo in their name. A note explains the change and shows the PSE. To calculate p-values, JMP uses a degrees of freedom for error of m/3, where m is the number of parameter estimates excluding the intercept.
Expanded Estimates
Use the Expanded Estimates option when there are nominal terms in the model and you want to see the full set of coefficients.
Figure 3.6 Comparison of Parameter Estimates and Expanded Estimates
Notice that the coefficient for Drug(f) appears in the Expanded Estimates report.
Statistical Details for Expanded Estimates
When you have nominal terms in your model, the platform constructs a set of dummy columns to represent the levels in the classification. Full details are shown in the appendix
Statistical Details in the Statistical Details appendix. For
n levels, there are
n - 1 dummy columns. Each dummy variable is a zero-or-one indicator for a particular level, except for the last level, which is coded -1 for all dummy variables. The following table shows an example of the
A1 and
A2 dummy columns when column
A has levels
A1,
A2, and
A3.
These columns do not appear in the report, but they help conceptualize the fitting process. The parameter estimates are the coefficients fit to these columns. In this case, there are two, labeled A[A1] and A[A2]. This coding causes the parameter estimates to be interpreted as how much the response for each level differs from the average across all levels. Suppose, however, that you want the coefficient for the last level, A[A3]. The coefficient for the last level is the negative of the sum across the other levels, because the sum across all levels is constrained to be zero. Although many other codings are possible, this coding has proven to be practical and interpretable.
However, you probably do not want to hand calculate the estimate for the last level. The Expanded Estimates option in the Estimates menu calculates these missing estimates and shows them in a text report. You can verify that the mean (or sum) of the estimates across a classification is zero.
Keep in mind that the Expanded Estimates option with high-degree interactions of two-level factors produces a lengthy report. For example, a five-way interaction of two-level factors produces only one parameter but has 25 = 32 expanded coefficients, which are all the same except for sign changes.
Indicator Parameterization Estimates
This option displays the estimates using the Indicator Variable parameterization. To re-create the report in
Figure 3.7, follow the steps in
Show Prediction Expression with one exception: instead of selecting
Estimates > Show Prediction Expression, select
Estimates > Indicator Parameterization Estimates.
Figure 3.7 Indicator Parameterization Estimates
This parameterization is inspired by the PROC GLM parameterization. Some models match, but others, such as no-intercept models, models with missing cells, and mixture models, will most likely show differences.
Sequential Tests
Sequential Tests show the reduction in the residual sum of squares as each effect is entered into the fit. The sequential tests are also called Type I sums of squares (Type I SS). One benefit of the Type I SS is that they are independent and sum to the regression SS. One disadvantage is that they depend on the order of terms in the model; each effect is adjusted only for the preceding effects in the model.
The following models are considered appropriate for the Type I hypotheses:
• balanced analysis of variance models specified in proper sequence (that is, interactions do not precede main effects in the effects list, and so on)
• purely nested models specified in the proper sequence
• polynomial regression models specified in the proper sequence.
Custom Test
If you want to test a custom hypothesis, select
Custom Test from the
Estimates menu. To jointly test several linear functions, click on
Add Column. This displays the window shown to the left in
Figure 3.8 for constructing the test in terms of all the parameters. After filling in the test, click
Done. The dialog then changes to a report of the results, as shown on the right in
Figure 3.8.
The space beneath the Custom Test title bar is an editable area for entering a test name. Use the Custom Test dialog as follows:
Parameter
lists the names of the model parameters. To the right of the list of parameters are columns of zeros corresponding to these parameters. Click a cell here, and enter a new value corresponding to the test that you want.
Add Column
adds a column of zeros so that you can test jointly several linear functions of the parameters. Use the Add Column button to add as many columns to the test as you want.
The last line in the Parameter list is labeled
=. Enter a constant into this box to test the linear constraint against. For example, to test the hypothesis
β0=1, enter a 1 in the = box. In
Figure 3.8, the constant is equal to zero.
When you finish specifying the test, click Done to see the test performed. The results are appended to the bottom of the dialog.
When the custom test is done, the report lists the test name, the function value of the parameters tested, the standard error, and other statistics for each test column in the dialog. A joint F-test for all columns shows at the bottom.
Caution: The test is always done with respect to residual error. If you have random effects in your model, this test might not be appropriate if you use EMS instead of REML.
Note: If you have a test within a classification effect, consider using the contrast dialog (which tests hypotheses about the least squares means) instead of a custom test.
Figure 3.8 The Custom Test Dialog and Test Results for Age Variable
Joint Factor Tests
This option appears when interaction effects are present. For each main effect in the model, JMP produces a joint test on all the parameters involving that main effect.
Example of a Joint Factor Tests Report
1. Open the Big Class.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select weight and click Y.
4. Select age, sex, and height and click Macros > Factorial to degree.
5. Click Run.
6. From the red triangle next to Response weight, select Estimates > Joint Factor Tests.
Figure 3.9 Joint Factor Tests for Big Class model
Note that age has 15 degrees of freedom; it is testing the five parameters for age, the five parameters for age*sex, and the five parameters for height*age. All parameters tested to be zero.
Inverse Prediction
To find the value of
x for a given
y requires
inverse prediction, sometimes called
calibration. The
Inverse Prediction option on the
Estimates menu lets you ask for the specific value of one independent (
X) variable given a specific value of a dependent variable and other
x values (
Figure 3.11). The inverse prediction computation includes confidence limits (fiducial limits) on the prediction.
Example of Inverse Prediction
1. Open the Fitness.jmp sample data table.
2. Select Analyze > Fit Y by X.
3. Select Oxy and click Y, Response.
4. Select Runtime and click X, Factor.
When there is only a single X, as in this example, the Fit Y by X platform can give you a visual approximation of the inverse prediction values.
5. Click OK.
6. From the red triangle menu, select Fit Line.
Use the crosshair tool to approximate inverse prediction.
7. Select Tools > Crosshairs.
8. Click on the Oxy axis at about 50 and then drag the cursor so that the crosshairs intersect with the prediction line.
Figure 3.10 Bivariate Fit for Fitness.jmp
Figure 3.10 shows which value of
Runtime gives an
Oxy value of 50, intersecting the
Runtime axis at about 9.779, which is an approximate inverse prediction. However, to see the exact prediction of
Runtime, use the Fit Model dialog as follows:
1. From the Fitness.jmp sample data table, select Analyze > Fit Model.
2. Select Oxy and click Y.
3. Add Runtime as the single model effect.
4. Click Run.
5. From the red triangle menu next to Response Oxy, select Estimates > Inverse Prediction.
Figure 3.11 Inverse Prediction Given by the Fit Model Platform
7. Click OK.
The Inverse Prediction report (
Figure 3.12) gives the exact predictions for each
Oxy value specified with upper and lower 95% confidence limits. The exact prediction for
Runtime when
Oxy is 50 is 9.7935. This value is close to the approximate prediction of 9.779 found in
Figure 3.10.
Figure 3.12 Inverse Prediction Given by the Fit Model Platform
Note: The fiducial confidence limits are formed by Fieller’s method. Sometimes this method results in a degenerate (outside) interval, or an infinite interval, for one or both sides of an interval. When this happens for both sides, Wald intervals are used. If it happens for only one side, the Fieller method is still used and a missing value is returned. See the
Statistical Details appendix.
The inverse prediction option also predicts a single x value when there are multiple effects in the model. To predict a single x, you supply one or more y values of interest. You also set the x value that you predict is missing. By default, the other x values are set to the regressor’s means but can be changed to any desirable value.
Example Predicting a Single X Value with Multiple Model Effects
1. From the Fitness.jmp sample data table, select Analyze > Fit Model.
2. Select Oxy and click Y.
3. Add Runtime, RunPulse, and RstPulse as effects.
4. Click Run.
5. From the red triangle menu next to Response Oxy, select Estimates > Inverse Prediction.
Figure 3.13 Inverse Prediction Dialog for a Multiple Regression Model
7. Delete the value for Runtime, because you want to predict that value.
8. Click OK.
Figure 3.14 Inverse Prediction for a Multiple Regression Model
Cox Mixtures
Note: This option is available only for mixture models.
In mixture designs, the model parameters cannot easily be used to judge the effects of the mixture components. The Cox Mixture model (a reparameterized and constrained version of the Scheffé model) produces parameter estimates. You can then derive factor effects and the response surface shape relative to a reference point in the design space. See Cornell (1990) for a complete discussion.
Example of Cox Mixtures
1. Open the Five Factor Mixture.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y1 and click Y.
4. Select X1, X2, and X3.
5. Click on Macros > Mixture Response Surface.
6. Click Run.
7. From the red triangle menu next to Response Y1, select Estimates > Cox Mixtures.
Figure 3.15 Cox Mixtures Window
8. Type the reference mixture points shown in
Figure 3.15. Note that if the components of the reference point do not equal one, then the values are scaled so that they do equal one.
9. Click OK.
The report shows the parameter estimates along with standard errors, hypothesis tests, and the reference mixture.
Parameter Power
Suppose that you want to know how likely it is that your experiment will detect some difference at a given α-level. The probability of getting a significant test result is called the power. The power is a function of the unknown parameter values tested, the sample size, and the unknown residual error variance.
Alternatively, suppose that you already did an experiment and the effect was not significant. You think that the effect might have been significant if you had more data. How much more data do you need?
JMP offers the following calculations of statistical power and other details related to a given hypothesis test:
• LSV, the least significant value, is the value of some parameter or function of parameters that would produce a certain p-value alpha.
• LSN, the least significant number, is the number of observations that would produce a specified p-value alpha if the data has the same structure and estimates as the current sample.
• Power values are also available. See
The Power for details.
The LSV, LSN, and power values are important measuring sticks that should be available for all test statistics, especially when the test statistics are not significant. If a result is not significant, you should at least know how far from significant the result is in the space of the estimate (rather than in the probability). YOu should also know how much additional data is needed to confirm significance for a given value of the parameters.
Sometimes a novice confuses the role of the null hypotheses, thinking that failure to reject the null hypothesis is equivalent to proving it. For this reason, it is recommended that the test be presented in these other aspects (power and LSN) that show the test’s sensitivity. If an analysis shows no significant difference, it is useful to know the smallest difference that the test is likely to detect (LSV).
The power details provided by JMP are for both prospective and retrospective power analyses. In the planning stages of a study, a prospective analysis helps determine how large your sample size must be to obtain a desired power in tests of hypothesis. During data analysis, however, a retrospective analysis helps determine the power of hypothesis tests that have already been conducted.
Technical details for power, LSN, and LSV are covered in the section
Power Calculations in the Power Calculations appendix.
Calculating retrospective power at the actual sample size and estimated effect size is somewhat non-informative, even controversial [Hoenig and Heisey, 2001]. The calculation does not provide additional information about the significance test, but rather shows the test in a different perspective. However, we believe that many studies fail to detect a meaningful effect size due to insufficient sample size. There should be an option to help guide for the next study, for specified effect sizes and sample sizes.
For more information, see John M. Hoenig and Dinnis M. Heisey, (2001) “The Abuse of Power: The Pervasive Fallacy of Power Calculations for Data Analysis.”, American Statistician (v55 No 1, 19-24).
Power Options
Power options are available only for continuous-response models. Power and other test details are available in the following contexts:
• If you want the 0.05 level details for all parameter estimates, select Estimates > Parameter Power from the report’s red triangle menu. This option produces the LSV, LSN, and adjusted power for an alpha of 0.05 for each parameter in the linear model.
• If you want the details for an F-test for a certain effect, select Power Analysis from the effect’s red triangle menu.
• If you want the details for a Contrast, create the contrast and then select Power Analysis from the effect’s red triangle menu.
• If you want the details for a custom test, create the test that you want with the Custom Test option from the platform popup menu and then select Power Analysis option in the popup menu next to the Custom Test.
In all cases except the first, you enter information in the Power Details window for the calculations that you want.
Example of Power Analysis
The Power Details window (
Figure 3.17) shows the contexts and options for test detailing. After you fill in the values, the results are added to the end of the report.
1. Open the Big Class.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select weight and click Y.
4. Add age, sex, and height as the effects.
5. Click Run.
6. From the red triangle next to age, select Power Analysis.
Figure 3.17 Power Details Window
7. Replace the
Delta values with 3,6, and 1 as shown in
Figure 3.17.
8. Replace the
Number values with 20, 60, and 10 as shown in
Figure 3.17.
9. Select Solve for Power and Solve for Least Significant Number.
10. Click Done.
Power Details Window
In the Alpha, Sigma, Delta, and Number columns, you can enter a single value, two values, or the start, stop, and increment for a sequence of values. Power calculations are done on all possible combinations of the values that you specify.
Effect Size
The power is the probability that an F achieves its α-critical value given a noncentrality parameter related to the hypothesis. The noncentrality parameter is zero when the null hypothesis is true (that is, when the effect size is zero). The noncentrality parameter λ can be factored into the three components that you specify in the JMP Power window as
λ = (nδ2)/σ2 .
Power increases with λ. This means that the power increases with sample size n and raw effect size δ, and decreases with error variance σ2. Some books (Cohen 1977) use standardized rather than raw Effect Size, Δ = δ/σ, which factors the noncentrality into two components: λ = nΔ2.
Delta (δ) is initially set to the value implied by the estimate given by the square root of SSH/n, where SSH is the sum of squares for the hypothesis. If you use this estimate for delta, you might want to correct for bias by asking for the Adjusted Power.
In the special case for a balanced one-way layout with k levels:
Because JMP uses parameters of the following form:
![](http://imgdetail.ebookreading.net/design/8/9781612902166/9781612902166__jmp-10-modeling__9781612902166__images__MMM_03_SLS_11.png)
with
![](http://imgdetail.ebookreading.net/design/8/9781612902166/9781612902166__jmp-10-modeling__9781612902166__images__MMM_03_SLS_12.png)
the delta for a two-level balanced layout appears as follows:
Text Reports for Power Analysis
The Power Analysis option calculates power as a function of every combination of α, σ, δ, and n values that you specify in the Power Details window.
• For every combination of α, σ, and δ, the least significant number is calculated.
• For every combination of α, σ, and n, the least significant value is calculated.
Perform the example shown in
Figure 3.17 to produce the results in
Figure 3.18.
Figure 3.18 The Power Analysis Tables
If you check Adjusted Power and Confidence Interval in the Power Details window, the Power report includes the AdjPower, LowerCL, and UpperCL columns.
Plot of Power by Sample Size
To see a plot of power by sample size, select the Power Plot option from the red triangle menu at the bottom of the Power report. See
Figure 3.18. JMP plots the Power by N columns from the Power table. The plot in
Figure 3.19 is a result of plotting the Power table in
Figure 3.18.
Figure 3.19 Plot of Power by Sample Size
The Least Significant Value (LSV)
After a single-degree-of-freedom hypothesis test is performed, you often want to know how sensitive the test was. In other words, you want to know how small would a significant effect be at some p-value alpha.
The LSV provides a significant measuring stick on the scale of the parameter rather than on a probability scale. It shows the sensitivity of the design and data. LSV also encourages proper statistical intuition concerning the null hypothesis by highlighting how small a value would be detected as significant by the data.
• The LSV is the value that the parameter must be greater than or equal to in absolute value, in order to give the p-value of the significance test a value less than or equal to alpha.
• The LSV is the radius of the confidence interval for the parameter. A 1–alpha confidence interval is derived by taking the parameter estimate plus or minus the LSV.
• The absolute value of the parameter or function of the parameters tested is equal to the LSV, if and only if the p-value for its significance test is exactly alpha.
Compare the absolute value of the parameter estimate to the LSV. If the absolute parameter estimate is bigger, it is significantly different from zero. If the LSV is bigger, the parameter is not significantly different from zero.
The Least Significant Number (LSN)
The LSN or least significant number is the number of observations needed to decrease the variance of the estimates enough to achieve a significant result with the given values of alpha, sigma, and delta (the significance level, the standard deviation of the error, and the effect size, respectively).If you need more data points (a larger sample size) to achieve significance, the LSN indicates how many more data points are necessary.
Note: LSN is not a recommendation of how large a sample to take because the probability of significance (power) is only about 0.5 at the LSN.
The LSN has these characteristics:
• If the LSN is less than the actual sample size n, then the effect is significant. This means that you have more data than you need to detect the significance at the given alpha level.
• If the LSN is greater than n, the effect is not significant. In this case, if you believe that more data will show the same standard errors and structural results as the current sample, the LSN suggests how much data you would need to achieve significance.
• If the LSN is equal to n, then the p-value is equal to the significance level alpha. The test is on the border of significance.
• Power (described next) calculated when n = LSN is always greater than or equal to 0.5.
The Power
The power is the probability of getting a significant result. It is a function of the sample size n, the effect size δ, the standard deviation of the error σ, and the significance level α. The power tells you how likely your experiment will detect a difference at a given α-level.
Power has the following characteristics:
• If the true value of the parameter is the hypothesized value, the power should be alpha, the size of the test. You do not want to reject the hypothesis when it is true.
• If the true value of the parameters is not the hypothesized value, you want the power to be as great as possible.
• The power increases with the sample size, as the error variance decreases, and as the true parameter gets farther from the hypothesized value.
The Adjusted Power and Confidence Intervals
You typically substitute sample estimates in power calculations, because power is a function of unknown population quantities (Wright and O’Brien 1988). If you regard these sample estimates as random, you can adjust them to have a more proper expectation.
You can also construct a confidence interval for this adjusted power. However, the confidence interval is often very wide. The adjusted power and confidence intervals can be computed only for the original
δ, because that is where the random variation is. For details about adjusted power see
Computations for Adjusted Power in the Computations for Adjusted Power appendix
Correlation of Estimates
The Correlation of Estimates command on the Estimates menu creates a correlation matrix for all parameter estimates.
Example of Correlation of Estimates
1. Open the Tiretread.jmp data table.
2. Select Analyze > Fit Model.
3. Select ABRASION and click Y.
4. Select SILICA and SILANE and click Add.
5. Select SILICA and SILANE and click Cross.
6. Click Run.
7. From the Response red triangle menu, select Estimates > Correlation of Estimates.
Figure 3.20 Correlation of Estimates Report
Effect Screening
The Effect Screening options on the Response red triangle menu examine the sizes of the effects.
Scaled Estimates and the Coding Of Continuous Terms
The parameter estimates are highly dependent on the scale of the factor. When you convert a factor from grams to kilograms, the parameter estimates change by a multiple of a thousand. When you apply the same change to a squared (quadratic) term, the scale changes by a multiple of a million.
To learn more about the effect size, examine the estimates in a more scale-invariant fashion. This means converting from an arbitrary scale to a meaningful one. Then the sizes of the estimates relate to the size of the effect on the response. There are many approaches to doing this. In JMP, the Effect Screening > Scaled Estimates command on the report’s red triangle menu gives coefficients corresponding to factors that are scaled to have a mean of zero and a range of two. If the factor is symmetrically distributed in the data, then the scaled factor has a range from -1 to 1. This factor corresponds to the scaling used in the design of experiments (DOE) tradition. Therefore, for a simple regressor, the scaled estimate is half of the predicted response change as the regression factor travels its whole range.
Scaled estimates are important in assessing effect sizes for experimental data that contains uncoded values. If you use coded values (–1 to 1), then the scaled estimates are no different from the regular estimates. Scaled estimates also take care of the issues for polynomial (crossed continuous) models, even if they are not centered by the Center Polynomials command on the launch window’s red triangle menu.
You also do not need scaled estimates if your factors have the Coding column property. In that case, they are converted to uncoded form when the model is estimated and the results are already in an interpretable form for effect sizes.
Example of Scaled Estimates
1. Open the Drug.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select y and click Y.
4. Add Drug and x as the effects.
5. Click Run.
6. From the red triangle menu next to Response y, select Effect Screening > Scaled Estimates.
As noted in the report, the estimates are parameter-centered by the mean and scaled by range/2.
Figure 3.21 Scaled Estimates
Plot Options
The Normal, Bayes, and Pareto Plot options correct for scaling and for correlations among the estimates. These three Effect Screening options point out the following:
1. The process of fitting can be thought of as converting one set of realizations of random values (the response values) into another set of realizations of random values (the parameter estimates). If the design is balanced with an equal number of levels per factor, these estimates are independent and identically distributed, just as the responses are.
2. If you are fitting a screening design that has many effects and only a few runs, you expect that only a few effects are active. That is, a few effects have a sizable impact and the rest of them are inactive (they are estimating zeros). This is called the assumption of effect sparsity.
3. Given points 1 and 2 above, you can think of screening as a way to determine which effects are inactive with random values around zero and which ones are outliers (not part of the distribution of inactive effects).
Therefore, you treat the estimates themselves as a set of data to help you judge which effects are active and which are inactive. If there are few runs, with little or no degrees of freedom for error, then there are no classical significance tests, and this approach is especially needed.
The final Effect Screening options provide more reports:
• The Effect Screening report contains the Lenth PSE (pseudo-standard error) value.
• The Parameter Estimate Population report highlights significant factors.
You can also look at the model parameters from different angles using scaled estimates and three plots. See
Parameter Estimate Population Report for details.
Normal Plot
The
Normal Plot option displays a Parameter Estimate Population report (see
Parameter Estimate Population Report) and shows a normal plot of these parameter estimates (Daniel 1959). The estimates are on the vertical axis, and the normal quantiles on the horizontal axis.
Example of a Normal Plot
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select F, Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Effect Screening > Normal Plot.
If all effects are due to random noise, they tend to follow a straight line with slope σ, the standard error. In the plot, the line with slope equal to the Lenth’s PSE estimate is blue.
The Normal plot helps you identify effects that deviate from the normal lines. Estimates that deviate substantially are labeled.
Half-Normal Plot
Below the Normal Plot report title, select
Half Normal Plot from the list to plot the absolute values of the estimates against the normal quantiles for the absolute value normal distribution (shown in
Figure 3.23).
Figure 3.23 Half Normal Plot
Bayes Plot
Another approach to resolving which effects are important (sometimes referred to as active contrasts) is computing posterior probabilities using a Bayesian approach. This method, due to Box and Meyer (1986), assumes that the estimates are a mixture from two distributions. Some portion of the effects is assumed to come from pure random noise with a small variance. The remaining terms are assumed to come from a contaminating distribution that has a variance K times larger than the error variance.
An effect’s prior probability is the chance you give that effect of being nonzero (or being in the contaminating distribution). These priors are usually set to equal values for each effect. 0.2 is a commonly recommended prior probability value. The K contamination coefficient is often set at 10. This value indicates that the contaminating distribution has a variance that is 10 times the error variance.
The Bayes plot is done with respect to normalized estimates (JMP lists as Orthog t-Ratio), which have been transformed to be uncorrelated and have equal variance.
Example of a Bayes Plot
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select F, Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Effect Screening > Bayes Plot.
Figure 3.24 Bayes Plot Specifications
8. Click Go to start the calculation of the posterior probabilities.
Figure 3.25 Bayes Plot Report
The Std Err Est value is set to 0 for a saturated model with no estimate of error. If there is an estimate of standard error (the root mean square error), this value is set to 1 because the estimates have already been transformed and scaled to unit variance. If you specify a different value, it should be done as a scale factor of the RMSE estimate.
The resulting posterior probabilities are listed and plotted with bars as shown in
Figure 3.25. An overall posterior probability is also shown for the outcome that the sample is uncontaminated.
Bayes Plot for Factor Activity
JMP includes a script that helps you determine which factors are active in the design.
1. Open the Reactor.jmp sample data table.
1. Open the BayesPlotforFactors.jsl file in the Sample Scripts folder.
2. Select Edit > Run Script.
3. Select Y and click Y, Response.
4. Select F, Ct, A, T, and Cn and click X, Factor.
5. Click OK.
Figure 3.26 Bayes Plot for Factor Activity
In this case, we specified that the highest order interaction to consider is two. Therefore, all possible models that include (up to) second-order interactions are constructed. Based on the value assigned to Prior Probability (see the Controls section of the plot), a posterior probability is computed for each of the possible models. The probability for a factor is the sum of the probabilities for each of the models where it was involved.
In this example, we see that
Ct,
T, and
Cn are active and that
A and
F are not. These results match those of the Bayes Plot shown in
Figure 3.25.
Note: If the ridge parameter were zero (not allowed), all the models would be fit by least squares. As the ridge parameter gets large, the parameter estimates for any model shrink toward zero. Details on the ridge parameter (and why it cannot be zero) are explained in Box and Meyer (1993).
Pareto Plot
The
Pareto Plot selection gives plots of the absolute values of the orthogonalized estimates showing their composition relative to the sum of the absolute values. The estimates are orthogonalized to be uncorrelated and standardized to have equal variances by default. If your data set has estimates that are correlated and/or have unequal variances, then your data is transformed, by default, to have equal variances and to be uncorrelated. However, you have the option of undoing the transformations. (See
Parameter Estimate Population Report.) In this case, the Pareto Plot represents your selection of equal variances or unequal variances and uncorrelated or correlated estimates.
Example of a Pareto Plot
1. Open the Reactor.jmp sample data table.
For this data set, the estimates have equal variances and are not correlated.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select F, Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Effect Screening > Pareto Plot.
Figure 3.27 Pareto Plot for Reactor.jmp
Parameter Estimate Population Report
Most inferences about effect size first assume that the estimates are uncorrelated and have equal variances. This is true for fractional factorials and many classical experimental designs. However, these assumptions are not true for some designs.
The last three Effect Screening options in the report’s red triangle menu show a plot and the Parameter Estimate Population report. The Parameter Population report finds the correlation of the estimates and tells you whether the estimates are uncorrelated and have equal variances.
• If the estimates are correlated, a normalizing transformation can be applied make them uncorrelated and have equal variances.
• If an estimate of error variance is unavailable, the relative standard error for estimates is calculated by setting the error variance to one.
• If the estimates are uncorrelated and have equal variances, then the following notes appear under the Effect Screening report title:
– The parameter estimates have equal variances.
– The parameter estimates are not correlated.
If the estimates are correlated and/or have unequal variances, then each of these two notes can appear as list items. The selected list item shows that JMP has transformed the estimates.
To undo both transformations, click these list items.
The transformation to uncorrelate the estimates is the same as that used to calculate sequential sums of squares. The estimates measure the additional contribution of the variable after all previous variables have been entered into the model.
Example of Data with Equal Variances and Uncorrelated Estimates
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Effect Screening > Normal Plot.
8. Open the Parameter Estimate Population report.
Figure 3.28 Parameter Estimate Population Report for Reactor.jmp
Saturated Models
Screening experiments often involve fully saturated models, where there are not enough degrees of freedom to estimate error. Because of this, neither standard errors for the estimates, nor t-ratios, nor p-values can be calculated in the traditional way.
For these cases, JMP uses the relative standard error, corresponding to a residual standard error of 1. In cases where all the variables are identically coded (say, [–1,1] for low and high levels), these relative standard errors are identical.
JMP also displays a Pseudo-t-ratio, calculated as follows:
using Lenth’s PSE (pseudo standard-error) and degrees of freedom for error (DFE) equal to one-third the number of parameters. The value for Lenth’s PSE is shown at the bottom of the report.
Example of a Saturated Model
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Select the following five columns: F, Ct, A, T, and Cn.
5. Click on the Macros button and select Full Factorial.
6. Click Run.
The parameter estimates are presented in sorted order, with smallest
p-values listed first. The sorted parameter estimates are presented in
Figure 3.29.
Figure 3.29 Saturated Report
In cases where the relative standard errors are different (perhaps due to unequal scaling), a similar report appears. However, there is a different value for Lenth’s PSE for each estimate.
Description of Lenth’s Method
An estimate of standard error is calculated using the method of Lenth (1989) and shows in the Effect Screening report (shown above). This estimate, called the pseudo standard error, is formed by taking 1.5 times the median absolute value of the estimates after removing all the estimates greater than 3.75 times the median absolute estimate in the complete set of estimates.
Factor Profiling
Assuming that the prediction equation is estimated well, you still must explore the equation itself to answer a number of questions:
• What type of curvature does the response surface have?
• What are the predicted values at the corners of the factor space?
• Would a transformation on the response produce a better fit?
The tools described in this section explore the prediction equation to answer these questions assuming that the equation is correct enough to work with.
The
Profiler (or Prediction Profiler) shows prediction traces for each X variable (
Figure 3.30). The vertical dotted line for each X variable shows its
current value or
current setting. Use the Profiler to change one variable at a time and examine the effect on the predicted response.
Figure 3.30 Illustration of Prediction Traces
If the variable is nominal, the x-axis identifies categories.
For each X variable, the value above the factor name is its current value. You change the current value by clicking in the graph or by dragging the dotted line where you want the new current value to be.
• The horizontal dotted line shows the current predicted value of each Y variable for the current values of the X variables.
• The black lines within the plots show how the predicted value changes when you change the current value of an individual X variable. The 95% confidence interval for the predicted values is shown by a dotted curve surrounding the prediction trace (for continuous variables) or an error bar (for categorical variables).
Use the interactive Contour Profiler for optimizing response surfaces graphically. The Contour Profiler shows contours for the fitted model for two factors at a time. The report also includes a surface plot.
Example Using the Contour Profiler
1. Open the Tiretread.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select ABRASION, MODULUS, ELONG, and HARDNESS and click Y.
4. Select SILICA, SILANE, and SULFUR and click Macros > Response Surface.
5. Click Run.
6. From the red triangle menu next to Least Squares Fit, select Profilers > Contour Profiler.
Figure 3.31 Contour Profiler
The Mixture Profiler shows response contours of mixture experiment models on a ternary plot. Use the Mixture Profiler when three or more factors in your experiment are components in a mixture. The Mixture Profiler enables you to visualize and optimize the response surfaces of your experiment.
Example Using the Mixture Profiler
1. Open the Plasticizer.jmp sample data table.
2. Select Analyze > Fit Model.
The response (Y) and predictor variables (p1, p2, and p3) are automatically populated. The predictor variables use the Macros > Mixture Response Surface option.
3. Click Run.
4. From the red triangle menu next to Response Y, select Factor Profiling > Mixture Profiler.
Figure 3.32 Mixture Profiler
You modify plot axes for the factors by selecting different radio buttons at the top left of the plot. The Lo and Hi Limit columns at the upper right of the plot let you enter constraints for both the factors and the response.
The Surface Profiler shows a three-dimensional surface plot of the response surface.
Example Using the Surface Profiler
1. Open the Odor.jmp sample data table.
2. From the red triangle next to Model, select Run Script.
3. Click Run.
4. From the red triangle menu next to Response odor, select Factor Profiling > Surface Profiler.
Interaction Plots
The Interaction Plots option shows a matrix of interaction plots when there are interaction effects in the model.
Example of Interaction Plots
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Factor Profiling > Interaction Plots.
Figure 3.34 Interaction Plots
In an Interaction Plot, evidence of interaction shows as nonparallel lines. For example, in the T*Cn plot in the bottom row of plots, the effect of Cn is very small at the low values of temperature. However, the Cn effect diverges widely for the high values of temperature.
Cube Plots
The Cube Plots option displays a set of predicted values for the extremes of the factor ranges. These values appear on the vertices of cubes (
Figure 3.35). If a factor is nominal, then the vertices are the first and last level.
Example of Cube Plots
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Factor Profiling > Cube Plots.
To change the layout so that the factors are mapped to different cube coordinates, click one of the factor names in the first cube and drag it to the desired axis. For example, in
Figure 3.35, if you click
T and drag it over
Ct, then
T and
Ct (and their corresponding coordinates) exchange places. When there is more than one response, the multiple responses are shown stacked at each vertex.
Box Cox Y Transformations
Sometimes a transformation on the response fits the model better than the original response. A commonly used transformation raises the response to some power. Box and Cox (1964) formalized and described this family of power transformations. The formula for the transformation is constructed so that it provides a continuous definition and the error sums of squares are comparable.
![](http://imgdetail.ebookreading.net/design/8/9781612902166/9781612902166__jmp-10-modeling__9781612902166__images__MMM_03_SLS_18.png)
where
![](http://imgdetail.ebookreading.net/design/8/9781612902166/9781612902166__jmp-10-modeling__9781612902166__images__MMM_03_SLS_19.png)
is the geometric mean
The plot shown here illustrates the effect of this family of power transformations on Y.
Figure 3.36 Power Transformations on Y
The Box-Cox Y Transformation option fits transformations from λ = –2 to 2 in increments of 0.2. This transformation also plots the sum of squares error (SSE) across the λ power.
Example of Box-Cox Y Transformation
1. Open the Reactor.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select Y and click Y.
4. Make sure that the Degree box has a 2 in it.
5. Select F, Ct, A, T, and Cn and click Macros > Factorial to Degree.
6. Click Run.
7. From the red triangle menu next to Response Y, select Factor Profiling > Box Cox Y Transformation.
Figure 3.37 Box Cox Y Transformation
Figure 3.37 shows the best fit when
λ is between 1.0 and 1.5. The best transformation is found on the plot by finding the lowest point on the curve.
Leverage Plots
To graphically view the significance of the model or focus attention on whether an effect is significant, you want to display the data by focusing on the hypothesis for that effect. You might say that you want more of an X-ray picture showing the inside of the data rather than a surface view from the outside. The leverage plot gives this view of your data; it offers maximum insight into how the fit carries the data.
The effect in a model is tested for significance by comparing the sum of squared residuals to the sum of squared residuals of the model with that effect removed. Residual errors that are much smaller when the effect is included in the model confirm that the effect is a significant contribution to the fit.
The graphical display of an effect’s significance test is called a leverage plot. See Sall (1990). This type of plot shows for each point what the residual would be both with and without that effect in the model. Leverage plots are found in the Row Diagnostics submenu of the Fit Model report.
A leverage plot is constructed as illustrated in
Figure 3.38. The distance from a point to the line of fit shows the actual residual. The distance from the point to the horizontal line of the mean shows what the residual error would be without the effect in the model. In other words, the mean line in this leverage plot represents the model where the hypothesized value of the parameter (effect) is constrained to zero.
Historically, leverage plots are referred to as a partial-regression residual leverage plot by Belsley, Kuh, and Welsch (1980) or an added variable plot by Cook and Weisberg (1982).
The term leverage is used because a point exerts more influence on the fit if it is farther away from the middle of the plot in the horizontal direction. At the extremes, the differences of the residuals before and after being constrained by the hypothesis are greater and contribute a larger part of the sums of squares for that effect’s hypothesis test.
The fitting platform produces a leverage plot for each effect in the model. In addition, there is one special leverage plot titled Whole Model that shows the actual values of the response plotted against the predicted values. This Whole Model leverage plot dramatizes the test that all the parameters (except intercepts) in the model are zero. The same test is reported in the Analysis of Variance report.
Figure 3.38 Illustration of a General Leverage Plot
The leverage plot for the linear effect in a simple regression is the same as the traditional plot of actual response values and the regressor.
Example of a Leverage Plot for a Linear Effect
1. Open the Big Class.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select weight and click Y.
4. Select height and click Add.
5. Click Run.
Figure 3.39 Whole Model and Effect Leverage Plots
The plot on the left is the Whole Model test for all effects, and the plot on the right is the leverage plot for the effect height.
The points on a leverage plot for simple regression are actual data coordinates, and the horizontal line for the constrained model is the sample mean of the response. But when the leverage plot is for one of multiple effects, the points are no longer actual data values. The horizontal line then represents a partially constrained model instead of a model fully constrained to one mean value. However, the intuitive interpretation of the plot is the same whether for simple or multiple regression. The idea is to judge if the line of fit on the effect’s leverage plot carries the points significantly better than does the horizontal line.
Figure 3.38 is a general diagram of the plots in
Figure 3.39. Recall that the distance from a point to the line of fit is the actual residual and that the distance from the point to the mean is the residual error if the regressor is removed from the model.
Confidence Curves
The leverage plots are shown with confidence curves. These indicate whether the test is significant at the 5% level by showing a confidence region for the line of fit. If the confidence region between the curves contains the horizontal line, then the effect is not significant. If the curves cross the line, the effect is significant. Compare the examples shown in
Figure 3.40.
Figure 3.40 Comparison of Significance Shown in Leverage Plots
Interpretation of X Scales
If the modeling type of the regressor is continuous, then the
x-axis is scaled like the regressor and the slope of the line of fit in the leverage plot is the parameter estimate for the regressor. (See the left illustration in
Figure 3.41.)
If the effect is nominal or ordinal, or if a complex effect like an interaction is present instead of a simple regressor, then the
x-axis cannot represent the values of the effect directly. In this case the
x-axis is scaled like the
y-axis, and the line of fit is a diagonal with a slope of 1. The whole model leverage plot is a version of this. The
x-axis turns out to be the predicted response of the whole model, as illustrated by the right-hand plot in
Figure 3.41.
Figure 3.41 Leverage Plots for Simple Regression and Complex Effects
The influential points in all leverage plots are the ones far out on the x-axis. If two effects in a model are closely related, then these effects as a whole do not have much leverage. This problem is called collinearity. By scaling regressor axes by their original values, collinearity is shown by shrinkage of the points in the x direction.
See the appendix
Statistical Details for the details of leverage plot construction.
Save Columns
Each Save Columns option generates one or more new columns in the current data table, where <colname> is the name of the response variable.
Note: If you are using the Graph option to invoke the Profiler, then you should first save the columns Prediction Formula and StdErr Pred Formula to the data table. Then, place both of these formulas into the Y,Prediction Formula role in the Profiler launch window. The resulting window asks whether you want to use the PredSE colname to make confidence intervals for the Pred Formula colname, instead of making a separate profiler plot for the PredSE colname.
Effect Options
Use the red triangle options next to an effect to request reports, plots, and tests for the effect. You can close results or dismiss the results by deselecting the item in the menu.
LSMeans Table
Least squares means are predicted values from the specified model across the levels of a categorical effect where the other model factors are controlled by being set to neutral values. The neutral values are the sample means (possibly weighted) for regressors with interval values, and the average coefficient over the levels for unrelated nominal effects.
Least squares means are the values that let you see which levels produce higher or lower responses, holding the other variables in the model constant. Least squares means are also called adjusted means or population marginal means. Least squares means can differ from simple means when there are other effects in the model.
Least squares means are the statistics that are compared when effects are tested. They might not reflect typical real-world values of the response if the values of the factors do not reflect prevalent combinations of values in the real world. Least squares means are useful as comparisons in experimental situations.
Example of a Least Squares Means Table
1. Open the Big Class.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select height and click Y.
4. Select age and click Add.
5. Click Run.
Figure 3.42 Least Squares Mean Table
A Least Squares Means table with standard errors is produced for all categorical effects in the model. For main effects, the Least Squares Means table also includes the sample mean. It is common for the least squares means to be closer together than the sample means. For further details about least squares means, see the appendix
Statistical Details.
LSMeans Plot
The LSMeans Plot option plots least squares means (LSMeans) plots for nominal and ordinal main effects and interactions.
Example of an LS Means Plot
1. Open the Popcorn.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select yield and click Y.
4. Select popcorn, oil amt, and batch and click Add.
5. Select popcorn in the Construct Model Effects section and batch in the Select Columns section.
6. Click Cross to obtain the popcorn*batch interaction.
7. Click Run.
8. Hold down the CTRL key. From the red triangle menu for one of the effects, select LSMeans Plot.
Tip: Holding down the CTRL key broadcasts the command to all of the effects.
Figure 3.43 LSMeans Plots for Main Effects and Interactions
Figure 3.43 shows the effect plots for main effects and the two-way interaction.
To transpose the factors of the LSMeans plot for a two-factor interaction, hold down the SHIFT key and select the
LSMeans Plot option. (In this example, you must first deselect the
LSMeans Plot option and then select it again, holding down the SHIFT key.)
Figure 3.44 shows both the default and the transposed factors plots.
Figure 3.44 LSMeans Plot Comparison with Transposed Factors
LSMeans Contrast
A contrast is a set of linear combinations of parameters that you want to jointly test to be zero. JMP builds contrasts in terms of the least squares means of the effect. By convention, each column of the contrast is normalized to have sum zero and an absolute sum equal to two. If a contrast involves a covariate, you can specify the value of the covariate at which to test the contrast.
Example of the LSMeans Contrast Option
To illustrate using the LSMeans Contrast option:
1. Open the Big Class.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select height and click Y.
4. Select age and click Add.
5. Click Run.
6. From the red triangle menu next to age, select LSMeans Contrast.
Figure 3.45 LSMeans Contrast Specification Window
This Contrast Specification window shows the name of the effect and the names of the levels in the effect. Beside the levels is an area enclosed in a rectangle that has a column of numbers next to boxes of + and - signs.
To construct a contrast, click the + and - boxes beside the levels that you want to compare. If possible, the window normalizes each time to make the sum for a column zero and the absolute sum equal to two each time you click; it adds to the plus or minus score proportionately.
For example, to form a contrast that compares the first two age levels with the second two levels:
1. Click + for the ages 12 and 13
2. Click - for ages 14 and 15
If you wanted to do more comparisons, you would click the New Column button for a new column to define the new contrast.
Figure 3.46 LSMeans Contrast Specification for Age
3. Click Done.
The contrast is estimated, and the Contrast report appears.
Figure 3.47 LSMeans Contrast Report
The Contrast report shows the following:
• Contrast as a function of the least squares means
• Estimates and standard errors of the contrast for the least squares means
• t-tests for each column of the contrast
• F-test for all columns of the contrast tested jointly
• Parameter Function report that shows the contrast expressed in terms of the parameters. In this example, the parameters are for the ordinal variable, age.
LSMeans Student’s t and LSMeans Tukey’s HSD
The LSMeans Student’s t and LSMeans Tukey’s HSD options give multiple comparison tests for model effects. For details about these tests, see the Basic Analysis and Graphing book.
The red triangle options that appear in each report window show or hide optional reports.
Test Slices
The Test Slices option, which is enabled for interaction effects, is a quick way to do many contrasts at the same time. For each level of each classification column in the interaction, it makes comparisons among all the levels of the other classification columns in the interaction. For example, if an interaction is A*B*C, then there is a slice called A=1, which tests all the B*C levels when A=1. There is another slice called A=2, and so on, for all the levels of B, and C. This is a way to detect the importance of levels inside an interaction.
Restricted Maximum Likelihood (REML) Method
Random effects are those where the effect levels are chosen randomly from a larger population of levels. These random effects represent a sample from the larger population. In contrast, the levels of fixed effects are of direct interest rather than representative. If you have both random and fixed (nonrandom) effects in a model, it is called a mixed model.
Note: It is very important to declare random effects. Otherwise, the test statistics produced from the fitted model are calculated with the wrong assumptions.
Typical random effects can be any of the following:
• Subjects in a repeated measures experiment, where the subject is measured at several times.
• Plots in a split plot experiment, where an experimental unit is subdivided and multiple measurements are taken, usually with another treatment applied to the subunits.
• Measurement studies, where multiple measurements are taken in order to study measurement variation.
• Random coefficients models, where the random effect is built with a continuous term crossed with categories.
The Fit Model platform in JMP fits mixed models using these modern methods, now generally regarded as best practice:
• REML estimation method (Restricted Maximum Likelihood)
• Kenward-Roger tests
For historical interest only, the platform also offers the Method of Moments (EMS), but this is no longer a recommended method, except in special cases where it is equivalent to REML.
If you have a model where all of the effects are random, you can also fit it in the Variability Chart platform.
The REML method for fitting mixed models is now the mainstream, state-of-the-art method, supplanting older methods.
In the days before availability of powerful computers, researchers needed to restrict their interest to situations in which there were computational short cuts to obtain estimates of variance components and tests on fixed effects in a mixed model. Most books today introduce mixed models using these short cuts that work on balanced data. See McCulloch, Searle, and Neuhaus (2008), Poduri (1997), and Searle, Casella, and McCulloch(1992). The Method of Moments provided a way to calculate what the expected value of Mean Squares (EMS) were in terms of the variance components, and then back-solve to obtain the variance components. It was also possible using these techniques to obtain expressions for test statistics that had the right expected value under the null hypotheses that were synthesized from mean squares.
If your model satisfies certain conditions (that is, it has random effects that contain the terms of the fixed effects that they provide random structure for) then you can use the EMS choice to produce these traditional analyses. However, since the newer REML method produces identical results to these models, but is considerably more general, the EMS method is never recommended.
The REML approach was pioneered by Patterson and Thompson in 1974. See also Wolfinger, Tobias, and Sall (1994) and Searle, Casella, and McCulloch(1992). The reason to prefer REML is that it works without depending on balanced data, or shortcut approximations, and it gets all the tests right, even contrasts that work across interactions. Most packages that use the traditional EMS method are either not able to test some of these contrasts, or compute incorrect variances for them.
Introduction to Random Effects
Levels in random effects are randomly selected from a larger population of levels. For the purpose of testing hypotheses, the distribution of the effect on the response over the levels is assumed to be normal, with mean zero and some variance (called a variance component).
In one sense, every model has at least one random effect, which is the effect that makes up the residual error. The units making up individual observations are assumed to be randomly selected from a much larger population, and the effect sizes are assumed to have a mean of zero and some variance, σ2.
The most common model that has random effects other than residual error is the repeated measures or split plot model.
Types of Effects in a Split Plot Model, lists the types of effects in a split plot model. In these models the experiment has two layers. Some effects are applied on the whole plots or subjects of the experiment. Then these plots are divided or the subjects are measured at different times and other effects are applied within those subunits. The effects describing the whole plots or subjects are one random effect, and the subplots or repeated measures are another random effect. Usually the subunit effect is omitted from the model and absorbed as residual error.
Each of these cases can be treated as a layered model, and there are several traditional ways to fit them in a fair way. The situation is treated as two different experiments:
1. The whole plot experiment has whole plot or subjects as the experimental unit to form its error term.
2. Subplot treatment has individual measurements for the experimental units to form its error term (left as residual error).
The older, traditional way to test whole plots is to do any one of the following:
• Take means across the measurements and fit these means to the whole plot effects.
• Form an F-ratio by dividing the whole plot mean squares by the whole plot ID mean squares.
• Organize the data so that the split or repeated measures form different columns and do a MANOVA model, and use the univariate statistics.
These approaches work if the structure is simple and the data are complete and balanced. However, there is a more general model that works for any structure of random effects. This more generalized model is called the mixed model, because it has both fixed and random effects.
Another common situation that involves multiple random effects is in measurement systems where there are multiple measurements with different parts, different operators, different gauges, and different repetitions. In this situation, all the effects are regarded as random.
Generalizability
Random effects are randomly selected from a larger population, where the distribution of their effect on the response is assumed to be a realization of a normal distribution with a mean of zero and a variance that can be estimated.
Often, the exact effect sizes are not of direct interest. It is the fact that they represent the larger population that is of interest. What you learn about the mean and variance of the effect tells you something about the general population from which the effect levels were drawn. That is different from fixed effects, where you only know about the levels you actually encounter in the data.
Unrestricted Parameterization for Variance Components in JMP
Note: Read this section only if you are concerned about matching the results of certain textbooks.
There are two different statistical traditions for parameterizing the variance components: the unrestricted and the restricted approaches. JMP and SAS use the unrestricted approach. In this approach, while the estimated effects always sum to zero, the true effects are not assumed to sum to zero over a particular random selection made of the random levels. This is the same assumption as for residual error. The estimates make the residual errors have mean zero, and the true mean is zero. But a random draw of data using the true parameters is some random event that might not have a mean of exactly zero.
You need to know about this assumption because many statistics textbooks use the restricted approach. Both approaches have been widely taught for 50 years. A good source that explains both sides is Cobb (1998, section 13.3).
Negative Variances
Note: Read this section only when you are concerned about negative variance components.
Though variances are always positive, it is possible to have a situation where the unbiased estimate of the variance is negative. This happens in experiments when an effect is very weak, and by chance the resulting data causes the estimate to be negative. This usually happens when there are few levels of a random effect that correspond to a variance component.
JMP can produce negative estimates for both REML and EMS. For REML, there are two check boxes in the model launch window: Unbounded Variance Components and Estimate Only Variance Components. Unchecking the box beside Unbounded Variance Components constrains the estimate to be nonnegative. We recommend that you do not uncheck this if you are interested in fixed effects. Constraining the variance estimates leads to bias in the tests for the fixed effects. If, however, you are interested only in variance components, and you do not want to see negative variance components, then checking the box beside Estimate Only Variance Components is appropriate.
If you remain uncomfortable about negative estimates of variances, please consider that the random effects model is statistically equivalent to the model where the variance components are really covariances across errors within a whole plot. It is not hard to think of situations in which the covariance estimate can be negative, either by random happenstance, or by a real process in which deviations in some observations in one direction would lead to deviations in the other direction in other observations. When random effects are modeled this way, the covariance structure is called compound symmetry.
So, consider negative variance estimates as useful information. If the negative value is small, it can be considered happenstance in the case of a small true variance. If the negative value is larger (the variance ratio can get as big as 0.5), it is a troubleshooting sign that the rows are not as independent as you had assumed, and some process worth investigating is happening within blocks.
Scripting Note
The JSL option for Unbounded Variance Components is No Bounds( Boolean ). Setting this option to true (1) is equivalent to checking the Unbounded Variance Components option.
Random Effects BLUP Parameters
Random effects have a dual character. In one perspective, they appear like residual error, often the error associated with a whole-plot experimental unit. In another respect, they are like fixed effects with a parameter to associate with each effect category level. As parameters, you have extra information about them—they are derived from a normal distribution with mean zero and the variance estimated by the variance component. The effect of this extra information is that the estimates of the parameters are shrunk toward zero. The parameter estimates associated with random effects are called BLUPs (Best Linear Unbiased Predictors). Some researchers consider these BLUPs as parameters of interest, and others consider them by-products of the method that are not interesting in themselves. In JMP, these estimates are available, but in an initially closed report.
BLUP parameter estimates are used to estimate random-effect least squares means, which are therefore also shrunken toward the grand mean, at least compared to what they would be if the effect were treated as a fixed effect. The degree of shrinkage depends on the variance of the effect and the number of observations per level in the effect. With large variance estimates, there is little shrinkage. If the variance component is small, then more shrinkage takes place. If the variance component is zero, the effect levels are shrunk to exactly zero. It is even possible to obtain highly negative variance components where the shrinkage is reversed. You can consider fixed effects as a special case of random effects where the variance component is very large.
If the number of observations per level is large, the estimate shrinks less. If there are very few observations per level, the estimates shrink more. If there are infinite observations, there is no shrinkage and the answers are the same as fixed effects.
The REML method balances the information about each individual level with the information about the variances across levels.
For example, suppose that you have batting averages for different baseball players. The variance component for the batting performance across player describes how much variation is usual between players in their batting averages. If the player only plays a few times and if the batting average is unusually small or large, then you tend not to trust that estimate, because it is based on only a few at-bats; the estimate has a high standard error. But if you mixed it with the grand mean, that is, shrunk the estimate toward the grand mean, you would trust the estimate more. For players that have a long batting record, you would shrink much less than those with a short record.
You can run this example and see the results for yourself. The example batting average data are in the Baseball.jmp sample data file. To compare the Method of Moments (EMS) and REML, run the model twice. Assign Batting as Y and Player as an effect. Select Player in the Construct Model Effects box, and select Random Effect from the Attributes pop-up menu.
Run the model and select REML (Recommended) from the Method popup menu. The results show the best linear unbiased prediction (BLUP) for each level of random effects.
Run the model again with EMS (Traditional) as Method. The least square mean for fixed affects and each level of random effects and is the same.
Comparison of Estimates between Method of Moments and REML, summarizes the estimates between Method of Moments and REML across a set of baseball players in this simulated example. Note that Suarez, with only three at-bats, is shrunk more than the others with more at-bats.
REML and Traditional Methods Agree on the Standard Cases
It turns out that in balanced designs, the REML F-test values for fixed effects is the same as with the Method of Moments (Expected Means Squares) approach. The degrees of freedom could differ in some cases. There are a number of methods of obtaining the degrees of freedom for REML F-tests; the one that JMP uses is the smallest degrees of freedom associated with a containing effect.
F-Tests in Mixed Models
Note: This section details the tests produced with REML.
The REML method obtains the variance components and parameter estimates, but there are a few additional steps to obtain tests on fixed effects in the model. The objective is to construct the F statistic and associated degrees of freedom to obtain a p-value for the significance test.
Historically, in simple models using the Method of Moments (EMS), standard tests were derived by construction of quadratic forms that had the right expectation under the null hypothesis. Where a mean square had to be synthesized from a linear combination of mean squares to have the right expectation, Satterthwaite's method could be used to obtain the degrees of freedom to get the p-value. Sometimes these were fractional degrees of freedom, just as you might find in a modern (Aspin-Welch) Student's t-test.
With modern computing power and recent methods, we have much improved techniques to obtain the tests. First, Kackar, and Harville (1984) found a way to estimate a bias-correction term for small samples. This was refined by Kenward and Roger (1997) to correct further and obtain the degrees of freedom that gave the closest match of an F-distribution to the distribution of the test statistic. These are not easy calculations. Consequently, they can take some time to perform for larger models.
If you have a simple balanced model, the results from REML-Kenward-Roger agree with the results from the traditional approach, provided that the estimates are not bounded at zero.
• These results do not depend on analyzing the syntactic structure of the model. There are no rules about finding containing effects. The method does not care if your whole plot fixed effects are purely nested in whole plot random effects. It gets the right answer regardless.
• These results do not depend on having categorical factors. It handles continuous (random coefficient) models just as easily.
• These methods produce different (and better) results than older versions of JMP (that is, earlier than JMP 6) that implemented older, less precise, technology to do these tests.
• These methods do not depend on having positive variance components. Negative variance components are not only supported, but need to be allowed in order for the tests to be unbiased.
Our goal in implementing these methods was not just to handle general cases, but to handle cases without the user needing to know very much about the details. Just declare which effects are random, and everything else is automatic. It is particularly important that engineers learn to declare random effects, because they have a history of performing inadvertent split-plot experiments where the structure is not identified.
Specifying Random Effects
Models with Random Effects use the same Fit Model dialog as other models. To specify a random effect, highlight it in the Construct Model Effects list and select Attributes > Random Effect. This appends &Random to the effect name in the model effect list.
Split Plot
The most common type of layered design is a balanced split plot, often in the form of repeated measures across time. One experimental unit for some of the effects is subdivided (sometimes by time period) and other effects are applied to these subunits.
Example of a Split Plot
Consider the data in the Animals.jmp sample data table (the data are fictional). The study collected information about differences in the seasonal hunting habits of foxes and coyotes. Each season for one year, three foxes and three coyotes were marked and observed periodically. The average number of miles that they wandered from their dens during different seasons of the year was recorded (rounded to the nearest mile). The model is defined by the following aspects:
• The continuous response variable called miles
• The species effect with values fox or coyote
• The season effect with values fall, winter, spring, and summer
• An animal identification code called subject, with nominal values 1, 2, and 3 for both foxes and coyotes
There are two layers to the model:
1. The top layer is the between-subject layer, in which the effect of being a fox or coyote (species effect) is tested with respect to the variation from subject to subject. The bottom layer is the within-subject layer, in which the repeated-measures factor for the four seasons (season effect) is tested with respect to the variation from season to season within a subject. The within-subject variability is reflected in the residual error.
2. The season effect can use the residual error for the denominator of its F-statistics. However, the between-subject variability is not measured by residual error and must be captured with the subject within species (subject[species]) effect in the model. The F-statistic for the between-subject effect species uses this nested effect instead of residual error for its F-ratio denominator.
To create the split plot for this data, follow these steps:
1. Open the Animals.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select miles and click Y.
4. Select species and subject and click Add.
5. In the Select Columns list, select species.
6. In the Construct Model Effects list, select subject.
7. Click Nest.
This adds the subject within species (subject[species]) effect to the model.
8. Select the nested effect subject[species].
9. Select Attributes > Random Effect.
This nested effect is now identified as an error term for the species effect and shows as subject[species]&Random.
10. In the Select Columns list, select season and click Add.
Figure 3.48 Fit Model Dialog
When you assign any effect as random from the Attributes popup menu, the Method options (REML and EMS) appear at the top right of the dialog, with REML selected as the default.
A new option,
Convergence Settings, is available from the red triangle menu next to Model Specification. For details, see
Model Specification Options.
11. Click Run.
Figure 3.49 Partial Report of REML Analysis
REML Results
A nice feature of REML is that the report does not need qualification (
Figure 3.49). The estimates are all properly shrunk and the standard errors are properly scaled (SAS Institute Inc. 1996). The variance components are shown as a ratio to the error variance, and as a portion of the total variance.
There is no special table of synthetic test creation, because all the adjustments are automatically taken care of by the model itself. There is no table of expected means squares, because the method does not need this.
If you have random effects in the model, the analysis of variance report is not shown. This is because the variance does not partition in the usual way, nor do the degrees of freedom attribute in the usual way, for REML-based estimates. You can obtain the residual variance estimate from the REML report rather than from the analysis of variance report.
To obtain the standard deviations (square-root of the variance components), right-click on the REML Variance Component Estimates report, and select Columns > Sqrt Variance Component.
The Variance Component Estimates table shows 95% confidence intervals for the variance components using the Satterthwaite (1946) approximation. If the Unbounded Variance Components options is used, the confidence intervals are +/-Wald style. You can right-click on the Variance Components Estimates table to toggle on the Norm KHC (Kackar-Harville correction). This value is an approximation of the magnitude of the increase in the mean squared errors of the estimators for the mixed model. See Kackar and Harville (1984) for a discussion of approximating standard errors in mixed models.
REML Save Columns Options
When you use the REML method, additional options appear in the Save Columns menu. All of these options allow the random effects predicted values to participate in the formulas, rather than using their expected value (zero).
REML Profiler Option
When you use REML method and select
Factor Profiling > Profiler, a new option,
Conditional Predictions, appears on the red triangle menu next to Prediction Profiler. For details, see
Profiler Options.
Method of Moments Results
Note: This section is only of use in matching historical results.
We no longer recommend the Method of Moments, but we understand the need to support it for teaching use, in order to match the results of many textbooks still in use.
You have the option of choosing the EMS (Traditional) approach from the Method popup menu on the Fit Model dialog. This is also called the Method of Moments method.
Results from the steps for the Method of Moments are as follows:
• For each effect, the coefficients of the expected mean squares for that effect are calculated. This is a linear combination of the variance components and fixed effect values that describes the expected value of the mean square for that effect. All effects also have a unit coefficient on the residual variance.
• The coefficients of expected mean squares for all the random effects, including the residual error, are gathered into a matrix, and this is used to solve for variance components for each random effect.
• For each effect to be tested, a denominator for that effect is synthesized using the terms of the linear combination of mean squares in the numerator that do not contain the effect to be tested or other fixed effects. Thus, the expectation is equal for those terms common to the numerator and denominator. The remaining terms in the numerator then constitute the effect test.
• Degrees of freedom for the synthesized denominator are constructed using Satterthwaite’s method.
• The effect tests use the synthetic denominator.
JMP handles random effects like the SAS GLM procedure with a
Random statement and the
Test option.
Figure 3.50, shows example results.
Caution: Standard errors for least squares means and denominators for contrast F-tests also use the synthesized denominators. Contrasts using synthetic denominators might not be appropriate, especially in crossed effects compared at common levels. The leverage plots and custom tests are done with respect to the residual, so they might not be appropriate.
Caution: Crossed and nested relationships must be declared explicitly. For example, if knowing a subject ID also identifies the group that contains the subject (that is, if each subject is in only one group), then subject must be declared as nested within group. In that situation, the nesting must be explicitly declared to define the design structure.
Note: JMP cannot fit a layered design if the effect for a layer’s error term cannot be specified under current effect syntax. An example of this is a design with a Latin Square on whole plots for which the error term would be Row*Column–Treatment. Fitting such special cases with JMP requires constructing your own F-tests using sequential sums of squares from several model runs.
For the Animals example above, the EMS reports are as follows.
Figure 3.50 Report of Method of Moments Analysis for Animals Data
The random submatrix from the EMS table is inverted and multiplied into the mean squares to obtain variance component estimates. These estimates are usually (but not necessarily) positive. The variance component estimate for the residual is the Mean Square Error.
Note that the CV of the variance components is initially hidden in the Variance Components Estimates report. To reveal it, right-click (or hold down CTRL and click on the Macintosh) and select Columns > CV from the menu that appears.
Singularity Details
When there are linear dependencies between model effects, the Singularity Details report appears.
Figure 3.51 Singularity Details Report
Examples with Statistical Details
The examples in this section are based on the following example:
1. Open the Drug.jmp sample data table.
2. Select Analyze > Fit Model.
3. Select y and click Y.
4. Select Drug and click Add.
5. Click Run.
One-Way Analysis of Variance with Contrasts
In a one-way analysis of variance, a different mean is fit to each of the different sample (response) groups, as identified by a nominal variable. To specify the model for JMP, select a continuous Y and a nominal X variable, such as Drug. In this example, Drug has values a, d, and f. The standard least squares fitting method translates this specification into a linear model as follows: The nominal variables define a sequence of dummy variables, which have only values 1, 0, and –1. The linear model is written as follows:
where:
– yi is the observed response in the ith trial
– x1i is the level of the first predictor variable in the ith trial
– x2i is the level of the second predictor variable in the ith trial
– β0, β1, and β2 are parameters for the intercept, the first predictor variable, and the second predictor variable, respectively
– ει are the independent and normally distributed error terms in the ith trial
As shown here, the first dummy variable denotes that Drug=a contributes a value 1 and Drug=f contributes a value –1 to the dummy variable:
The second dummy variable is given the following values:
The last level does not need a dummy variable because in this model, its level is found by subtracting all the other parameters. Therefore, the coefficients sum to zero across all the levels.
The estimates of the means for the three levels in terms of this parameterization are as follows:
Solving for βi yields the following:
![](http://imgdetail.ebookreading.net/design/8/9781612902166/9781612902166__jmp-10-modeling__9781612902166__images__MMM_03_SLS_37.png)
(the average over levels)
Therefore, if regressor variables are coded as indicators for each level minus the indicator for the last level, then the parameter for a level is interpreted as the difference between that level’s response and the average response across all levels. See the appendix
Statistical Details for additional information about the interpretation of the parameters for nominal factors.
Figure 3.52 shows the Parameter Estimates and the Effect Tests reports from the one-way analysis of the drug data.
Figure 3.1, at the beginning of the chapter, shows the Least Squares Means report and LS Means Plot for the
Drug effect.
Figure 3.52 Parameter Estimates and Effect Tests for Drug.jmp
The Drug effect can be studied in more detail by using a contrast of the least squares means, as follows:
1. From the red triangle menu next to Drug, select LSMeans Contrast.
2. Click the + boxes for drugs
a and
d, and the - box for drug
f to define the contrast that compares the average of drugs
a and
d to
f (shown in
Figure 3.53).
3. Click Done.
Figure 3.53 Contrast Example for the Drug Experiment
The Contrast report shows that the Drug effect looks more significant using this one-degree-of-freedom comparison test. The LSMean for drug f is clearly significantly different from the average of the LSMeans of the other two drugs.
Analysis of Covariance
An analysis of variance model with an added regressor term is called an analysis of covariance. Suppose that the data are the same as above, but with one additional term, x3i, in the formula as a new regressor. Both x1i and x2i continue to be dummy variables that index over the three levels of the nominal effect. The model is written as follows:
There is an intercept plus two effects: one is a nominal main effect using two parameters, and the other is an interval covariate regressor using one parameter.
Re-run the Snedecor and Cochran
Drug.jmp example, but add the
x to the model effects as a covariate. Compared with the main effects model (
Drug effect only), the
R2 increases from 22.8% to 67.6%, and the standard error of the residual reduces from 6.07 to 4.0. As shown in
Figure 3.54, the
F-test significance probability for the whole model decreases from 0.03 to less than 0.0001.
Figure 3.54 ANCOVA Drug Results
Sometimes you can investigate the functional contribution of a covariate. For example, some transformation of the covariate might fit better. If you happen to have data where there are exact duplicate observations for the regressor effects, it is possible to partition the total error into two components. One component estimates error from the data where all the x values are the same. The other estimates error that can contain effects for unspecified functional forms of covariates, or interactions of nominal effects. This is the basis for a lack of fit test. If the lack of fit error is significant, then the fit model platform warns that there is some effect in your data not explained by your model. Note that there is no significant lack of fit error in this example, as seen by the large probability value of 0.7507.
The covariate, x, has a substitution effect with respect to Drug. It accounts for much of the variation in the response previously accounted for by the Drug variable. Thus, even though the model is fit with much less error, the Drug effect is no longer significant. The effect previously observed in the main effects model now appears explainable to some extent in terms of the values of the covariate.
The least squares means are now different from the ordinary mean because they are adjusted for the effect of x, the covariate, on the response, y. Now the least squares means are the predicted values that you expect for each of the three values of Drug, given that the covariate, x, is held at some constant value. The constant value is chosen for convenience to be the mean of the covariate, which is 10.7333.
So, the prediction equation gives the least squares means as follows:
fit equation:
-2.696 - 1.185*Drug[a - f] - 1.0761*Drug[d - f] + 0.98718*x
for a:
-2.696 - 1.185*(1) -1.0761*(0) + 0.98718*(10.7333) = 6.71
for d:
-2.696 - 1.185*(0) -1.0761*(1) + 0.98718*(10.7333) = 6.82
for f:
-2.696 - 1.185*(-1) -1.0761*(-1) + 0.98718*(10.7333) = 10.16
Figure 3.55 shows a leverage plot for each effect. Because the covariate is significant, the leverage values for
Drug are dispersed somewhat from their least squares means.
Figure 3.55 Comparison of Leverage Plots for Drug Test Data
Analysis of Covariance with Separate Slopes
This example is a continuation of the Drug.jmp example presented in the previous section. The example uses data from Snedecor and Cochran (1967, p. 422). A one-way analysis of variance for a variable called Drug, shows a difference in the mean response among the levels a, d, and f, with a significance probability of 0.03.
The lack of fit test for the model with main effect Drug and covariate x is not significant. However, for the sake of illustration, this example includes the main effects and the Drug*x effect. This model tests whether the regression on the covariate has separate slopes for different Drug levels.
This specification adds two columns to the linear model (call them x4i and x5i) that allow the slopes for the covariate to be different for each Drug level. The new variables are formed by multiplying the dummy variables for Drug by the covariate values, giving the following formula:
Coding of Analysis of Covariance with Separate Slopes, shows the coding of this Analysis of Covariance with Separate Slopes. (The mean of
X is 10.7333.)
A portion of the report is shown in
Figure 3.56. The Regression Plot shows fitted lines with different slopes. The Effect Tests report gives a p-value for the interaction of 0.56. This is not significant statistically, indicating the model does not need to have different slopes.
Figure 3.56 Plot with Interaction