22
GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈ ≈≈ page 1 MULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab multiple regression. It includes descriptions of the Minitab commands, and the Minitab output is heavily annotated. Minitab output appears in Courier type face: like this. Commentary is in Times New Roman: like this. The comments will also cover some interpretations. Letters in square brackets, such as [a], identify endnotes which will give details of the calculations and explanations. The endnotes begin on page 13. Output from Minitab sometimes will be edited to reduce empty space or to improve page layout. This document was prepared with Minitab 14. The data set used in this document can be found on the Stern network as file X:\SOR\B011305\M\SWISS.MTP. The data set concerns fertility rates in 47 Swiss cantons (provinces) in the year 1888. The dependent variable will be Fert, the fertility rate, and all the other variables will function as independent variables. The data are found in Data Analysis and Regression , by Mosteller and Tukey, pages 550-551. This document was prepared by the Statistics Group of the I.O.M.S. Department. If you find this document to be helpful, we’d like to know! If you have comments that might improve this presentation, please let us know also. Please send e-mail to [email protected]. Revision date 12 AUG 2004 gs2004

MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

Embed Size (px)

Citation preview

Page 1: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 1

MULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab multiple regression. It includes descriptions of the Minitab commands, and the Minitab output is heavily annotated. Minitab output appears in Courier type face: like this. Commentary is in Times New Roman: like this. The comments will also cover some interpretations. Letters in square brackets, such as [a], identify endnotes which will give details of the calculations and explanations. The endnotes begin on page 13. Output from Minitab sometimes will be edited to reduce empty space or to improve page layout. This document was prepared with Minitab 14. The data set used in this document can be found on the Stern network as file X:\SOR\B011305\M\SWISS.MTP. The data set concerns fertility rates in 47 Swiss cantons (provinces) in the year 1888. The dependent variable will be Fert, the fertility rate, and all the other variables will function as independent variables. The data are found in Data Analysis and Regression, by Mosteller and Tukey, pages 550-551. This document was prepared by the Statistics Group of the I.O.M.S. Department. If you find this document to be helpful, we’d like to know! If you have comments that might improve this presentation, please let us know also. Please send e-mail to [email protected]. Revision date 12 AUG 2004 gs2004

Page 2: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 2

Data was brought into the program through File ⇒ Open Worksheet ⇒. Minitab’s default for Files of type: is (*.mtw; *.mpj), so you will want to change this to *.mtp to obtain the file. On the Stern network, this file is in the folder X:\SOR\B011305\M, and the file name is SWISS.MTP. The listing below shows the data set, as copied directly from Minitab’s data window.

Fert Ag Army Ed Catholic Mort 0.802[a] 0.170 0.15 0.12 9.96 0.222 0.831 0.451 0.06 0.09 84.84 0.222 0.925 0.397 0.05 0.05 93.40 0.202 0.858 0.365 0.12 0.07 33.77 0.203 0.769 0.435 0.17 0.15 5.16 0.206 0.761 0.353 0.09 0.07 90.57 0.266 0.838 0.702 0.16 0.07 92.85 0.236 0.924 0.678 0.14 0.08 97.16 0.249 0.824 0.533 0.12 0.07 97.67 0.210 0.829 0.452 0.16 0.13 91.38 0.244 0.871 0.645 0.14 0.06 98.61 0.245 0.641 0.620 0.21 0.12 8.52 0.165 0.669 0.675 0.14 0.07 2.27 0.191 0.689 0.607 0.19 0.12 4.43 0.227 0.617 0.693 0.22 0.05 2.82 0.187 0.683 0.726 0.18 0.02 24.20 0.212 0.717 0.340 0.17 0.08 3.30 0.200 0.557 0.194 0.26 0.28 12.11 0.202 0.543 0.152 0.31 0.20 2.15 0.108 0.651 0.730 0.19 0.09 2.84 0.200 0.655 0.598 0.22 0.10 5.23 0.180 0.650 0.551 0.14 0.03 4.52 0.224 0.566 0.509 0.22 0.12 15.14 0.167 0.574 0.541 0.20 0.06 4.20 0.153 0.725 0.712 0.12 0.01 2.40 0.210 0.742 0.581 0.14 0.08 5.23 0.238 0.720 0.635 0.06 0.03 2.56 0.180 0.605 0.608 0.16 0.10 7.72 0.163 0.583 0.268 0.25 0.19 18.46 0.209 0.654 0.495 0.15 0.08 6.10 0.225 0.755 0.859 0.03 0.02 99.71 0.151 0.693 0.849 0.07 0.06 99.68 0.198 0.773 0.897 0.05 0.02 100.00 0.183 0.705 0.782 0.12 0.06 98.96 0.194 0.794 0.649 0.07 0.03 98.22 0.202 0.650 0.759 0.09 0.09 99.06 0.178 0.922 0.846 0.03 0.03 99.46 0.163 0.793 0.631 0.13 0.13 96.83 0.181 0.704 0.384 0.26 0.12 5.62 0.203 0.657 0.077 0.29 0.11 13.79 0.205 0.727 0.167 0.22 0.13 11.22 0.189 0.644 0.176 0.35 0.32 16.92 0.230 0.776 0.376 0.15 0.07 4.97 0.200 0.676 0.187 0.25 0.07 8.65 0.195 0.350 0.012 0.37 0.53 42.34 0.180 0.447 0.466 0.16 0.29 50.43 0.182 0.428 0.277 0.22 0.29 58.33 0.193

Page 3: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 3

The item below is Minitab’s Project Manager window. You can get this to appear by clicking on the ⓘⓘⓘⓘ icon on the toolbar.

[b]

The following section gives basic statistical facts. It is obtained by Stat ⇒ Basic Statistics ⇒ Display Descriptive Statistics ⇒. All variables were requested. The request can be done by listing each variable by name (Fert Ag Army Ed Catholic Mort) or by listing the column numbers (C1-C6) or by clicking on the names in the variable listing. Descriptive Statistics: Fert, Ag, Army, Ed, Catholic, Mort [c][d] [e] [f] Variable N N* Mean SE Mean StDev Minimum Q1 Median Q3 Fert 47 0 0.7014 0.0182 0.1249 0.3500 0.6440 0.7040 0.7930 Ag 47 0 0.5066 0.0331 0.2271 0.0120 0.3530 0.5410 0.6780 Army 47 0 0.1649 0.0116 0.0798 0.0300 0.1200 0.1600 0.2200 Ed 47 0 0.1098 0.0140 0.0962 0.0100 0.0600 0.0800 0.1200 Catholic 47 0 41.14 6.08 41.70 2.15 5.16 15.14 93.40 Mort 47 0 0.19943 0.00425 0.02913 0.10800 0.18100 0.20000 0.22200 Variable Maximum Fert 0.9250 Ag 0.8970 Army 0.3700 Ed 0.5300 Catholic 100.00 Mort 0.26600

The next listing shows the correlations. It is obtained through

Stat ⇒ Basic Statistics ⇒ Correlation ⇒ and then listing all the variable names. This listing has de-selected the feature Display p-values.

Page 4: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 4

Correlations: Fert, Ag, Army, Ed, Catholic, Mort Fert Ag Army Ed Catholic Ag 0.353 Army -0.646 -0.687[g] Ed -0.664 -0.640 0.698 Catholic 0.464 0.401 -0.573 -0.154 Mort 0.417 -0.061 -0.114 -0.099 0.175

Cell Contents: Pearson correlation

The linear regression of dependent variable Fert on the independent variables can be started through

Stat ⇒ Regression ⇒ Regression ⇒ Set up the panel to look like this:

Observe that Fert was selected as the dependent variable (response) and all the others were used as independent variables (predictors). If you click OK you will see the basic regression results. For the sake of illustration, we’ll show some additional features.

Click the Options…button and then select Variance inflation factors. The choice Fit intercept is the default and should already be selected; if it is not, please select it. The Fit intercept option should be de-selected only in extremely special situations.

We recommend that you routinely examine the variance inflation factors if strong collinearity is suspected. The Durbin-Watson statistic was not used here because the data are not time-sequenced.

Page 5: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 5

Click the Graphs… button and select the indicated choices:

Examining the Residuals versus fits plot is now part of routine statistical practice. The other selections can show some interesting clues as well. Here we will use the Four in one option, as it shows the residual versus fitted plot, along with the other three as well. The Residuals versus order plot will not be useful, because the data are not time-ordered.

Some of the choices made here reflect features of this data set or particular desires of the analyst. Only the Residuals versus fits should be regarded as mandatory. (This is included in the Four in one choice.) Here the Regular form of the residuals was desired; other choices would be just as reasonable.

Click the Storage…button and select Hi (leverages).

This provides a very thorough regression job.

The model corresponding to this request is

Ferti = β0 + βAg Agi + βArmy Armyi + βEd Edi

+ βCatholic Catholici + βMort Morti + εi

Page 6: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 6

Regression Analysis: Fert versus Ag, Army, Ed, Catholic, Mort The regression equation is [h] Fert = 0.669 - 0.172 Ag - 0.258 Army - 0.871 Ed + 0.00104 Catholic + 1.08 Mort Predictor Coef SE Coef T P VIF Constant[i] 0.6692[j] 0.1071[k] 6.25[l] 0.000[m] [n] Ag -0.17211[ø] 0.07030 -2.45 0.019 2.3[p]

Army -0.2580 0.2539 -1.02[q] 0.315[r] 3.7

Ed -0.8709 0.1830 -4.76 0.000 2.8

Catholic 0.0010412 0.0003526 2.95 0.005 1.9

Mort 1.0770 0.3817 2.82 0.007 1.1

S = 0.0716537[s] R-Sq = 70.7%[t] R-Sq(adj) = 67.1% [u] Analysis of Variance [v] Source DF[w] SS[2a] MS[2e] F[2i] P[2j]

Regression 5[x] 0.50729[2b] 0.10146[2f] 19.76 0.000

Residual Error 41[y] 0.21050[2c] 0.00513[2g]

Total 46[z] 0.71780[2d] [2h]

Source DF Seq SS[2k] Ag 1 0.08948 Army 1 0.22104 Ed 1 0.08918 Catholic 1 0.06671 Mort 1 0.04088 Unusual Observations[2l] Obs Ag Fert Fit SE Fit Residual St Resid 6[2m] 0.353 0.7610 0.9050[2n] 0.0319[2ø]-0.1440[2p] -2.24R [2q] 37 0.846 0.9220 0.7688 0.0270 0.1532 2.31R 45 0.012 0.3500 0.3480 0.0484 0.0020 0.04 X[2r] 47 0.277 0.4280 0.5807 0.0244 -0.1527 -2.27R R denotes an observation with a large standardized residual X denotes an observation whose X value gives it large influence. Many graphs were requested in this run. The Four in one panel examines the behavior of the residuals because they provide clues as to the appropriateness of the assumptions made on the εi terms in the model. The most important of these is the residuals versus fitted plot, the plot at the upper right on the next page. The normal probability plot and the histogram of the residuals are used to assess whether or not the noise terms are approximately normally distributed. Since the data points are not time-ordered, we will not use the plot of the residuals versus the order of the data.

Page 7: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 7

Residual

Per

cen

t

0.20.10.0-0.1-0.2

99

90

50

10

1

Fitted Value

Res

idu

al

0.900.750.600.450.30

0.1

0.0

-0.1

Residual

Freq

uen

cy

0.120.060.00-0.06-0.12

10.0

7.5

5.0

2.5

0.0

Observation OrderR

esid

ual

454035302520151051

0.1

0.0

-0.1

Normal Probability Plot of the Residuals Residuals Versus the Fitted Values

Histogram of the Residuals Residuals Versus the Order of the Data

Residual Plots for Fert

[2s]

Many users choose also to examine the plots of the residuals against each of the predictor variables. These were requested for this run, but this document will show only the plot of the residuals against the variable Mort.

Mort

Res

idua

l

0.280.260.240.220.200.180.160.140.120.10

0.15

0.10

0.05

0.00

-0.05

-0.10

-0.15

Residuals Versus Mort(response is Fert)

[2t]

Page 8: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 8

Finally, recall that we had requested the high leverage points through Stat ⇒ Regression ⇒ Regression ⇒ Storage ⇒ and then selecting Hi (leverages). These will show up in a new column, called HI1, in the data window. This column can be used in plots, or it can simply be examined. What shows below is that column, copied out of the data window, and restacked to save space.

Case HI1 Case HI1 Case HI1 1 0.156817 [2u] 17 0.068535 33 0.108342 2 0.122585 18 0.101750 34 0.098006 3 0.173683 19 0.351208 35 0.076759 4 0.079616 20 0.111375 36 0.091772 5 0.072190 21 0.074258 37 0.142462 6 0.198332 22 0.082771 38 0.081257 7 0.143082 23 0.064105 39 0.076831 8 0.141458 24 0.109214 40 0.226297 9 0.079940 25 0.100362 41 0.099816 10 0.106823 26 0.125696 42 0.205322 11 0.136769 27 0.180591 43 0.073667 12 0.083193 28 0.079051 44 0.172191 13 0.083926 29 0.053282 45 0.455836[2v] 14 0.109909 30 0.077062 46 0.210670 15 0.125512 31 0.173359 47 0.115954 16 0.106312 32 0.092047

There is a commonly-used threshold of concern, as discussed in [2u]. Minitab will automatically mark points that exceed this threshold; see [2l] and [2r]. Most users to not request that the leverage, or Hi, values be computed. Many problems on multiple regression end up with model building. This is an effort to select only those independent variables that contributed to the regression explanation in a reasonable way. There is no easy unique solution to the model building exercise, and this document cannot provide perfect guidance. This document will illustrate Best Subsets Regression and Stepwise Regression.

Page 9: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 9

Minitab provides Best Subsets Regression through Stat ⇒ Regression ⇒ Best subsets. Fill in the resulting panel as follows:

The panel Predictors in all models is not used here. If for some reason you want all considered models to contain Ag (C2), then C2 would appear in this panel and C3 - C6 would appear in the Free predictors panel. Minitab will not permit any variable to be listed in both panels.

The output from this step follows:

Best Subsets Regression: Fert versus Ag, Army, Ed, Catholic, Mort Response is Fert C a t h A o M r l o Mallows A m E i r Vars R-Sq R-Sq(adj) C-p S g y d c t 1 44.1 42.8 35.2 0.094460 X 1 41.7 40.4 38.5 0.096420 X 2 57.5 55.5 18.5 0.083314 X X 2 56.5 54.5 19.8 0.084261 X X 3 66.3 63.9 8.2 0.075054 X X X 3 64.2 61.7 11.0 0.077278 X X X 4 69.9 67.1 5.0 0.071682 X X X X 4 66.4 63.2 10.0 0.075794 X X X X 5 70.7 67.1 6.0 0.071654 X X X X X

The Vars column of the output is 1, 1, 2, 2, 3, 3, 4, 4, 5. Many people find it easier to deal with this display when just one model of each level is chosen. Accordingly, try Options, and then next to Models of each size to print, type 1. This will give simpler output that follows.

Page 10: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 10

Best Subsets Regression: Fert versus Ag, Army, Ed, Catholic, Mort Response is Fert [2w] C [2x] a t h A o M [3b] r l o [2y] Mallows [3c] A m E i r Vars R-Sq R-Sq(adj) C-p S g y d c t 1 44.1[2z] 42.8[3a] 35.2 0.094460 X 2 57.5 55.5 18.5 0.083314 X X 3 66.3 63.9 8.2 0.075054 X X X 4 69.9 67.1 5.0 0.071682 X X X X 5 70.7 67.1 6.0[3d] 0.071654 X X X X X

Most users select the simplest model (meaning smallest value under Vars) for which Cp ≈ p, as in note [3b]. Others will select the simplest model at which R2 hits a plateau or at which S hits a plateau. Another tool used for model building is stepwise regression, accessed the Stat ⇒ Regression ⇒ Stepwise. Set up the initial panel as follows:

The panel Predictors to include in every model is not used here. If for some reason you want all considered models to contain Ag (C2), then C2 would appear in this panel while still listing C2 - C6 in the Predictors panel. Minitab requires that variables listed in the Predictors to include in every model panel also be listed in the panel above. This is different from the procedure for Best subsets.

Page 11: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 11

Clicking Methods gets to this panel:

This shows the default settings. The top line has radio buttons for the selection method, either Use alpha values or Use F values. The results are usually the same, but the alpha values method is recommended here. Many users change Alpha to enter and Alpha to remove to 0.05; this will not make a difference for most problems, but this change tends to make Stepwise agree more closely with Best subsets. The stepwise method, using the forward and backward default will start with a model using only the variables in the panel Predictors to use include in every model. If this panel is empty, as it usually is, the procedure begins with a model with no predictors. It then adds on that variable among the available predictors to maximize the R2 value, provided the p-value (see note [r]) is less than or equal to the value for Alpha to enter. The stepwise procedure continues to add on the next-best variable (in terms of R2) provided its p-value is less than or equal to Alpha to enter. When no variables can be added on, the procedure stops. Every time a variable is added to the model, the p-values for all the formerly-selected variables are examined. If any of these p-values exceed Alpha to remove, the corresponding variables are removed.

You are invited to try both Forward selection and Backward elimination. In terms of the final selected model, these two methods often do not agree with Stepwise (forward and backward) and generally do not agree with each other.

Page 12: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 12

The output for this problem follows. Note that Alpha to enter and Alpha to remove were both set to 0.05. (Leaving these values at the default 0.15 would have produced exactly the same results for this problem.)

Stepwise Regression: Fert versus Ag, Army, Ed, Catholic, Mort Alpha-to-Enter: 0.05 Alpha-to-Remove: 0.05 [3e] Response is Fert on 5 predictors, with N = 47 [3f] [3g] Step 1 2 3 4 Constant [3h] 0.7961 0.7423 0.4868 0.6210 Ed [3i] -0.86[3j] -0.79 -0.76 -0.98 T-Value -5.95 -6.10 -6.50 -6.62 P-Value 0.000[3k] 0.000 0.000 0.000

Catholic [3l] 0.00111 0.00096 0.00125 T-Value 3.72 3.53 4.31 P-Value 0.001[3k] 0.001 0.000 Mort 1.30 1.08 T-Value 3.35 2.82 P-Value 0.002[3k] 0.007 Ag -0.155 T-Value -2.27 P-Value 0.029[3k] S 0.0945 0.0833[3m] 0.0751 0.0717 R-Sq 44.06 57.45[3n] 66.25 69.93 R-Sq(adj) 42.82 55.52[3ø] 63.90 67.07 Mallows C-p 35.2 18.5[3p] 8.2 5.0

Page 13: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 13

ENDNOTES: [a] This is the first line of the data listing. The line numbers (1 through 47) are not shown here, although they do appear in the Minitab data window. The numbers across this row indicate that this first canton had Fert1 = 0.802, Ag1 = 0.17, Army1 = 0.15, and so on. [b] Minitab’s Project Manager window shows the variable names for the columns, and also some basic accounting. We see that each variable has 47 values, with none missing. Minitab data sets can also have Constants and Matrices, although this set has none. Descriptions are saved only with project (*.MPJ) files. [c] The symbol N refers to the sample size, after removing missing data. In this data set, there are 47 cantons and all information is complete. We have 47 pieces of information for each variable. In some data sets, the column of N-values might list several different numbers. [d] N* is the number of missing values. In this set of data, all variables are complete.

[e] This is the standard error of the mean. It’s computed for each variable as SD

N ,

where N is the number of non-missing values. Here you can confirm that for variable

Fert, 01249

470 0182

..≈ .

[f] Minitab computes the quartiles Q1 and Q3 by an interpolation method. If the sample size is n, then Q1 is the observation at rank position (n+1)/4. If (n+1)/4 is not an integer, then Q1 is obtained as a weighted average of the values at the surrounding integer positions. For instance, if (n+1)/4 = 6.75, then Q1 is 3

4 of the distance between the 6th and 7th values. The procedure for finding Q3 works from rank position 3(n+1)/4. [g] The value -0.687 is the correlation between Army and Ag. It is also the correlation between Ag and Army. Since correlations are symmetric, it is not necessary to print the entire correlation matrix. The correlation between a variable and itself is 1.000 (also not printed). Generally we like to see strong correlations (say above +0.9 or below -0.9) involving the dependent variable, here Fert. We prefer not to have strong correlations among the other variables. [h] This is the estimated model or fitted equation. Some people like to place the “hat” on the dependent variable Fert as Fêrt to denote estimation or fitting. Note that the numbers are repeated in the Coef column below. The letter b is used for estimated values; thus b0 = 0.669, bAg = -0.172, and so on.

Page 14: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 14

[i] The term Constant refers to the inclusion of β0 in the regression model. This is sometimes called the intercept. [j] This is the estimated value of β0 and is often called b0. [k] This is estimated standard deviation of the value in the Coef column; this is also called the standard error. In this instance, we believe that the estimated value, 0.6692, is good to within a standard error of 0.1071. We’re about 95% confident that the true value of β0 is in the interval 0.6692 ± 2(0.1071), which is the interval (0.4550 , 0.8834).

[l] The value of T, also called Student’s t, is CoefSE Coef

; here that arithmetic is

0.66920.1071

≈ 6.25. This is the number of estimated standard deviations that the estimate,

0.6692, is away from zero. The phrase “estimated standard deviations” refers to the distribution of the sample coefficient, and not to the standard deviation in the regression model. This T can be regarded as a test of H0: β0 = 0 versus H1: β0 ≠ 0. Here T is outside the interval (-2,+2) and we should certainly believe that β0 (the true-but-unknown population value) is different from zero. Some users believe that the intercept should not be subjected to a statistical test; indeed some software does not provide a T value or a P (see the next item) for the Constant line. [m] The column P (for p-value) is the result of subjecting the data to a statistical test as to whether β0 = 0 or β0 ≠ 0. There is a precise technical definition, but crudely P is the smallest Type I error probability that you might make in deciding that β0 ≠ 0. Very small values of P suggest that β0 ≠ 0, while larger values indicate that you should maintain β0 = 0. The typical cutoff between these actions is 0.05; thus P ≤ 0.05 causes you to decide that the population parameter, here β0, is really different from zero. The p-value is never exactly zero, but it sometimes prints as 0.000. The p-value is directly related to T; values of T far outside the interval (-2,+2) lead to small P. When P ≤ 0.05, we say that the estimated value is statistically significant.

As an important side note, many people believe that the Constant β0 should never be subjected to statistical tests. According to this point of view, we should not even ask whether β0 = 0 or β0 ≠ 0; indeed we should not even list T and P in the Constant line. This side note applies only to the Constant.

[n] The VIF, variance inflation factor, comes as the result of a special request. The VIF does not apply to the Constant. See also item [p].

Page 15: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 15

[ø] The value -0.1721 is bAG, the estimated value for βAG. In fact, the Coef column can be used to write the fitted equation [h]. Some people like to write the standard errors in parentheses under the estimated coefficients in the fitted equation in this fashion:

Fêrt = 0.669 - 0.172 Ag - 0.258 Army - 0.871 Ed (0.107) (0.070) (0.254) (0.183) + 0.00104 Catholic + 1.08 Mort (0.00035) (0.38)

You may also see this kind of display using in parentheses the values of T, so indicate for your readers exactly what you are doing. This is precisely the relationship that is used to determine the n = 47 fitted values:

Fêrti = 0.669 - 0.172 Agi - 0.258 Armyi - 0.871 Edi + 0.00104 Catholici + 1.08 Morti

The differences between the observed and fitted values are the residuals. The ith residual, usually denoted ei, is Ferti - Fêrti. [p] The VIF, variance inflation factor, measures how much of the standard error, SE Coef, can be accounted for by inter-relation of one independent variable with all of the other independent variables. The VIF can never be less than 1. If some VIF values are large (say 10 or more), then you have a collinearity problem. Plausible solutions to the collinearity are Stepwise Regression and Best Subsets Regression. A related concept

is the tolerance; these are related through Tolerance = 1VIF

.

[q] The value of T for the variable Army tests the hypothesis H0: βArmy = 0 versus H1: βArmy ≠ 0. This T is in the interval (-2,+2), suggesting that H0 is correct, so you might consider repeating the problem without using Army in the model. [r] The P for variable Army exceeds 0.05. This is consistent with the previous comment, and it suggests repeating the problem without using Army in the model. [s] This is one of the most important numbers in the regression output. It’s called standard error of estimate or standard error of regression. It is the estimate of σ, the standard deviation of the noise terms (the εi’s). A common notation is sε . In this data set, that value comes out to 0.07165. It is useful to compare this to 0.1249, which was the standard deviation of Fert on the Descriptive Statistics list. The original “noise” level in Fert was 0.1249 (without doing any regression); the “noise” left over after the regression was 0.07165. Item [s] is the square root of item [2g], the residual mean square, whose value is 0.00513; observe that 0 00513. ≈ 0.07162.

Page 16: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 16

[t] This is the heavily-cited R2 ; it’s generally given as a percent, here 70.7%, but it might also be given as decimal 0.707. The formal statement used is “The percent of the variation in Fert that is explained by the regression is 70.7%.” Technically, R2 is the ratio of two sums of squares; it is the ratio of item [2b], the regression sum of squares, to

item [2d], the total sum of squares. Observe that 0 507290 71780..

≈ 0.7067 = 70.67%. Large

values of R2 are considered to be good. [u] This is an adjustment made to R2 to account for the sample size and the number of

independent variables and is given by Rn

n kRadj

2 211

11= −

−− −

−( ) . In this formula, n is

the number of points (here 47) and k is the number of predictors used (here 5). Thus R2

has the neat interpretation in [t], and we have 2

2

Fert

1adjsR

= −

. Here sε is 0.07165 and

sFert = 0.1249 and given in [s]. Most users prefer R2 over 2adjR .

[v] This is the analysis of variance table. The work is based on the algebraic identity

2 2 2

1 1 1

ˆ ˆ( ) ( ) ( )n n n

i i i ii i i

y y y y y y= = =

− = − + −∑ ∑ ∑

in which yi denotes the value of the dependent variable for point i, and $yi denotes the fitted value for point i. Since Fert is the dependent variable, we identify $yi with with Fêrti as in [h] and [ø]. The three sums of squares in this equation are, respectively, SStotal , SSregression , and SSresidual error . These have other names or abbreviations. For instance

SStotal is often written as SStot .

SSregression is often written as SSreg and sometimes as SSfit or SSmodel .

SSresidual error is often written as SSresidual or SSresid or SSres or SSerror or SSerr .

[w] The DF stands for degrees of freedom. This is an accounting of the dimensions of the problem, and the numbers in this column add up to the indicated total as 5 + 41 = 46. See the next three notes.

[x] The Regression line in the analysis of variance table refers to the sum 2

1

ˆ( )n

ii

y y=

−∑

which appeared in [v]. The degrees of freedom for this calculation is k, the number of independent variables. Here k = 5.

Page 17: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 17

[y] The Residual Error line in the analysis of variance table refers to the sum

( $ )y yi ii

n

−=∑ 2

1 which appeared in [v]. The degrees of freedom for this calculation is

n - 1 - k, where n is the number of data points and k is the number of independent variables. Here n = 47 and k = 5, so that 47 - 1 - 5 = 41 appears in this position.

[z] The Total line in the analysis of variance table refers to the sum 2

1( )

n

ii

y y=

−∑ which

appears in [v]. The degrees of freedom for this calculation is n - 1, where n is the number of data points. Here n = 47, so that 46 appears in this position. [2a] The SS stands for sum of squares. This column gives the numbers corresponding to the identity described in [v]. The values in this column add up to the indicated total as 0.50729 + 0.21050 = 0.71780.

[2b] This is the sum 2

1

ˆ( )n

ii

y y=

−∑ which appeared in [v]. This is SSregression.

[2c] This is the sum ( $ )y yi ii

n

−=∑ 2

1 which appeared in [v]. This is SSresidual error.

[2d] This is the sum 2

1( )

n

ii

y y=

−∑ which appears in [v]. This is SStotal, the total sum of

squares. It involves only the dependent variable, so it says nothing about the regression. The regression is successful if the regression sum of squares is large relative to the residual error sum of squares. The F statistic, item [ii], is the appropriate measure of success. [2e] The MS stands for Mean Squares. Each value is obtained by dividing the corresponding Sum Squares by its Degrees of Freedom. [2f] This is MSregression = 0.50729 ÷ 5 ≈ 0.10146. In a successful regression, this is large relative to item [2g], the residual mean square. [2g] This is MSresidual error = 0.21050 ÷ 41 ≈ 0.00513. [2h] There is no mean square entry in this position. The computation SStotal ÷ (n - 1) would nonetheless be useful, since it is the sample variance of the dependent variable. [2i] This is the F statistic. It is computed as MSregression ÷ MSresidual error , meaning 0.10146 ÷ 0.00513 ≈ 19.76. To test at significance level α, this is to be compared to , 1k n kF α

− − , the upper α point from the F distribution with k and n - 1 degrees of freedom, obtained from statistical tables or from Minitab. The F statistic is a formal test of the null hypothesis that all the independent variable coefficients are zero against the alternative that they are not. In our example, we would write

Page 18: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 18

H0: βAg = 0, βArmy = 0, βEd = 0, βCatholic = 0, βMort = 0 H1: at least one of βAg, βArmy, βEd, βCatholic, βMort is not zero

If the F statistic is larger than , 1k n kF α

− − , then the null hypothesis is rejected. Otherwise, we accept H0 (or reserve judgment). In this instance, using significance level α = 0.05, we find , 1k n kF α

− − = F5 410 05,. = 2.4434 from a statistical table. Since 19.76 > 2.4434, we

would reject H0 at the 0.05 level of significance; we would describe this regression as statistically significant.

Here’s how to use Minitab to get the values of , 1k n kF α− − .

Calc ⇒ Probability Distributions ⇒ F ⇒ Select Inverse cumulative probability, choose Numerator degrees of freedom: 5 Denominator degrees of freedom: 41 Input constant: 0.95

Then click OK . [2j] This is the p-value associated with the F statistic. This is the result of subjecting the data to a statistical test of H0 versus H1 in item [2i]. The p-value noted in item [m] is not cleanly related to this. [2k] The Seq SS column is constructed by fitting the predictor variables in the order given and noting the change in SSregression. Logically, this means here five regressions: Fert on Ag has SSreg = 0.08948 Fert on Ag, Army has SSreg = 0.22104 + 0.08948 = 0.31052 Fert on Ag, Army, Ed has SSreg = 0.08918 + 0.31052 = 0.39970 Fert on Ag, Army, Ed, Catholic has SSreg = 0.06671 + 0.39970 = 0.46641 Fert on Ag, Army, Ed, Catholic, Mort has SSreg = 0.04088 + 0.46641 = 0.50729 The final value 0.50729 is SSreg for the whole regression, item [bb]. Naming the predictor variables in another order would produce different results. This arithmetic can only be interesting if the order of the variables is interesting. In this case, there is no reason to have any interest in these values.

Page 19: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 19

[2l] Minitab will list for you data points which are “unusual observations” and are worthy (perhaps) of special attention. There are two concerns, unusual residuals and high influence.

Unfortunately, Minitab has too low a threshold of concern regarding the residuals, as it will list any standardized residual below -2 or above +2. Virtually every data set has points with this property, so that nothing “unusual” is involved. A more reasonable concern would be for residuals below -2.5 or above +2.5. Indeed, in large data sets, one might move the thresholds of concern to -3 and +3. The (-2, 2) thresholds cannot be reset, so you will have to live with the output.

Extreme residuals occur with points for which the regression model does not fit well. It is always worth examining these points. It is generally not necessary to remove these points from the regression. The geometric configuration of the predictor variables might indicate that some points unduly influence the regression results. These points are said to have high influence or high leverage. Whether such points should be removed from the regression is a difficult question. This section of the Minitab listing does not really provide enough guidance as to the degree of influence. You can get additional information on high influence points through Storage and then asking for Hi (leverages) when the regression is initiated. See also point [2u]. [2m] The unusual observations are identified by their case numbers (here 6, 37, 45, and 47), by their values on the first-named predictor variable, and by their values on the dependent variable. [2n] The fitted values refer to the calculation suggested in item [ø], and here 0.9050 is the value of Fêrt6, the fitted value for point 6. [2ø] The SE Fit refers to the estimated standard deviation of the fitted value in the previous column. This is only marginally interesting. [2p] This gives the actual residual; Here -0.1440 = Fert6 - Fêrt6 = 0.7610 - 0.9050. [2q] Since the actual residuals bear the units of the dependent variable, they are hard to appraise. Thus, we use the standardized residuals. These can be thought of as approximate z-scores, so that about 5% of them should be outside the interval (-2, 2). [2r] This marks a large influence point. See item [2l].

Page 20: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 20

[2s] The purpose of the residual versus fitted plot is to check for possible violations of regression assumptions, particularly non-homogeneous residuals. This pathology reveals itself through a pattern in which the spread of the residuals changes in moving from left to right on the plot. The residual versus fitted plot will sometimes reveal curvature as well. Large positive and large negative residuals will be seen on this plot. The detection and relieving of these pathologies are subtle processes which go beyond the content of this document. The appearance of this plot is not materially influenced by the particular choice made for standardizing the residuals; this was done here from the Stat ⇒⇒⇒⇒ Regression ⇒ Regression ⇒ Graphs panel by choosing Residuals for plots: as Regular. [2t] The purpose of plotting the residuals against the predictor variables (in turn) is to check for non-linearity. This particular plot shows no problems. [2u] The leverage value for the first canton in the data set is 0.156817. This is computed as the (1, 1) entry of the matrix X(X´X)-1X where X is the n-by-(k + 1) matrix whose n rows represent the n data points. The k + 1 columns consist of the constant column (containing a 1 in each position) and one column for each of the k predictor variables. The dependent variable Fert is not involved in this arithmetic. The leverage value for the jth canton is the (j, j) entry of this matrix. A commonly accepted cutoff

marking off “high” leverage points is ( )3 1kn+

, which is here ( )3 5 147

+ ≈ 0.3830. Only

the leverage value for canton 45 is larger than this; see [2r] and [2v]. [2v] The leverage value for the 45th canton is 0.455836. This is clearly a high leverage point. There is some cause for concern, because high leverage points can distort the estimation process. A quick look at the data set will show that this canton, the third from the bottom on the data list, has very unusual values for Ag, Army, and Ed. A thorough analysis of this data set would probably include another run in which canton 45 is deleted. [2w] Best subsets repeats the name of the dependent variable. [2x] The independent variable names are printed vertically, and X marks are used to show which variables are used in which models. In the model corresponding to Vars = 1, only Ed is used. In the model corresponding to Vars = 2, Ed and Catholic are used. It is not necessarily true that all the variables used for Vars = k will be used for Vars = k + 1. [2y] This column indicates the number of variables taken from the panel Free predictors. The best model (in terms of R2) that uses just one variable from this panel is Ferti = β0 + βED EDi + εi . In describing the information for [2x] to [3d] it will be convenient to use k as the entry in the Vars column.

Page 21: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 21

[2z] The model Ferti = β0 + βED EDi + εi produces R2 = 44.1%, and no other model using one variable from the panel Free predictors can do better. Similarly, the model Ferti = β0 + βED EDi + βCATHOLIC CATHOLICi + εi gives R2 = 57.5%, and no other model using two variables from the panel Free predictors can do better.

[3a] The 2adjR value is computed from R2 as R

nn k

Radj2 21

11

1= −−

− −−( ) , as in note [u].

In making this calculation, the value k should be interpreted as (value listed in the Vars column) + (number of variables in panel Predictors in all models)

In the example used here, there are no variables in the panel Predictors in all models.

For the Vars = 3 line, the calculation is 2adjR = 47 11 (1 0.663)

47 1 3−− −

− −≈ 0.639 = 63.9%.

[3b] Mallows’ Cp identifies p through

p = 1 + (value listed in the Vars column) + (number of variables in panel Predictors in all models)

and then

Residual SUM of squares for model in this line ( 2 )Residual MEAN square using all the independent variablespC n p= − −

Minitab does not provide the component parts of this calculation. If the model in this line fits reasonably well, compared to the model using all available independent variables, then Cp should be small. Any model that produces Cp < p should be considered an excellent fit, but most users are very happy with a model that has Cp only a little larger than p. [3c] This is the estimate of the noise standard deviation, as noted in [s]. As the value of Vars increases, the value of S usually (but not always) decreases. [3d] When all available variables are used, Cp = p automatically. This problem has 5 available variables; when all are used p = 1 + 5 + 0 = 6, as in note [3b]. Thus Cp = 6.0 is guaranteed for this position. [3e] The values for Alpha to enter and Alpha to remove are echoed on the output. [3f] The sample size and number of variables are echoed on the output.

Page 22: MULTIPLE LINEAR REGRESSION IN MINITAB - New …pages.stern.nyu.edu/~churvich/Undergrad/Minitab/Multiple.pdfMULTIPLE LINEAR REGRESSION IN MINITAB This document shows a complicated Minitab

GUIDE TO MINITAB REGRESSION ≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈≈

≈≈

page 22

[3g] This display is to be read columnwise. The fitted regression at Step = 1 is found in the indicated column. [3h] This refers to the estimate of β0 , the intercept. This appears in every fitted model. Under Options the user is allowed to fit models without an intercept. Removing the intercept is not recommended. [3i] The variable Ed is selected first, and it is the only variable with information listed under the column for Step = 1. [3j] This is the estimated coefficient when only Ed is used as a predictor. The fitted model is Fêrt = 0.7961 - 0.86 Ed. See note [h]. [3k] The p-value for the most-recently selected variable must always be less than or equal to Alpha to enter. In this example, the p-values for all the variables previously selected are also below this threshold, so there are no steps in which variables are removed. [3l] The variable Catholic is selected next. The corresponding fitted model is Fêrt = 0.7423 - 0.79 Ed + 0.00111 Catholic. [3m] The model Ferti = β0 + βEd Edi + βCatholic Catholici + εi has an estimated noise standard deviation of 0.0833. The values in the row S will decrease as variables are added. (If the addition of a variable caused S to decrease, that variable could not be entered.) [3n] The model Ferti = β0 + βEd Edi + βCatholic Catholici + εi has an R2 value of 57.45%. The values in the row R-Sq will increase as variable are added (and will decrease whenever a variable is removed). [3ø] The model Ferti = β0 + βEd Edi + βCatholic Catholici + εi has an adjusted R-squared value of 55.52%. The values in the row R-Sq(adj) will increase as variables are added. (If the addition of a variable caused R-Sq(adj) to increase, that variable could not be entered.) [3p] The model Ferti = β0 + βEd Edi + βCatholic Catholici + εi has a Cp value of 18.5. See note [3b]. It usually happens that the last (rightmost) column of a stepwise regression produces Cp very close to, or even below, p.