Upload
-
View
229
Download
0
Embed Size (px)
Citation preview
7/27/2019 First Five Chapter Gujrati
1/21
BASIC ECONOMETRICS: FOURTH EDITION
Damodar N. Gujarati
PART SINGLE-EQUATION REGRESSION MODELS 15
1 The Nature of Regression Analysis
2 Two-Variable Regression Analysis: Some Basic Ideas
3 Two-Variable Regression Model: The Problem of Estimation
4 Classical Normal Linear Regression Model (CNLRM)
5 Two-Variable Regression: Interval Estimation and
Hypothesis Testing
6 Extensions of the Two-Variable Linear Regression Model
7 Multiple Regression Analysis: The Problem of Estimation
8 Multiple Regression Analysis: The Problem of Inference
9 Dummy Variable Regression Models
PART II RELAXING THE ASSUMPTIONS OF THE
CLASSICAL MODEL
10 Multicollinearity: What Happens if the Regressors Are Correlated
11 Heteroscedasticity: What Happens if the Error
Variance Is Nonconstant?
12 Autocorrelation: What Happens if the Error Terms Are Correlated
13 Econometric Modeling: Model Specification and
Diagnostic Testing
vi BRIEF CONTENTS
PART III TOPICS IN ECONOMETRICS
14 Nonlinear Regression Models
15 Qualitative Response Regression Models
7/27/2019 First Five Chapter Gujrati
2/21
16 Panel Data Regression Models
17 Dynamic Econometric Models: Autoregressive and
Distributed-Lag Models
PART IV SIMULTANEOUS-EQUATION MODELS
18 Simultaneous-Equation Models
19 The Identification Problem
20 Simultaneous-Equation Methods
21 Time Series Econometrics: Some Basic Concepts
22 Time Series Econometrics: Forecasting
INTRODUCTION
I.1 WHAT IS ECONOMETRICS?
Literally interpreted, econometrics means economic measurement. Al-though measurement is an
important part of econometrics, the scope of econometrics is much broader, as can be seen from the
following quotations: Econometrics, the result of a certain outlook on the role of economics, consists of
the application of mathematical statistics to economic data to lend empirical sup-port to the models
constructed by mathematical economics and to obtain numerical results.
. . . econometrics may be defined as the quantitative analysis of actual economic phenomena based on
the concurrent development of theory and observation, related by appropriate methods of inference.
Econometrics may be defined as the social science in which the tools of economic theory, mathematics,
and statistical inference are applied to the analysis of economic phenomena. Econometrics is concerned
with the empirical determination of economic laws.
METHODOLOGY OF ECONOMETRICS
How do econometricians proceed in their analysis of an economic problem? That is, what is theirmethodology? Although there are several schools of thought on econometric methodology, we present
here the traditional or classical methodology, which still dominates empirical research in economics and
other social and behavioral sciences. Broadly speaking, traditional econometric methodology proceeds
along the following lines:
1. Statement of theory or hypothesis.
7/27/2019 First Five Chapter Gujrati
3/21
2. Specification of the mathematical model of the theory
3. Specification of the statistical, or econometric, model
4. Obtaining the data
5. Estimation of the parameters of the econometric model
6. Hypothesis testing
7. Forecasting or prediction
8. Using the model for control or policy purposes.
FIGURE I.4 Anatomy of econometric modeling. Please see Page 10.
CHAPTER ONE: The Nature of Regression Analysis
1.1Historical origin of the term regression
1.2The modern interpretation of regression
1.3 STATISTICAL VERSUS DETERMINISTIC RELATIONSHIPS
1.4 REGRESSION VERSUS CAUSATION
1.5 REGRESSION VERSUS CORRELATION
1.6 TERMINOLOGY AND NOTATION
Before we proceed to a formal analysis of regression theory, let us dwell briefly on the matter of
terminology and notation. In the literature the terms dependent variable and explanatory variable are
described variously. A representative list is:
Dependent variable Explanatory variable
Explained variable Independent variable
Predictand Predictor
Regressand Regressor
Response Stimulus
Endogenous Exogenous
Outcome Covariate
7/27/2019 First Five Chapter Gujrati
4/21
Controlled variable Control variable
1.7 THE NATURE AND SOURCES OF DATA FOR ECONOMIC ANALYSIS10
The success of any econometric analysis ultimately depends on the avail-ability of the appropriate data.
It is therefore essential that we spend some time discussing the nature, sources, and limitations of the
data that one may encounter in empirical analysis.
Types of Data
Three types of data may be available for empirical analysis: time series, cross-section, and pooled (i.e.,
combination of time series and cross-section) data.
Time Series Data Table I.2: Please see Page 32.
Cross-section Data Table 1.1 Pages 27
7/27/2019 First Five Chapter Gujrati
5/21
CHAPTER TWO: TWO-VARIABLE REGRESSION ANALYSIS: SOME BASIC IDEAS 41
2.1 A hypothetical Example
2.2 THE CONCEPT OF POPULATION REGRESSION FUNCTION (PRF)
2.3 THE MEANING OF THE TERM LINEAR
Since this text is concerned primarily with linear models, it is essential to know what the term linear
really means, for it can be interpreted in two different ways.
Linearity in the Variables
Linearity in the Parameters
2.4 STOCHASTIC SPECIFICATION OF PRF
2.5 THE SIGNIFICANCE OF THE STOCHASTIC DISTURBANCE TERM
As noted in Section 2.4, the disturbance term ui is a surrogate for all those variables that are omitted
from the model but that collectively affect Y. The obvious question is: Why not introduce these variables
into the model explicitly? Stated otherwise, why not develop a multiple regression model with as many
variables as possible? The reasons are many.
1. Vagueness of theory: The theory, if any, determining the behavior of Y may be, and often is,
incomplete. We might know for certain that weekly income X influences weekly consumption
expenditure Y, but we might be ignorant or unsure about the other variables affecting Y. Therefore, ui
maybe used as a substitute for all the excluded or omitted variables from the model.
2. Unavailability of data: Even if we know what some of the excluded variables are and therefore
consider a multiple regression rather than a simple regression, we may not have quantitative
information about these.
3. Core variables vs peripheral variables
4. Intrinsic randomness in human behavior
5. Poor proxy variables
6. Principles of parsimony
7. Wrong functional form
2.7 AN ILLUSTRATIVE EXAMPLE Please see page 51.
7/27/2019 First Five Chapter Gujrati
6/21
7/27/2019 First Five Chapter Gujrati
7/21
CHAPTER THREE: TWO-VARIABLE REGRESSION MODEL
3.1 THE METHOD OF ORDINARY LEAST SQUARES
3.2 THE CLASSICAL LINEAR REGRESSION MODEL:
THE ASSUMPTIONS UNDERLYING THE METHOD OF LEAST SQUARES
Assumption 1: Linear regression model. The regression model is linear in the parameters, as shown in
Yi = 1 2Xi ui
This assumption means that the dependent variable can be calculated as a linear function of a set of
independent variables, plus a disturbance term. This is shown in the above equation that states that:
the regression model is linear in the unknown coefficients 1, 2 so that Yi = 1 2Xi ui , for i =
1,2,3.,n.
Assumption 2: X values are fixed in repeated sampling. Values taken by the regressor X are consideredfixed in repeated samples. More technically, X is assumed to be nonstochastic. the manner in which Yi
are generated. To see why this requirement is needed, look at the PRF: Yi = 1 2 Xi ui . It shows that
Yi depends on both Xi and ui . Therefore, unless we are specific about how Xi and ui are created or
generated, there is no way we can make any statistical inference about the Yi and also, as we shall see,
about 1 and 2. Thus, the assumptions made about the Xi variable(s) and the error term are extremely
critical to the valid interpretation of the regression estimates.
Assumption 3 states that the mean value of ui , conditional upon the given Xi , is zero. Geometrically,
this assumption can be pictured as in Figure 3.3, which shows a few values of the variable X and the Y
populations associated with each of them. As shown, each Y population corresponding to a given X is
distributed around its mean value (shown by the circled points on the PRF) with some Y values above
the mean and some below it. The distances above and below the mean values are nothing but the ui ,
and what (3.2.1) requires is that the average or mean value of these deviations corresponding to any
given X should be zero.
This assumption should not be difficult to comprehend in view of the discussion in Section 2.4 .All that
this assumption says is that the factors not explicitly included in the model, and therefore subsumed in
ui , do not systematically affect the mean value of Y; so to speak, the positive ui 9For illustration, we are
assuming merely that the us are distributed symmetrically as shown in Figure 3.3. But shortly we will
assume that the us are distributed normally.
Assumption 4: Homoscedasticity or equal variance of ui. Given the value of X, the variance of ui is the
same for all observations. That is, the conditional variances of ui are identical. Symbolically, we have
var (ui | Xi) = E *ui E(ui | Xi)+2
= E(ui2 | Xi ) because of Assumption 3
7/27/2019 First Five Chapter Gujrati
8/21
= 2 (3.2.2)
where var stands for variance.
In passing, note that the assumption E(ui | Xi ) = 0 implies that E(Yi | Xi ) = i 2 Xi . (Why?) Therefore,
the two assumptions are equivalent.
Assumption 5: No autocorrelation between the disturbances. Given any two X values, Xi and Xj (i = j ),
the correlation between any two ui and uj (i = j ) is zero. Symbolically, cov (ui,uj | Xi,Xj) = E *ui E (ui)+ |
Xi -*uj E(uj)+ | Xj -
= E(ui | Xi)(uj | Xj) (why?)
= 0 (3.2.5)
Assumption 6: Zero covariance between ui and Xi , or E(ui/Xi) = 0. Formally, cov (ui, Xi) = E *ui E(ui)+*Xi
E(Xi)+
= E *ui (Xi E(Xi))+ since E(ui) = 0
= E (uiXi) E(Xi)E(ui) since E(Xi) is non-stochastic (3.2.6)
= E(uiXi) since E(ui) = 0
= 0 by assumption
Assumption 6 states that the disturbance u and explanatory variable X are uncorrelated. The rationale
for this assumption is as follows: When we expressed the PRF as in (2.4.2), we assumed that X and u
(which may represent the influence of all the omitted variables) have separate (and additive) influence
on Y. But if X and u are correlated, it is not possible to assess their individual effects on Y. Thus, if X and u
are positively correlated, X increases.
Assumption 7: The number of observations n must be greater than the number of parameters to be
estimated. Alternatively, the number of observations n must be greater than the number of explanatory
variables.
Assumption 8: Variability in X values. The X values in a given sample must not all be the same.
Technically, var (X ) must be a finite positive number.13
Assumption 9:The regression model is correctly specified. Alternatively, there is no specification bias or
error in the model used in empirical analysis.
Assumption 10: There is no perfect multicollinearity. That is, there are no perfect linear relationships
among the explanatory variables.
7/27/2019 First Five Chapter Gujrati
9/21
3.3 PRECISION OR STANDARD ERRORS OF LEAST-SQUARES ESTIMATES
Meaning of Degrees of Freedom: Please see page 77
3.4 PROPERTIES OF LEAST-SQUARES ESTIMATORS: THE GAUSSMARKOV THEOREM
As noted earlier, given the assumptions of the classical linear regression model, the least-squares
estimates possess some ideal or optimum proper-ties. These properties are contained in the well-known
GaussMarkov theorem. To understand this theorem, we need to consider the best linear unbiasedness
property of an estimator. For this the following conditions must hold:
1. It is linear, that is, a linear function of a random variable, such as the dependent variable Y in the
regression model.
2. It is unbiased, that is, its average or expected value, E( 2), is equal to the true value, 2.
3. It has minimum variance in the class of all such linear unbiased estimators; an unbiased estimator
with the least variance is known as an efficient estimator. In the regression context it can be proved that
the OLS estimators are BLUE. This is the gist of the famous GaussMarkov theorem, which can be stated
as follows:
GaussMarkov Theorem: Given the assumptions of the classical linear regression model, the least-
squares estimators, in the class of unbiased linear estimators, have minimum variance, that is, they are
BLUE. The proof of this theorem is sketched in Appendix 3A, Section 3A.6.
Question: Prove that OLS coefficient for the slope parameter in the simple linear regression model is
unbiased and efficient.
Answer: See Appendix A.
3.5 THE COEFFICIENT OF DETERMINATION2r : A MEASURE OF GOODNESS OF FIT: The regular
coefficient of determination,2r is a measure of the closeness of fit in the multiple regression. However,
2r , cannot be used as a means of comparing two different equations containing different numbers ofexplanatory variables. This is because when additional explanatory variables are added, the proportion
of variation in Y (Dependent variable) explained by the Xs (Independent variable),2r , will always be
increased. Therefore, we will always obtain a higher2r regardless of the importance or not of the
additional regressor. For this reason we need a different measure that will take into account the number
of explanatory variables included in each model. This measure is called the adjusted2r , because it is
adjusted for the number of regressors (or adjusted for the degrees of freedom)
Please see page 81-87 for details.
7/27/2019 First Five Chapter Gujrati
10/21
Adjusted2r : Adjusted
2r =)(
)1(1
)1/(
)/(1
knTSS
nRSS
nTSS
knRSS
RSS =Residual sum of squares; TSS =Total sum of squares.
TSS = RSS + ESS ; ESS = explained sum of squares.
General criteria for model selection:
As we know that increasing the number of explanatory variables in multiple regression model will
decrease the RSS, and2r will therefore increase. However, the cost of that is a loss in terms of degrees
of freedom. A different method-apart from adjusted2r - of allowing for number of Xs to change when
assessing goodness of fit is to use different criteria for model comparison such as:
Akaike Information Criterion (AIC); Schwartz Information Criterion (SIC); Schwartz Bayesian Criterion
(SBC), Hannan and Quin Critirion (HQC).
The general guideline to select a model: In comparing two or more models, the model with the lowest
AIC is preferred when used AIC criterion. Similarly the model with the lowest SIC is preferred when used
SIC criterion.
For details please see page: 530-539.
3.6 A NUMERICAL EXAMPLE: Please see this example at page 89.
Illustrative Example: Please see 90.
7/27/2019 First Five Chapter Gujrati
11/21
Some examples of Regression
Data Source: wage.wf1
SUMMARY
OUTPUT TABLE 1
Regression Statistics
Multiple R 0.385548014
R Square 0.148647271
Adjusted R Square 0.14579676
Standard Error 0.388465077
Observations 900
ANOVA
df SS MS FSignificance
F
Regression 3 23.60801004 7.869336682 52.14758055 4.53E-31
Residual 896 135.2109838 0.150905116
Total 899 158.8189938
Coefficients Standard Error t Stat P-value Lower 95%
Up
95
Intercept 5.528329355 0.112794574 49.01236961 8.6993E-256 5.306957 5.74
EDUC 0.07311662 0.006635679 11.01871044 1.44273E-26 0.060093 0.0
EXPER 0.015357835 0.003425312 4.483630929 8.2905E-06 0.008635 0.0TENURE 0.012964063 0.002630727 4.927939405 9.89741E-07 0.007801 0.01
14758055.52150905116.0
869336682.7
)/(
)1/(
/
/
knRSS
kESS
dfRSS
dfESSF
OR)/()1(
)1/(2
2
knR
kRF
The p-value of obtaining an Fvalue of as much as 52.14758055 or greater is zero, leading to therejection of the hypothesis that together educ, exper, and tenure have no effect on lnwage. If
you were to use the conventional 5% level-of-significance value, the critical F value for 3 df in
the numenator and 896 df in the denominator is about 3.84. Obviously the observed F of
52.14758055 far exceeds the critical Fvalues of 3.84 and thereby we reject the null hypothesis
of 0: 3210 H .
7/27/2019 First Five Chapter Gujrati
12/21
Note that t-testing procedure is based on the assumption that the error termi
u follows the
normal distribution. Although we cannot directly observei
u , we can observe their proxy, the
iu , that is, the residuals. For our lnwage regression, the histogram of the residuals is as shown
in the above figure. From the histogarmit seems that that the residuals are not normally
distributed. In our case, the JB value is 35 with a p- value of 0. It seems that the error term is
not normally distributed.
For our example, the skewness value is -0.23 and the kutosis value is 3.84. Recall that for a
normally distributed variablr the skewness and kurtosis values are, respectively, 0 and 3.
0
20
40
60
80
100
120
-1.5 -1.0 -0.5 0.0 0.5 1.0
Series: RESIDSample 1 900Observations 900
Mean 1.16e-15Median 0.023735Maximum 1.332397Minimum -1.837909Std. Dev. 0.387816Skewness -0.236952Kurtosis 3.841932
Jarque-Bera 35.00382Probability 0.000000
7/27/2019 First Five Chapter Gujrati
13/21
0
50
100
150
200
250
300
350
400
9 10 11 12 13 14 15 16 17 18
Series: EDUCSample 1 900Observations 900
Mean 13.48000Median 12.00000Maximum 18.00000Minimum 9.000000Std. Dev. 2.200374Skewness 0.541434Kurtosis 2.261056
Jarque-Bera 64.44902Probability 0.000000
7/27/2019 First Five Chapter Gujrati
14/21
Some examples of Regression
Data Source: wage.wf1
itenureereducwage 3210 expln
As may be seen from Table 1.1, all three variables have positive coefficients. These are all above
the rule of thumb critical t-value of 2, hence all are significant. So, it may be said that wages
will increase as education, experience and tenure increases. Despite the significance of these
variables, the adjusted 2r is quite low (0.145) as there are probably other variables that affect wage.
For example, education increases by 1 year, on average, wage would increase by Taka 0.07. If education
was zero, the average wage would be Taka 5.52 (which is the constant). The2r value of about 0.15
means that 15% of the variation in lnwage is explained by educ, exper, and tenure.
Testing hypothesis: Suppose we want to test null hypothesis that there is no relationship between
lnwage and education , that is, the true slope coefficient 01 . The estimated value of 1 is 0.073117.
If the null hypothesis were true, what is the probability of obtaining a value of 0.073117? Under the null
hypothesis, we observe that the t- value is 11.01871 and the p value of obtaining such a t- value is
practically zero. In other words, we can reject the null hypothesis resoundingly.
Dependent Variable: LNWAGE
Method: Least Squares
Date: 11/05/12 Time: 18:49
Sample: 1 900Included observations: 900
Table 1.1: Results from the wage equation
Variable Coefficient Std. Error t-Statistic Prob.
C 5.528329 0.112795 49.01237 0.0000
EDUC 0.073117 0.006636 11.01871 0.0000
EXPER 0.015358 0.003425 4.483631 0.0000
TENURE 0.012964 0.002631 4.927939 0.0000
R-squared 0.148647 Mean dependent var 6.786164
Adjusted R-squared 0.145797 S.D. dependent var 0.420312
S.E. of regression 0.388465 Akaike info criterion 0.951208
Sum squared resid 135.2110 Schwarz criterion 0.972552Log likelihood -424.0434 Hannan-Quinn criter. 0.959361
F-statistic 52.14758 Durbin-Watson stat 1.750376
Prob(F-statistic) 0.000000
7/27/2019 First Five Chapter Gujrati
15/21
.
Equation: Untitled
Test Statistic Value df Probability
t-statistic 0.498655 896 0.6181
F-statistic 0.248656 (1, 896) 0.6181
Chi-square 0.248656 1 0.6180
Null Hypothesis: C(3)=C(4)
Null Hypothesis Summary:
Descriptive statistics:Table 2
EDUC EXPER WAGE TENURE
Mean 13.48000 11.59222 964.2644 7.265556
Median 12.00000 11.00000 912.5000 7.000000
Maximum 18.00000 23.00000 3078.000 22.00000
Minimum 9.000000 1.000000 115.0000 0.000000
Std. Dev. 2.200374 4.379564 405.1624 5.080611
Skewness 0.541434 0.072955 1.203925 0.422778
Kurtosis 2.261056 2.437964 5.746906 2.199197
Jarque-Bera 64.44902 12.64404 500.3712 50.85935
Probability 0.000000 0.001796 0.000000 0.000000
Sum 12132.00 10433.00 867838.0 6539.000
Sum Sq. Dev. 4352.640 17243.35 1.48E+08 23205.53
Observations 900 900 900 900
7/27/2019 First Five Chapter Gujrati
16/21
Normalized Restriction (= 0) Value Std. Err.
C(3) - C(4) 0.002394 0.004800
Restrictions are linear in coefficients.
Redundant Variables Test
Equation: UNTITLED
Specification: LNWAGE C EDUC EXPER TENURE
Redundant Variables: TENURE
Value df Probability
t-statistic 4.927939 896 0.0000
F-statistic 24.28459 (1, 896) 0.0000
Likelihood ratio 24.06829 1 0.0000
F-test summary:
Sum of Sq. dfMean
Squares
Test SSR 3.664668 1 3.664668
Restricted SSR 138.8757 897 0.154822
Unrestricted SSR 135.2110 896 0.150905
Unrestricted SSR 135.2110 896 0.150905
LR test summary:
Value df
Restricted LogL -436.0776 897
Unrestricted LogL -424.0434 896
Restricted Test Equation:
Dependent Variable: LNWAGE
Method: Least SquaresDate: 11/05/12 Time: 18:53
Sample: 1 900
Included observations: 900
Variable Coefficient Std. Error t-Statistic Prob.
C 5.537798 0.114233 48.47827 0.0000
EDUC 0.075865 0.006697 11.32741 0.0000
EXPER 0.019470 0.003365 5.786278 0.0000
R-squared 0.125573 Mean dependent var 6.786164
Adjusted R-squared 0.123623 S.D. dependent var 0.420312
S.E. of regression 0.393475 Akaike info criterion 0.975728
Sum squared resid 138.8757 Schwarz criterion 0.991736
Log likelihood -436.0776 Hannan-Quinn criter. 0.981843
F-statistic 64.40718 Durbin-Watson stat 1.770020
Prob(F-statistic) 0.000000
7/27/2019 First Five Chapter Gujrati
17/21
Dependent Variable: LNWAGE
Method: Least Squares
Date: 11/05/12 Time: 18:55
Sample: 1 900
Included observations: 900
Variable Coefficient Std. Error t-Statistic Prob.
C 6.697589 0.040722 164.4699 0.0000
EXPER -0.002011 0.003239 -0.621069 0.5347
TENURE 0.015400 0.002792 5.516228 0.0000
R-squared 0.033285 Mean dependent var 6.786164
Adjusted R-squared 0.031130 S.D. dependent var 0.420312
S.E. of regression 0.413718 Akaike info criterion 1.076062
Sum squared resid 153.5327 Schwarz criterion 1.092070
Log likelihood -481.2281 Hannan-Quinn criter. 1.082178
F-statistic 15.44241 Durbin-Watson stat 1.662338
Prob(F-statistic) 0.000000
Omitted Variables Test
Equation: UNTITLED
Specification: LNWAGE C EXPER TENURE
Omitted Variables: EDUC
Value df Probability
t-statistic 11.01871 896 0.0000
F-statistic 121.4120 (1, 896) 0.0000
Likelihood ratio 114.3693 1 0.0000
F-test summary:
Sum of Sq. dfMean
Squares
Test SSR 18.32169 1 18.32169Restricted SSR 153.5327 897 0.171162
Unrestricted SSR 135.2110 896 0.150905
Unrestricted SSR 135.2110 896 0.150905
LR test summary:
Value df
Restricted LogL -481.2281 897
Unrestricted LogL -424.0434 896
Unrestricted Test Equation:
Dependent Variable: LNWAGE
Method: Least Squares
Date: 11/05/12 Time: 18:56
Sample: 1 900
Included observations: 900
Variable Coefficient Std. Error t-Statistic Prob.
C 5.528329 0.112795 49.01237 0.0000
EXPER 0.015358 0.003425 4.483631 0.0000
TENURE 0.012964 0.002631 4.927939 0.0000
7/27/2019 First Five Chapter Gujrati
18/21
EDUC 0.073117 0.006636 11.01871 0.0000
R-squared 0.148647 Mean dependent var 6.786164
Adjusted R-squared 0.145797 S.D. dependent var 0.420312
S.E. of regression 0.388465 Akaike info criterion 0.951208
Sum squared resid 135.2110 Schwarz criterion 0.972552
Log likelihood -424.0434 Hannan-Quinn criter. 0.959361
F-statistic 52.14758 Durbin-Watson stat 1.750376
CLASSICAL NORMAL LINEAR REGRESSION MODEL (CNLRM): CHAPTER 4
4.2 THE NORMALITY ASSUMPTION FOR ui
The classical normal linear regression model assumes that each ui is distributed normally with
Mean: E(ui ) = 0 (4.2.1)
Variance: E*ui E(ui )+2 = E u2 = 2 (4.2.2)
cov (ui, uj): E*(ui E(ui)+*uj E(uj )+-= E(ui uj ) = 0 i j (4.2.3)
The assumptions given above can be more compactly stated as ui N(0, 2) (4.2.4)
Why the normality assumption is required? Please see page 109
4.3: Properties of the OLS estimators under the normality assumption:
See page: 110-112.
7/27/2019 First Five Chapter Gujrati
19/21
CHAPTER FIVE: TWO VARIABLE REGRESSION: INTERVAL ESTIMATION AND HYPOTHESIS TESTING
5.1: Statistical prerequisites
5.2 Interval Estimation: Some basic ideas
5.3 Confidence Intervals for regression coefficients
5.4 CONFIDENCE INTERVAL FOR 2
5.5 HYPOTHESIS TESTING: GENERAL COMMENTS
Having discussed the problem of point and interval estimation, we shall now consider the topic of
hypothesis testing. In this section we discuss briefly some general aspects of this topic; Appendix A gives
some additional details. The problem of statistical hypothesis testing may be stated simply as follows: Is
a given observation or finding compatible with some stated hypothesis or not? The word compatible,
as used here, means sufficiently close to the hypothesized value so that we do not reject the stated
hypothesis.
5.6 HYPOTHESIS TESTING: THE CONFIDENCE-INTERVAL APPROACH
Two-Sided or Two-Tail Test
To illustrate the confidence-interval approach, once again we revert to the
Consumptionincome example. As we know, the estimated marginal propen-sity to consume (MPC),
2, is 0.5091. Suppose we postulate that
H0: 2 = 0.3
H1: 2 = 0.3
that is, the true MPC is 0.3 under the null hypothesis but it is less than or greater than 0.3 under the
alternative hypothesis. The null hypothesis is a simple hypothesis, whereas the alternative hypothesis is
composite; actually it is what is known as a two-sided hypothesis. Very often such a two-sided
alternative hypothesis reflects the fact that we do not have a strong a priori or theoretical expectation
about the direction in which the alternative hypothesis should move from the null hypothesis.
Is the observed 2 compatible with H0? To answer this question, let us refer to the confidence interval
(5.3.9). We know that in the long run intervals like (0.4268, 0.5914) will contain the true 2 with 95
percent probability. Consequently, in the long run (i.e., repeated sampling) such intervals provide a
range or limits within which the true 2 may lie with a confidence coefficient of, say, 95%. Thus, the
confidence interval provides a set of plausible null hypotheses. Therefore, if 2 under H0 falls within the
100(1 )% confidence interval, we do not reject the null hypothesis; if it lies outside the interval, we
may reject it.7 This range is illustrated schematically in Figure 5.2. Always bear in mind that there is a
100 percent chance that the confidence interval does not contain 2 under H0 even though the
7/27/2019 First Five Chapter Gujrati
20/21
hypothesis is correct. In short, there is a 100 percent chance of committing a Type I error. Thus, if =
0.05, there is a 5 percent chance that we could reject the null hypothesis even though it is true.
Following this rule, for our hypothetical example, H0: 2 = 0.3 clearly lies outside the 95% confidence
interval given in (5.3.9). Therefore, we can reject the hypothesis that MPC is 0.3, with 95% confidence.
One-Sided or One-Tail Test
Sometimes we have a strong a priori or theoretical expectation (or expectations based on some previous
empirical work) that the alternative hypothe-sis is one-sided or unidirectional rather than two-sided, as
just discussed. Thus, for our consumptionincome example, one could postulate that
H0: 2 0.3 and H1: 2 > 0.3
Perhaps economic theory or prior empirical work suggests that the mar-ginal propensity to consume is
greater than 0.3. Although the procedure to test this hypothesis can be easily derived from (5.3.5), the
actual mechanics are better explained in terms of the test-of-significance approach.
5.7 HYPOTHESIS TESTING:
THE TEST-OF-SIGNIFICANCE APPROACH
Testing the Significance of Regression Coefficients: The t Test An alternative but complementary
approach to the confidence-interval method of testing statistical hypotheses is the test-of-significance
approach developed along independent lines by R. A. Fisher and jointly by Neyman and Pearson. Broadly
speaking, a test of significance is a procedure by which sample results are used to verify the truth or
falsity of a null hypothesis. The key idea behind tests of significance is that of a test statistic (estimator)
and the sampling distribution of such a statistic under the null hypothesis. The decision to accept or
reject H0 is made on the basis of the value of the test statistic obtained from the data at hand.
As an illustration, recall that under the normality assumption the variable
t = 2 2/se ( 2)
=2
22 )( ix )/ please see page 129-132
5.8 HYPOTHESIS TESTING: SOME PRACTICAL ASPECTS
The Meaning of Accepting or Rejecting a Hypothesis If on the basis of a test of significance, say, the t
test, we decide to accept the null hypothesis, all we are saying is that on the basis of the sample
evidence we have no reason to reject it; we are not saying that the null hypothesis is true beyond any
doubt. Why? To answer this, let us revert to our consumptionincome example and assume that H0: 2
(MPC) = 0.50. Now the estimated value of the MPC is 2 = 0.5091 with a se ( 2) = 0.0357. Then
7/27/2019 First Five Chapter Gujrati
21/21
on the basis of the t test we find that t = (0.5091 0.50)/0.0357 = 0.25, which is insignificant, say, at
= 5%. Therefore, we say accept H0. But now let us assume H0: 2 = 0.48. Applying the t test, we
obtain t = (0.5091 0.48)/0.0357 = 0.82, which too is statistically insignificant. So now we say accept
this H0.
Please read page 139:The choice between confidence interval and test of significance approaches to
hypothesis testing
5.9 REGRESSION ANALYSIS AND ANALYSIS OF VARIANCE
In this section we study regression analysis from the point of view of the analysis of variance and
introduce the reader to an illuminating and complementary way of looking at the statistical inference
problem.
5.12 EVALUATING THE RESULTS OF REGRESSION ANALYSIS
Please read page 147: Normality tests that include (1) Histogram of residuals; (2) Normal probability
plot (NPP), a graphical device; and (3) Jarque-Bera test.