3. Hypothesis tests and confidence intervals in multiple regression

3. Hypothesis tests and confidence intervals in multiple regression Contents of previous section: • Definition of the multiple regression model • OLS ...
Author: Adele Fields
1 downloads 1 Views 271KB Size
3. Hypothesis tests and confidence intervals in multiple regression Contents of previous section: • Definition of the multiple regression model • OLS estimation of the coefficients • Measures-of-fit (based on estimation results) • Some problems in the regression model (omitted-variable bias, multicollinearity) Now: • Statistical inference based on OLS estimation (hypothesis tests, confidence intervals) 39

3.1. Standard errors for the OLS estimators Recall: • OLS estimators are subject to sampling uncertainty • Given the OLS assumptions on Slide 18 the OLS estimators are normally distributed in large samples, that is βˆj ∼ N (βj , σβ2ˆ ) j

for j = 0, . . . , k

Now: • How can we estimate the (unknown) r OLS estimator’s variance σ 2ˆ and its standard deviation σ 2ˆ ≡ σβˆ βj

βj

j

40

Definition 3.1: (Standard error) We call an appropriately defined estimator of the standard deviation σβˆ the standard error of βˆj and denote it by SE(βˆj ). j

Natural question: • What constitutes a good estimator of σβˆ ? j Answer: • The analytical formula of a good estimator crucially hinges on whether the errors ui are homoskedastic or heteroskedastic (see Definition 2.2 on Slide 8) 41

3.1.1. Homoskedasticity / heteroskedasticity Important notes: • The way we defined the terms ’homoskedasticity’ and ’heteroskedasticity’ in Definition 2.2 on Slide 8 implies that ’homoskedasticity’ is a special case of ’heteroskedasticity’ (’heteroskedasticity’ is more general than ’homoskedasticity’) • Since the OLS assumptions on Slide 18 place no restrictions on the conditional variance of the error terms ui, they apply to both the general case of ’heteroskedasticity’ and the special case of ’homoskedasticity’ −→ Theorem 2.4 on Slide 19 is valid under both concepts 42

Corollary 3.2: (To Theorem 2.4, Slide 19) Given the OLS assumptions on Slide 18, the OLS estimators are unbiased, consistent, and normally distributed in large samples (asymptotically normal) irrespective of whether the error terms are heteroskedastic or homoskedastic. Classical econometrics: • In classical econometrics the default assumption is that the error terms are homoskedastic • Given our OLS assumptions plus homoskedasticity, the OLS estimators are efficient (optimal) among all alternative linear and unbiased estimators of the regression coefficients β0 , . . . , β k (Gauss-Markov theorem) 43

Classical econometrics: [continued] • Under heteroskedasticity there are more efficient estimators than OLS, namely the so-called (feasible) Generalized Least Squares (GLS) estimators (see the lectures Econometrics I+II) Mathematical aspects: • There exist specific formulas for the standard errors SE(βˆj ) both under heteroskedasticity and homoskedasticity • Since homoskedasticity is a special case of heteroskedasticity the standard errors under homoskedasticity have a simpler structural form (homoskedasticity-only standard errors) 44

Mathematical aspects: [continued] • The homoskedasticity-only standard errors are valid only under homoskedasticity, but lead to invalid statistical inference under heteroskedasticity • The more general standard errors under heteroskedasticity were proposed by Eicker (1967), Huber (1967), and White (1980) (Eicker-Huber-White standard errors) • The Eicker-Huber-White standard errors produce valid statistical inference irrespective of whether the error terms are heteroskedastic or homoskedastic (heteroskedasticity-robust standard errors) 45

Homoskedasticity-only and heteroskedasticity-robust standard errors for the house-prices dataset Dependent Variable: SALEPRICE Method: Least Squares Date: 28/02/12 Time: 09:43 Sample: 1 546 Included observations: 546 White heteroskedasticity-consistent standard errors & covariance

Dependent Variable: SALEPRICE Method: Least Squares Date: 07/02/12 Time: 16:50 Sample: 1 546 Included observations: 546 Variable

Coefficient

Std. Error

t-Statistic

Prob.

Variable

Coefficient

Std. Error

t-Statistic

Prob.

C LOTSIZE BEDROOMS BATHROOMS STOREYS

-4009.550 5.429174 2824.614 17105.17 7634.897

3603.109 0.369250 1214.808 1734.434 1007.974

-1.112803 14.70325 2.325153 9.862107 7.574494

0.2663 0.0000 0.0204 0.0000 0.0000

C LOTSIZE BEDROOMS BATHROOMS STOREYS

-4009.550 5.429174 2824.614 17105.17 7634.897

3668.048 0.459424 1262.593 2263.281 917.6527

-1.093102 11.81735 2.237153 7.557690 8.320029

0.2748 0.0000 0.0257 0.0000 0.0000

R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood F-statistic Prob(F-statistic)

0.535547 0.532113 18265.23 1.80E+11 -6129.993 155.9529 0.000000

Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Hannan-Quinn criter. Durbin-Watson stat

 

68121.60 26702.67 22.47250 22.51190 22.48790 1.482942

R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood F-statistic Prob(F-statistic)

0.535547 0.532113 18265.23 1.80E+11 -6129.993 155.9529 0.000000

Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Hannan-Quinn criter. Durbin-Watson stat

68121.60 26702.67 22.47250 22.51190 22.48790 1.482942

 

46

Practical issues: • Heteroskedasticity arises in many econometric applications −→ It is prudent to assume heteroskedastic errors unless you have compelling reasons to believe otherwise • Rule of thumb: to be on the safe side, always use heteroskedasticity-robust standard errors • Many software packages (like EViews) report homoskedasticity-only standard errors as their default setting • It is up to the user to activate the option of heteroskedasticity-robust standard errors (in EViews: use the command ls(cov=white)) 47

3.1.2. Autocorrelated errors Problem: • Particularly in time-series regressions (that is when the index i represents distinct points in time) we often encounter autocorrelated error terms: Corr(ui, uj ) = 6 0

for some i 6= j

• Under autocorrelation the OLS coefficient estimators are still consistent, but the usual OLS standard errors become inconsistent −→ Statistical inference based on the usual OLS standard errors becomes invalid 48

Solution: • Standard errors should be computed using a heteroskedasticity- and autocorrelation-consistent (HAC) estimator of the variance • Such HAC standard errors become relevant in the Sections 6 and 9 • A well-known (special) HAC estimator is the so-called NeweyWest variance estimator (see Newey and West, 1987) • For a more formal discussion of HAC estimators see Stock and Watson (2011, Section 15.4) 49

3.2. Hypothesis tests and confidence intervals for a single coefficient Testing problem: • Consider one of the k regressors, say Xj , and the corresponding regression coefficient βj • We aim at testing the two-sided problem that the unknown βj takes on some specific value βj,0 • In technical terms: H0 : βj = βj,0

vs.

H1 : β j = 6 βj,0 50

Testing procedure: • Compute the standard error of βˆj , SE(βˆj ) • Compute the so-called t-statistic:

βˆj − βj,0 t= SE(βˆj )

• Compute the p-value:

Œ‘ Œ Œ actŒ p-value = 2 · Φ − Œt Œ , 

(3.1)

(3.2)

where tact is the value of the t-statistic actually computed and Φ(·) is the cdf of the standard normal distribution • Reject H0 at the 5% significance level if p-value < 0.05 (or, equivalently, if |tact| > 1.96) 51

Remarks: • Our testing procedure makes use of the result that the sampling distribution of the OLS estimator βˆj is approximately normal for moderate and large sample sizes • Under H0 the mean of this distribution is βj,0 −→ The t-statistic (3.1) is approximately N (0, 1) distributed • The phrasing ’t-statistic’ stems from the fact that for finite sample sizes and under some additional (classical) assumptions on the multiple regression model the t-statistic (3.1) follows the t-distribution with n − k − 1 degrees of freedom • Given these restrictive assumptions the p-values should be computed from the quantiles of the tn−k−1-distribution 52

Remarks: [continued] • Since these additional assumptions are rarely met in realworld applications and since sample sizes are typically moderate or even large, we base inference on the p-values (3.2) computed from the normal distribution

Attention: • Many software packages (like EViews) assume the validity of the classical assumptions and report p-values based on the tn−k−1-distribution in the default setting −→ p-values should be corrected manually (see the following example) 53

p-values for the house-prices dataset based on the t- and the normal distribution, respectively

Dependent Variable: SALEPRICE Method: Least Squares Date: 28/02/12 Time: 09:43 Sample: 1 546 Included observations: 546 White heteroskedasticity-consistent standard errors & covariance

Dependent Variable: SALEPRICE Method: Least Squares Date: 28/02/12 Time: 09:43 Sample: 1 546 Included observations: 546 White heteroskedasticity-consistent standard errors & covariance Variable

Coefficient

Std. Error

t-Statistic

Prob.

Variable

Coefficient

Std. Error

t-Statistic

Prob.

C LOTSIZE BEDROOMS BATHROOMS STOREYS

-4009.550 5.429174 2824.614 17105.17 7634.897

3668.048 0.459424 1262.593 2263.281 917.6527

-1.093102 11.81735 2.237153 7.557690 8.320029

0.2748 0.0000 0.0257 0.0000 0.0000

C LOTSIZE BEDROOMS BATHROOMS STOREYS

-4009.550 5.429174 2824.614 17105.17 7634.897

3668.048 0.459424 1262.593 2263.281 917.6527

-1.093102 11.81735 2.237153 7.557690 8.320029

0.2743 0.0000 0.0253 0.0000 0.0000

R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood F-statistic Prob(F-statistic)

 

0.535547 0.532113 18265.23 1.80E+11 -6129.993 155.9529 0.000000

Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Hannan-Quinn criter. Durbin-Watson stat

68121.60 26702.67 22.47250 22.51190 22.48790 1.482942

R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood F-statistic Prob(F-statistic)

0.535547 0.532113 18265.23 1.80E+11 -6129.993 155.9529 0.000000

Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Hannan-Quinn criter. Durbin-Watson stat

68121.60 26702.67 22.47250 22.51190 22.48790 1.482942

 

54

Remarks: [continued] • A (1 − α) two-sided confidence interval for the coefficient βj is an interval that contains the true value of βj with a (1 − α) probability • It contains the true value of βj in 100·(1−α)% of all possible randomly drawn samples • Equivalently, it is the set of values of βj,0 that cannot be rejected by an α-level hypothesis test H0 : βj = βj,0 vs. H1 : βj 6= βj,0 • When the sample size is large, we approximate the (1 − α) confidence interval for βj by h

i

βˆj − u1−α/2 · SE(βˆj ), βˆj + u1−α/2 · SE(βˆj ) , (3.3) where uα denotes the α-quantile of the N (0, 1)-distribution 55

3.3. Tests of joint hypotheses Now: • Testing hypotheses on two or more regression coefficients (joint hypotheses) • Example: H0 : β1 = 0 and β2 = 0

vs.

H1 : β1 6= 0 and/or β2 6= 0

(two restrictions)

56

General form of joint hypothesis: • Consider the k + 1 regression coefficients β0, β1, . . . , βk and k + 1 prespecified real numbers β0,0, β1,0, . . . , βk,0 • For q out of the k + 1 coefficients we consider joint null and alternative hypotheses of the form H0 :

βj = βj,0, βm = βm,0, . . . ,

H1 :

one or more of the q restrictions under H0 does or do not hold

for a total of q restrictions,

57

Remarks: • A special case is given by considering the k restrictions with β1,0 = 0, β2,0 = 0, . . . , βk,0 = 0, that is H0 :

β1 = 0, . . . , βk = 0

H1 :

at least one of the βm is nonzero for m = 1, . . . , k

(overall-significance test of the regression model) • It is tempting to conduct the test by using the usual tstatistics to test the k restrictions one at a time: Test #1: H0 : β1 = 0 vs. H1 : β1 6= 0 Test #2: H0 : β2 = 0 vs. H1 : β2 6= 0 ...

...

...

6 0 Test #k: H0 : βk = 0 vs. H1 : βk = 58

Remarks: [continued] • This approach is unreliable: ... testing a series of single hypotheses is not equivalent to testing those same hypotheses jointly. The intuitive reason for this is that in a joint test of several hypotheses any single hypothesis is ’affected’ by the information in the other hypotheses. (Gujarati and Porter, 2009, p. 238) Solution: • Joint-hypotheses testing on the basis of the F -statistic • EViews-example: overall-significance test for the house-prices dataset 59

F -test (overall-significance test) for the house-prices dataset Dependent Variable: SALEPRICE Method: Least Squares Date: 28/02/12 Time: 09:43 Sample: 1 546 Included observations: 546 White heteroskedasticity-consistent standard errors & covariance Variable

Coefficient

Std. Error

t-Statistic

Prob.

C LOTSIZE BEDROOMS BATHROOMS STOREYS

-4009.550 5.429174 2824.614 17105.17 7634.897

3668.048 0.459424 1262.593 2263.281 917.6527

-1.093102 11.81735 2.237153 7.557690 8.320029

0.2743 0.0000 0.0253 0.0000 0.0000

R-squared Adjusted R-squared S.E. of regression Sum squared resid Log likelihood F-statistic Prob(F-statistic)

0.535547 0.532113 18265.23 1.80E+11 -6129.993 155.9529 0.000000

Mean dependent var S.D. dependent var Akaike info criterion Schwarz criterion Hannan-Quinn criter. Durbin-Watson stat

68121.60 26702.67 22.47250 22.51190 22.48790 1.482942

 

60

Form and null-distribution of the F -statistic: • The exact formula of the F -statistic used for testing the (general) problem of q restrictions given on Slide 57 depends on the specific assumptions imposed on the multiple regression model • Also, the exact null-distribution of the F -statistic (F -distribution with exactly specified degree-of-freedom parameters n1 and n2) also depends on the these assumptions • In EViews, the F -statistic and its null-distribution are computed under the classical assumptions of (1) normally distributed and (2) homoskedastic error terms ui • In contrast, Stock and Watson (2011) do not assume normally distributed errors ui and consider a heteroskedasticityrobust F -statistic (Stock and Watson, 2011, pp. 748-749) 61

3.4. Testing single restrictions involving multiple coefficients General F -testing: • Consider again the multiple regression model: Yi = β0 + β1 · X1i + β2 · X2i + . . . + βk · Xki + ui

(3.4)

• We aim at testing hypotheses involving some linear restrictions on the parameters of the k-variable model such as H0 : β2 = β3 vs. H1 : β2 6= β3 H0 : β3 + β4 + β5 = 3 vs. H1 : β3 + β4 + β5 6= 3 H0 : β3 = β4 = β5 = 0 vs. 6 0 H 1 : β3 = 6 0 and/or β4 6= 0 and/or β5 = (the regressors X3, X4, X5 are absent from the model) 62

General F -testing: [continued] • All these hypotheses can be tested using a general F -statistic • This general testing strategy distinguishes sharply between the so-called unrestricted regression model (3.4) and the restricted regression obtained from plugging the restriction specified under H0 into the unrestricted regression (3.4) • The general F -statistic then compares the sum of squared residuals obtained from the unrestricted regression (3.4), denoted by SSRUR, with the sum of squared residuals obtained from the restricted regression, denoted by SSRR (see Gujarati and Porter, 2009, pp. 249-254)

63

General F -testing: [continued] • As in Section 3.3., the exact null-distribution of this general F -statistic (F -distribution with exactly specified degree-offreedom parameters) again depends on the assumptions imposed on the multiple regression model (in particular on the normality/nonnormality and the homoskedasticity/heteroskedasticity of the error terms ui) • EViews provides a fully-fledged framework for performing these general F -tests under the classical assumptions of normally distributed and homoskedastic error terms −→ A thorough discussion will be given in the class 64

3.5. Confidence sets for multiple coefficients Definition 3.3: (Confidence set) A 95% confidence set for two or more coefficients is the set of numbers that contains the true population values of these coefficients in 95% of randomly drawn samples. Remarks: • A confidence set is the generalization to two or more coefficients of a confidence interval for a single coefficient • Recall Formula (3.3) on Slide 55 for constructing a confidence interval for the single coefficient βj 65

Remarks: [continued] • Instead of using Formula (3.3), an equivalent way of constructing a say 95% confidence interval for the single coefficient βj consists in determining the set of all values βj,0 that cannot be rejected by a two-sided hypothesis test H0 : βj = βj,0

vs.

H1 : βj = 6 βj,0

at the 5% significance level based on the t-statistic (3.1) on Slide 51 • This approach can be extended to the case of multiple coefficients using the general F -testing approach described in Section 3.3 66

Example: [continued] • Suppose you are interested in constructing a confidence set for the two coefficients βj and βm (for j, m = 0, . . . , k, j = 6 m) • In line with Slide 57, consider testing a joint null hypothesis with the 2 restrictions H0 : βj = βj,0, βm = βm,0 at the 5% level using the appropriate F -statistic • The set of all pairs (βj,0, βm,0) for which you cannot reject H0 at the 5% level constitutes a 95% confidence set for βj and βm

67

Remarks: • In line with the confidence-interval formula (3.3), there are also analytical formulas for constructing confidence sets for multiple coefficients (not to be discussed here) • EViews provides a fully-fledged framework for constructing confidence intervals and confidence sets (see class for details)

Example: • Confidence intervals and two-dimensional confidence sets for the house-prices dataset in EViews 68

Single-coefficient confidence intervals for the house-prices dataset

Coefficient Confidence Intervals Date: 20/03/12 Time: 12:39 Sample: 1 546 Included observations: 546 90% CI

95% CI

99% CI

Variable

Coefficient

Low

High

Low

High

Low

High

C LOTSIZE BEDROOMS BATHROOMS STOREYS

-4009.550 5.429174 2824.614 17105.17 7634.897

-10053.30 4.672192 744.2712 13376.02 6122.903

2034.202 6.186155 4904.956 20834.33 9146.891

-11214.91 4.526700 344.4288 12659.28 5832.298

3195.812 6.331647 5304.799 21551.07 9437.496

-13491.26 4.241586 -439.1222 11254.71 5262.813

5472.162 6.616761 6088.350 22955.64 10006.98

 

69

95% two-dimensional confidence sets for the house-prices dataset 6.5

LOTSIZE

6.0 5.5 5.0

BEDROOMS

4.5

4,000

2,000

0

BATHROOMS

22,000 20,000 18,000 16,000 14,000 12,000

STOREYS

9,000 8,000 7,000

70

6,000

-10,000 -5,000 C

0

4.5

5.0

5.5

LOTSIZE

6.0

6.5

0

2,000

4,000

BEDROOMS

12,000

16,000

20,000

BATHROOMS

 

Remarks: • The vertical and horizontal dotted lines show the corresponding 95% confidence intervals for the single coefficients βj , βm • The orientation of the ellipse indicates the estimated correlation between the OLS estimators βˆj and βˆm • If the OLS estimators βˆj and βˆm were independent, the ellipses would be exact circles

71

Suggest Documents