伍德里奇计量经济学第六版答案Chapter-12
- 格式:pdf
- 大小:575.34 KB
- 文档页数:13
计量经济学各章作业习题(后附答案)《计量经济学》习题集第⼀章绪论⼀、单项选择题1、变量之间的关系可以分为两⼤类,它们是【】A 函数关系和相关关系B 线性相关关系和⾮线性相关关系C 正相关关系和负相关关系D 简单相关关系和复杂相关关系2、相关关系是指【】A 变量间的依存关系B 变量间的因果关系C 变量间的函数关系D 变量间表现出来的随机数学关系3、进⾏相关分析时,假定相关的两个变量【】A 都是随机变量B 都不是随机变量C ⼀个是随机变量,⼀个不是随机变量D 随机或⾮随机都可以4、计量经济研究中的数据主要有两类:⼀类是时间序列数据,另⼀类是【】A 总量数据B 横截⾯数据C平均数据 D 相对数据5、下⾯属于截⾯数据的是【】A 1991-2003年各年某地区20个乡镇的平均⼯业产值B 1991-2003年各年某地区20个乡镇的各镇⼯业产值C 某年某地区20个乡镇⼯业产值的合计数D 某年某地区20个乡镇各镇⼯业产值6、同⼀统计指标按时间顺序记录的数据列称为【】A 横截⾯数据B 时间序列数据C 修匀数据D原始数据7、经济计量分析的基本步骤是【】A 设定理论模型→收集样本资料→估计模型参数→检验模型B 设定模型→估计参数→检验模型→应⽤模型C 个体设计→总体设计→估计模型→应⽤模型D 确定模型导向→确定变量及⽅程式→估计模型→应⽤模型8、计量经济模型的基本应⽤领域有【】A 结构分析、经济预测、政策评价B 弹性分析、乘数分析、政策模拟C 消费需求分析、⽣产技术分析、市场均衡分析D 季度分析、年度分析、中长期分析9、计量经济模型是指【】A 投⼊产出模型B 数学规划模型C 包含随机⽅程的经济数学模型D 模糊数学模型10、回归分析中定义【】A 解释变量和被解释变量都是随机变量B 解释变量为⾮随机变量,被解释变量为随机变量C 解释变量和被解释变量都是⾮随机变量D 解释变量为随机变量,被解释变量为⾮随机变量11、下列选项中,哪⼀项是统计检验基础上的再检验(亦称⼆级检验)准则【】A. 计量经济学准则 B 经济理论准则C 统计准则D 统计准则和经济理论准则12、理论设计的⼯作,不包括下⾯哪个⽅⾯【】A 选择变量B 确定变量之间的数学关系C 收集数据D 拟定模型中待估参数的期望值13、计量经济学模型成功的三要素不包括【】A 理论B 应⽤C 数据D ⽅法14、在经济学的结构分析中,不包括下⾯那⼀项【】C ⽐较静⼒分析D ⽅差分析⼆、多项选择题1、⼀个模型⽤于预测前必须经过的检验有【】A 经济准则检验B 统计准则检验C 计量经济学准则检验D 模型预测检验E 实践检验2、经济计量分析⼯作的四个步骤是【】A 理论研究B 设计模型C 估计参数D 检验模型E 应⽤模型3、对计量经济模型的计量经济学准则检验包括【】A 误差程度检验B 异⽅差检验C 序列相关检验D 超⼀致性检验E 多重共线性检验4、对经济计量模型的参数估计结果进⾏评价时,采⽤的准则有【】A 经济理论准则B 统计准则C 经济计量准则D 模型识别准则E 模型简单准则三、名词解释1、计量经济学2、计量经济学模型3、时间序列数据4、截⾯数据5、弹性1、简述经济计量分析⼯作的程序。
班级:金融学×××班姓名:××学号:×××××××C6.9 NBASAL.RAW points=β0+β1exper+β2exper2+β3age+β4coll+u 解:(ⅰ)按照通常的格式报告结果。
由上图可知:points=35.22+2.364exper−0.077exper2−1.074age−1.286coll6.9870.4050.02350.295 (0.451)n=269,R2=0.1412,R2=0.1282。
(ⅱ)保持大学打球年数和年龄不变,从加盟的第几个年份开始,在NBA打球的经历实际上将降低每场得分?这讲得通吗?由上述估计方程可知,转折点是exper的系数与exper2系数的两倍之比:exper∗= β12β2= 2.364[2×−0.077]=15.35,即从加盟的第15个到第16个年份之间,球员在NBA打球的经历实际上将降低每场得分。
实际上,在模型所用的数据中,269名球员中只有2位的打球年数超过了15年,数据代表性不大,所以这个结果讲不通。
(ⅲ)为什么coll具有负系数,而且统计显著?一般情况下,NBA运动员的球员都会在读完大学之前被选拔出,甚至从高中选出,所以这些球员在大学打球的时间少,但每场得分却很高,所以coll具有负系数。
同时,coll的t统计量为-2.85,所以coll统计显著。
(ⅳ)有必要在方程中增加age的二次项吗?控制exper和coll之后,这对年龄效应意味着什么?增加age的二次项后,原估计模型变成:points=73.59+2.864exper−0.128exper2−3.984age+0.054age2−1.313coll35.930.610.05 2.690.05 (0.45)n=269,R2=0.1451,R2=0.1288。
Chapter 12 Instrumental Variables Regression1) Estimation of the IV regression modelA) requires exact identification.B) allows only one endogenous regressor, which is typically correlated with the error term.C) requires exact identification or overidentification.D) is only possible if the number of instruments is the same as the number of regressors.Answer: C2) Two Stage Least Squares is calculated as follows; in the first stage:A) Y is regressed on the exogenous variables only. The predicted value of Y is then regressed on the instrumental variables.B) the unknown coefficients in the reduced form equation are estimated by OLS, and the predicted values are calculated. In the second stage, Y is regressed on these predicted values and the other exogenous variables.C) the exogenous variables are regressed on the instruments. The predicted value of the exogenous variables is then used in the second stage, together with the instruments, to predict the dependent variable.D) the unknown coefficients in the reduced form equation are estimated by weighted least squares, and the predicted values are calculated. In the second stage, Y is regressed on these predicted values and the other exogenous variables.Answer: B3) The conditions for a valid instruments do not include the following:A) each instrument must be uncorrelated with the error term.B) each one of the instrumental variables must be normally distributed.C) at least one of the instruments must enter the population regression of X on the Z s and the W s.D) perfect multicollinearity between the predicted endogenous variables and the exogenous variables must be ruled out.Answer: B4) The IV regression assumptions include all of the following with the exception ofA) the error terms must be normally distributed.B) E(ui W1i,…, W ri) = 0.C) Large outliers are unlikely: the X's, W's, Z's, and Y's all have nonzero, finite fourth moments.D) (X1i,…, X ki, W1i,…,W ri, Z1i, … Z mi, Y i) are i.i.d. draws from their joint distribution.Answer: A5) The rule-of-thumb for checking for weak instruments is as follows: for the case of a single endogenous regressor,A) a first stage F must be statistically significant to indicate a strong instrument.B) a first stage F> 1.96 indicates that the instruments are weak.C) the t-statistic on each of the instruments must exceed at least 1.64.D) a first stage F< 10 indicates that the instruments are weak.Answer: D6) The J-statisticA) tells you if the instruments are exogenous.B) provides you with a test of the hypothesis that the instruments are exogenous for the case of exact identification.C) is distributed where m-k is the degree of overidentification.D) is distributed where m-k is the number of instruments minus the number of regressors. Answer: C7) In the case of the simple regression model Y i=β0+β1X i+u i, i= 1, …, n, when X and u are correlated, thenA) the OLS estimator is biased in small samples only.B) OLS and TSLS produce the same estimate.C) X is exogenous.D) the OLS estimator is inconsistent.Answer: D8) The following will not cause correlation between X and u in the simple regression model:A) simultaneous causality.B) omitted variables.C) irrelevance of the regressor.D) errors in variables.Answer: C9) The distinction between endogenous and exogenous variables isA) that exogenous variables are determined inside the model and endogenous variables are determined outside the model.B) dependent on the sample size: for n> 100, endogenous variables become exogenous.C) depends on the distribution of the variables: when they are normally distributed, they are exogenous, otherwise they are endogenous.D) whether or not the variables are correlated with the error term.Answer: D10) The two conditions for a valid instrument areA) corr(Z i , X i ) = 0 and corr(Z i , u i ) ≠ 0.B) corr(Z i , X i ) = 0 and corr(Z i , u i ) = 0.C) corr(Z i , X i ) ≠ 0 and corr(Z i , u i ) = 0.D) corr(Z i , X i ) ≠ 0 and corr(Z i , u i ) ≠ 0.Answer: C11) Instrument relevanceA) means that the instrument is one of the determinants of the dependent variable.B) is the same as instrument exogeneity.C) means that some of the variance in the regressor is related to variation in the instrument.D) is not possible since X and u are correlated and Z and u are not correlated.Answer: C12) Consider a competitive market where the demand and the supply depend on the current price of the good. Then fitting a line through the quantity -price outcomes willA) give you an estimate of the demand curve.B) estimate neither a demand curve nor a supply curve.C) enable you to calculate the price elasticity of supply.D) give you the exogenous part of the demand in the first stage of TSLS.Answer: B13) When there is a single instrument and single regressor, the TSLS estimator for the slope can be calculated as follows: A) 1ˆTSLS ZY ZXS S β=. B) 12ˆTSLS XY X S S β=. C) 1ˆTSLS ZX ZY S S β=. D) 12ˆTSLS ZY ZS S β=. Answer: A14) The TSLS estimator isA) consistent and has a normal distribution in large samples.B) unbiased.C) efficient in small samples.D) F -distributed.Answer: A15) The reduced form equation for XA) regresses the endogenous variable X on the smallest possible subset of regressors.B) relates the endogenous variable X to all the available exogenous variables, both those included in the regression of interest and the instruments.C) uses the predicted values of X from the first stage as a regressor in the original equation.D) uses smaller standard errors, such as homoskedasticity-only standard errors, for inference. Answer: B16) When calculating the TSLS standard errorsA) you do not have to worry about heteroskedasticity, since it was eliminated in the first stageB) you can use the standard errors reported by OLS estimation of the second stage regression.C) the critical values from the standard normal table should be adjusted for the proper degrees of freedom.D) you should use heteroskedasticity-robust standard errors.Answer: D17) Having more relevant instrumentsA) is a problem because instead of being just identified, the regression now becomes overidentified.B) is like having a larger sample size in that the more information is available for use in the IV regressions.C) typically results in larger standard errors for the TSLS estimator.D) is not as important for inference as having the same number of endogenous variables as instruments. Answer: B18) Weak instruments are a problem becauseA) the TSLS estimator may not be normally distributed, even in large samples.B) they result in the instruments not being exogenous.C) the TSLS estimator cannot be computed.D) you cannot predict the endogenous variables any longer in the first stage.Answer: A19) (Requires Appendix material) The relationship between the TSLS slope and the corresponding population parameter is:A) ()()()()11111ˆ1n i i TSLS i n i i i Z Z u n Z Z X X n ββ==−−=−−∑∑. B) ()()()()11111ˆ1ni TSLS i n i i i Z Z n Z Z X X n ββ==−−=−−∑∑. C) ()()()111211ˆ1ni i TSLS i n i i Z Z u n Z Z n ββ==−−=−∑∑. D) ()()()()11111ˆ1ni i TSLS i n i i i X X u n Z Z X X n ββ==−−=−−∑∑. Answer: A20) If the instruments are not exogenous,A) you cannot perform the first stage of TSLS.B) then, in order to conduct proper inference, it is essential that you use heteroskedasticity -robust standard errors.C) your model becomes overidentified.D) then TSLS is inconsistent.Answer: D21) In the case of exact identificationA) you can use the J -statistic in a test of overidentifying restrictions.B) you cannot use TSLS for estimation purposes.C) you must rely on your personal knowledge of the empirical problem at hand to assess whether the instruments are exogenous.D) OLS and TSLS yield the same estimate.Answer: C22) To calculate the J-statistic you regress theA) squared values of the TSLS residuals on all exogenous variables and the instruments. The statistic is then the number of observations times the regression R2.B) TSLS residuals on all exogenous variables and the instruments. You then multiply the homoskedasticity-only F-statistic from that regression by the number of instruments.C) OLS residuals from the reduced form on the instruments. The F-statistic from this regression is the J-statistic.D) TSLS residuals on all exogenous variables and the instruments. You then multiply the heteroskedasticity-robust F-statistic from that regression by the number of instruments.Answer: B23) (Requires Chapter 8) When using panel data and in the presence of endogenous regressorsA) the TSLS does not exist.B) you do not have to worry about the validity of instruments, since there are so many fixed effects.C) the OLS estimator is consistent.D) application of the TSLS estimator is straightforward if you use two time periods and difference the data.Answer: D24) In practice, the most difficult aspect of IV estimation isA) finding instruments that are both relevant and exogenous.B) that you have to use two stages in the estimation process.C) calculating the J-statistic.D) finding instruments that are exogenous. Relevant instruments are easy to find.Answer: A25) Consider a model with one endogenous regressor and two instruments. Then the J-statistic will be largeA) if the number of observations are very large.B) if the coefficients are very different when estimating the coefficients using one instrument at a time.C) if the TSLS estimates are very different from the OLS estimates.D) when you use homoskedasticity-only standard errors.Answer: B26) Let W be the included exogenous variables in a regression function that also has endogenous regressors (X). The W variables canA) be control variablesB) have the property E(u i|W i) = 0C) make an instrument uncorrelated with uD) all of the aboveAnswer: D27) The logic of control variables in IV regressionsA) parallels the logic of control variables in OLSB) only applies in the case of homoskedastic errors in the first stage of two stage least squares estimationC) is different in a substantial way from the logic of control variables in OLS since there are two stages in estimationD) implies that the TSLS is efficientAnswer: A28) For W to be an effective control variable in IV estimation, the following condition must holdA) E(u i) = 0B) E(u i|Z i,W i) = E(u i|W i)C) E(u i u j) ≠ 0D) there must be an intercept in the regressionAnswer: B29) The IV estimator can be used to potentially eliminate bias resulting fromA) multicollinearity.B) serial correlation.C) errors in variables.D) heteroskedasticity.Answer: C30) Instrumental Variables regression uses instruments toA) establish the Mozart Effect.B) increase the regression R2.C) eliminate serial correlation.D) isolate movements in X that are uncorrelated with u.Answer: D31) Endogenous variablesA) are correlated with the error term.B) always appear on the LHS of regression functions.C) cannot be regressors.D) are uncorrelated with the error term.Answer: A32) Consider the following two equations to describe labor markets in various sectors of the economyN d=β0+β1+uN s=γ0+γ1+vN d=N s=NA) W/P is exogenous, n is endogenousB) Both n and W/P are endogenousC) n is exogenous, W/P is endogenousD) the parameters cannot be estimated because it would require two equations to be estimated at the same time (simultaneously)Answer: B。
CHAPTER 11TEACHING NOTESMuch of the material in this chapter is usually postponed, or not covered at all, in an introductory course. However, as Chapter 10 indicates, the set of time series applications that satisfy all of the classical linear model assumptions might be very small. In my experience, spurious time series regressions are the hallmark of many student projects that use time series data. Therefore, students need to be alerted to the dangers of using highly persistent processes in time series regression equations. (Spurious regression problem and the notion of cointegration are covered in detail in Chapter 18.)It is fairly easy to heuristically describe the difference between a weakly dependent process and an integrated process. Using the MA(1) and the stable AR(1) examples is usually sufficient.When the data are weakly dependent and the explanatory variables are contemporaneously exogenous, OLS is consistent. This result has many applications, including the stable AR(1) regression model. When we add the appropriate homoskedasticity and no serial correlation assumptions, the usual test statistics are asymptotically valid.The random walk process is a good example of a unit root (highly persistent) process. In a one-semester course, the issue comes down to whether or not to first difference the data before specifying the linear model. While unit root tests are covered in Chapter 18, just computing the first-order autocorrelation is often sufficient, perhaps after detrending. The examples in Section 11.3 illustrate how different first-difference results can be from estimating equations in levels. Section 11.4 is novel in an introductory text, and simply points out that, if a model is dynamically complete in a well-defined sense, it should not have serial correlation. Therefore, we need not worry about serial correlation when, say, we test the efficient market hypothesis. Section 11.5 further investigates the homoskedasticity assumption, and, in a time series context, emphasizes that what is contained in the explanatory variables determines what kind of hetero-skedasticity is ruled out by the usual OLS inference. These two sections could be skipped without loss of continuity.129130SOLUTIONS TO PROBLEMS11.1 Because of covariance stationarity, 0γ = Var(x t ) does not depend on t , so sd(x t+hany h ≥ 0. By definition, Corr(x t ,x t +h ) = Cov(x t ,x t+h )/[sd(x t )⋅sd(x t+h)] = 0/.h h γγγ=11.2 (i) E(x t ) = E(e t ) – (1/2)E(e t -1) + (1/2)E(e t -2) = 0 for t = 1,2, Also, because the e t are independent, they are uncorrelated and so Var(x t ) = Var(e t ) + (1/4)Var(e t -1) + (1/4)Var(e t -2) = 1 + (1/4) + (1/4) = 3/2 because Var (e t ) = 1 for all t .(ii) Because x t has zero mean, Cov(x t ,x t +1) = E(x t x t +1) = E[(e t – (1/2)e t -1 + (1/2)e t -2)(e t +1 – (1/2)e t + (1/2)e t -1)] = E(e t e t+1) – (1/2)E(2t e ) + (1/2)E(e t e t -1) – (1/2)E(e t -1e t +1) + (1/4(E(e t-1e t ) – (1/4)E(21t e -) + (1/2)E(e t-2e t+1) – (1/4)E(e t-2e t ) +(1/4)E(e t-2e t-1) = – (1/2)E(2t e ) – (1/4)E(21t e -) = –(1/2) – (1/4) = –3/4; the third to last equality follows because the e t are pairwise uncorrelated and E(2t e ) = 1 for all t . Using Problem 11.1 and the variance calculation from part (i),Corr(x t x t+1) = – (3/4)/(3/2) = –1/2.Computing Cov(x t ,x t+2) is even easier because only one of the nine terms has expectation different from zero: (1/2)E(2t e ) = ½. Therefore, Corr(x t ,x t+2) = (1/2)/(3/2) = 1/3.(iii) Corr(x t ,x t+h ) = 0 for h >2 because, for h > 2, x t+h depends at most on e t+j for j > 0, while x t depends on e t+j , j ≤ 0.(iv) Yes, because terms more than two periods apart are actually uncorrelated, and so it is obvious that Corr(x t ,x t+h ) → 0 as h → ∞.11.3 (i) E(y t ) = E(z + e t ) = E(z ) + E(e t ) = 0. Var(y t ) = Var(z + e t ) = Var(z ) + Var(e t ) + 2Cov(z ,e t ) = 2z σ + 2e σ + 2⋅0 = 2z σ + 2e σ. Neither of these depends on t .(ii) We assume h > 0; when h = 0 we obtain Var(y t ). Then Cov(y t ,y t+h ) = E(y t y t+h ) = E[(z + e t )(z + e t+h )] = E(z 2) + E(ze t+h ) + E(e t z ) + E(e t e t+h ) = E(z 2) = 2z σ because {e t } is an uncorrelated sequence (it is an independent sequence and z is uncorrelated with e t for all t . From part (i) we know that E(y t ) and Var(y t ) do not depend on t and we have shown that Cov(y t ,y t+h ) depends on neither t nor h . Therefore, {y t } is covariance stationary.(iii) From Problem 11.1 and parts (i) and (ii), Corr(y t ,y t+h ) = Cov(y t ,y t+h )/Var(y t ) = 2z σ/(2z σ + 2e σ) > 0.(iv) No. The correlation between y t and y t+h is the same positive value obtained in part (iii) now matter how large is h . In other words, no matter how far apart y t and y t+h are, theircorrelation is always the same. Of course, the persistent correlation across time is due to the presence of the time-constant variable, z .13111.4 Assuming y 0 = 0 is a special case of assuming y 0 nonrandom, and so we can obtain the variances from (11.21): Var(y t ) = 2e σt and Var(y t+h ) = 2e σ(t + h ), h > 0. Because E(y t ) = 0 for all t (since E(y 0) = 0), Cov(y t ,y t+h ) = E(y t y t+h ) and, for h > 0,E(y t y t+h ) = E[(e t + e t-1 + e 1)(e t+h + e t+h -1 + + e 1)]= E(2t e ) + E(21t e -) + + E(21e ) = 2e σt ,where we have used the fact that {e t } is a pairwise uncorrelated sequence. Therefore, Corr(y t ,y t+h ) = Cov(y t ,y t+ht11.5 (i) The following graph gives the estimated lag distribution:wage inflation has its largest effect on price inflation nine months later. The smallest effect is at the twelfth lag, which hopefully indicates (but does not guarantee) that we have accounted for enough lags of gwage in the FLD model.(ii) Lags two, three, and twelve have t statistics less than two. The other lags are statistically significant at the 5% level against a two-sided alternative. (Assuming either that the CLMassumptions hold for exact tests or Assumptions TS.1' through TS.5' hold for asymptotic tests.)132(iii) The estimated LRP is just the sum of the lag coefficients from zero through twelve:1.172. While this is greater than one, it is not much greater, and the difference from unity could be due to sampling error.(iv) The model underlying and the estimated equation can be written with intercept α0 and lag coefficients δ0, δ1, , δ12. Denote the LRP by θ0 = δ0 + δ1 + + δ12. Now, we can write δ0 = θ0 - δ1 - δ2 - - δ12. If we plug this into the FDL model we obtain (with y t = gprice t and z t = gwage t )y t = α0 + (θ0 - δ1 - δ2 - - δ12)z t + δ1z t -1 + δ2z t -2 + + δ12z t -12 + u t= α0 + θ0z t + δ1(z t -1 – z t ) + δ2(z t -2 – z t ) + + δ12(z t -12 – z t ) + u t .Therefore, we regress y t on z t , (z t -1 – z t ), (z t -2 – z t ), , (z t -12 – z t ) and obtain the coefficient and standard error on z t as the estimated LRP and its standard error.(v) We would add lags 13 through 18 of gwage t to the equation, which leaves 273 – 6 = 267 observations. Now, we are estimating 20 parameters, so the df in the unrestricted model is df ur =267. Let 2ur R be the R -squared from this regression. To obtain the restricted R -squared, 2r R , weneed to reestimate the model reported in the problem but with the same 267 observations used toestimate the unrestricted model. Then F = [(2ur R -2r R )/(1 - 2ur R )](247/6). We would find thecritical value from the F 6,247 distribution.[Instructor’s Note: As a computer exercise, you might have the students test whether all 13 lag coefficients in the population model are equal. The restricted regression is gprice on (gwage + gwage -1 + gwage -2 + gwage -12), and the R -squared form of the F test, with 12 and 259 df , can be used.]11.6 (i) The t statistic for H 0: β1 = 1 is t = (1.104 – 1)/.039 ≈ 2.67. Although we must rely on asymptotic results, we might as well use df = 120 in Table G.2. So the 1% critical value against a two-sided alternative is about 2.62, and so we reject H 0: β1 = 1 against H 1: β1 ≠ 1 at the 1% level. It is hard to know whether the estimate is practically different from one withoutcomparing investment strategies based on the theory (β1 = 1) and the estimate (1ˆβ= 1.104). But the estimate is 10% higher than the theoretical value.(ii) The t statistic for the null in part (i) is now (1.053 – 1)/.039 ≈ 1.36, so H 0: β1 = 1 is no longer rejected against a two-sided alternative unless we are using more than a 10% significance level. But the lagged spread is very significant (contrary to what the expectations hypothesis predicts): t = .480/.109≈ 4.40. Based on the estimated equation, when the lagged spread is positive, the predicted holding yield on six-month T-bills is above the yield on three-month T-bills (even if we impose β1 = 1), and so we should invest in six-month T-bills.(iii) This suggests unit root behavior for {hy3t }, which generally invalidates the usual t -testing procedure.133(iv) We would include three quarterly dummy variables, say Q2t , Q3t , and Q4t , and do an F test for joint significance of these variables. (The F distribution would have 3 and 117 df .)11.7 (i) We plug the first equation into the second to gety t – y t-1 = λ(0γ + 1γx t + e t – y t-1) + a t ,and, rearranging,y t = λ0γ + (1 - λ)y t-1 + λ1γx t + a t + λe t ,≡ β0 + β1y t -1 + β2 x t + u t ,where β0 ≡ λ0γ, β1 ≡ (1 - λ), β2 ≡ λ1γ, and u t ≡ a t + λe t .(ii) An OLS regression of y t on y t-1 and x t produces consistent, asymptotically normal estimators of the βj . Under E(e t |x t ,y t-1,x t-1, ) = E(a t |x t ,y t-1,x t-1, ) = 0 it follows that E(u t |x t ,y t-1,x t-1, ) = 0, which means that the model is dynamically complete [see equation (11.37)]. Therefore, the errors are serially uncorrelated. If the homoskedasticity assumption Var(u t |x t ,y t-1) = σ2 holds, then the usual standard errors, t statistics and F statistics are asymptotically valid.(iii) Because β1 = (1 - λ), if 1ˆβ= .7 then ˆλ= .3. Further, 2ˆβ= 1ˆˆλγ, or 1ˆγ= 2ˆβ/ˆλ= .2/.3 ≈ .67.11.8 (i) Sequential exogeneity does not rule out correlation between, say, 1t u - and tj x for anyregressors j = 1, 2, …, k . The differencing generally induces correlation between the differenced errors and the differenced regressors. To see why, consider a single explanatory variable, x t . Then 1t t t u u u -∆=- and 1t t t x x x -∆=-. Under sequential exogeneity, t u is uncorrelated with t x and 1t x -, and 1t u - is uncorrelated with 1t x -. But 1t u - can be correlated with t x , which means that t u ∆ and t x ∆ are generally correlated. In fact, under sequential exogeneity, it is always true that 1Cov(,)Cov(,)t t t t x u x u -∆∆=-.(ii) Strict exogeneity of the regressors in the original equation is sufficient for OLS on the first-differenced equation to be consistent. Remember, strict exogeneity implies that theregressors in any time period are uncorrelated with the errors in any time period. Of course, we could make the weaker assumption: for any t , u t is uncorrelated with 1,t j x -, t j x , and 1,t j x + for all j = 1, 2, …, k . The strengthening beyond sequential exogeneity is the assumption that u t is uncorrelated w ith all of next period’s outcomes on all regressors. In practice, this is probably similar to just assuming strict exogeneity.(iii) If we assume sequential exogeneity in a static model, the condition can be written as。
伍德里奇计量经济学导论第六版英文课件Woodridge's Introduction to Econometrics, 6th Edition, is a comprehensive textbook that covers the fundamentals of econometrics in a clear and concise manner. The accompanying courseware is designed to help students further understand and apply the concepts discussed in the book.The courseware includes PowerPoint slides, practice quizzes, and interactive exercises to enhance students' learning experience. The slides cover the key topics in each chapter, providing visual aids to help students grasp complex concepts. The quizzes allow students to test their understanding of the material and receive immediate feedback on their performance. The interactive exercises provide hands-on practice withreal-world data sets, helping students develop their econometric skills.In addition to the courseware, students have access to online resources such as supplementary readings, video tutorials, and self-assessment tools. These resources are designed to support students in their learning journey and provide additional assistance when needed.Overall, Woodridge's Introduction to Econometrics, 6th Edition, is a valuable resource for students studying econometrics. The comprehensive courseware offers a range of tools to support students in their learning, making it easier for them to understand and apply the concepts discussed in the textbook. With its clear explanations and practical exercises, this courseware is an essential companion for students looking to excel in econometrics.。
CHAPTER 1TEACHING NOTESYou have substantial latitude about what to emphasize in Chapter 1. I find it useful to talk about the economics of crime example (Example 1.1) and the wage example (Example 1.2) so that students see, at the outset, that econometrics is linked to economic reasoning, even if the economics is not complicated theory.I like to familiarize students with the important data structures that empirical economists use, focusing primarily on cross-sectional and time series data sets, as these are what I cover in a first-semester course. It is probably a good idea to mention the growing importance of data sets that have both a cross-sectional and time dimension.I spend almost an entire lecture talking about the problems inherent in drawing causal inferences in the social sciences. I do this mostly through the agricultural yield, return to education, and crime examples. These examples also contrast experimental and nonexperimental (observational) data. Students studying business and finance tend to find the term structure of interest rates example more relevant, although the issue there is testing the implication of a simple theory, as opposed to inferring causality. I have found that spending time talking about these examples, in place of a formal review of probability and statistics, is more successful (and more enjoyable for the students and me).CHAPTER 2TEACHING NOTESThis is the chapter where I expect students to follow most, if not all, of the algebraic derivations. In class I like to derive at least the unbiasedness of the OLS slope coefficient, and usually I derive the variance. At a minimum, I talk about the factors affecting the variance. To simplify the notation, after I emphasize the assumptions in the population model, and assume random sampling, I just condition on the values of the explanatory variables in the sample. Technically, this is justified by random sampling because, for example, E(u i|x1,x2,…,x n) = E(u i|x i) by independent sampling. I find that students are able to focus on the key assumption SLR.4 and subsequently take my word about how conditioning on the independent variables in the sample is harmless. (If you prefer, the appendix to Chapter 3 does the conditioning argument carefully.) Because statistical inference is no more difficult in multiple regression than in simple regression, I postpone inference until Chapter 4. (This reduces redundancy and allows you to focus on the interpretive differences between simple and multiple regression.)You might notice how, compared with most other texts, I use relatively few assumptions to derive the unbiasedness of the OLS slope estimator, followed by the formula for its variance. This is because I do not introduce redundant or unnecessary assumptions. For example, once SLR.4 is assumed, nothing further about the relationship between u and x is needed to obtain the unbiasedness of OLS under random sampling.CHAPTER 3TEACHING NOTESFor undergraduates, I do not work through most of the derivations in this chapter, at least not in detail. Rather, I focus on interpreting the assumptions, which mostly concern the population. Other than random sampling, the only assumption that involves more than population considerations is the assumption about no perfect collinearity, where the possibility of perfect collinearity in the sample (even if it does not occur in the population) should be touched on. The more important issue is perfect collinearity in the population, but this is fairly easy to dispense with via examples. These come from my experiences with the kinds of model specification issues that beginners have trouble with.The comparison of simple and multiple regression estimates – based on the particular sample at hand, as opposed to their statistical properties – usually makes a strong impression. Sometimes I do not bother with the “partialling out” interpretation of multiple regression.As far as statistical properties, notice how I treat the problem of including an irrelevant variable: no separate derivation is needed, as the result follows form Theorem 3.1.I do like to derive the omitted variable bias in the simple case. This is not much more difficult than showing unbiasedness of OLS in the simple regression case under the first four Gauss-Markov assumptions. It is important to get the students thinking about this problem early on, and before too many additional (unnecessary) assumptions have been introduced.I have intentionally kept the discussion of multicollinearity to a minimum. This partly indicates my bias, but it also reflects reality. It is, of course, very important for students to understand the potential consequences of having highly correlated independent variables. But this is often beyond our control, except that we can ask less of our multiple regression analysis. If two or more explanatory variables are highly correlated in the sample, we should not expect to precisely estimate their ceteris paribus effects in the population.I find extensive treatments of multicollinearity, where one “tests” or somehow “solves” the multicollinearity problem, to be misleading, at best. Even the organization of some texts gives the impression that imperfect multicollinearity is somehow a violation of the Gauss-Markov assumptions: they include multicollinearity in a chapter or part of the book devoted to “violation of the basic assumptions,” or something like that. I have noticed that master’s students who have had some undergraduate econometrics are often confused on the multicollinearity issue. It is very important that students not confuse multicollinearity among the included explanatory variables in a regression model with the bias caused by omitting an important variable.I do not prove the Gauss-Markov theorem. Instead, I emphasize its implications. Sometimes, and certainly for advanced beginners, I put a special case of Problem 3.12 on a midterm exam, where I make a particular choice for the function g(x). Rather than have the students directly compare the variances, they shouldappeal to the Gauss-Markov theorem for the superiority of OLS over any other linear, unbiased estimator.CHAPTER 4TEACHING NOTESAt the start of this chapter is good time to remind students that a specific error distribution played no role in the results of Chapter 3. That is because only the first two moments were derived under the full set of Gauss-Markov assumptions. Nevertheless, normality is needed to obtain exact normal sampling distributions (conditional on the explanatory variables). I emphasize that the full set of CLM assumptions are used in this chapter, but that in Chapter 5 we relax the normality assumption and still perform approximately valid inference. One could argue that the classical linear model results could be skipped entirely, and that only large-sample analysis is needed. But, from a practical perspective, students still need to know where the t distribution comes from because virtually all regression packages report t statistics and obtain p -values off of the t distribution. I then find it very easy tocover Chapter 5 quickly, by just saying we can drop normality and still use t statistics and the associated p -values as being approximately valid. Besides, occasionally students will have to analyze smaller data sets, especially if they do their own small surveys for a term project.It is crucial to emphasize that we test hypotheses about unknown population parameters. I tell my students that they will be punished if they write something likeH 0:1ˆ = 0 on an exam or, even worse, H 0: .632 = 0. One useful feature of Chapter 4 is its illustration of how to rewrite a population model so that it contains the parameter of interest in testing a single restriction. I find this is easier, both theoretically and practically, than computing variances that can, in some cases, depend on numerous covariance terms. The example of testing equality of the return to two- and four-year colleges illustrates the basic method, and shows that the respecified model can have a useful interpretation. Of course, some statistical packages now provide a standard error for linear combinations of estimates with a simple command, and that should be taught, too.One can use an F test for single linear restrictions on multiple parameters, but this is less transparent than a t test and does not immediately produce the standard error needed for a confidence interval or for testing a one-sided alternative. The trick of rewriting the population model is useful in several instances, including obtaining confidence intervals for predictions in Chapter 6, as well as for obtaining confidence intervals for marginal effects in models with interactions (also in Chapter6).The major league baseball player salary example illustrates the differencebetween individual and joint significance when explanatory variables (rbisyr and hrunsyr in this case) are highly correlated. I tend to emphasize the R -squared form of the F statistic because, in practice, it is applicable a large percentage of the time, and it is much more readily computed. I do regret that this example is biased toward students in countries where baseball is played. Still, it is one of the better examplesof multicollinearity that I have come across, and students of all backgrounds seem to get the point.CHAPTER 5TEACHING NOTESChapter 5 is short, but it is conceptually more difficult than the earlier chapters, primarily because it requires some knowledge of asymptotic properties of estimators. In class, I give a brief, heuristic description of consistency and asymptotic normality before stating the consistency and asymptotic normality of OLS. (Conveniently, the same assumptions that work for finite sample analysis work for asymptotic analysis.) More advanced students can follow the proof of consistency of the slope coefficient in the bivariate regression case. Section E.4 contains a full matrix treatment of asymptotic analysis appropriate for a master’s level course.An explicit illustration of what happens to standard errors as the sample size grows emphasizes the importance of having a larger sample. I do not usually cover the LM statistic in a first-semester course, and I only briefly mention the asymptotic efficiency result. Without full use of matrix algebra combined with limit theorems for vectors and matrices, it is very difficult to prove asymptotic efficiency of OLS.I think the conclusions of this chapter are important for students to know, even though they may not fully grasp the details. On exams I usually include true-false type questions, with explanation, to test the students’ understanding of asymptotics. [For example: “In large samples we do not have to worry about omitted variable bias.” (False). Or “Even if the error term is not normally distributed, in large samples we can still compute approximately valid confidence intervals under the Gauss-Markov assumptions.” (True).]CHAPTER6TEACHING NOTESI cover most of Chapter 6, but not all of the material in great detail. I use the example in Table 6.1 to quickly run through the effects of data scaling on the important OLS statistics. (Students should already have a feel for the effects of data scaling on the coefficients, fitting values, and R-squared because it is covered in Chapter 2.) At most, I briefly mention beta coefficients; if students have a need for them, they can read this subsection.The functional form material is important, and I spend some time on more complicated models involving logarithms, quadratics, and interactions. An important point for models with quadratics, and especially interactions, is that we need to evaluate the partial effect at interesting values of the explanatory variables. Often, zero is not an interesting value for an explanatory variable and is well outside the range in the sample. Using the methods from Chapter 4, it is easy to obtain confidence intervals for the effects at interesting x values.As far as goodness-of-fit, I only introduce the adjusted R-squared, as I think using a slew of goodness-of-fit measures to choose a model can be confusing to novices (and does not reflect empirical practice). It is important to discuss how, if we fixate on a high R-squared, we may wind up with a model that has no interesting ceteris paribus interpretation.I often have students and colleagues ask if there is a simple way to predict y when log(y) has been used as the dependent variable, and to obtain a goodness-of-fit measure for the log(y) model that can be compared with the usual R-squared obtained when y is the dependent variable. The methods described in Section 6.4 are easy to implement and, unlike other approaches, do not require normality.The section on prediction and residual analysis contains several important topics, including constructing prediction intervals. It is useful to see how much wider the prediction intervals are than the confidence interval for the conditional mean. I usually discuss some of the residual-analysis examples, as they have real-world applicability.CHAPTER 7TEACHING NOTESThis is a fairly standard chapter on using qualitative information in regression analysis, although I try to emphasize examples with policy relevance (and only cross-sectional applications are included.).In allowing for different slopes, it is important, as in Chapter 6, to appropriately interpret the parameters and to decide whether they are of direct interest. For example, in the wage equation where the return to education is allowed to depend on gender, the coefficient on the female dummy variable is the wage differential between women and men at zero years of education. It is not surprising that we cannot estimate this very well, nor should we want to. In this particular example we would drop the interaction term because it is insignificant, but the issue of interpreting the parameters can arise in models where the interaction term is significant.In discussing the Chow test, I think it is important to discuss testing for differences in slope coefficients after allowing for an intercept difference. In many applications, a significant Chow statistic simply indicates intercept differences. (See the example in Section 7.4 on student-athlete GPAs in the text.) From a practical perspective, it is important to know whether the partial effects differ across groups or whether a constant differential is sufficient.I admit that an unconventional feature of this chapter is its introduction of the linear probability model. I cover the LPM here for several reasons. First, the LPM is being used more and more because it is easier to interpret than probit or logit models. Plus, once the proper parameter scalings are done for probit and logit, the estimated effects are often similar to the LPM partial effects near the mean or median values of the explanatory variables. The theoretical drawbacks of the LPM are often of secondary importance in practice. Computer Exercise C7.9 is a good one to illustrate that, even with over 9,000 observations, the LPM can deliver fitted values strictly between zero and one for all observations.If the LPM is not covered, many students will never know about using econometrics to explain qualitative outcomes. This would be especially unfortunate for students who might need to read an article where an LPM is used, or who might want to estimate an LPM for a term paper or senior thesis. Once they are introduced to purpose and interpretation of the LPM, along with its shortcomings, they can tackle nonlinear models on their own or in a subsequent course.A useful modification of the LPM estimated in equation (7.29) is to drop kidsge6 (because it is not significant) and then define two dummy variables, one for kidslt6 equal to one and the other for kidslt6 at least two. These can be included in place of kidslt6 (with no young children being the base group). This allows a diminishing marginal effect in an LPM. I was a bit surprised when a diminishing effect did not materialize.CHAPTER 8TEACHING NOTESThis is a good place to remind students that homoskedasticity played no role in showing that OLS is unbiased for the parameters in the regression equation. In addition, you probably should mention that there is nothing wrong with the R-squared or adjusted R-squared as goodness-of-fit measures. The key is that these are estimates of the population R-squared, 1 – [Var(u)/Var(y)], where the variances are the unconditional variances in the population. The usual R-squared, and the adjusted version, consistently estimate the population R-squared whether or not Var(u|x) = Var(y|x) depends on x. Of course, heteroskedasticity causes the usual standard errors, t statistics, and F statistics to be invalid, even in large samples, with or without normality.By explicitly stating the homoskedasticity assumption as conditional on the explanatory variables that appear in the conditional mean, it is clear that only heteroskedasticity that depends on the explanatory variables in the model affects the validity of standard errors and test statistics. The version of the Breusch-Pagan test in the text, and the White test, are ideally suited for detecting forms of heteroskedasticity that invalidate inference obtained under homoskedasticity. If heteroskedasticity depends on an exogenous variable that does not also appear in the mean equation, this can be exploited in weighted least squares for efficiency, but only rarely is such a variable available. One case where such a variable is available is when an individual-level equation has been aggregated. I discuss this case in the text but I rarely have time to teach it.As I mention in the text, other traditional tests for heteroskedasticity, such as the Park and Glejser tests, do not directly test what we want, or add too many assumptions under the null. The Goldfeld-Quandt test only works when there is a natural way to order the data based on one independent variable. This is rare in practice, especially for cross-sectional applications.Some argue that weighted least squares estimation is a relic, and is no longer necessary given the availability of heteroskedasticity-robust standard errors and test statistics. While I am sympathetic to this argument, it presumes that we do not care much about efficiency. Even in large samples, the OLS estimates may not be preciseenough to learn much about the population parameters. With substantial heteroskedasticity we might do better with weighted least squares, even if the weighting function is misspecified. As discussed in the text on pages 288-289, one can, and probably should, compute robust standard errors after weighted least squares. For asymptotic efficiency comparisons, these would be directly comparable to the heteroskedasiticity-robust standard errors for OLS.Weighted least squares estimation of the LPM is a nice example of feasible GLS, at least when all fitted values are in the unit interval. Interestingly, in the LPM examples in the text and the LPM computer exercises, the heteroskedasticity-robust standard errors often differ by only small amounts from the usual standard errors. However, in a couple of cases the differences are notable, as in Computer ExerciseC8.7.CHAPTER 9TEACHING NOTESThe coverage of RESET in this chapter recognizes that it is a test for neglected nonlinearities, and it should not be expected to be more than that. (Formally, it can be shown that if an omitted variable has a conditional mean that is linear in the included explanatory variables, RESET has no ability to detect the omitted variable. Interested readers may consult my chapter in Companion to Theoretical Econometrics, 2001, edited by Badi Baltagi.) I just teach students the F statistic version of the test.The Davidson-MacKinnon test can be useful for detecting functional form misspecification, especially when one has in mind a specific alternative, nonnested model. It has the advantage of always being a one degree of freedom test.I think the proxy variable material is important, but the main points can be made with Examples 9.3 and 9.4. The first shows that controlling for IQ can substantially change the estimated return to education, and the omitted ability bias is in the expected direction. Interestingly, education and ability do not appear to have an interactive effect. Example 9.4 is a nice example of how controlling for a previous value of the dependent variable – something that is often possible with survey and nonsurvey data – can greatly affect a policy conclusion. Computer Exercise 9.3 is also a good illustration of this method.I rarely get to teach the measurement error material, although the attenuation bias result for classical errors-in-variables is worth mentioning.The result on exogenous sample selection is easy to discuss, with more details given in Chapter 17. The effects of outliers can be illustrated using the examples. I think the infant mortality example, Example 9.10, is useful for illustrating how a single influential observation can have a large effect on the OLS estimates.With the growing importance of least absolute deviations, it makes sense to at least discuss the merits of LAD, at least in more advanced courses. Computer Exercise 9.9 is a good example to show how mean and median effects can be very different, even though there may not be “outliers” in the usual sense.CHAPTER 10TEACHING NOTESBecause of its realism and its care in stating assumptions, this chapter puts a somewhat heavier burden on the instructor and student than traditional treatments of time series regression. Nevertheless, I think it is worth it. It is important that students learn that there are potential pitfalls inherent in using regression with time series data that are not present for cross-sectional applications. Trends, seasonality, and high persistence are ubiquitous in time series data. By this time, students should have a firm grasp of multiple regression mechanics and inference, and so you can focus on those features that make time series applications different fromcross-sectional ones.I think it is useful to discuss static and finite distributed lag models at the same time, as these at least have a shot at satisfying the Gauss-Markov assumptions.Many interesting examples have distributed lag dynamics. In discussing the time series versions of the CLM assumptions, I rely mostly on intuition. The notion of strict exogeneity is easy to discuss in terms of feedback. It is also pretty apparent that, in many applications, there are likely to be some explanatory variables that are not strictly exogenous. What the student should know is that, to conclude that OLS is unbiased – as opposed to consistent – we need to assume a very strong form of exogeneity of the regressors. Chapter 11 shows that only contemporaneous exogeneity is needed for consistency.Although the text is careful in stating the assumptions, in class, after discussing strict exogeneity, I leave the conditioning on X implicit, especially when I discuss the no serial correlation assumption. As this is a new assumption I spend some time on it. (I also discuss why we did not need it for random sampling.)Once the unbiasedness of OLS, the Gauss-Markov theorem, and the sampling distributions under the classical linear model assumptions have been covered – which can be done rather quickly – I focus on applications. Fortunately, the students already know about logarithms and dummy variables. I treat index numbers in this chapter because they arise in many time series examples.A novel feature of the text is the discussion of how to compute goodness-of-fit measures with a trending or seasonal dependent variable. While detrending or deseasonalizing y is hardly perfect (and does not work with integrated processes), it is better than simply reporting the very high R-squareds that often come with time series regressions with trending variables.CHAPTER 11TEACHING NOTESMuch of the material in this chapter is usually postponed, or not covered at all, in an introductory course. However, as Chapter 10 indicates, the set of time series applications that satisfy all of the classical linear model assumptions might be very small. In my experience, spurious time series regressions are the hallmark of manystudent projects that use time series data. Therefore, students need to be alerted to the dangers of using highly persistent processes in time series regression equations. (Spurious regression problem and the notion of cointegration are covered in detail in Chapter 18.)It is fairly easy to heuristically describe the difference between a weakly dependent process and an integrated process. Using the MA(1) and the stable AR(1) examples is usually sufficient.When the data are weakly dependent and the explanatory variables are contemporaneously exogenous, OLS is consistent. This result has many applications, including the stable AR(1) regression model. When we add the appropriate homoskedasticity and no serial correlation assumptions, the usual test statistics are asymptotically valid.The random walk process is a good example of a unit root (highly persistent) process. In a one-semester course, the issue comes down to whether or not to first difference the data before specifying the linear model. While unit root tests are covered in Chapter 18, just computing the first-order autocorrelation is often sufficient, perhaps after detrending. The examples in Section 11.3 illustrate how different first-difference results can be from estimating equations in levels.Section 11.4 is novel in an introductory text, and simply points out that, if a model is dynamically complete in a well-defined sense, it should not have serial correlation. Therefore, we need not worry about serial correlation when, say, we test the efficient market hypothesis. Section 11.5 further investigates the homoskedasticity assumption, and, in a time series context, emphasizes that what is contained in the explanatory variables determines what kind of heteroskedasticity is ruled out by the usual OLS inference. These two sections could be skipped without loss of continuity.CHAPTER 12TEACHING NOTESMost of this chapter deals with serial correlation, but it also explicitly considers heteroskedasticity in time series regressions. The first section allows a review of what assumptions were needed to obtain both finite sample and asymptotic results. Just as with heteroskedasticity, serial correlation itself does not invalidate R-squared. In fact, if the data are stationary and weakly dependent, R-squared and adjustedR-squared consistently estimate the population R-squared (which is well-defined under stationarity).Equation (12.4) is useful for explaining why the usual OLS standard errors are not generally valid with AR(1) serial correlation. It also provides a good starting point for discussing serial correlation-robust standard errors in Section 12.5. The subsection on serial correlation with lagged dependent variables is included to debunk the myth that OLS is always inconsistent with lagged dependent variables and serial correlation. I do not teach it to undergraduates, but I do to master’s student s.Section 12.2 is somewhat untraditional in that it begins with an asymptotic t test for AR(1) serial correlation (under strict exogeneity of the regressors). It may seem heretical not to give the Durbin-Watson statistic its usual prominence, but I do believe the DW test is less useful than the t test. With nonstrictly exogenous regressors I cover only the regression form of Durbin’s test, as the h statistic is asymptotically equivalent and not always computable.Section 12.3, on GLS and FGLS estimation, is fairly standard, although I try to show how comparing OLS estimates and FGLS estimates is not so straightforward. Unfortunately, at the beginning level (and even beyond), it is difficult to choose a course of action when they are very different.I do not usually cover Section 12.5 in a first-semester course, but, because some econometrics packages routinely compute fully robust standard errors, students can be pointed to Section 12.5 if they need to learn something about what the corrections do.I do cover Section 12.5 for a master’s level course in applied econometrics (after the first-semester course).I also do not cover Section 12.6 in class; again, this is more to serve as a reference for more advanced students, particularly those with interests in finance. One important point is that ARCH is heteroskedasticity and not serial correlation, something that is confusing in many texts. If a model contains no serial correlation, the usual heteroskedasticity-robust statistics are valid. I have a brief subsection on correcting for a known form of heteroskedasticity and AR(1) errors in models with strictly exogenous regressors.CHAPTER 13TEACHING NOTESWhile this chapter falls under “Advanced Topics,” most of this chapter requires no more sophistication than the previous chapters. (In fact, I would argue that, with the possible exception of Section 13.5, this material is easier than some of the time series chapters.)Pooling two or more independent cross sections is a straightforward extension of cross-sectional methods. Nothing new needs to be done in stating assumptions, except possibly mentioning that random sampling in each time period is sufficient. The practically important issue is allowing for different intercepts, and possibly different slopes, across time.The natural experiment material and extensions of the difference-in-differences estimator is widely applicable and, with the aid of the examples, easy to understand.Two years of panel data are often available, in which case differencing across time is a simple way of removing g unobserved heterogeneity. If you have covered Chapter 9, you might compare this with a regression in levels using the second year of data, but where a lagged dependent variable is included. (The second approach only requires collecting information on the dependent variable in a previous year.) These often give similar answers. Two years of panel data, collected before and after a policy change, can be very powerful for policy analysis.。
CHAPTER 5TEACHING NOTESChapter 5 is short, but it is conceptually more difficult than the earlier chapters, primarily because it requires some knowledge of asymptotic properties of estimators. In class, I give a brief, heuristic description of consistency and asymptotic normality before stating the consistency and asymptotic normality of OLS. (Conveniently, the same assumptions that work for finite sample analysis work for asymptotic analysis.) More advanced students can follow the proof of consistency of the slope coefficient in the bivariate regression case. Section E.4 contains a full matrix treatment of asymptotic analysis appropriate for a master’s level course.An explicit illustration of what happens to standard errors as the sample size grows emphasizes the importance of having a larger sample. I do not usually cover the LM statistic in a first-semester course, and I only briefly mention the asymptotic efficiency result. Without full use of matrix algebra combined with limit theorems for vectors and matrices, it is difficult to prove asymptotic efficiency of OLS.I think the conclusions of this chapter are important for students to know, even though they may not fully grasp the details. On exams I usually include true-false type questions, with explanation, to test the students’ understanding of asymptotics. [For example: “In large samples we do not have to worry about omitted variable bias.” (False). Or “Even if the error term is not normally distributed, in large samples we can still compute approximately valid confidence intervals under the Gauss-Markov assumptions.” (True).]SOLUTIONS TO PROBLEMS5.1 Write y = 0β + 1βx 1 + u , and take the expected value: E(y ) = 0β + 1βE(x 1) + E(u ), or µy = 0β + 1βµx since E(u ) = 0, where µy = E(y ) and µx = E(x 1). We can rewrite this as 0β = µy - 1βµx . Now, 0ˆβ = y - 1ˆβ1x . Taking the plim of this we have plim(0ˆβ) = plim(y - 1ˆβ1x ) = plim(y ) – plim(1ˆβ)⋅plim(1x ) = µy - 1βµx , where we use the fact that plim(y ) = µy and plim(1x ) = µx by the law of large numbers, and plim(1ˆβ) = 1β. We have also used the parts of Property PLIM.2 from Appendix C.5.2 A higher tolerance of risk means more willingness to invest in the stock market, so 2β > 0.By assumption, funds and risktol are positively correlated. Now we use equation (5.5), where δ1 > 0: plim(1β) = 1β + 2βδ1 > 1β, so 1β has a positive inconsistency (asymptotic bias). This makes sense: if we omit risktol from the regression and it is positively correlated with funds , some of the estimated effect of funds is actually due to the effect of risktol .5.3 The variable cigs has nothing close to a normal distribution in the population. Most people do not smoke, so cigs = 0 for over half of the population. A normally distributed randomvariable takes on no particular value with positive probability. Further, the distribution of cigs is skewed, whereas a normal random variable must be symmetric about its mean.5.4 Write y = 0β + 1βx + u , and take the expected value: E(y ) = 0β + 1βE(x ) + E(u ), or µy = 0β + 1βµx , since E(u ) = 0, where µy = E(y ) and µx = E(x ). We can rewrite this as 0β = µy - 1βµx . Now, 0β = y - 1βx . Taking the plim of this we have plim(0β) = plim(y - 1βx ) = plim(y ) – plim(1β)⋅plim(x ) = µy - 1βµx , where we use the fact that plim(y ) = µy and plim(x ) = µx by the law of large numbers, and plim(1β) = 1β. We have also used the parts of the Property PLIM.2 from Appendix C.SOLUTIONS TO COMPUTER EXERCISESC5.1 (i) The estimated equation iswage = -2.87 + .599 educ + .022 exper + .169 tenure (0.73) (.051) (.012) (.022)n = 526, R 2 = .306, ˆσ= 3.085.Below is a histogram of the 526 residual, ˆi u, i = 1, 2 , ..., 526. The histogram uses 27 bins, which is suggested by the formula in the Stata manual for 526 observations. For comparison, the normal distribution that provides the best fit to the histogram is also plotted.log()wage = .284 + .092 educ+ .0041 exper+ .022 tenure(.104) (.007) (.0017) (.003)n = 526, R2 = .316, ˆ = .441.The histogram for the residuals from this equation, with the best-fitting normal distributionoverlaid, is given below:(iii) The residuals from the log(wage) regression appear to be more normally distributed.Certainly the histogram in part (ii) fits under its comparable normal density better than in part (i),and the histogram for the wage residuals is notably skewed to the left. In the wage regressionthere are some very large residuals (roughly equal to 15) that lie almost five estimated standard deviations (ˆσ = 3.085) from the mean of the residuals, which is identically zero, of course. Residuals far from zero does not appear to be nearly as much of a problem in the log(wage)regression.C5.2 (i) The regression with all 4,137 observations iscolgpa= 1.392 - .01352 hsperc+ .00148 sat(0.072) (.00055) (.00007)n = 4,137, R2 = .273.(ii) Using only the first 2,070 observations givescolgpa= 1.436 - .01275 hsperc+ .00147 sat(0.098) (.00072) (.00009)n = 2,070, R2 = .283.(iii) The ratio of the standard error using 2,070 observations to that using 4,137 observationsis about 1.31. From (5.10) we compute ≈ 1.41, which is somewhat above the C5.3 We first run the regression colgpa on cigs, parity, and faminc using only the 1,191 observations with nonmissing observations on motheduc and fatheduc. After obtaining these residuals,iu, these are regressed on cigs i, parity i, faminc i, motheduc i, and fatheduc i, where, of course, we can only use the 1,197 observations with nonmissing values for both motheduc andfatheduc. The R-squared from this regression,2uR, is about .0024. With 1,191 observations, thechi-square statistic is (1,191)(.0024) ≈ 2.86. The p-value from the 22χ distribution is about .239, which is very close to .242, the p-value for the comparable F test.C5.4 (i) The measure of skewness for inc is about 1.86. When we use log(inc), the skewness measure is about .360. Therefore, there is much less skewness in log of income, which means inc is less likely to be normally distributed. (In fact, the skewness in income distributions is a well-documented fact across many countries and time periods.)(ii) The skewness for bwght is about -.60. When we use log(bwght), the skewness measure is about -2.95. In this case, there is much more skewness after taking the natural log.(iii) The example in part (ii) clearly shows that this statement cannot hold generally. It is possible to introduce skewness by taking the natural log. As an empirical matter, for many economic variables, particularly dollar values, taking the log often does help to reduce or eliminate skewness. But it does not have to.(iv) For the purposes of regression analysis, we should be studying the conditional distributions; that is, the distributions of y and log(y) conditional on the explanatory variables,1, ...,kx x. If we think the mean is linear, as in Assumptions MLR.1 and MLR.3, then this is equivalent to studying the distribution of the population error, u. In fact, the skewness measure studied in this question often is applied to the residuals from and OLS regression.C5.5 (i) The variable educ takes on all integer values from 6 to 20, inclusive. So it takes on 15 distinct values. It is not a continuous random variable, nor does it make sense to think of it as approximately continuous. (Contrast a variable such as hourly wage, which is rounded to two decimal places but takes on so many different values it makes sense to think of it as continuous.) (ii) With a discrete variable, usually a histogram has bars centered at each outcome, with the height being the fraction of observations taking on the value. Such a histogram, with a normal distribution overlay, is given below.Even discounting the discreteness, the best fitting normal distribution (matching the sample mean and variance) fits poorly. The focal point at educ = 12 clearly violates the notion of a smooth bell-shaped density.(iv) Given the findings in part (iii), the error term in the equation201234educ motheduc fatheduc abil abil u βββββ=+++++cannot have a normal distribution independent of the explanatory variables. Thus, MLR.6 is violated. In fact, the inequality 0educ ≥ means that u is not even free to vary over all values given motheduc , fatheduc , and abil . (It is likely that the homoskedasticity assumption fails, too, but this is less clear and does not follow from the nature of educ .)(v) The violation of MLR.6 means that we cannot perform exact statistical inference; we must rely on asymptotic analysis. This in itself does not change how we perform statistical inference: without normality, we use exactly the same methods, but we must be aware that our inference holds only approximately.d e n s i t y。
计量经济学第六版第二章计算机课后答案1.计量经济学是一门什么样的学科?答:计量经济学的英文单词是Econometrics,本意是“经济计量”,研究经济问题的计量方法,因此有时也译为“经济计量学”。
将Econometrics译为“计量经济学”是为了强调它是现代经济学的一门分支学科,不仅要研究经济问题的计量方法,还要研究经济问题发展变化的数量规律。
可以认为,计量经济学是以经济理论为指导,以经济数据为依据,以数学、统计方法为手段,通过建立、估计、检验经济模型,揭示客观经济活动中存在的随机因果关系的一门应用经济学的分支学科。
2.计量经济学与经济理论、数学、统计学的联系和区别是什么?答:计量经济学是经济理论、数学、统计学的结合,是经济学、数学、统计学的交叉学科(或边缘学科)。
计量经济学与经济学、数学、统计学的联系主要是计量经济学对这些学科的应用。
计量经济学对经济学的应用主要体现在以下几个方面:第一,计量经济学模型的选择和确定,包括对变量和经济模型的选择,需要经济学理论提供依据和思路;第二,计量经济分析中对经济模型的修改和调整,如改变函数形式、增减变量等,需要有经济理论的指导和把握;第三,计量经济分析结果的解读和应用也需要经济理论提供基础、背景和思路。
计量经济学对统计学的应用,至少有两个重要方面:一是计量经济分析所采用的数据的收集与处理、参数的估计等,需要使用统计学的方法和技术来完成;一是参数估计值、模型的预测结果的可靠性,需要使用统计方法加以分析、判断。
计量经济学对数学的应用也是多方面的,首先,对非线性函数进行线性转化的方法和技巧,是数学在计量经济学中的应用;其次,任何的参数估计归根结底都是数学运算,较复杂的参数估计方法,或者较复杂的模型的参数估计,更需要相当的数学知识和数学运算能力,另外,在计量经济理论和方法的研究方面,需要用到许多的数学知识和原理。
计量经济学与经济学、数学、统计学的区别也很明显,经济学、数学、统计学中的任何一门学科,都不能替代计量经济学,这三门学科简单地合起来,也不能替代计量经济学。
CHAPTER 12TEACHING NOTESMost of this chapter deals with serial correlation, but it also explicitly considers heteroskedasticity in time series regressions. The first section allows a review of what assumptions were needed to obtain both finite sample and asymptotic results. Just as with heteroskedasticity, serial correlation itself does not invalidate R-squared. In fact, if the data are stationary and weakly dependent, R-squared and adjusted R-squared consistently estimate the population R-squared (which is well-defined under stationarity).Equation (12.4) is useful for explaining why the usual OLS standard errors are not generally valid with AR(1) serial correlation. It also provides a good starting point for discussing serial correlation-robust standard errors in Section 12.5. The subsection on serial correlation with lagged dependent variables is included to debunk the myth that OLS is always inconsistent with lagged dependent variables and serial correlation. I do not teach it to undergraduates, but I do to Section 12.2 is somewhat untraditional in that it begins with an asymptotic t test for AR(1) serial correlation (under strict exogeneity of the regressors). It may seem heretical not to give the Durbin-Watson statistic its usual prominence, but I do believe the DW test is less useful than the tthe h statistic is asymptotically equivalent and not always computable.Section 12.3, on GLS and FGLS estimation, is fairly standard, although I try to show how comparing OLS estimates and FGLS estimates is not so straightforward. Unfortunately, at the beginning level (and even beyond), it is difficult to choose a course of action when they are very different.I do not usually cover Section 12.5 in a first-semester course, but, because some econometrics packages routinely compute fully robust standard errors, students can be pointed to Section 12.5 if they need to learn something about what the corrections do. I do cover Section 12.5 for a-semester course).I also do not cover Section 12.6 in class; again, this is more to serve as a reference for more advanced students, particularly those with interests in finance. One important point is that ARCH is heteroskedasticity and not serial correlation, something that is confusing in many texts. If a model contains no serial correlation, the usual heteroskedasticity-robust statistics are valid. I have a brief subsection on correcting for a known form of heteroskedasticity and AR(1) errors in models with strictly exogenous regressors.143SOLUTIONS TO PROBLEMS12.1 We can reason this from equation (12.4) because the usual OLS standard error is anthe AR(1) parameter, , tends to be positive in time series regression models. Further, the independent variables tend to be positive correlated, so (x t)(x t+j) which is what generally appears in (12.4) when the {x t} do not have zero sample average tends to be positive for most t and j. With multiple explanatory variables the formulas are more complicated but have similar features.If < 0, or if the {x t} is negatively autocorrelated, the second term in the last line of (12.4)could be negative, in which case the true standard deviation of12.2 This statement implies that we are still using OLS to estimate the j. But we are not using OLS; we are using feasible GLS (without or with the equation for the first time period). In other words, neither the Cochrane-Orcutt nor the Prais-Winsten estimators are the OLS estimators (and they usually differ from each other).12.3 (i) Because U.S. presidential elections occur only every four years, it seems reasonable to think the unobserved shocks that is, elements in u t in one election have pretty much dissipated four years later. This would imply that {u t} is roughly serially uncorrelated.(ii) The t statistic for H0: = 0 is .068/.240 .28, which is very small. Further, the estimate = .068 is small in a practical sense, too. There is no reason to worry about serial correlation in this example.(iii) Because the test based on is only justified asymptotically, we would generally be concerned about using the usual critical values with n = 20 in the original regression. But any kind of adjustment, either to obtain valid standard errors for OLS as in Section 12.5 or a feasible GLS procedure as in Section 12.3, relies on large sample sizes, too. (Remember, FGLS is not even unbiased, whereas OLS is under TS.1 through TS.3.) Most importantly, the estimate of is practically small, too. With so close to zero, FGLS or adjusting the standard errors would yield similar results to OLS with the usual standard errors.12.4 This is false, and a source of confusion in several textbooks. (ARCH is often discussed as a way in which the errors can be serially correlated.) As we discussed in Example 12.9, the errors in the equation return t = 0 + 1return t-1 + u t are serially uncorrelated, but there is strong evidence of ARCH; see equation (12.51).12.5 (i) There is substantial serial correlation in the errors of the equation, and the OLS standard errors almost certainly underestimate the true standard deviation in . This makes the usual confidence interval for EZ and t statistics invalid.144(ii) We can use the method in Section 12.5 to obtain an approximately valid standard error. [See equation (12.43).] While we might use g = 2 in equation (12.42), with monthly data we might want to try a somewhat longer lag, maybe even up to g = 12.12.6 With the strong heteroskedasticity in the errors it is not too surprising that the robust standard error for differs from the OLS standard error by a substantial amount: the robust standard error is almost 82% larger. Naturally, this reduces the t statistic. The robust t statistic is .059/.069 .86, which is even less significant than before. Therefore, we conclude that, once heteroskedasticity is accounted for, there is very little evidence that return t-1 is useful for predicting return t.12.7 (i) The usual Prais-Winsten standard errors will be incorrect because the usual transformation will not fully eliminate the serial correlation in u t. The regression of the OLS residuals on a single lag can be expected to consistently estimate the correlation coefficient, but the transformed variables will have serial correlation (where is the correlation between u t and u t-1 under stationarity).(ii) Equation (12.32) shows the transformed equation used by the Prais-Winsten method (without the first time period, which is unimportant for this discussion). If the AR(1) model is misspecified then the error term in (12.32), e t, is still serially correlated. But the Newey-West method works, in theory, for any kind of serial correlation. Therefore, we can just apply Newey-West directly to (12.32). Note that if we use the first observation our estimates of the are still the P-W estimates. We are simply obtaining standard errors that are robust to other forms of serial correlation. If we drop the first observation then we obtain the Cochrane-Orcutt estimates with standard errors robust to arbitrary serial correlation.It may make sense to try to improve over OLS assuming strict exogeneity, of course by then we might want our inference to be robust to more general forms of serial correlation.(iii) If we drop the homoskedasticity assumption then the e t in equation (12.32) exhibit heteroskedasticity as well as serial correlation. But Newey-West handles both problems at the same time. So we need do nothing different if we think the u t were heteroskedastic. (Of course, as discussed in Section 12.6, we could model the heteroskedasticity, too, and try to adjust for both. See pages 438 and 439. Even if we undertake such an exercise we can still use Newey-West standard errors on the transformed equation.)SOLUTIONS TO COMPUTER EXERCISESC12.1 Regressing on , using the 69 available observations, gives .292 andse() .118. The t statistic is about 2.47, and so there is significant evidence of positive AR(1) serial correlation in the errors (even though the variables have been differenced). This means we should view the standard errors reported in equation (11.27) with some suspicion.145。