课件八高研一下2econometrics-i_第1页
课件八高研一下2econometrics-i_第2页
课件八高研一下2econometrics-i_第3页
课件八高研一下2econometrics-i_第4页
课件八高研一下2econometrics-i_第5页
已阅读5页,还剩30页未读 继续免费阅读

下载本文档

版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领

文档简介

1、Econometrics IProfessor William GreeneStern School of BusinessDepartment of EconomicsEconometrics IPart 6 Finite Sample Properties of Least SquaresTerms of ArtEstimates and estimatorsProperties of an estimator - the sampling distribution“Finite sample” properties as opposed to “asymptotic” or “large

2、 sample” propertiesApplication: Health Care Panel DataGerman Health Care Usage Data, 7,293 Individuals, Varying Numbers of PeriodsData downloaded from Journal of Applied Econometrics Archive. There are altogether 27,326 observations. The number of observations ranges from 1 to 7. (Frequencies are: 1

3、=1525, 2=2158, 3=825, 4=926, 5=1051, 6=1000, 7=987).Variables in the file are DOCVIS = number of doctor visits in last three months HOSPVIS = number of hospital visits in last calendar year DOCTOR = 1(Number of doctor visits 0) HOSPITAL= 1(Number of hospital visits 0) HSAT = health satisfaction, cod

4、ed 0 (low) - 10 (high) PUBLIC = insured in public health insurance = 1; otherwise = 0 ADDON = insured by add-on insurance = 1; otherswise = 0 HHNINC = household nominal monthly net e in German marks / 10000. (4 observations with e=0 were dropped) HHKIDS = children under age 16 in the household = 1;

5、otherwise = 0 EDUC = years of schooling AGE = age in years MARRIED = marital statusFor now, treat this sample as if it were a cross section, and as if it were the full population.Population RegressionSampling DistributionA sampling experiment: Draw 25 observations at random from the population. Comp

6、ute the regression. Repeat 100 times. Display estimates.matrix ; beduc=init(100,1,0)$proc$draw ; n=25 $regress; quietly ; lhs=hhninc ; rhs = one,educ $matrix ; beduc(i)=b(2) $sample;all$endproc$execute ; i=1,100 $histogram;rhs=beduc $How should we interpret this variation in the regression slope?The

7、 Statistical Context of Least Squares EstimationThe sample of data from the population: Data generating process is y = x + The stochastic specification of the regression model: Assumptions about the random .Endowment of the stochastic properties of the model upon the least squares estimator. The est

8、imator is a function of the observed (realized) data. Least SquaresDeriving the Propertiesb = a parameter vector + a linear combination of the disturbances, each times a vector.Therefore, b is a vector of random variables. We analyze it as such.The assumption of nonstochastic regressors. How it is u

9、sed at this point.We do the analysis conditional on an X, then show that results do not depend on the particular X in hand, so the result must be general i.e., independent of X. Properties of the LS Estimator: b is unbiasedExpected value and the property of unbiasedness.Eb|X = E + (XX)-1X|X = + (XX)

10、-1XE|X = + 0Eb = EXEb|X = Eb.(The law of iterated expectations.)Sampling ExperimentMeans of Repetitions b|xPartitioned RegressionA Crucial Result About Specification: y = X11 + X22 + Two sets of variables. What if the regression is computed without the second set of variables?What is the expectation

11、 of the short regression estimator? Eb1|(y = X11 + X22 + ) b1 = (X1X1)-1X1yThe Left Out Variable Formula“Short” regression means we regress y on X1 when y = X11 + X22 + and 2 is not 0(This is a VVIR!) b1 = (X1X1)-1X1y = (X1X1)-1X1(X11 + X22 + ) = (X1X1)-1X1X11 + (X1X1)-1X1 X22 + (X1X1)-1X1) Eb1 = 1

12、+ (X1X1)-1X1X22ApplicationThe (truly) short regression estimator is biased.Application: Quantity = 1Price + e + If you regress Quantity on Price and leave out e. What do you get?Application: Left out VariableLeave out e. What do you get? In time series data, 1 0 (usually)CovPrice, e 0 in time series

13、 data.So, the short regression will overestimate the price coefficient. It will be pulled toward and even past zero.Simple Regression of G on a constant and PGPrice Coefficient should be negative.Estimated Demand EquationShouldnt the Price Coefficient be Negative?Multiple Regression of G on Y and PG

14、. The Theory Works!-Ordinary least squares regression .LHS=G Mean = 226.09444 Standard deviation = 50.59182 Number of observs. = 36Model size Parameters = 3 Degrees of freedom = 33Residuals Sum of squares = 1472.79834 Standard error of e = 6.68059Fit R-squared = .98356 Adjusted R-squared = .98256Mod

15、el test F 2, 33 (prob) = 987.1(.0000)-+-Variable| Coefficient Standard Error t-ratio P|T|t Mean of X-+-Constant| -79.7535* 8.67255 -9.196 .0000 Y| .03692* .00132 28.022 .0000 9232.86 PG| -15.1224* 1.88034 -8.042 .0000 2.31661-+-The Extra Variable FormulaA Second Crucial Result About Specification: y

16、 = X11 + X22 + but 2 really is 0.Two sets of variables. One is superfluous. What if the regression is computed with it anyway?The Extra Variable Formula: (This is a VIR!) Eb1.2| 2 = 0 = 1The long regression estimator in a short regression is unbiased.)Extra variables in a model do not induce biases.

17、 Why not just include them? Variance of bAssumption about disturbances:i has zero mean and is uncorrelated with every other j Vari|X = 2. The variance of i does not depend on any data in the sample. Variance of the Least Squares EstimatorVariance of the Least Squares EstimatorSpecification Errors-1O

18、mitting relevant variables: Suppose the correct model is y = X11 + X22 + . I.e., two sets of variables. Compute least squares omitting X2. Some easily proved results:Varb1 is smaller than Varb1.2. (The latter is the northwest submatrix of the full covariance matrix. The proof uses the residual maker

19、 (again!). I.e., you get a smaller variance when you omit X2. (One interpretation: Omitting X2 amounts to using extra information (2 = 0). Even if the information is wrong (see the next result), it reduces the variance. (This is an important result.)Omitted Variables(No free lunch) Eb1 = 1 + (X1X1)-

20、1X1X22 1. So, b1 is biased.(!) The bias can be huge. Can reverse the sign of a price coefficient in a “demand equation.” b1 may be more “precise.” Precision = Mean squared error = variance + squared bias. Smaller variance but positive bias. If bias is small, may still favor the short regression.(Fre

21、e lunch?) Suppose X1X2 = 0. Then the bias goes away. Interpretation, the information is not “right,” it is irrelevant. b1 is the same as b1.2.Specification Errors-2 Including superfluous variables: Just reverse the results.Including superfluous variables increases variance. (The cost of not using in

22、formation.)Does not cause a bias, because if the variables in X2 are truly superfluous, then 2 = 0, so Eb1.2 = 1. Linear RestrictionsContext: How do linear restrictions affect the properties of the least squares estimator? Model: y = X + Theory (information) R - q = 0Restricted least squares estimat

23、or: b* = b - (XX)-1RR(XX)-1R-1(Rb - q)Expected value: Eb* = - (XX)-1RR(XX)-1R-1(Rb - q)Variance: 2(XX)-1 - 2 (XX)-1RR(XX)-1R-1 R(XX)-1 = Varb a nonnegative definite matrix VarbImplication: (As before) nonsample information reduces the variance of the estimator.InterpretationCase 1: Theory is correct

24、: R - q = 0 (the restrictions do hold). b* is unbiased Varb* is smaller than Varb How do we know this?Case 2: Theory is incorrect: R - q 0 (the restrictions do not hold). b* is biased what does this mean? Varb* is still smaller than VarbRestrictions and InformationHow do we interpret this important

25、result? The theory is information Bad information leads us away from the truth Any information, good or bad, makes us more certain of our answer. In this context, any information reduces variance.What about ignoring the information? Not using the correct information does not lead us away from the tr

26、uth Not using the information foregoes the variance reduction - i.e., does not use the ability to reduce uncertainty.Gauss-Markov TheoremA theorem of Gauss and Markov: Least Squares is the minimum variance linear unbiased estimator (MVLUE) 1. Linear estimator2. Unbiased: Eb|X = Theorem: Varb*|X Varb

27、|X is nonnegative definite for any other linear and unbiased estimator b* that is not equal to b.Definition: b is efficient in this class of estimators.Implications of Gauss-MarkovTheorem: Varb*|X Varb|X is nonnegative definite for any other linear and unbiased estimator b* that is not equal to b. Implies:bk = the kth particular element of b.Varbk|X = the kth diagonal element of Varb|XVarbk|X Varbk*|X for each coefficient.cb = any linear combination of

温馨提示

  • 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
  • 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
  • 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
  • 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
  • 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
  • 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
  • 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

评论

0/150

提交评论