課件八高研一下2econometrics-i_第1頁(yè)
課件八高研一下2econometrics-i_第2頁(yè)
課件八高研一下2econometrics-i_第3頁(yè)
課件八高研一下2econometrics-i_第4頁(yè)
課件八高研一下2econometrics-i_第5頁(yè)
已閱讀5頁(yè),還剩30頁(yè)未讀, 繼續(xù)免費(fèi)閱讀

下載本文檔

版權(quán)說(shuō)明:本文檔由用戶(hù)提供并上傳,收益歸屬內(nèi)容提供方,若內(nèi)容存在侵權(quán),請(qǐng)進(jìn)行舉報(bào)或認(rèn)領(lǐng)

文檔簡(jiǎn)介

1、Econometrics IProfessor William GreeneStern School of BusinessDepartment of EconomicsEconometrics IPart 6 Finite Sample Properties of Least SquaresTerms of ArtEstimates and estimatorsProperties of an estimator - the sampling distribution“Finite sample” properties as opposed to “asymptotic” or “l(fā)arge

2、 sample” propertiesApplication: Health Care Panel DataGerman Health Care Usage Data, 7,293 Individuals, Varying Numbers of PeriodsData downloaded from Journal of Applied Econometrics Archive. There are altogether 27,326 observations. The number of observations ranges from 1 to 7. (Frequencies are: 1

3、=1525, 2=2158, 3=825, 4=926, 5=1051, 6=1000, 7=987).Variables in the file are DOCVIS = number of doctor visits in last three months HOSPVIS = number of hospital visits in last calendar year DOCTOR = 1(Number of doctor visits 0) HOSPITAL= 1(Number of hospital visits 0) HSAT = health satisfaction, cod

4、ed 0 (low) - 10 (high) PUBLIC = insured in public health insurance = 1; otherwise = 0 ADDON = insured by add-on insurance = 1; otherswise = 0 HHNINC = household nominal monthly net e in German marks / 10000. (4 observations with e=0 were dropped) HHKIDS = children under age 16 in the household = 1;

5、otherwise = 0 EDUC = years of schooling AGE = age in years MARRIED = marital statusFor now, treat this sample as if it were a cross section, and as if it were the full population.Population RegressionSampling DistributionA sampling experiment: Draw 25 observations at random from the population. Comp

6、ute the regression. Repeat 100 times. Display estimates.matrix ; beduc=init(100,1,0)$proc$draw ; n=25 $regress; quietly ; lhs=hhninc ; rhs = one,educ $matrix ; beduc(i)=b(2) $sample;all$endproc$execute ; i=1,100 $histogram;rhs=beduc $How should we interpret this variation in the regression slope?The

7、 Statistical Context of Least Squares EstimationThe sample of data from the population: Data generating process is y = x + The stochastic specification of the regression model: Assumptions about the random .Endowment of the stochastic properties of the model upon the least squares estimator. The est

8、imator is a function of the observed (realized) data. Least SquaresDeriving the Propertiesb = a parameter vector + a linear combination of the disturbances, each times a vector.Therefore, b is a vector of random variables. We analyze it as such.The assumption of nonstochastic regressors. How it is u

9、sed at this point.We do the analysis conditional on an X, then show that results do not depend on the particular X in hand, so the result must be general i.e., independent of X. Properties of the LS Estimator: b is unbiasedExpected value and the property of unbiasedness.Eb|X = E + (XX)-1X|X = + (XX)

10、-1XE|X = + 0Eb = EXEb|X = Eb.(The law of iterated expectations.)Sampling ExperimentMeans of Repetitions b|xPartitioned RegressionA Crucial Result About Specification: y = X11 + X22 + Two sets of variables. What if the regression is computed without the second set of variables?What is the expectation

11、 of the short regression estimator? Eb1|(y = X11 + X22 + ) b1 = (X1X1)-1X1yThe Left Out Variable Formula“Short” regression means we regress y on X1 when y = X11 + X22 + and 2 is not 0(This is a VVIR!) b1 = (X1X1)-1X1y = (X1X1)-1X1(X11 + X22 + ) = (X1X1)-1X1X11 + (X1X1)-1X1 X22 + (X1X1)-1X1) Eb1 = 1

12、+ (X1X1)-1X1X22ApplicationThe (truly) short regression estimator is biased.Application: Quantity = 1Price + e + If you regress Quantity on Price and leave out e. What do you get?Application: Left out VariableLeave out e. What do you get? In time series data, 1 0 (usually)CovPrice, e 0 in time series

13、 data.So, the short regression will overestimate the price coefficient. It will be pulled toward and even past zero.Simple Regression of G on a constant and PGPrice Coefficient should be negative.Estimated Demand EquationShouldnt the Price Coefficient be Negative?Multiple Regression of G on Y and PG

14、. The Theory Works!-Ordinary least squares regression .LHS=G Mean = 226.09444 Standard deviation = 50.59182 Number of observs. = 36Model size Parameters = 3 Degrees of freedom = 33Residuals Sum of squares = 1472.79834 Standard error of e = 6.68059Fit R-squared = .98356 Adjusted R-squared = .98256Mod

15、el test F 2, 33 (prob) = 987.1(.0000)-+-Variable| Coefficient Standard Error t-ratio P|T|t Mean of X-+-Constant| -79.7535* 8.67255 -9.196 .0000 Y| .03692* .00132 28.022 .0000 9232.86 PG| -15.1224* 1.88034 -8.042 .0000 2.31661-+-The Extra Variable FormulaA Second Crucial Result About Specification: y

16、 = X11 + X22 + but 2 really is 0.Two sets of variables. One is superfluous. What if the regression is computed with it anyway?The Extra Variable Formula: (This is a VIR!) Eb1.2| 2 = 0 = 1The long regression estimator in a short regression is unbiased.)Extra variables in a model do not induce biases.

17、 Why not just include them? Variance of bAssumption about disturbances:i has zero mean and is uncorrelated with every other j Vari|X = 2. The variance of i does not depend on any data in the sample. Variance of the Least Squares EstimatorVariance of the Least Squares EstimatorSpecification Errors-1O

18、mitting relevant variables: Suppose the correct model is y = X11 + X22 + . I.e., two sets of variables. Compute least squares omitting X2. Some easily proved results:Varb1 is smaller than Varb1.2. (The latter is the northwest submatrix of the full covariance matrix. The proof uses the residual maker

19、 (again!). I.e., you get a smaller variance when you omit X2. (One interpretation: Omitting X2 amounts to using extra information (2 = 0). Even if the information is wrong (see the next result), it reduces the variance. (This is an important result.)Omitted Variables(No free lunch) Eb1 = 1 + (X1X1)-

20、1X1X22 1. So, b1 is biased.(!) The bias can be huge. Can reverse the sign of a price coefficient in a “demand equation.” b1 may be more “precise.” Precision = Mean squared error = variance + squared bias. Smaller variance but positive bias. If bias is small, may still favor the short regression.(Fre

21、e lunch?) Suppose X1X2 = 0. Then the bias goes away. Interpretation, the information is not “right,” it is irrelevant. b1 is the same as b1.2.Specification Errors-2 Including superfluous variables: Just reverse the results.Including superfluous variables increases variance. (The cost of not using in

22、formation.)Does not cause a bias, because if the variables in X2 are truly superfluous, then 2 = 0, so Eb1.2 = 1. Linear RestrictionsContext: How do linear restrictions affect the properties of the least squares estimator? Model: y = X + Theory (information) R - q = 0Restricted least squares estimat

23、or: b* = b - (XX)-1RR(XX)-1R-1(Rb - q)Expected value: Eb* = - (XX)-1RR(XX)-1R-1(Rb - q)Variance: 2(XX)-1 - 2 (XX)-1RR(XX)-1R-1 R(XX)-1 = Varb a nonnegative definite matrix VarbImplication: (As before) nonsample information reduces the variance of the estimator.InterpretationCase 1: Theory is correct

24、: R - q = 0 (the restrictions do hold). b* is unbiased Varb* is smaller than Varb How do we know this?Case 2: Theory is incorrect: R - q 0 (the restrictions do not hold). b* is biased what does this mean? Varb* is still smaller than VarbRestrictions and InformationHow do we interpret this important

25、result? The theory is information Bad information leads us away from the truth Any information, good or bad, makes us more certain of our answer. In this context, any information reduces variance.What about ignoring the information? Not using the correct information does not lead us away from the tr

26、uth Not using the information foregoes the variance reduction - i.e., does not use the ability to reduce uncertainty.Gauss-Markov TheoremA theorem of Gauss and Markov: Least Squares is the minimum variance linear unbiased estimator (MVLUE) 1. Linear estimator2. Unbiased: Eb|X = Theorem: Varb*|X Varb

27、|X is nonnegative definite for any other linear and unbiased estimator b* that is not equal to b.Definition: b is efficient in this class of estimators.Implications of Gauss-MarkovTheorem: Varb*|X Varb|X is nonnegative definite for any other linear and unbiased estimator b* that is not equal to b. Implies:bk = the kth particular element of b.Varbk|X = the kth diagonal element of Varb|XVarbk|X Varbk*|X for each coefficient.cb = any linear combination of

溫馨提示

  • 1. 本站所有資源如無(wú)特殊說(shuō)明,都需要本地電腦安裝OFFICE2007和PDF閱讀器。圖紙軟件為CAD,CAXA,PROE,UG,SolidWorks等.壓縮文件請(qǐng)下載最新的WinRAR軟件解壓。
  • 2. 本站的文檔不包含任何第三方提供的附件圖紙等,如果需要附件,請(qǐng)聯(lián)系上傳者。文件的所有權(quán)益歸上傳用戶(hù)所有。
  • 3. 本站RAR壓縮包中若帶圖紙,網(wǎng)頁(yè)內(nèi)容里面會(huì)有圖紙預(yù)覽,若沒(méi)有圖紙預(yù)覽就沒(méi)有圖紙。
  • 4. 未經(jīng)權(quán)益所有人同意不得將文件中的內(nèi)容挪作商業(yè)或盈利用途。
  • 5. 人人文庫(kù)網(wǎng)僅提供信息存儲(chǔ)空間,僅對(duì)用戶(hù)上傳內(nèi)容的表現(xiàn)方式做保護(hù)處理,對(duì)用戶(hù)上傳分享的文檔內(nèi)容本身不做任何修改或編輯,并不能對(duì)任何下載內(nèi)容負(fù)責(zé)。
  • 6. 下載文件中如有侵權(quán)或不適當(dāng)內(nèi)容,請(qǐng)與我們聯(lián)系,我們立即糾正。
  • 7. 本站不保證下載資源的準(zhǔn)確性、安全性和完整性, 同時(shí)也不承擔(dān)用戶(hù)因使用這些下載資源對(duì)自己和他人造成任何形式的傷害或損失。

評(píng)論

0/150

提交評(píng)論