This preview has intentionally blurred sections. Sign up to view the full version.
View Full Document
Unformatted text preview: 5 three possible alternative hypotheses: (a) β … 0: women are paid different hourly wages than men, either higher or lower; (b) β > 0: women are paid higher hourly wages than men; (c) β < 0: women are paid lower hourly wages than men. The last hypothesis is usually what is meant by “gender discrimination.” A test for the null hypothesis β = 0 against one of these alternative hypotheses can be based on the estimate of β , ˆ β provided that we know how is related to β . ˆ β It will be shown below that and are indeed reasonable approximations of α and β , ˆ α ˆ β respectively, possessing particular desirable properties. In general an estimator of an unknown parameter is a function of the data that serves as an approximation of the parameter involved. It follows from (1) that and are functions of ˆ α ˆ β the data, Because and will be used as approximations of α and β , ( Y 1 , X 1 ),...,( Y n , X n ). ˆ α ˆ β respectively, and were obtained by minimizing the squared errors, we will call and the ˆ α ˆ β Ordinary 2 Least Squares (OLS) estimators of α and β , respectively. 3.1 Unbiasedness The first property of and is that they are unbiased estimators of α and β : ˆ α ˆ β Proposition 1 . Under Assumptions II and IV the OLS estimators and are unbiased , which ˆ α ˆ β means that = E [ˆ α ] α and E [ ˆ β ] ' β . This result follows from the fact that we can write $ " ' " % j n j ' 1 1 n & ¯ X ( X j & ¯ X ) ' n i ' 1 ( X i & ¯ X ) 2 . U j , $ $ ' $ % ' n j ' 1 ( X j & ¯ X ) U j ' n i ' 1 ( X i & ¯ X ) 2 . (3) See the Appendix. 6 3.2 The variances of and ˆ α ˆ β . Our next issue concerns the variances of and . For deriving these variances the ˆ α ˆ β following two lemmas are convenient. Lemma 1 . Let U 1 , U 2 ,...,U n be independent random variables with zero mathematical expectation ( thus E ( U j ) = 0) and variance σ 2 . ( Thus E [( U j E ( U j )) 2 ] = E ( U j 2 ) = σ 2 ). Let v 1 , v 2 ,..., v n and w 1 , w 2 ,..., w n be given constants. Then E [( ' n j ' 1 v j U j )( ' n j ' 1 w j U j )] ' σ 2 ' n j ' 1 v j w j . Proof . See the Appendix. Note that if we choose in Lemma 1 then it reads: v j ' w j for j ' 1,2,..., n Lemma 2 . Let U 1 , U 2 ,...,U n be independent random variables with zero mathematical expectation and variance σ 2 . Let w 1 , w 2 ,..., w n be given constants. Then E [( ' n j ' 1 w j U j ) 2 ] ' σ 2 ' n j ' 1 w 2 j . Using (3) and Lemmas 1 and 2 it can be shown that Proposition 2 . Under the assumptions I  IV, var( $ " ) ' F 2 ' n j ' 1 X 2 j n ' n j ' 1 ( X j & ¯ X ) 2 ' F 2 $ " , say , var( $ $ ) ' F 2 ' n j ' 1 ( X j & ¯ X ) 2 ' F 2 $ $ , say , and cov( $ " , $ $ ) ' & F 2 ¯ X ' n j ' 1 ( X j & ¯ X ) 2 ....
View
Full Document
 Fall '10
 H.Bierens
 Economics, Econometrics, Normal Distribution, Regression Analysis, Null hypothesis, Statistical hypothesis testing, j Uj, j'1

Click to edit the document details