ST5215: Advanced Statistical Theory Chen Zehua Department of Statistics & Applied Probability Tuesday, October 14, 2014 Chen Zehua ST5215: Advanced Statistical Theory Lecture 17: Methods for the derivation of UMVUE The 1st method for deriving UMVUE: solving for h I Find a sufficient and complete statistic T and its distribution. I Try some function h to see if E [h(T )] is related to ϑ. I Solve for h such that E [h(T )] = ϑ for all P. Example 3.1 (cont.) Let X1 , ..., Xn be i.i.d. from the uniform distribution on (0, θ), θ > 0. The order statistic X (n) is sufficient and complete with Lebesgue p.d.f. nθ−n x n−1 I(0,θ) (x). Consider ϑ = g (θ), where g is a differentiable function on (0, ∞). Chen Zehua ST5215: Advanced Statistical Theory Example 3.1 (cont.) I An unbiased estimator h(X(n) ) of ϑ must satisfy n Z θ g (θ) = n θ h(x)x n−1 dx for all θ > 0. 0 I Differentiating both sizes of the previous equation and applying the result of differentiation of an integral lead to nθn−1 g (θ) + θn g 0 (θ) = nh(θ)θn−1 . I Hence, the UMVUE of ϑ is h(X(n) ) = g (X(n) ) + n−1 X(n) g 0 (X(n) ). I In particular, if ϑ = θ, then the UMVUE of θ is (1 + n−1 )X(n) . Chen Zehua ST5215: Advanced Statistical Theory Example 3.2 Let X1 , ..., Xn be i.i.d. from Pthe Poisson distribution P(θ) with an unknown θ > 0. T (X ) = ni=1 Xi is sufficient and complete for θ > 0 and has the Poisson distribution P(nθ). SupposeP that ϑ = g (θ), where g is a smooth function such that j g (x) = ∞ j=0 aj x , x > 0. An unbiased estimator h(T ) of ϑ must satisfy (for any θ > 0): ∞ X h(t)nt t=0 t! θt = e nθ g (θ) ∞ X nk ∞ X θ aj θj k! j=0 k=0 ∞ X X n k aj θt . = k! = t=0 Chen Zehua k j,k:j+k=t ST5215: Advanced Statistical Theory Example 3.2 (continued) Thus, a comparison of coefficients in front of θt leads to h(t) = t! nt X j,k:j+k=t nk aj , k! i.e., h(T ) is the UMVUE of ϑ. In particular, if ϑ = θr for some fixed integer r ≥ 1, then ar = 1 and ak = 0 if k 6= r and ( 0 t<r h(t) = t! t≥r nr (t−r )! Chen Zehua ST5215: Advanced Statistical Theory Example 3.5 Let X1 , ..., Xn be i.i.d. from a power series distribution, i.e., P(Xi = x) = γ(x)θx /c(θ), x = 0, 1, 2, ..., with a known function γ(x) ≥ 0 and an unknown parameter θ > 0. It turns out that the joint distribution of X = (X1 , ..., Xn ) is in an exponential Pnfamily with a sufficient and complete statistic T (X ) = i=1 Xi . Furthermore, the distribution of T is also in a power series family, i.e., P(T = t) = γn (t)θt /[c(θ)]n , t = 0, 1, 2, ..., where γn (t) is the coefficient of θt in the power series expansion of [c(θ)]n . This result can help us to find the UMVUE of ϑ = g (θ). Chen Zehua ST5215: Advanced Statistical Theory Example 3.5 (continued) For example, by comparing both sides of ∞ X h(t)γn (t)θt = [c(θ)]n−p θr , t=0 we conclude that the UMVUE of θr /[c(θ)]p is ( 0 T <r h(T ) = γn−p (T −r ) T ≥ r, γn (T ) where r and p are nonnegative integers. In particular, the case of p = 1 produces the UMVUE γ(r )h(T ) of the probability P(X1 = r ) = γ(r )θr /c(θ) for any nonnegative integer r . Chen Zehua ST5215: Advanced Statistical Theory The 2nd method for deriving UMVUE: conditioning I Find an unbiased estimator of ϑ, say U(X ). I Conditioning on a sufficient and complete statistic T (X ): E [U(X )|T ] is the UMVUE of ϑ. I The distribution of T is not needed. We only need to work out the conditional expectation E [U(X )|T ]. I From the uniqueness of the UMVUE, it does not matter which U(X ) is used. Thus, U(X ) should be chosen so as to make the calculation of E [U(X )|T ] as easy as possible. Chen Zehua ST5215: Advanced Statistical Theory Example 3.3 Let X1 , ..., Xn be i.i.d. from the exponential distribution E (0, θ) with p.d.f. fθ (x) = 1θ e −x/θ I(0,∞) (x). Consider the estimation of ϑ = 1 − Fθ (t). X¯ is sufficient and complete for θ > 0. I(t,∞) (X1 ) is unbiased for ϑ, E [I(t,∞) (X1 )] = P(X1 > t) = ϑ. Hence T (X ) = E [I(t,∞) (X1 )|X¯ ] = P(X1 > t|X¯ ) is the UMVUE of ϑ. If the conditional distribution of X1 given X¯ is available, then we can calculate P(X1 > t|X¯ ) directly. By Basu’s theorem (Theorem 2.4), X1 /X¯ and X¯ are independent. By Proposition 1.10(vii), P(X1 > t|X¯ = x¯) = P(X1 /X¯ > t/X¯ |X¯ = x¯) = P(X1 /X¯ > t/¯ x ). Chen Zehua ST5215: Advanced Statistical Theory To compute this unconditional probability, we need the distribution ! of X n n X X1 Xi = X1 X1 + Xi . i=1 i=2 Using theP transformation technique discussed in §1.3.1 and the fact that ni=2 Xi is independentPof X1 and has a gamma distribution, we obtain that X1 / ni=1 Xi has the Lebesgue p.d.f. (n − 1)(1 − x)n−2 I(0,1) (x). Hence Z 1 t n−1 P(X1 > t|X¯ = x¯) = (n − 1) (1 − x)n−2 dx = 1 − n¯ x t/(n¯ x) and the UMVUE of ϑ is t n−1 T (X ) = 1 − ¯ . nX Chen Zehua ST5215: Advanced Statistical Theory Example 3.4 (cont.) Let X1 , ..., Xn be i.i.d. from N(µ, σ 2 ) with unknown µ ∈ R and σ 2 > 0. Let c be a fixed constant and c −µ . ϑ = P(X1 ≤ c) = Φ σ Consider the UMVUE of ϑ. I I Since I(−∞,c) (X1 ) is an unbiased estimator of ϑ, the UMVUE of ϑ is E [I(−∞,c) (X1 )|T ] = P(X1 ≤ c|T ). By Basu’s theorem, the ancillary statistic Z (X ) = (X1 − X¯ )/S is independent of T = (X¯ , S 2 ). Chen Zehua ST5215: Advanced Statistical Theory Example 3.4 (continued) I Then, by Proposition 1.10(vii), c − X¯ 2 T = (¯ x, s ) P X1 ≤ c|T = (¯ x, s ) = P Z ≤ S c − x¯ =P Z ≤ . s 2 I It can be shown that Z has the Lebesgue p.d.f. √ f (z) = √ I nΓ n−1 2 π(n − 1)Γ 1− n−2 2 nz 2 (n − 1)2 (n/2)−2 I(0,(n−1)/√n) (|z|) Hence the UMVUE of ϑ is Z P(X1 ≤ c|T ) = Chen Zehua (c−X¯ )/S √ −(n−1)/ n f (z)dz ST5215: Advanced Statistical Theory Example 3.4 (continued) Now consider the estimation of 1 0 c −µ ϑ= Φ , σ σ the Lebesgue p.d.f. of X1 evaluated at a fixed c, where Φ0 is the first-order derivative of Φ. I I I By the previous result, the conditional p.d.f. of X1 given x . X¯ = x¯ and S 2 = s 2 is s −1 f x−¯ s Let fT be the joint p.d.f. of T = (X¯ , S 2 ). Then Z Z 1 c − x¯ c − X¯ 1 ϑ= f f fT (t)dt = E . s s S S Hence the UMVUE of ϑ is 1 c − X¯ f . S S Chen Zehua ST5215: Advanced Statistical Theory Example Let X1 , ..., Xn be i.i.d. with Lebesgue p.d.f. fθ (x) = θx −2 I(θ,∞) (x), where θ > 0 is unknown. Suppose that ϑ = P(X1 > t) for a constant t > 0. The smallest order statistic X(1) is sufficient and complete for θ. Hence, the UMVUE of ϑ is P(X1 > t|X(1) ) = P(X1 > t|X(1) = x(1) ) X1 t X = x =P > (1) X(1) X(1) (1) X1 t =P > X = x(1) X(1) x(1) (1) X1 =P >s X(1) (Basu’s theorem), where s = t/x(1) . If s ≤ 1, this probability is 1. Chen Zehua ST5215: Advanced Statistical Theory Consider s > 1 and assume θ = 1 in the calculation: X n X1 X1 P >s = P > s, X(1) = Xi X(1) X(1) i=1 n X X1 > s, X(1) = Xi = P X(1) i=2 X1 = (n − 1)P > s, X(1) = Xn X(1) = (n − 1)P (X1 > sXn , X2 > Xn , ..., Xn−1 > Xn ) Z n Y 1 dx1 · · · dxn = (n − 1) 2 x1 >sxn ,x2 >xn ,...,xn−1 >xn i=1 xi # Z ∞ "Z ∞ n−1 Y Z ∞ 1 1 1 = (n − 1) dxi dx1 2 dxn 2 2 xn x1 1 sxn i=2 xn xi Z ∞ (n − 1)x(1) 1 = (n − 1) dxn = n+1 nt sxn 1 Chen Zehua ST5215: Advanced Statistical Theory Example (continued) This shows that the UMVUE of P(X1 > t) is ( (n−1)X(1) X(1) < t nt h(X(1) ) = 1 X(1) ≥ t Use the method of finding h: n The Lebesgue p.d.f. of X(1) is xnθ n+1 I(θ,∞) (x). Let the UMVUE be h(X(1) ) where, for all θ > 0, h satisfies Z ∞ nθn Pθ (X1 > t) = h(x)dx. x n+1 θ If θ ≥ t, then Pθ (X1 > t) = 1. Hence , for all θ ≥ t, Z ∞ nθn h(x)dx, 1= x n+1 θ which implies that h(x) = 1 when x ≥ t. Chen Zehua ST5215: Advanced Statistical Theory If θ < t, ∞ nθn dx x n+1 θ Z t Z ∞ Z t nθn nθn nθn θn = h(x) n+1 dx + dx = h(x) dx + x x n+1 x n+1 tn θ t θ Since P(X1 > t) = θ/t, we have Z t Z t θ nθn θn 1 n 1 = h(x) n+1 dx + n i.e. n−1 = h(x) n+1 dx + n t x t tθ x t θ θ Differentiating both sizes w.r.t. θ leads to n n−1 − n = −h(θ) n+1 tθ θ Hence, for any X(1) < t, Z E [h(X(1) )] = h(X(1) ) = Chen Zehua h(x) (n − 1)X(1) . nt ST5215: Advanced Statistical Theory A necessary and sufficient condition for UMVUE Theorem 3.2 Let U be the set of all unbiased estimators of 0 with finite variances and T be an unbiased estimator of ϑ with E (T 2 ) < ∞. (i) A necessary and sufficient condition for T (X ) to be a UMVUE of ϑ is that E [T (X )U(X )] = 0 for any U ∈ U and any P ∈ P. (ii) Suppose that T = h(T˜ ), where T˜ is a sufficient statistic for P ∈ P and h is a Borel function. Let UT˜ be the subset of U consisting of Borel functions of T˜ . Then a necessary and sufficient condition for T to be a UMVUE of ϑ is that E [T (X )U(X )] = 0 for any U ∈ UT˜ and any P ∈ P. Chen Zehua ST5215: Advanced Statistical Theory Proof of Theorem 3.2(i) Suppose that T is a UMVUE of ϑ. Then Tc = T + cU, where U ∈ U and c is a fixed constant, is also unbiased for ϑ and, thus, Var(Tc ) ≥ Var(T ) c ∈ R, P ∈ P, which is the same as c 2 Var(U) + 2cCov(T , U) ≥ 0 c ∈ R, P ∈ P. This is impossible unless Cov(T , U) = E (TU) = 0 for any P ∈ P. Suppose now E (TU) = 0 for any U ∈ U and P ∈ P. Let T0 be another unbiased estimator of ϑ with Var(T0 ) < ∞. Then T − T0 ∈ U and, hence, E [T (T − T0 )] = 0 P ∈ P, which with the fact that ET = ET0 implies that Var(T ) = Cov(T , T0 ) P ∈ P. Note that [Cov(T , T0 )]2 ≤ Var(T )Var(T0 ). Hence Var(T ) ≤ Var(T0 ) for any P ∈ P. Chen Zehua ST5215: Advanced Statistical Theory Proof of Theorem 3.2(ii) It suffices to show that E (TU) = 0 for any U ∈ UT˜ and P ∈ P implies that E (TU) = 0 for any U ∈ U and P ∈ P Let U ∈ U. Then E (U|T˜ ) ∈ UT˜ and the result follows from the fact that T = h(T˜ ) and E (TU) = E [E (TU|T˜ )] = E [E (h(T˜ )U|T˜ )] = E [h(T˜ )E (U|T˜ )]. Theorem 3.2 can be used I to find a UMVUE, I to check whether a particular estimator is a UMVUE, and I to show the nonexistence of any UMVUE. If there is a sufficient statistic, then by Rao-Blackwell’s theorem, we only need to focus on functions of the sufficient statistic and, hence, Theorem 3.2(ii) is more convenient to use. Chen Zehua ST5215: Advanced Statistical Theory Corollary 3.1 (i) Let Tj be a UMVUE of ϑj , j = 1, ..., k, where k is a fixed positive Pinteger. P Then kj=1 cj Tj is a UMVUE of ϑ = kj=1 cj ϑj for any constants c1 , ..., ck . (ii) Let T1 and T2 be two UMVUE’s of ϑ. Then T1 = T2 a.s. P for any P ∈ P. Example 3.7 Let X1 , ..., Xn be i.i.d. from the uniform distribution on the interval (0, θ). In Example 3.1, (1 + n−1 )X(n) is shown to be the UMVUE for θ when the parameter space is Θ = (0, ∞). Suppose now that Θ = [1, ∞). Then X(n) is not complete, although it is still sufficient for θ. Thus, Theorem 3.1 does not apply to X(n) . Chen Zehua ST5215: Advanced Statistical Theory Example 3.7 (continued) We now use Theorem 3.2(ii) to find a UMVUE of θ. Let U(X(n) ) be an unbiased estimator of 0. Since X(n) has the Lebesgue p.d.f. nθ−n x n−1 I(0,θ) (x), Z 1 Z θ 0= U(x)x n−1 dx + U(x)x n−1 dx for all θ ≥ 1. 0 1 This implies that U(x) = 0 a.e. Lebesgue measure on [1, ∞) and Z 1 U(x)x n−1 dx = 0. 0 Consider T = h(X(n) ). To have E (TU) = 0, we must have Z 1 h(x)U(x)x n−1 dx = 0. 0 Thus, we may consider the following function: c 0≤x ≤1 h(x) = bx x > 1, where c and b are some constants. Chen Zehua ST5215: Advanced Statistical Theory Example 3.7 (continued) From the previous discussion, E [h(X(n) )U(X(n) )] = 0, θ ≥ 1. Since E [h(X(n) )] = θ, we obtain that θ = cP(X(n) ≤ 1) + bE [X(n) I(1,∞) (X(n) )] = cθ−n + [bn/(n + 1)](θ − θ−n ). Thus, c = 1 and b = (n + 1)/n. The UMVUE of θ is then 1 0 ≤ X(n) ≤ 1 h(X(n) ) = −1 (1 + n )X(n) X(n) > 1. This estimator is better than (1 + n−1 )X(n) , which is the UMVUE when Θ = (0, ∞) and does not make use of the information about θ ≥ 1. When Θ = (0, ∞), this estimator is not unbiased. In fact, h(X(n) ) is complete and sufficient for θ ∈ [1, ∞). Chen Zehua ST5215: Advanced Statistical Theory Example 3.7 (continued) It suffices to show that 1 0 ≤ X(n) ≤ 1 X(n) X(n) > 1. is complete and sufficient for θ ∈ [1, ∞). The sufficiency follows from the fact that the joint p.d.f. of X1 , ..., Xn is 1 1 I(0,θ) (X(n) ) = n I(0,θ) (g (X(n) )). n θ θ If E [f (g (X(n) )] = 0 for all θ > 1, then Z θ Z 1 Z θ n−1 n−1 0= f (g (x))x dx = f (1)x dx + f (x)x n−1 dx g (X(n) ) = 0 0 1 for all θ > 1. Letting θ → 1 we obtain that f (1) = 0. Then Z θ 0= f (x)x n−1 dx 1 for all θ > 1, which implies f (x) = 0 a.e. for x > 1. Hence, g (X(n) ) is complete. Chen Zehua ST5215: Advanced Statistical Theory Example 3.8 Let X be a sample (of size 1) from the uniform distribution U(θ − 12 , θ + 21 ), θ ∈ R. There is no UMVUE of ϑ = g (θ) for any nonconstant function g . Note that an unbiased estimator U(X ) of 0 must satisfy Z θ+ 1 2 U(x)dx = 0 for all θ ∈ R. θ− 21 Differentiating both sides of the previous equation and applying the result of differentiation of an integral lead to U(x) = U(x + 1) a.e. m, where m is the Lebesgue measure on R. If T is a UMVUE of g (θ), then T (X )U(X ) is unbiased for 0 and, hence, T (x)U(x) = T (x + 1)U(x + 1) a.e. m, where U(X ) is any unbiased estimator of 0. Chen Zehua ST5215: Advanced Statistical Theory Example 3.8 (continued) Since this is true for all U, T (x) = T (x + 1) Since T is unbiased for g (θ), Z θ+ 1 2 g (θ) = T (x)dx a.e. m. for all θ ∈ R. θ− 12 Differentiating both sides of the previous equation and applying the result of differentiation of an integral, we obtain that g 0 (θ) = T θ + 12 − T θ − 12 = 0 a.e. m. Chen Zehua ST5215: Advanced Statistical Theory
© Copyright 2025 ExpyDoc