Section 9 (cont.): Review November 20th, 2014 Lesson 20 In this lesson, we will examine a few more properties of transformations of random variables. Mainly we will examine sums of random variables. We will also work several examples. Lesson 20 We list several properties of sums of random variables: If X1 and X2 are random variables, and Y = X1 + X2 , then E [Y ] = E [X1 ] + E [X2 ] and Var [Y ] = Var [X1 ] + Var [X2 ] + 2Cov [X1 , X2 ]. If X1 , . . . , Xn are random variables, and Y = then E [Y ] = n X Pn i=1 Xi , E [Xi ] i=1 and Var [Y ] = n X i=1 Var [Xi ] + 2 n X n X i=1 j=i+1 Lesson 20 Cov [Xi , Xj ]. If X1 , . . . , Xn are mutually independent random variables, then Var [Y ] = n X Var [Xi ] i=1 and MY (t) = n Y MXi (t) = MX1 (t) · MX2 (t) · · · MXn (t). i=1 If X1 , . . . , Xn and Y1 , . . . , Ym are random variables and a1 , . . . , an , b, c1 , . . . , cm , d are constants, then Cov n X ai Xi i=1 + b, m X cj Yj + d = j=1 n X m X i=1 j=1 Lesson 20 ai cj Cov [Xi , Yj ]. Example (9.8) Given n independent random variables X1 , . . . , Xn each having the same variance of σ 2 , and defining U = 2X1 + X2 + · · · + Xn−1 and V = X2 + · · · + Xn−1 + 2Xn , find the coefficient of correlation between U and V . Lesson 20 Example (9.9) Independent random variables X , Y , and Z are identically distributed. Let W = X + Y . The moment generating function of W is MW (t) = (.7 + .3e t )6 . Find the moment generating function of V = X + Y + Z . Lesson 20 Suppose that X is a random variable with mean µ and standard variance σ and suppose that X1 , . . . , Xn are n independent random variables with the same distribution as X . Let Yn = X1 + · · · + Xn . Then E [Yn ] = nµ and Var [Yn ] = nσ 2 . As n increases, the distribution of Yn approaches a normal distribution N(nµ, nσ 2 ). This is called the Central Limit Theorem. This theorem is the justification for using normal distributions to approximate the distribution of a sum of random variables. Lesson 20 Suppose that X1 , . . . , Xk are independent random variables P and Y = ki=1 Xi . distribution of Xi binomial B(ni , p) Poisson λi geometric p negative binomial ri , p normal N(µi , σi2 ) exponential with mean µ gamma with αi , β distribution of Y P binomial B( ni , p) P Poisson λi negative binomial k, p P negative binomial ri , p P P N( µi , σi2 ) gamma with α = k, β = 1/µ P gamma with αi , β Lesson 20 Example (9.10) Smith estimates his chance of winning a particular hand of blackjack at a casino is .45, his probability of losing is .5, and his probability of breaking even on a hand is .05. He is playing at a $10 table, which means that on each play, he either wins $10, loses $10 or breaks even, with the stated probabilities. Smith plays 100 times. What is the approximate probability that he has won money on the 100 plays of the game in total (assume separate hands are independent)? Lesson 20 Example (9.11) If the number of typographical errors per page typed by a certain typist follows a Poisson distribution with a mean of λ, find the probability that the total number of errors in 10 randomly selected pages is 10. Lesson 20
© Copyright 2024 ExpyDoc