Lecture 19: More Than Two Random Variables 0/ 17
Definition If X 1 , X 2 , . . . , X n are discrete random variables defined on the same sample space then their joint pmf is the function P X 1 , X 2 ,..., X n ( x 1 , x 2 , . . . , x n ) = P ( X 1 = x 1 , . . . , X n = x n ) If X 1 , X 2 , . . . , X n are continuous then their joint pdf is the function f X 1 , X 2 ,..., X n ( x 1 , x 2 , . . . , x n ) such that 1/ 17 Lecture 19: More Than Two Random Variables
Definition (Cont.) for any region A in R n � � P (( X 1 , X 2 , . . . , X n ) ∈ A ) = f X 1 , X 2 ,..., X n ( x 1 , . . . , x n ) dx 1 , . . . , dx n . . . A � ��������������������������������������������������� �� ��������������������������������������������������� � n-fold multiple integral Definition The discrete random variables X 1 , X 2 , . . . , X n are independent if P X 1 ,..., X n ( x 1 , . . . , x n ) = P X 1 ( x 1 ) P X 2 ( x 2 ) . . . P X n ( x n ) . Equivalently P ( X 1 = x 1 , . . . , X n = x n ) = P ( X 1 = x 1 ) . . . P ( X n = x n ) 2/ 17 Lecture 19: More Than Two Random Variables
The continuous random variables X 1 , X 2 , . . . , X n are independent if f X 1 , X 2 ,..., X n ( x 1 , x 2 , . . . , x n ) = f X 1 ( x 1 ) f X 2 ( x 2 ) . . . f X n ( x n ) Definition X 1 , X 2 , . . . , X n are pairwise independent if each pair X i , X j ( i � j ) is independent. We will now see Pairwise independence � = ⇒ Independence of random variables ⇐ = of random variables 3/ 17 Lecture 19: More Than Two Random Variables
First we will prove ⇐ = Theorem X 1 , X 2 , . . . , X n independent ⇒ X 1 , X 2 , . . . , X n are pairwise independent. From now on we will restrict to the case n = 3 so we have THREE random variables X , Y , Z . 4/ 17 Lecture 19: More Than Two Random Variables
How do we get P X , Y ( x , y ) P X , Y , Z ( x , y , z ) from Answer (left to you to prove) � P X , Y ( x , y ) = P X , Y , Z ( x , y , z ) ( # ) all z Now we can prove X , Y , Z independent. = ⇒ X , Y independent 5/ 17 Lecture 19: More Than Two Random Variables
Since X , Y , Z are independent we have P X , Y , Z ( x , y , z ) = P X ( x ) P Y ( y ) P Z ( z ) ( ## ) Now play the RHS of ( ## ) into the RHS of ( # ) 6/ 17 Lecture 19: More Than Two Random Variables
This proves X and Y are independent Identical proofs prove the pairs X , Z and Y , Z are independent. Now we construct X , Y , Z (actually X A , X B , X C ) so that each pair is independent but the triple X , Y , Z is not independent . 7/ 17 Lecture 19: More Than Two Random Variables
The multinomial coefficient � � n The multinomial coefficient is defined by k 1 , k 2 , ··· , k r � � n n ! = k 1 , k 2 , · · · , k r k 1 ! k 2 ! · · · k r ! Suppose an experiment has r outcomes denoted 1 , 2 , 3 , · · · , r with probabilities p 1 , p 2 , · · · , p r respectively. Repeat the experiment n times and assume the trials are independent. � � n k 1 , k 2 , ··· , k r 8/ 17 Lecture 19: More Than Two Random Variables
A Variation on the Cool Counter example Lets go back to the “cool counter example”, Lecture 16, page 18 of three events A , B , C which are pairwise independent but no independent so P ( A ∩ B ∩ C ) � P ( A ) P ( B ) P ( C ) The idea is to convert the three events to random variables X A , X B , X C so that X A = 1 on A and O on A ’ etc. 9/ 17 Lecture 19: More Than Two Random Variables
In fact we won’t need the corner points ( − 1 , − 1 ) , ( − 1 , 1 ) , ( 1 , − 1 ) and ( 1 , 1 ) we put S 1 = ( 0 , 1 ) , S 2 = ( − 1 , 0 ) , S 3 = ( 0 , 1 ) , S 4 = ( 1 , 0 ) and retain their probabilities so P ( { S j } ) = 1 4 , 1 ≤ j ≤ 4 10/ 17 Lecture 19: More Than Two Random Variables
We define A = { s 1 , s 2 } B = { s 1 , s 3 } C = { s 1 , s 4 } We define X A , X B , X C on S by 1 , if s j ∈ A X A ( s j ) = 0 , if s j � A 11/ 17 Lecture 19: More Than Two Random Variables
1 , if s j ∈ B X B ( s j ) = 0 , if s j � B 1 , if s j ∈ C X C ( s j ) = 0 , if s j � C P ( X A = 1 ) = P ( { S 1 , S 2 } ) = 1 So 2 P ( X A = 0 ) = P ( { S 3 , S 4 } ) = 1 2 and similarly for X B and X C . So X A , X B and X C are Bernoulli random variables 12/ 17 Lecture 19: More Than Two Random Variables
Let’s compute the joint pmf of X A and X B . We know the margin ❍❍❍❍❍ X B 0 1 X A ❍ 0 1 / 2 1 1 / 2 1 / 1 / 2 2 The subset where X A = 1 is the subset { s 1 , s 2 } so we write an equality of events ( X A = 1 ) = { s 1 , s 2 } Similarly ( X A = 0 ) = { s 3 , s 4 } ( X B = 1 ) = { s 1 , x 3 } , ( X B = 0 ) = { s 2 , s 4 } ( X C = 1 ) = { s 1 , s 4 } , ( X C = 0 ) = { s 2 , s 3 } 13/ 17 Lecture 19: More Than Two Random Variables
Hence ( X A = 0 ) ∩ ( X B = 0 ) = { S 4 } P ( X A = 0 , X B = 0 ) = 1 so 4 ( X A = 0 ) ∩ ( X B = 1 ) = { S 3 } P ( X A = 0 , X B = 1 ) = 1 so 4 ( X A = 1 ) ∩ ( X B = 0 ) = { S 2 } P ( X A = 1 , X B = 0 ) = 1 4 ( X A = 1 ) ∩ ( X B = 1 ) = { S 1 } P ( X A = 1 , X B = 1 ) = P ( { S 1 } ) = 1 4 etc. 14/ 17 Lecture 19: More Than Two Random Variables
So the joint proof of X A and X B is ❍❍❍❍❍ X B 0 1 X A ❍ 0 1 / 1 / 1 / 4 4 2 1 1 / 1 / 1 / 4 4 2 1 / 1 / 2 2 so X A and X B are independent. The same is true for X A and X C and X B and χ C . Now we show the triple X A , X B and X C is NOT independent. 15/ 17 Lecture 19: More Than Two Random Variables
We will show P ( X A = 1 , X B = 1 , X C = 1 ) � P ( X A = 1 ) P ( X B = 1 ) P ( X C = 1 ) � 1 � � 1 � � 1 � = 1 The RHS = 2 2 2 8 The left-hand side is the probability of the event ( X A = 1 ) ∩ ( X B = 1 ) ∩ ( X C = 1 ) = { S 1 , S 2 } ∩ { S 1 , S 3 } ∩ { S 1 , S 4 } = { S 1 } . 16/ 17 Lecture 19: More Than Two Random Variables
So P ( X A = 1 , X B = 1 , X C = 1 ) = P ( { S 1 } ) = 1 4 so LHS = 1 4 RHS = 1 8 Remark This counter example is more or less the some as the “cool counter example”. We just replaced (more or less) A , B , C by their “characteristic functions”. 17/ 17 Lecture 19: More Than Two Random Variables
Recommend
More recommend