cs70 jean walrand lecture 36
play

CS70: Jean Walrand: Lecture 36. Gaussian and CLT CS70: Jean - PowerPoint PPT Presentation

CS70: Jean Walrand: Lecture 36. Gaussian and CLT CS70: Jean Walrand: Lecture 36. Gaussian and CLT Warning: CS70: Jean Walrand: Lecture 36. Gaussian and CLT Warning: This lecture is also rated R. CS70: Jean Walrand: Lecture 36. Gaussian and


  1. Expectation of function of RV Definition The expectation of a function of a random variable is defined as

  2. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx .

  3. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ .

  4. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] n

  5. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] = ∑ h ( n δ ) f X ( n δ ) δ n n

  6. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, � ∞ E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] = ∑ h ( n δ ) f X ( n δ ) δ = − ∞ h ( x ) f X ( x ) dx . n n � g ( x ) dx ≈ ∑ n g ( n δ ) δ with g ( x ) = h ( x ) f X ( x ) . Indeed,

  7. Expectation of function of RV Definition The expectation of a function of a random variable is defined as � ∞ E [ h ( X )] = − ∞ h ( x ) f X ( x ) dx . Justification: Say X = n δ w.p. f X ( n δ ) δ . Then, � ∞ E [ h ( X )] = ∑ h ( n δ ) Pr [ X = n δ ] = ∑ h ( n δ ) f X ( n δ ) δ = − ∞ h ( x ) f X ( x ) dx . n n � g ( x ) dx ≈ ∑ n g ( n δ ) δ with g ( x ) = h ( x ) f X ( x ) . Indeed, Fact Expectation is linear.

  8. Variance Definition: The variance of a continuous random variable X is defined as

  9. Variance Definition: The variance of a continuous random variable X is defined as E (( X − E ( X )) 2 ) var [ X ] =

  10. Variance Definition: The variance of a continuous random variable X is defined as E (( X − E ( X )) 2 ) = E ( X 2 ) − ( E ( X )) 2 var [ X ] =

  11. Variance Definition: The variance of a continuous random variable X is defined as E (( X − E ( X )) 2 ) = E ( X 2 ) − ( E ( X )) 2 var [ X ] = � ∞ � � ∞ � 2 − ∞ x 2 f ( x ) dx − = − ∞ xf ( x ) dx .

  12. Motivation for Gaussian Distribution Key fact:

  13. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution.

  14. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem.

  15. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.)

  16. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled.

  17. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled. This explains why the Gaussian distribution

  18. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled. This explains why the Gaussian distribution (the bell curve)

  19. Motivation for Gaussian Distribution Key fact: The sum of many small independent RVs has a Gaussian distribution. This is the Central Limit Theorem. (See later.) Examples: Binomial and Poisson suitably scaled. This explains why the Gaussian distribution (the bell curve) shows up everywhere.

  20. Normal Distribution. For any µ and σ , a normal (aka Gaussian )

  21. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf

  22. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √

  23. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 .

  24. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 .

  25. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 . Note: Pr [ | Y − µ | > 1 . 65 σ ] = 10 %;

  26. Normal Distribution. For any µ and σ , a normal (aka Gaussian ) random variable Y , which we write as Y = N ( µ , σ 2 ) , has pdf 1 2 πσ 2 e − ( y − µ ) 2 / 2 σ 2 . f Y ( y ) = √ Standard normal has µ = 0 and σ = 1 . Note: Pr [ | Y − µ | > 1 . 65 σ ] = 10 %; Pr [ | Y − µ | > 2 σ ] = 5 % .

  27. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) .

  28. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . √

  29. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √

  30. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] =

  31. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]]

  32. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]]

  33. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ

  34. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ f X ( y − µ ) dy = σ σ

  35. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ f X ( y − µ ) dy σ = 1 σ f X ( y − µ = ) dy σ σ

  36. Scaling and Shifting Theorem Let X = N ( 0 , 1 ) and Y = µ + σ X . Then Y = N ( µ , σ 2 ) . 2 π exp {− x 2 1 Proof: f X ( x ) = 2 } . Now, √ f Y ( y ) dy = Pr [ Y ∈ [ y , y + dy ]] = Pr [ µ + σ X ∈ [ y , y + dy ]] = Pr [ σ X ∈ [ y − µ , y − µ + dy ]] Pr [ X ∈ [ y − µ , y − µ + dy = σ ]] σ σ f X ( y − µ ) dy σ = 1 σ f X ( y − µ = ) dy σ σ 2 πσ 2 exp {− ( y − µ ) 2 1 √ = } dy . 2 σ 2

  37. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ

  38. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 .

  39. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,....

  40. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √

  41. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry.

  42. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. E [ X 2 ] var [ X ] =

  43. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π

  44. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 1 � √ = − 2 } 2 π

  45. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � √ √ = − 2 } = 2 } dx by IBP 2 π 2 π

  46. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � √ √ = − 2 } = 2 } dx by IBP 2 π 2 π � = f X ( x ) dx

  47. Expectation, Variance. Theorem If Y = N ( µ , σ 2 ) , then E [ Y ] = µ and var [ Y ] = σ 2 . Proof: It suffices to show the result for X = N ( 0 , 1 ) since Y = µ + σ X ,.... 2 π exp {− x 2 1 Thus, f X ( x ) = 2 } . √ First note that E [ X ] = 0 , by symmetry. exp {− x 2 1 � E [ X 2 ] = x 2 var [ X ] = √ 2 } dx 2 π xd exp {− x 2 exp {− x 2 1 1 � � √ √ = − 2 } = 2 } dx by IBP 2 π 2 π � = f X ( x ) dx = 1 .

  48. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.”

  49. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 .

  50. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n .

  51. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n .

  52. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . E ( A ′ n )

  53. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ )

  54. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 .

  55. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ n )

  56. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ 1 n ) = σ 2 / n Var ( A n )

  57. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ 1 n ) = σ 2 / n Var ( A n ) = 1 .

  58. Central limit theorem. Law of Large Numbers: For any set of independent identically distributed random variables, X i , A n = 1 n ∑ X i “tends to the mean.” Say X i have expecation µ = E ( X i ) and variance σ 2 . Mean of A n is µ , and variance is σ 2 n . n = A n − µ Let A ′ σ / √ n . 1 E ( A ′ n ) = σ / √ n ( E ( A n ) − µ ) = 0 . Var ( A ′ 1 n ) = σ 2 / n Var ( A n ) = 1 . Central limit theorem: As n goes to infinity the distribution of A ′ n approaches the standard normal distribution. � α 1 ∞ e − x 2 / 2 dx . Pr [ A ′ √ n ≤ α ] → 2 π

  59. Coins and normal.. Let X 1 , X 2 ,... be i.i.d. B ( p ) .

Recommend


More recommend