Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Choosing the Summary Statistics and the Acceptance Rate in Approximate Bayesian Computation (ABC) Michael G.B. Blum Laboratoire TIMC-IMAG, CNRS, Grenoble COMPSTAT 2010; Thursday, August 26
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion A typical application of ABC in population genetics Estimating the time T since the out-of-Africa migration Recent Out-of-Africa Single Origin Population Past N A T Present non-Africa Africa (a) Model of human (b) Data origins
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Flowchart of ABC Different ¡values ¡ Simulated ¡DNA ¡ Simula'ons ¡ of ¡the ¡parameter ¡ sequences ¡ T ¡ Most ¡probable ¡ Observed ¡DNA ¡ ABC ¡ values ¡for ¡T ¡ sequences ¡
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Rejection algorithm for targeting p ( φ | S ) Generate a parameter φ according to the prior distribution 1 π ; Simulate data D ′ according to the model p ( D ′ | φ ) ; 2 Compute the summary statistic S ′ from D ′ and accept the 3 simulation if d ( S , S ′ ) < δ . Potential problem : the curse of dimensionality limits the number of statistics that rejection-ABC can handle.
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Regression-adjustment for ABC Beaumont, Zhang and Balding; Genetics 2002 Local linear regression φ i | S i = m ( S i ) + ǫ i , with a linear function for m . Adjustment i = ˆ φ ∗ m ( S ) + ˜ ǫ i , ˆ m is found with weighted least-squares.
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Regression-adjustment for ABC Weighted least-squares n � { φ i − ( β 0 + ( S i − S ) T β 1 ) } 2 W i , i = 1 where W i ∝ K ( || S − S i || /δ ) . Adjustment ǫ i = φ i − ( S i − S ) T ˆ i = ˆ β 0 β 1 φ ∗ LS + ˜ LS .
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Regression-adjustment for ABC φ i * φ i Csilléry, Blum, Gaggiotti and François; TREE 2010
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Asymptotic theorem for ABC Blum; JASA 2010 If there is a local homoscedastic relationship between φ 1 and S , Bias with regression adjustment < Bias with rejection only But 2 Rate of convergence of the MSE = θ ( n − 4 / ( d + 5 ) ) d = dimension of the summary statistics n = number of simulations
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion A Gaussian example to illustrate potential pitfalls with ABC Toy example 1 : Estimation of σ 2 σ 2 Inv χ 2 ( d . f . = 1 ) ∼ N ( 0 , σ 2 ) µ ∼ N = 50 Summary statistics ( S 1 , . . . , S 5 ) = (¯ x N , s 2 N , u 1 , u 2 , u 3 ) u j ∼ N ( 0 , 1 ) , j = 1 , 2 , 3
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion A Gaussian example to illustrate potential pitfalls with ABC 1 summary statistic 5 summary statistics 150 150 ● ● Empirical Variance Empirical Variance ● ● Accepted ● ● ● ● ● ● ● ● ● ● ● ● ● Rejected ● ● ● 100 100 ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 50 ● 50 ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0 ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0 ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ●● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0.1 1.0 10.0 100.0 0.1 1.0 10.0 100.0 σ 2 σ 2
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Local Bayesian linear regression Hjort; Book chapter 2003 Prior for the regression coefficients β β ∼ N ( 0 , α − 1 I p + 1 ) The Maximum a posteriori minimizes the regularized weighted least-squares problem n ( φ i − ( S i − S ) T β ) 2 W i + α 1 � 2 β T β. E ( β ) = 2 τ 2 i = 1
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Local Bayesian linear regression Posterior distribution of the regression coefficients β ∼ N ( β MAP , V ) , τ − 2 VX T W δ φ β MAP = V − 1 ( α I p + 1 + τ − 2 X T W δ X ) . = Regression-adjustment for ABC i = φ i − ( S i − S ) T ˆ β 1 φ ∗ MAP .
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion The evidence function as an omnibus criterion Empirical Bayes /Evidence approximation � � � p ( φ | τ 2 , α, p δ ) = Π n i = 1 p ( φ i | β, τ 2 ) W i p ( β | α ) d β, α is the precision hyperparameter τ is the variance of the residuals p δ is the percentage of accepted simulations. Maximizing the evidence for choosing p δ 1 choosing the set of summary statistics 2
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion The evidence function as an omnibus criterion A closed-formed formula p + 1 log α − N W 2 log τ 2 − E ( β MAP ) log p ( φ | τ 2 , α, p δ ) = 2 − 1 2 log | V − 1 | − N W 2 log 2 π, where N W = � W i .
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion The evidence function as an omnibus criterion The evidence as a function of the tolerance rate ( α,τ ) log p ( φ | τ 2 , α, p δ ) . log p ( φ | p δ ) = max The evidence as a function of the set of summary statistics ( α,τ, p δ ) log p ( φ | τ 2 , α, p δ ) . log p ( φ | S ) = max
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Iterative algorithm for maximizing the evidence w.r.t. α and τ Updating the value of the hyperparameter γ α = , β T MAP β MAP where γ is the effective number of summary statistics. γ = ( p + 1 ) − α Tr ( V ) . � n i = 1 ( φ i − ( S i − S ) T β ) 2 W i τ 2 = . N W − γ
Introduction Standard algorithms Potential pitfalls with ABC Local Bayesian linear regression Conclusion Using the evidence for choosing p δ Toy example 2 φ ∼ U − c , c , c ∈ R , e φ � � 1 + e φ , σ 2 = ( . 05 ) 2 S ∼ N , Log evidence ● ● ● ● 100 ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● ● 0 ● ● −150 ● 0.003 0.010 0.100 1.000 Acceptance rate
Recommend
More recommend