computational information games a minitutorial part i
play

Computational Information Games A minitutorial Part I Houman Owhadi - PowerPoint PPT Presentation

Computational Information Games A minitutorial Part I Houman Owhadi ICERM June 5, 2017 DARPA EQUiPS / AFOSR award no FA9550-16-1-0054 (Computational Information Games) Probabilistic Numerical Methods Statistical Inference approaches to


  1. Computational Information Games A minitutorial Part I Houman Owhadi ICERM June 5, 2017 DARPA EQUiPS / AFOSR award no FA9550-16-1-0054 (Computational Information Games)

  2. Probabilistic Numerical Methods Statistical Inference approaches to numerical approximation and algorithm design http://probabilistic-numerics.org/ http://oates.work/samsi

  3. 3 approaches to inference and to dealing with uncertainty 3 approaches to Numerical Approximation

  4. Game theory John Von Neumann John Nash J. Von Neumann. Zur Theorie der Gesellschaftsspiele. Math. Ann. , 100(1):295 – 320, 1928 J. Von Neumann and O. Morgenstern. Theory of Games and Economic Behavior . Princeton University Press, Princeton, New Jersey, 1944. N. Nash. Non-cooperative games. Ann. of Math. , 54(2), 1951.

  5. Deterministic zero sum game Player II 3 -2 Player I Player I’s payoff -2 1 How should I & II play the (repeated) game?

  6. Worst case approach Player II 3 -2 Player I -2 1 II should play blue and lose 1 in the worst case

  7. Worst case approach Player II 3 -2 Player I -2 1 I should play red and lose 2 in the worst case

  8. No saddle point Player II 3 -2 Player I -2 1

  9. Average case (Bayesian) approach Player II 1/2 -1/2 3 -2 -2 1

  10. Mixed strategy (repeated game) solution Player II 3 -2 -2 1 II should play red with probability 3/8 and win 1/8 on average

  11. Mixed strategy (repeated game) solution Player I 3 -2 -2 1 I should play red with probability 3/8 and lose 1/8 on average

  12. Game theory Optimal strategies are mixed strategies Player II Optimal way to q 1 − q play is at random p 3 -2 John Von Neumann Player I 1 − p -2 1 Saddle point

  13. The optimal mixed strategy is determined by the loss matrix Player II p 5 -2 Player I 1 − p -2 1 II should play red with probability 3/10 and win 1/8 on average

  14. Bayesian/probabilistic approach not new but appears to have remained overlooked Pioneering work “ “ These concepts and techniques These concepts and techniques have attracted little attention have attracted little attention among numerical analysts” (Larkin, 1972) among numerical analysts” (Larkin, 1972)

  15. Bayesian Numerical Analysis P. Diaconis A. O’ Hagan J. E. H. Shaw

  16. Information based complexity J. F. Traub H. Wozniakowski G. W. Wasilkowski E. Novak

  17. Compute P. Diaconis Numerical Analysis Approach

  18. Compute Bayesian Approach

  19. E.g.

  20. E.g. E.g.

  21. Q − div( a ∇ u ) = g, x ∈ Ω , (1) x ∈ ∂ Ω , u = 0 , ∂ Ω is piec. Lip. Ω ⊂ R d a i,j ∈ L ∞ ( Ω ) a unif. ell. Approximate the solution space of (1) with a finite dimensional space

  22. Numerical Homogenization Approach Work hard to find good basis functions Harmonic Coordinates Babuska, Caloz, Osborn, 1994 Allaire Brizzi 2005; Owhadi, Zhang 2005 Kozlov, 1979 [Hou, Wu: 1997]; [Efendiev, Hou, Wu: 1999] MsFEM [Fish - Wagiman, 1993] Variational Multiscale Method, Orthogonal decomposition Nolen, Papanicolaou, Pironneau, 2008 Projection based method Engquist, E, Abdulle, Runborg, Schwab, et Al. 2003-... HMM Flux norm Berlyand, Owhadi 2010; Symes 2012 Harmonic continuation

  23. Bayesian Approach − div( a ∇ u ) = g, x ∈ Ω , x ∈ ∂ Ω , u = 0 , Proposition Put a prior on g Compute E u ( x ) fi nite no of observations

  24. Bayesian approach Replace g by ξ ξ : White noise Gaussian fi eld with covariance function Λ ( x, y ) = δ ( x − y ) R ¡ ¢ ⇔ ∀ f ∈ L 2 ( Ω ), 0 , k f k 2 Ω f ( x ) ξ ( x ) dx is N L 2 ( Ω )

  25. x 1 Let Ω x 1 , . . . , x N ∈ Ω x N Theorem x i a = I d [Harder-Desmarais, 1972] [Duchon 1976, 1977,1978] a i,j ∈ L ∞ ( Ω ) [Owhadi-Zhang-Berlyand 2013]

  26. Standard deviation of the statistical error bounds/controls the worst case error Theorem

  27. Summary The Bayesian approach leads to old and new quadrature rules. Statistical errors seem to imply/control deterministic worst case errors Questions • Why does it work? • How far can we push it? • What are its limitations? • How can we make sense of the process of randomizing a known function?

  28. L g u Given u fi nd g Direct Problem Given g fi nd u Inverse Problem u and g live in in fi nite dimensional spaces Direct computation is not possible

  29. L g u Reduced operator ∈ R m Inverse Problem R m g m u m Numerical implementation requires computation with partial information. φ 1 , . . . , φ m ∈ B ∗ 1 u m = ([ φ 1 , u ] , . . . , [ φ m , u ]) u m ∈ R m Missing information u ∈ B 1

  30. Fast Solvers Multigrid Methods Multigrid: [Fedorenko, 1961, Brandt, 1973, Hackbusch, 1978] Multiresolution/Wavelet based methods [Brewster and Beylkin, 1995, Beylkin and Coult, 1998, Averbuch et al., 1998] Robust/Algebraic multigrid [Mandel et al., 1999,Wan-Chan-Smith, 1999, [Panayot - 2010] Xu and Zikatanov, 2004, Xu and Zhu, 2008], [Ruge-St¨ uben, 1987] Stabilized Hierarchical bases, Multilevel preconditioners [Vassilevski - Wang, 1997, 1998] [Chow - Vassilevski, 2003] [Panayot - Vassilevski, 1997] [Aksoylu- Holst, 2010] Low rank matrix decomposition methods Fast Multipole Method: [Greengard and Rokhlin, 1987] Hierarchical Matrix Method: [Hackbusch et al., 2002] [Bebendorf, 2008]:

  31. Common theme between these methods Computation is done with partial information over hierarchies of levels of complexity Restriction Interpolation To compute fast we need to compute with partial information

  32. The process of discovery of interpolation operators is based on intuition, brilliant insight, and guesswork Missing information Problem This is one entry point for statistical inference into Numerical analysis and algorithm design

  33. A simple approximation problem Based on the information that Φ : Known m × n Φ x = y rank m matrix ( m < n ) y : Known element of R m

  34. Worst case approach (Optimal Recovery) Problem

  35. Solution

  36. Average case approach (IBC) Problem

  37. Solution

  38. Adversarial game approach Player II Player I Min Min Max Max

  39. Loss function Player I Player II No saddle point of pure strategies

  40. Randomized strategy for player I Player I Player II Min Min Max Max

  41. Loss function Saddle point

  42. Canonical Gaussian field

  43. Equilibrium saddle point Player I Player II

  44. Statistical decision theory Abraham Wald A. Wald. Statistical decision functions which minimize the maximum risk. Ann. of Math. (2) , 46:265 – 280, 1945. A. Wald. An essentially complete class of admissible decision functions. Ann. Math. Statistics , 18:549 – 555, 1947. A. Wald. Statistical decision functions. Ann. Math. Statistics , 20:165 – 205, 1949.

  45. The game theoretic solution is equal to the worst case solution

  46. Generalization

  47. Examples L

  48. Canonical Gaussian field

  49. Canonical Gaussian field

  50. Canonical Gaussian field

  51. Examples L

  52. The recovery problem at the core of Algorithm Design and Numerical Analysis To compute fast we need to compute with partial information Restriction Interpolation Missing information Problem

  53. Player I Player II Max Max Min Min

  54. Examples Player II Player I Player II Player I

  55. Loss function Player I Player II No saddle point of pure strategies

  56. Randomized strategy for player I Player I Player II Min Min Max Max

  57. Loss function Theorem But

  58. Loss function Theorem Definition

  59. Theorem

  60. Theorem

  61. Game theoretic solution = Worst case solution Optimal Recovery Solution

  62. Optimal bet of player II Gamblets

  63. Gamblets = Optimal Recovery Splines Optimal Recovery Splines

  64. Dual bases

  65. Example ( − div( a ∇ u ) = g, x ∈ Ω , x ∈ ∂ Ω , u = 0 ,

  66. Your best bet on the value of u ψ i ψ i given the information that R R τ j u = 0 for j 6 = i τ i u = 1 and

  67. Example

  68. Example

  69. Example x 1 Ω φ i ( x ) = δ ( x − x i ) x m x i ψ i : Polyharmonic splines [Harder-Desmarais, 1972] [Duchon 1976, 1977,1978]

  70. Example a i,j ∈ L ∞ ( Ω ) x 1 Ω φ i ( x ) = δ ( x − x i ) x m x i ψ i : Rough Polyharmonic splines [Owhadi-Zhang-Berlyand 2013]

  71. Example

  72. Example

Recommend


More recommend