Small ball probabilities and metric entropy Frank Aurzada, TU Berlin Sydney, February 2012 MCQMC
Outline Small ball probabilities vs. metric entropy 1 Connection to other questions 2 Recent results for concrete examples 3
Outline Small ball probabilities vs. metric entropy 1 Connection to other questions 2 Recent results for concrete examples 3
Small ball probabilities Let ( X t ) t ≥ 0 be a stochastic process with X 0 = 0 Goal: find asymptotic rate of � � sup | X t | ≤ ε ≈ ? , with ε → 0 P 0 ≤ t ≤ 1 ε 1 X − ε In many examples, � � = e − κε − γ ( 1 + o ( 1 )) , P sup | X t | ≤ ε with ε → 0 0 ≤ t ≤ 1 with γ > 0 und κ > 0.
Small ball probabilities Let ( X t ) t ≥ 0 be a stochastic process with X 0 = 0 Goal: find asymptotic rate of � � sup | X t | ≤ ε ≈ ? , with ε → 0 P 0 ≤ t ≤ 1 ε 1 X − ε In many examples, � � = e − κε − γ ( 1 + o ( 1 )) , P sup | X t | ≤ ε with ε → 0 0 ≤ t ≤ 1 with γ > 0 und κ > 0.
Small ball probabilities Let ( X t ) t ≥ 0 be a stochastic process with X 0 = 0 Goal: find asymptotic rate of � � sup | X t | ≤ ε ≈ ? , with ε → 0 P 0 ≤ t ≤ 1 ε 1 X − ε Therefore, we study � � = κε − γ ( 1 + o ( 1 )) , φ X ( ε ) := − log P sup | X t | ≤ ε with ε → 0 0 ≤ t ≤ 1 the so-called small ball function of X . γ
Entropy numbers Let X be a centred Gaussian random variable with values in a sep. Banach space ( E , || . || ) : i.e. ∀ g ∈ E ′ . � X , g � Gaussian
Entropy numbers Let X be a centred Gaussian random variable with values in a sep. Banach space ( E , || . || ) : i.e. ∀ g ∈ E ′ . � X , g � Gaussian There is a linear operator u : L 2 [ 0 , 1 ] → E belonging to X such that � − 1 � E e i � X , g � = exp 2 || u ′ ( g ) || 2 g ∈ E ′ . , 2 Note: u ( L 2 [ 0 , 1 ]) is the RKHS of X
Entropy numbers Let X be a centred Gaussian random variable with values in a sep. Banach space ( E , || . || ) : i.e. ∀ g ∈ E ′ . � X , g � Gaussian There is a linear operator u : L 2 [ 0 , 1 ] → E belonging to X such that � − 1 � E e i � X , g � = exp 2 || u ′ ( g ) || 2 g ∈ E ′ . , 2 Note: u ( L 2 [ 0 , 1 ]) is the RKHS of X Example: X BM in E = C [ 0 , 1 ] � t ( uf )( t ) = f ( s ) d s ; u : L 2 [ 0 , 1 ] → C [ 0 , 1 ] . 0
Entropy numbers / small ball function On the one hand, we consider the small ball function: � � �� φ X ( ε ) = − log P [ || X || E ≤ ε ] = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1
Entropy numbers / small ball function On the one hand, we consider the small ball function: � � �� φ X ( ε ) = − log P [ || X || E ≤ ε ] = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 On the other hand, the entropy numbers of u : e n ( u ) := inf { ε > 0 | ∃ ε -net of 2 n − 1 points of u ( B L 2 [ 0 , 1 ] ) in E } , where B L 2 [ 0 , 1 ] is the unit ball in L 2 [ 0 , 1 ] (inverse of covering numbers).
Asymptotics We use the following notation Weak asymptotics: a ( ε ) a ( ε ) � b ( ε ) , ε → 0 means lim sup b ( ε ) < ∞ ε → 0 a ( ε ) ≈ b ( ε ) , ε → 0 means a ( ε ) � b ( ε ) and b ( ε ) � a ( ε )
Asymptotics We use the following notation Weak asymptotics: a ( ε ) a ( ε ) � b ( ε ) , ε → 0 means lim sup b ( ε ) < ∞ ε → 0 a ( ε ) ≈ b ( ε ) , ε → 0 means a ( ε ) � b ( ε ) and b ( ε ) � a ( ε ) Strong asymptotics: a ( ε ) a ( ε ) � b ( ε ) , ε → 0 means lim sup b ( ε ) = 1 ε → 0 a ( ε ) ∼ b ( ε ) , ε → 0 means a ( ε ) � b ( ε ) and b ( ε ) � a ( ε ) Similarly for n → ∞
The small ball – entropy connection Theorem (Kuelbs/Li’93, Li/Linde’99, A./Ibragimov/Lifshits/van Zanten’08) For r > 0 and δ ∈ R : φ X ( ε ) � ε − r | log ε | δ e n ( u ) � n − 1 / 2 − 1 / r ( log n ) δ/ r ⇔ φ X ( ε ) � ε − r | log ε | δ e n ( u ) � n − 1 / 2 − 1 / r ( log n ) δ/ r ⇔ where the first ⇐ requires φ X ( ε ) � φ ( 2 ε ) . Further, for δ > 0 and κ > 0, φ X ( ε ) � κ | log ε | δ − log e n ( u ) � κ − 1 /δ n 1 /δ ⇔ φ X ( ε ) � κ | log ε | δ − log e n ( u ) � κ − 1 /δ n 1 /δ . ⇔ small ball pr. ↔ entropy numbers (probabilistic) (functional analytic)
The small ball - entropy connection Example: X Riemann-Liouville process in C [ 0 , 1 ] � t ( t − s ) H − 1 / 2 f ( s ) d s ; ( uf )( t ) = u : L 2 [ 0 , 1 ] → C [ 0 , 1 ] . 0 one has φ X ( ε ) ≈ ε − 1 / H e n ( u ) ≈ n − 1 / 2 − H In particular for X BM, H = 1 / 2 φ X ( ε ) ≈ ε − 2 e n ( u ) ≈ n − 1
Outline Small ball probabilities vs. metric entropy 1 Connection to other questions 2 Recent results for concrete examples 3
Connections of small ball prob. to other questions In the setup of Gaussian processes, there are various connections to: entropy of function classes convergence rate of series representations coding quantities for the process approximation quantitites for the process Chung’s law of the iterated logarithm statistical problems ... Generally: the small ball rate increases the slower the better the process can be approximated the smoother the process is
Connections of small ball prob. to other questions approximation of stochastic processes coding, quantisation, law of the iterated quadrature logarithm n X ( n ) � = ξ i ψ i ( t ) → X t sup s ≤ t | X s | t N lim inf = c i =1 � f ( ˆ E [ f ( X )] ≈ X i ) q i b ( t ) error || X ( n ) − X || → 0 t → 0 i =1 � � = κε − γ (1 + o (1)) φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 path regularity functional analysis Gaussian process entropy numbers of linear operators n -times differentiable ⇒ γ ≤ 1 /n between Banach spaces other approximation quantities such as PDE problems Kolmogorov widths, etc.
Connections of small ball prob. to other questions approximation of stochastic processes coding, quantisation, law of the iterated quadrature logarithm n X ( n ) � = ξ i ψ i ( t ) → X t sup s ≤ t | X s | t N lim inf = c i =1 � f ( ˆ E [ f ( X )] ≈ X i ) q i b ( t ) error || X ( n ) − X || → 0 t → 0 i =1 � � = κε − γ (1 + o (1)) φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 path regularity functional analysis Gaussian process entropy numbers of linear operators n -times differentiable ⇒ γ ≤ 1 /n between Banach spaces other approximation quantities such as PDE problems Kolmogorov widths, etc.
Connections of small ball prob. to other questions approximation of stochastic processes coding, quantisation, law of the iterated quadrature logarithm n X ( n ) � = ξ i ψ i ( t ) → X t sup s ≤ t | X s | t N lim inf = c i =1 � f ( ˆ E [ f ( X )] ≈ X i ) q i b ( t ) error || X ( n ) − X || → 0 t → 0 i =1 � � = κε − γ (1 + o (1)) φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 path regularity functional analysis Gaussian process entropy numbers of linear operators n -times differentiable ⇒ γ ≤ 1 /n between Banach spaces other approximation quantities such as PDE problems Kolmogorov widths, etc.
Connection to smoothness of process Theorem (A.’11) Let ( X t ) t ∈ [ 0 , 1 ] be a centred Gaussian process and n an integer. If (a modif. of) X is n -times differentiable with X ( n ) ∈ L 2 [ 0 , 1 ] then � � � ε − 1 / n . φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 ε 1 X − ε
Connection to smoothness of process Theorem (A.’11) Let ( X t ) t ∈ [ 0 , 1 ] be a centred Gaussian process and n an integer. If (a modif. of) X is n -times differentiable with X ( n ) ∈ L 2 [ 0 , 1 ] then � � � ε − 1 / n . φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 ε 1 X − ε Now, what happens when n (above) is non-integer?
Connection to smoothness of process Define fractional differentiation: Let γ > 0 (recall X 0 = 0) � t X ( γ ) ( t − s ) γ − 1 x ( t ) d t . = x ( t ) if X t = t 0
Connection to smoothness of process Define fractional differentiation: Let γ > 0 (recall X 0 = 0) � t X ( γ ) ( t − s ) γ − 1 x ( t ) d t . = x ( t ) if X t = t 0 Theorem (A.’11) Let ( X t ) t ∈ [ 0 , 1 ] be a centred Gaussian process and γ > 1 / 2. If X ( γ ) exists and X ( γ ) ∈ L 2 [ 0 , 1 ] then � � � ε − 1 /γ . φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1
Connection to smoothness of process Define fractional differentiation: Let γ > 0 (recall X 0 = 0) � t X ( γ ) ( t − s ) γ − 1 x ( t ) d t . = x ( t ) if X t = t 0 Theorem (A.’11) Let ( X t ) t ∈ [ 0 , 1 ] be a centred Gaussian process and γ > 1 / 2. If X ( γ ) exists and X ( γ ) ∈ L 2 [ 0 , 1 ] then � � � ε − 1 /γ . φ X ( ε ) = − log P sup | X t | ≤ ε 0 ≤ t ≤ 1 ‘’Example”: Brownian motion X is γ -times “differentiable” (H¨ older), γ < 1 2 . � � ≈ ε − 2 = ε − 1 1 / 2 . − log P sup | X t | ≤ ε 0 ≤ t ≤ 1
Recommend
More recommend