CAP representations (The right(?) way for generic MR analysis of Internet data ) Amos Ron University of Wisconsin ∗ – Madison WISP: UCSD, November 2004 ∗ : Breaking news, 06/11/04: Wisconsin routed Minnesota 38:14, on its way to the national title. 1
Outline • Possible goals behind “generic analysis on Internet signals” • Why is that a non-trivial task? • Predictability and pyramidal algorithms • Performance of pyramidal representation • CAMP and my favorite pyramidal representation • What parameters to extract from the representation? 2
A mathematical view of Internet signals • Main features in the signal: – burst types – rate of their appearances • This is non-trivial (why? After all, nothing is easier than 1D signals...) – the amount of data may be overwhelming – there is no clear way to judge success • It is also a cultural problem. really? 3
maybe it is time to show some images? 4
5
d 4 d 3 d 2 d 1 6
7
Pyramidal algorithms I: MR representation h : Z Z → I R is a symmetric, normalized, filter: h ( k ) = h ( − k ), � Z h ( k ) = 1. k ∈ Z ↓ , ↑ are downsampling & upsampling: y ↓ ( k ) = y (2 k ) , k ∈ Z Z � 2 y ( k/ 2) , k even, y ↑ ( k ) = 0 , otherwise. Z s.t: j = −∞ ⊂ C Z ( y j ) ∞ y j = Cy j +1 := ( h ∗ y j ) ↓ , ∀ j. C is Compression or Coarsification y j +1 is then predicted from y j by y j +1 ≈ Py j := h ∗ ( y j ↑ ) . P is Prediction or subdivision 8
Pyramidal algorithms II: the detail coefficients • Define the detail coefficients: d j := ( I − PC ) y j = y j − P y j − 1 . • Replace y j by the pair ( y j − 1 , d j ). • Continue iteratively. C C C ✲ ✲ y m − 2 . . . . . . ✲ y m y m − 1 y 1 y 0 I − PC I − PC I − PC I − PC ❄ ❄ ❄ ❄ d m d m − 1 d m − 2 d 1 Reconstruction. Recovering y m from y 0 , d 1 , d 2 , . . . , d m is trivial: y 1 = d 1 + Py 0 , y 2 = d 2 + Py 1 and so on. 9
Wavelet pyramids, Mallat, 1987 Decompose the detail map I − PC : I − PC = RD D : y �→ ( h 1 ∗ y ) ↓ =: w 1 ,j − 1 , R : y �→ h 1 ∗ y ↑ with h 1 a real, symmetric, highpass: � Z h 1 ( k ) = 0. k ∈ Z C C C ✲ ✲ ✲ y m − 2 . . . . . . y m y m − 1 y 1 y 0 ❩❩❩❩❩ ❩❩❩❩❩ ❩❩❩❩❩ D D D ⑦ ⑦ ⑦ w 1 ,m − 1 w 1 ,m − 2 w 1 , 0 Note that we can recover y m from y 0 , w 1 , 0 , w 1 , 1 , . . . , w 1 ,m − 1 since y 1 = Rw 1 , 0 + Py 0 , y 2 = Rw 1 , 1 + Py 1 and so on. 10
Performance • Ability to predict. The best prediction are based on local averaging, and on nothing else =spline predictors • Time blurring: good prediction requires long averaging filter. That blurs spontaneous events. • Internet data exhibit different behaviour at “small” scales than other scales. Hence: non-stationary representation • Standard wavelet systems are mediocre for Internet data: they blur time, and create artifacts, in order to gain unnecessary properties (orthonormality). 11
Poor prediction 40 40 35 35 30 30 25 25 20 20 15 15 10 10 5 5 0 0 0 5 10 15 20 25 30 35 40 0 5 10 15 20 25 30 35 40 12
My favorite representation well, before we conducted any numerical tests Step I: Build an MR representation based on h 1 = 1 4(1 2 1) Step II: Define the detail coefficients by: − y j ( k +1)+2 y j ( k ) − y j ( k − 1) , k even, 4 d j ( k ) = y j ( k − 3) − 9 y j ( k − 1)+16 y j ( k ) − 9 y j ( k +1) − y j ( k +3) , k odd. 16 The “performance grade” here is 2 in the strict sense. (To compare, Haar’s grade is 1 in the non-strict sense, and 0 in the strict sense.) This is an example of a new class of high-performance representations called CAMP 13
what to analyse? what to extract? for p ≥ 1, the p -norm is � | a k | p ) 1 /p � a � p = ( k the best thing to analyse is the “compressibility” of the detail coefficients: choose a number N , then (1) replace the N “most important” detail coefficients by 0, to obtain a signal e N . (2) reconstruct using e N to obtain Y N . ‘ (3) define e p ( N ) := � Y N � p . (4) find the a parameter α such that e p ( N ) ≈ CN − α . 14
α ( p ) = the predictability parameter in the p -norm “most important”=? (1) Non-linear: choose the largest ones (2) Linear: go from coarse scale to fine scale. Output: this way we have two functions p �→ α ( p ) . Goal: learn how to judge properties of your signal based on these two functions it might be that the detail coefficients behave rather differently at different scale (small scale vs. large scale). 15
CAP representations Choose: • two refinable functions φ c , φ r with refinement filters h c , h r . • A third (Auxiliary-Alignment) lowpass filter h a . Decompose: Fix f : I R → C. Z, define y j ( k ) := 2 j/ 2 � f, ( φ c ) j,k � . For all k, j ∈ Z The CAP operators are: C : y �→ ( h c ∗ y ) ↓ , (Coarsification-Compression), A : y �→ Ay := h a ∗ y, (Alignment), P : y �→ Py := h r ∗ ( y ↑ ) , (Prediction-subdivision). Then Cy j +1 = y j , ∀ j . 16
The detail coefficients are: d j := ( A − PAC ) y j = Ay j − PAy j − 1 . This is the CAP representation with ( d j ) the CAP coefficients . C C C ✲ ✲ ✲ y m − 2 . . . . . . y m y m − 1 y 1 y 0 A − PAC A − PAC A − PAC A − PAC ❄ ❄ ❄ ❄ d m d m − 1 d m − 2 d 1 y m is recovered from y 0 , d 1 , d 2 , . . . , d m since Ay 1 = d 1 + PAy 0 , Ay 2 = d 2 + PAy 1 , . . . , Ay m = d m + PAy m − 1 and deconvolving A from Ay m . 17
Summary Do they W F CAP implemented by fast pyramid algorithms ? � � � provides good function space characterizations ? � � � avoid mother wavelets ? � very short filters, with no artifacts ? � � have simple constructions ? � � avoid redundant representations ? � Wavelet are non-redundant. Caplets are only slightly redundant in high dimensions. Their redundancy is non-essential. 18
CAMP representations: Compression-Alignment-Modified Prediction With CAP in hand, one can modify the process s.t.: • The filters are shorter • The performance (:= function space characterization) is the same Example: Assume h is interpolatory. Define the details as: � y j − h ∗ y j , Z d , on 2Z d j := y j − h ∗ ( y j ↓↑ ) , otherwise. Let φ be the refinable function of h . If φ ∈ C s + ǫ , c then the above detail characterize L s p . 19
0 1 / 8 1 / 8 Example (2D): h = . 1 / 8 1 / 4 1 / 8 1 / 8 1 / 8 0 There are four (hidden) filters, for computing d j : 0 − 1 / 8 − 1 / 8 0 0 0 , − 1 / 8 +3 / 4 − 1 / 8 − 1 / 2 +1 − 1 / 2 − 1 / 8 − 1 / 8 0 0 0 0 0 − 1 / 2 0 0 0 − 1 / 2 , 0 +1 0 0 +1 0 0 − 1 / 2 0 − 1 / 2 0 0 Those are 7 , 3 , 3 , 3-tap. There are four (hidden) “CAMPlets”, whose average area of support is about 2. The performance is on par with tensor 3/5, whose filters are 25 , 15 , 15-tap. Each supported in 3 × 3 square. 20
y j 21
d V d D j j d * d H j j Figure 1: First level ˜ d CAMP coefficients, organized by cosets. 22
Wisconsin From left, 1st row: Julia Velikina, Youngmi Hur, Yeon Kim, Narfi Stefansson. 2nd row: Thomas Hangelbroek, Sangnam Nam, Jeff Kline, Steven Parker. 23
Julia Velikina: undersampled MRI data Schepp−Logan phantom Conventional recon. from 90 projections, acceptable quality Conventional recon. from 23 projections, unacceptable quality TV−based recon. from 23 projections 24
Jeff Kline: new data representation in NMR 25
Steven Parker: redundant representation of acoustic signals Adpative framelet−based representation of a vibraphone recording 31 1,1 30 2,1 29 5,7 28 5,6 27 6,11 26 7,21 25 7,20 24 6,9 23 6,8 22 8,31 21 8,30 20 8,29 19 8,28 coefs scaled to V0 18 6,6 17 7,11 16 7,10 15 7,9 14 7,8 13 7,7 12 7,6 11 8,11 10 8,10 9 8,9 8 8,8 7 8,7 6 8,6 5 8,5 4 8,4 3 7,1 2 8,1 1 8,0 131072 original signal (time) 26
Narfi Stefansson: sparse framelet representations 27
6/10 61440 coefficients quartic spline 34452 coefficients cubic spline 34608 coefficients box15,box17,box18 35085 coefficients 28
FrameNet: on-line interactive framelet and wavelet analysis 29
30
31
32
Recommend
More recommend