fundamentals and
play

FUNDAMENTALS AND MULTIPLE USER APPLICATIONS (PART II) Max H. M. - PowerPoint PPT Presentation

INFORMATION THEORY FUNDAMENTALS AND MULTIPLE USER APPLICATIONS (PART II) Max H. M. Costa Unicamp July 2018 LAWCI Unicamp Summary o Introduction o Entropy, K-L Divergence, Mutual Information o Asymptotic Equipartition Property (AEP) o 1.


  1. Changing Z channel into BEC  Show that code {01, 10} can transform a Z channel into a BEC.  What is a lower bound to capacity of the Z channel?

  2. Typewriter type 2: ½  ½ ½ ½ Sum channel: 2 C = 2 C1 + 2 C2 where C 1 = C 2 = 0.322 C = 1,322 bits/channel use How many noise free symbols?

  3. Example: Noisy typewriter ½  A A ½ B B C C X Y D D • E • • Z Z  C = max (H(Y) – H(Y|X))  = 𝑚𝑝𝑕 2 26 − 𝑚𝑝𝑕 2 2 = 𝑚𝑝𝑕 2 13 bits/transmission  Achieved with uniform distribution on the inputs.

  4. Remark:  For this example, we can also achieve  𝐷 = 𝑚𝑝𝑕 2 13 bits/transmission with P(error)=0 and  n = 1 by transmitting alternating input symbols, i.e.,  X = {A C E … Z}.

  5. Differential Entropy  Let 𝑌 be a continuous random variable with density 𝑔 𝑦 and support 𝑇 . The differential entropy of 𝑌 is ℎ 𝑌 = − 𝑔 𝑦 log 𝑔 𝑦 𝑒𝑦 (if it exists). 𝑇 Note: Also written as ℎ 𝑔 .

  6. Examples: Uniform distribution  Let 𝑌 be uniform in the interval 0, 𝑏 . Then 1  𝑔 𝑦 = 𝑏 in the interval and 𝑔 𝑦 = 0 outside. 𝑏 1 1  ℎ 𝑌 = − 𝑏 𝑚𝑝𝑕 𝑏 𝑒𝑦 = log 𝑏 0  Note that ℎ 𝑌 can be negative for 𝑏 < 1.  However, 2 ℎ(𝑔) = 2 log 𝑏 = 𝑏 is the size of the support set, which is non-negative.

  7. Example: Gaussian distribution −𝑦 2 1  Let 𝑌 ~  𝑦 = 2  2 𝑓𝑦𝑞 ( 2  2 ) 𝑦 2 2  2 − 𝑚𝑜 2  2 ] 𝑒𝑦  Then ℎ 𝑌 = ℎ  = −  𝑦 [− 𝐹𝑌 2 1 2 𝑚𝑜 2   2  = 2  2 + 1 2 𝑚𝑜 ( 2  e  2 ) nats  = 1 2 𝑚𝑝𝑕 ( 2  e  2 ) bits  Changing the base we have ℎ 𝑌 =

  8. Relation of Differential and Discrete Entropies  Consider a quantization of X, denoted by X    Let X  = 𝑦 𝑗 i nside the 𝑗 th interval. Then 𝐼(𝑌  ) = - 𝑞 𝑗 𝑚𝑝𝑕 𝑞 𝑗 𝑗 = -  𝑔(𝑦 𝑗 ) 𝑚𝑝𝑕 𝑔(𝑦 𝑗 ) - 𝑚𝑝𝑕  𝑗  ℎ 𝑔 − log 

  9. Differential Entropy  So the two entropies differ by the log of the quantization level  .  We can define joint differential entropy, conditional differential entropy, K-L divergence and mutual information with some care to avoid infinite differential entropies.

  10. K-L divergence and Mutual Information  𝑔  𝐸(𝑔 g) = 𝑔 𝑚𝑝𝑕 𝑕 𝑔(𝑦,𝑧)  𝐽 𝑌; 𝑍 = 𝑔 𝑦, 𝑧 𝑚𝑝𝑕 𝑔 𝑦 𝑔(𝑧) 𝑒𝑦 𝑒𝑧  Thus , I(X;Y) = h(X) + h(Y) – h(X,Y). Note: h(X) can be negative, but I(X;Y) is always  0.

  11. Differential entropy of a Gaussian vector  Theorem: Let 𝒀 be a Gaussian n -dimensional vector with mean  and covariance matrix 𝐿. Then 2 log((2  𝑓) 𝑜 𝐿 ) 1  ℎ 𝒀 =  where 𝐿 denotes the determinant of 𝐿.  Proof: Algebraic manipulation.

  12. The Gaussian Channel Z~N (0, N I) X 𝑋 W Y Channel Channel + Decoder Encoder Power Constraint: EX 2 ≤P

  13. The Gaussian Channel  Z~N (0, N I) X Y 𝑋 W Channel Channel + Decoder Encoder Power constraint: EX 2 ≤P  W  {1,2,…, 2 𝑜𝑆 } = message set of rate R  X = (x 1 x 2 … x n ) = codeword input to channel  Y = (y 1 y 2 … y n ) = codeword output from channel = decoded message P(error) = P{ W  𝑋}  𝑋

  14. The Gaussian Channel  Using the channel n times: X n • Y n • • • • • • • • •

  15. The Gaussian Channel   C𝑏𝑞𝑏𝑑𝑗𝑢𝑧 𝐷 = max 𝐽(𝑌; 𝑍) f(x): EX 2 ≤P  𝐽 𝑌; 𝑍 = ℎ 𝑍 − ℎ 𝑍 𝑌 = ℎ 𝑍 − ℎ 𝑌 + 𝑎|𝑌 1 1  = ℎ 𝑍 − ℎ 𝑎 ≤ 2 log 2  e 𝑄 + 𝑂 − 2 log 2  e 𝑂 1 𝑄  = 2 log 1 + 𝑂 bits/transmission

  16. The Gaussian Channel   The capacity of the discrete time additive Gaussian channel: 1 𝑄  𝐷 = 2 log 1 + 𝑂 bits/transmission  achieved with X ~ N(0 , P).

  17. Bandlimited Gaussian Channel  Consider the channel with continuous waveform inputs x(t) 𝑈 1 𝑈 𝑦 2 𝑢 𝑒𝑢 ≤ 𝑄) and Bandwidth with power constraint ( 0 limited to W. The channel has white Gaussian noise with power spectral density N 0 /2 watt/Hz.  In the interval (0,T) we can specify the code waveform by 2WT samples (Nyquist criterion). We can transmit these samples over discrete time Gaussian channels with noise variance N 0 /2. This gives 𝑄  𝐷 = 𝑋 log ( 1+ 𝑂 0 𝑋 ) bit/second

  18. Bandlimited Gaussian Channel  𝑄  𝐷 = 𝑋 log ( 1+ 𝑂 0 𝑋 ) bit/second  Note: If W   𝑄 𝑂 0 𝑚𝑝𝑕 2 𝑓 bits/second.  we have C =

  19. Bandlimited Gaussian Channel 𝑆 𝑋 be the spectral density  in bits per second  Let per Hertz. Also let 𝑄 = 𝐹 𝑐 𝑆 where 𝐹 𝑐 is the available energy per information bit.  We get 𝑆 𝐷 𝐹 𝑐 𝑆  𝑋 ≤ 𝑋 = log ( 1+ 𝑂 0 𝑋 ) bit/second.  Thus 2  −1 𝐹 𝑐  𝑂 0 ≥  This relation defines the so called Shannon Bound.

  20. The Shannon Bound 2  −1 𝐹 𝑐  𝑂 0 ≥  𝐹 𝑐 𝐹 𝑐   𝑂 0 (dB) – 𝑂 0 Shannon Bound  0 0.69 -1.59 – 5 0.1 0.718 -1.44 – • 4 0.25 0.757 -1.21 0.5 0.828 -0.82 – 3 1 1 0 – • 2 2 1.5 1.76 𝐹 𝑐 – • 𝑂 0 (dB) 1 4 3.75 5.74  8 31.87 15.03        • -1 0 1 2 3 4 5 6

  21. Shannon’s Water Filling Solution

  22. Parallel Gaussian Channels  3 2.5 2 1

  23. Example of Water Filling  Channels with noise levels 2, 1 and 3.  Available power = 2 1 0.5 1 1.5 1 0  Capacity= 2 log (1+ 2 ) + 2 log (1+ 1 ) + 2 log (1+ 3 )  Level of noise + signal power = 2.5  No power allocated to the third channel.

  24. Parallel Gaussian Channels  3 2.5 2 1

  25. Differential capacity Discrete memoryless channel as a band limited channel

  26. Multiplex strategies (TDMA, FDMA)  P j  j 𝑘 = 1 (1 + 𝑄 𝐷 2 log Aggregate capacity: : 𝑂 )

  27. Multiplex strategies (non-orthonal CDMA) P j  1 𝑄 2 log (1 + 𝑂+ 𝑘−1 𝑄 ) j Discrete memoryless channel as a band limited channel M 𝑘 = 1 (1 + 𝑁𝑄 𝐷 2 log 𝑂 ) Aggregate capacity: : j=1

  28. TDMA or FDMA versus CDMA Orthogonal schemes:  Bandwidth limitation (2WT dimensions) Number of Users Non-orthogonal CDMA (log has no cap) Aggregate Power

  29. Multiple User Information Theory  Building Blocks:  Multiple Access Channels (MACs)  Broadcast Channels (BCs)  Interference Channels (IFCs)  Relay Channels (RCs)  Note: These channels have their discrete memoryless and Gaussian versions. For simplicity we will look at the Gaussian models.

  30. Multiple Access Channel  A well understood model.  Models the uplink channel in wireless comm. X 1 W 1 Encoder ^ ^ W 1 ,W 2 Y P(y|x 1 ,x 2 ) Decoder X 2 Encoder W 2 Capacity region obtained by Ahlswede (1971) and Liao (1972)

  31. Capacity region - MAC  C = closure of convex hull of { (R1,R2) s.t.  R 1 ≤ I(X 1 ; Y | X 2 ),  R 2 ≤ I(X 2 ; Y | X 1 ),  R 1 + R 2 ≤ I(X 1 , X 2 ; Y ) for all p 1 (x 1 ) . p 2 (x 2 ) }

  32. Multiple Access Channel (MAC)

  33. Broadcast Channel (Cover, 1972)  Still open in discrete memoryless case.  Models the downlink channel in wireless comm. Y 1 ^ W 1 Decoder X W 1 ,W 2 Encoder P(y 1 ,y 2 |x) Y 2 ^ W 2 Decoder Cover introduced superposition coding.

  34. Superposition coding  Message Clouds Y 1 can see the message, Y 2 can see the cloud center

  35. Gaussian Broadcast Channel

  36. Superposition coding (1-  )P N 2  P 1

  37. Superposition coding (1-  )P N 2  P 1

  38. Interference Channel  Gaussian Interference Channel - standard form  Brief history  Z-Interference channel  Symmetric Interference channel

  39. Standard Gaussian Interference Channel Power P1 ^ W 1 W 1 a b ^ W 2 W 2 Power P2

  40. Symmetric Gaussian Interference Channel Power P Power P

  41. Z-Gaussian Interference Channel

  42. Interference Channel: Strategies Things that we can do with interference: Ignore (take interference as noise (IAN) 1. Avoid (divide the signal space (TDM/FDM)) 2. Partially decode both interfering signals 3. Partially decode one, fully decode the other 4. Fully decode both (only good for strong inter- 5. ference , a≥1)

  43. Interference Channel: Brief history  Carleial (1975): Very strong interference does not reduce capacity (a 2 ≥ 1+P)  Sato (1981), Han and kobayashi (1981): Strong interference (a 2 ≥ 1) : IFC behaves like 2 MACs  Motahari, Khandani (2007), Shang, Kramer and Chen (2007), Annapureddy, Veeravalli (2007): Very weak interference (2a(1+a 2 P) ≤ 1 ) : Treat interference as noise – (IAN)

  44. Interference Ch.: History (continued)  Sason (2004): Symmetrical superposition to beat TDM  Etkin, Tse, Wang (2008): capacity to within 1 bit  Polyanskiy and Wu, 2016: Corner points established.

  45. Summary: Z interference Channels  Z-Gaussian Interference Channel as a degraded interference channel  Discrete Memoryless Channel as a band limited channel  Multiplex Region: growing Noisebergs  Overflow Region: back to superposition

  46. Z-Gaussian Interference Channel

  47. Degraded Gaussian Interference Channel

  48. Differential capacity Discrete memoryless channel as a band limited channel

  49. Interference x Broadcast Channels

  50. Superposition coding (1-  )P N 2  P 1

  51. Superposition coding (1-  )P N 2  P 1

  52. Degraded Interference Channel - One Extreme Point

  53. Degraded Interference Channel - Another Extreme Point

  54. Intermediary Points (Multiplex Region)

  55. Admissible region for (  , h)

  56. Intermediary Point (Overflow Region)

  57. Admissible region h Q 1 =1 Q 2 = 1 a = 0.5 N 2 = 3 

  58. The Z-Gaussian Interference Channel Rate Region R 2 Q 1 =1 Q 2 = 1 a = 0.5 N 2 = 3 R 1

  59. Admissible region h Q 1 =1 Q 2 = 1 a = 0.99 N 2 = 0.02 

  60. The Z-Gaussian Interference Channel Rate Region R 2 Q 1 =1 Q 2 = 1 a = 0.99 N 2 = 0.02 R 1

Recommend


More recommend