defining perceived information based on shannon s
play

Defining Perceived Information based on Shannons Communication - PowerPoint PPT Presentation

Defining Perceived Information based on Shannons Communication Theory Cryptarchi 2016 June 21-24, 2016 La Grande Motte, France Eloi de Chrisey, Sylvain Guilley, & Olivier Rioul Tlcom ParisTech, Universit Paris-Saclay, France.


  1. Defining Perceived Information based on Shannon’s Communication Theory Cryptarchi 2016 June 21-24, 2016 La Grande Motte, France Eloi de Chérisey, Sylvain Guilley, & Olivier Rioul Télécom ParisTech, Université Paris-Saclay, France.

  2. Contents Introduction Motivation Assumptions and Notations How to Define Perceived Information? Markov Chain From MAP to PI Application of Shannon’s Theory Minimum Number of Traces Worst Possible Case for Designers Link with Perceived Information Conclusion 2 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  3. Contents Introduction Motivation Assumptions and Notations How to Define Perceived Information? Markov Chain From MAP to PI Application of Shannon’s Theory Minimum Number of Traces Worst Possible Case for Designers Link with Perceived Information Conclusion 3 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  4. Motivation Consolidate the state of the art about Perceived Information (PI) metrics; Continue the work of Annelie Heuser presented last year at CryptArchi; Establish clear and coherent definitions for PI based on optimal distinguishers and Shannon’s theory ; 4 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  5. Motivation Consolidate the state of the art about Perceived Information (PI) metrics; Continue the work of Annelie Heuser presented last year at CryptArchi; Establish clear and coherent definitions for PI based on optimal distinguishers and Shannon’s theory ; Deduce tests in order to evaluate the success of an attack; 4 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  6. Motivation Consolidate the state of the art about Perceived Information (PI) metrics; Continue the work of Annelie Heuser presented last year at CryptArchi; Establish clear and coherent definitions for PI based on optimal distinguishers and Shannon’s theory ; Deduce tests in order to evaluate the success of an attack; Introduce communication channels in Side-Channel Analysis (SCA). Is Shannon’s channel capacity useful in SCA? 4 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  7. Assumptions and Notations What is an attack? Two phases: profiling phase & attacking phase. 5 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  8. Assumptions and Notations What is an attack? Two phases: profiling phase & attacking phase. Profiling phase : secret key ˆ q textbytes ˆ k is known. A vector of ˆ t is given and ˆ q traces ˆ x are measured; Attacking phase : secret key ˜ k is unknown. A vector of ˜ q textbytes ˜ t is given and ˜ q traces ˜ x are measured; 5 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  9. Assumptions and Notations What is an attack? Two phases: profiling phase & attacking phase. Profiling phase : secret key ˆ q textbytes ˆ k is known. A vector of ˆ t is given and ˆ q traces ˆ x are measured; Attacking phase : secret key ˜ k is unknown. A vector of ˜ q textbytes ˜ t is given and ˜ q traces ˜ x are measured; The leakages follow some unknown distribution P ; x , ˆ x , ˜ Estimate P based on either ˆ t or ˜ t . 5 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  10. Assumptions and Notations (Cont’d) Consider the following sets and variables. X and ˜ ˆ X for ˆ x and ˜ x . T and ˜ ˆ T for ˆ t and ˜ t . 6 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  11. Assumptions and Notations (Cont’d) Consider the following sets and variables. X and ˜ ˆ X for ˆ x and ˜ x . T and ˜ ˆ T for ˆ t and ˜ t . Random variable ˆ X , ˜ X , ˆ T and ˜ T . Random vectors ˆ X , ˜ X , ˆ T and ˜ T . Generic notation x (either profiling or attacking) 6 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  12. Leakage Model Noise y x ¯ k ∗ Algorithmic Distinguish Emanation k t t Recall our notational conventions: − profiling phase with a hat ˆ • . − attacking phase with a tilde ˜ • . 7 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  13. Leakage Equivalent Flow-Graph Y X ¯ Leakage K Model Distinguisher K side information T Markov Chain We have the following Markov Chain given T : → ¯ K − → Y − → X − K The attacker receives X . 8 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  14. Estimations of the Probability Distribution P Definition (Profiled Estimation: OffLine) ˆ q P ( x, t ) = 1 � ˆ ∀ x, t (1) 1 ˆ x i = x, ˆ t i = t ˆ q i =1 9 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  15. Estimations of the Probability Distribution P Definition (Profiled Estimation: OffLine) ˆ q P ( x, t ) = 1 � ˆ ∀ x, t (1) 1 ˆ x i = x, ˆ t i = t q ˆ i =1 Definition (On-the-fly Estimation: OnLine) q ˜ P ( x, t ) = 1 � ˜ ∀ x, t (2) 1 ˜ x i = x, ˜ t i = t ˜ q i =1 9 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  16. Optimal Distinguisher Theorem (Optimal Distinguisher) The optimal distinguisher [2] is the maximum a posteriori (MAP) distinguisher defined by x , ˜ x , ˜ D Opt (˜ t ) = arg max P ( k | ˜ t ) (3) 10 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  17. Optimal Distinguisher Theorem (Optimal Distinguisher) The optimal distinguisher [2] is the maximum a posteriori (MAP) distinguisher defined by x , ˜ x , ˜ D Opt (˜ t ) = arg max P ( k | ˜ t ) (3) As P is unknown, we may replace it by ˆ P in the distinguisher : t ) = arg max ˆ x , ˜ x , ˜ D (˜ P ( k | ˜ t ) (4) 10 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  18. Contents Introduction Motivation Assumptions and Notations How to Define Perceived Information? Markov Chain From MAP to PI Application of Shannon’s Theory Minimum Number of Traces Worst Possible Case for Designers Link with Perceived Information Conclusion 11 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  19. SCA Seen as a Markov Chain Theorem (SCA as a Markov Chain) The following is a Markov Chain: → ( ¯ ( K, T ) − → ( Y , T ) − → ( X , T ) − K, T ) In other words: as T is known everywhere we can put it at every stage. Therefore, Mutual Information I ( K, T ; X , T ) is a relevant quantity. 12 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  20. Mutual Information Theorem (i.i.d. Channel) For an i.i.d. channel, we have: I ( K, T ; X , T ) = q · I ( K, T ; X, T ) (5) The relevant quantity becomes I ( K, T ; X, T ) . Proof. Using independence, I ( K, T ; X , T ) = H ( X , T ) − H ( X , T | K, T ) = q · H ( X, T ) − H ( X | K, T ) = q · H ( X, T ) − qH ( X | K, T ) = q · I ( K, T ; X, T ) 13 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  21. The Role of Perceived Information Mutual Information I ( K, T ; X, T ) is important in order to evaluate the attack. We have: I ( K, T ; X, T ) = H ( K, T ) − H ( K, T | X, T ) (6) � �� � � �� � = H ( K )+ H ( T ) = H ( K | X,T ) 14 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  22. The Role of Perceived Information Mutual Information I ( K, T ; X, T ) is important in order to evaluate the attack. We have: I ( K, T ; X, T ) = H ( K, T ) − H ( K, T | X, T ) (6) � �� � � �� � = H ( K )+ H ( T ) = H ( K | X,T ) giving � � � I ( K, T ; X, T ) = H ( K ) + H ( T ) − P ( k ) P ( t ) P ( x | k, t ) log P ( k | x, t ) . k t x (7) 14 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  23. The Role of Perceived Information (Cont’d) Issues P ( k | x, t ) is unknown! It has to be estimated: ˆ P and ˜ P . How to use ˆ P and ˜ P in order to estimate the Mutual Information? 15 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  24. The Role of Perceived Information (Cont’d) Issues P ( k | x, t ) is unknown! It has to be estimated: ˆ P and ˜ P . How to use ˆ P and ˜ P in order to estimate the Mutual Information? Answer We define the Perceived Information as the estimation of Mutual Information using the MAP distinguisher. 15 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  25. Deriving the Perceived Information The MAP distinguishing rule is given by MAP = arg max ˆ x , ˜ P ( k | ˜ t ) q ˜ � ˆ = arg max P ( k | x i , t i ) i =1 � P ( k | x, t ) ˜ ˆ n x,t = arg max x,t � ˜ P ( x, t | k ) log ˆ = arg max P ( k | x, t ) x,t � � ˜ P ( x | k, t ) log ˆ ˜ = arg max P ( t | k ) P ( k | x, t ) x t 16 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

  26. The Role of Perceived Information (Cont’d) One obtains � � ˜ P ( x | k, t ) log ˆ ˜ MAP = arg max P ( t | k ) P ( k | x, t ) (8) t x Summming over P ( k ) and adding H ( K ) + H ( T ) yields the form � � � ˜ ˜ P ( x | k, t ) log ˆ H ( K ) + H ( T ) + P ( k ) P ( t ) P ( k | x, t ) k t x 17 Télécom ParisTech Defining PI Thanks to Shannon June 21-24, 2016

Recommend


More recommend