bounds on reliable boolean function computation with
play

Bounds on Reliable Boolean Function Computation with Noisy Gates - - PowerPoint PPT Presentation

Bounds on Reliable Boolean Function Computation with Noisy Gates - R. L. Dobrushin & S. I. Ortyukov, 1977 - N. Pippenger, 1985 - P . G acs & A. G al, 1994 Presenter: Da Wang 6.454 Graduate Seminar in Area I EECS, MIT Oct. 5,


  1. Bounds on Reliable Boolean Function Computation with Noisy Gates - R. L. Dobrushin & S. I. Ortyukov, 1977 - N. Pippenger, 1985 - P . G´ acs & A. G´ al, 1994 Presenter: Da Wang 6.454 Graduate Seminar in Area I EECS, MIT Oct. 5, 2011 0 / 33

  2. Question Given a network of noisy logic gates, what is the redundancy required if we want to compute the a Boolean function reliably? noisy: gates produce the wrong output independently with error probability no more than ε . reliably: the value computed by the entire circuit is correct with probability at least 1 − δ redundancy: minimum #gates needed for reliable computation in noisy circuit minimum #gates needed for reliable computation in noiseless circuit ◮ noisy/noiseless complexity ◮ may depend on the function of interest ◮ upper bound: achievability ◮ lower bound: converse 1 / 33

  3. Part I Lower Bounds for the Complexity of Reliable Boolean Circuits with Noisy Gates 2 / 33

  4. History of development [Dobrushin & Ortyukov 1977] ◮ Contains all the key ideas ◮ Proofs for a few lemmas are incorrect [Pippenger & Stamoulis & Tsitsiklis 1990] ◮ Pointed out the errors in [DO1977] ◮ Provide proofs for the case of computing the parity function [G´ acs & G´ al 1994] ◮ Follow the ideas in [DO1977] and provide correct proofs ◮ Also prove some stronger results In this talk We will mainly follow the presentation in [G´ acs & G´ al 1994]. 3 / 33

  5. Problem formulation System Model Boolean circuit C Gate g a function g : { 0 , 1 } n g → { 0 , 1 } a directed acycic graph ◮ n g : fan-in of the gate node ∼ gate edge ∼ in/out of a gate Basis Φ Assumptions a set of possible gate functions each gate g has constant number of fan-ins n g . e.g., Φ = { AND, OR, XOR } f can be represented by complete basis compositions of gate functions in for circuit C : Φ C Φ C . maximum fan-in in C : n (Φ C ) 4 / 33

  6. Problem formulation Error models ( ε, p ) Gate error Circuit error A gate fails if its output value C ( x ) : random variable for output for z ∈ { 0 , 1 } n g is different from of circuit C on input x . g ( z ) A circuit computes f with error gates fail independently with probability at most p if ◮ fixed probability ε P [ C ( x ) � = f ( x )] ≤ p used for lower bound proof ◮ probability at most ε for any input x . ε ∈ (0 , 1 / 2) 5 / 33

  7. Problem formulation Sensitivity of a Boolean function Let f : { 0 , 1 } n → { 0 , 1 } be a Boolean function with binary input vector x = ( x 1 , x 2 , . . . , x n ) . Let x l be a binary vector that differs from x only in the l -th bit, i.e., � x i i � = l x l i = i = l . ¬ x i f is sensitive to the l th bit on x if f ( x l ) � = f ( x ) . Sensitivity of f on x : #bits in x that f is sensitive to. ◮ “effecitive” input size Sensitivity of f : maximum over all x . 6 / 33

  8. Asymptotic notations f ( n ) = O ( g ( n )) : � f ( n ) � � � lim sup � < ∞ , � � g ( n ) n →∞ � f ( n ) = Ω ( g ( n )) : � � f ( n ) � � lim inf � ≥ 1 , � � g ( n ) n →∞ � f ( n ) = Θ ( g ( n )) : f ( n ) = O ( g ( n )) and f ( n ) = Ω ( g ( n )) 7 / 33

  9. Main results Theorem: number of gates for reliable computation ◮ Let ε and p be any constants such that ε ∈ (0 , 1 / 2) , p ∈ (0 , 1 / 2) . ◮ Let f be any Boolean function with sensitivity s . Under the error model ( ε, p ) , the number of gates of the curcuit is Ω ( s log s ) . Corollary: redundancy of noisy computation For any Boolean function of n variables and with O ( n ) noiseless complexity and Ω ( n ) sensitivity, the redundancy of noisy computation is Ω (log n ) . ◮ e.g., nonconstant symmetric function of n variables has redundancy Ω (log n ) 8 / 33

  10. Equivalence result for wire failures Lemma 3.1 in Dobrushin&Ortyukov ◮ Let ε ∈ (0 , 1 / 2) and δ ∈ [0 , ε/n (Φ C )] . ◮ Let y and t be the vector that a gate receives when the wire fail and does not fail respectively. For any gate g in the circuit C there exists unique values η g ( y , δ ) such that if ◮ the wires of C fails independently with error probability δ , and ◮ the gate g fails with probability η g ( y , δ ) when receiving input y , then the probability that the output of g is different from g ( t ) is equal to ε . Insights Independent gate failures can be “simulated” by independently wire failures and corresponding gate failures. These two failure modes are equivalent in the sense that the circuit C computes f with the same error probability. 9 / 33

  11. “Noisy-wires” version of the main result Theorem ◮ Let ε and p be any constants such that ε ∈ (0 , 1 / 2) , p ∈ (0 , 1 / 2) . ◮ Let f be any Boolean function with sensitivity s . Let C be a circuit such that ◮ its wires fail independently with fixed probability δ , and ◮ each gate fails independently with probability η g ( y , δ ) when receiving y . Suppose C computes f with error probability at most p . Then the number of gates of the curcuit is Ω ( s log s ) . 10 / 33

  12. Error analysis Function and circuit inputs Maximal sensitive set S for f s > 0 : sensitivity of f z : an input vector with s bits that f is sensitive to ◮ an input vector where f has maximum sensitivity S : the set of sensitive bits in z ◮ key object B l : edges originated from l -th input 1 m l � | B l | 2 e.g. ◮ l = 3 3 ◮ B l f ( z ) 4 ◮ m l = 3 11 / 33

  13. Error analysis Wire failures For β ⊂ B l , let H ( β ) be the event that w 1 for wires in B l , only those in β fail. w 2 input l Let w 3 � H ( β ) C ( z l ) = f ( z l ) � β l � arg max � � P β ⊂ B l B l = { w 1 , w 2 , w 3 } β = { w 2 } ◮ the best failing set for input z l Let H l � H ( B l \ β l ) Fact 1 � H ( β l ) C ( z l ) = f ( z l ) � � � P [ C ( z ) � = f ( z ) | H l ] = P Proof ◮ f is sensitive to z l ◮ ¬ z l ⇔ “flip” all wires in B l β l is the worst non-failing set for input z 12 / 33

  14. Error analysis Error probability given wire failures Fact 2 � H ( β l ) C ( z l ) = f ( z l ) � � � P ≥ 1 − p Proof C ( z l ) = f ( z l ) ◮ P � � ≥ 1 − p � H ( β ) ◮ β l maximizes P C ( z l ) = f ( z l ) � � � Fact 1 & 2 ⇒ Fact 3 For each l ∈ S , P [ C ( z ) � = f ( z ) | H l ] ≥ 1 − p where { H l , l ∈ S } are independent events. Furthermore, Lemma 4.3 in [G´ acs&G´ al 1994] shows � � � ≥ (1 − √ p ) 2 � � P C ( z ) � = f ( z ) H l � � � l ∈ S The error probability given H l or � l ∈ S H l is relatively large. 13 / 33

  15. Error analysis Bounds on wire failure probabilities Note p ≥ P [ C ( z ) � = f ( z )] � � � �� � � � ≥ P C ( z ) � = f ( z ) H l P H l � � � l ∈ S l ∈ S Fact 3 implies Fact 4 �� � p H l ≤ P (1 − √ p ) 2 l ∈ S which implies (via Lemma 4.1 in [G´ acs&G´ al 1994]), Fact 5 �� � � � � p H l ≥ 1 − P [ H l ] P (1 − √ p ) 2 l ∈ S l ∈ S 14 / 33

  16. Error analysis Bounds on the total number of sensitive wires Fact 6 P [ H l ] = (1 − δ ) | β l | δ m l −| β l | ≥ δ m l Fact 4 & 5 ⇒ p � δ m l 1 − 2 √ p ≥ l ∈ S � 1 /s �� δ m l ≥ s l ∈ S which leads to s 1 − 2 √ p s � � � m l ≥ log(1 /δ ) log p l ∈ S lower bound on the total number of “sensitive wires” 15 / 33

  17. Lower bound on number of gates Let N C be the total number of gates in C : � n (Φ C ) N C ≥ n g g � ≥ m l l ∈ S s 1 − 2 √ p s � � ≥ log(1 /δ ) log p Comments: The above proof is for p ∈ (0 , 1 / 4) The case p ∈ (1 / 4 , 1 / 2) can be shown similarly. 16 / 33

  18. Block Sensitivity Let x S be a binary vector that differs from x in the S subset of indicies, i.e., � x i i / ∈ S x S i = i ∈ S . ¬ x i f is (block) sensitive to S on x if f ( x S ) � = f ( x ) . Block sensitivity of f on x : the largest number b such that ◮ there exists b disjoint sets S 1 , S 2 , · · · , S b ◮ for all 1 ≤ i ≤ b , f is sensitive to S i on x Block sensitivity of f : maximum over all x . ◮ block sensitivity ≥ sensitivity Theorem based on block sensitivity ◮ Let ε and p be any constants such that ε ∈ (0 , 1 / 2) , p ∈ (0 , 1 / 2) . ◮ Let f be any Boolean function with block sensitivity b . Under the error model ( ε, p ) , the number of gates of the curcuit is Ω ( b log b ) . 17 / 33

  19. Discussions Lower bound for specific functions Given an explicit function f of n variables, is there a lower boudn that is stronger than Ω ( n log n ) ? Open problem for unrestricted circuit C with complete basis function f that have Ω ( n log n ) noiseless complexity for circuit C with some incomplete basis Φ 18 / 33

  20. Discussions Computation model Exponential blowup A noisy circuit with multiple levels The output of gates at level l goes to a gate at level l + 1 Level 0 has n inputs ◮ Level 0 has N 0 = n log n output gates ◮ Level 1 has N 0 inputs ◮ Level 1 has N 1 = N 0 log N 0 output gates, . . . Why? “The theorem is generally applicable only to the very first step of such a fault tolerant computation” If the input is not the original ones, we can choose them to make the sensitivity of a Boolean function to be 0. ◮ f ( x 1 , x 2 , x 3 , x 4 , x 1 ⊕ x 2 ⊕ x 4 , x 1 ⊕ x 3 ⊕ x 4 , x 2 ⊕ x 3 ⊕ x 4 ) ◮ Lower bound does not apply: sensitivity is 0. How about block sensitivity? Problem formulation issue on the lower bound for coded input ◮ coding is also computation! 19 / 33

Recommend


More recommend