fifty shades of adaptivity (in property testing) An Adaptivity Hierarchy Theorem for Property Testing Clément Canonne (Columbia University) July 9, 2017 Joint work with Tom Gur (Weizmann Institute UC Berkeley)
“property testing?”
approximate, randomized decision algorithms that make queries ∙ Big object: too big ∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, 2
randomized decision algorithms that make queries ∙ Big object: too big ∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, approximate, 2
decision algorithms that make queries ∙ Big object: too big ∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, approximate, randomized 2
∙ Big object: too big ∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, approximate, randomized decision algorithms that make queries 2
∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, approximate, randomized decision algorithms that make queries ∙ Big object: too big 2
∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, approximate, randomized decision algorithms that make queries ∙ Big object: too big ∙ Expensive access: pricey data 2
Need to infer information – one bit – from the data: quickly, or with very few lookups. why? Sublinear, approximate, randomized decision algorithms that make queries ∙ Big object: too big ∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge 2
why? Sublinear, approximate, randomized decision algorithms that make queries ∙ Big object: too big ∙ Expensive access: pricey data ∙ “Model selection”: many options ∙ Good Enough: a priori knowledge Need to infer information – one bit – from the data: quickly, or with very few lookups. 2
3
Must decide: x , or d x ? (and be correct on any x with probability at least 2 3) how? Known space (say, { 0 , 1 } N ) Property P ⊆ { 0 , 1 } N ) Query (oracle) access to unknown x ∈ { 0 , 1 } N Proximity parameter ε ∈ ( 0 , 1 ] 4
, or d x ? (and be correct on any x with probability at least 2 3) how? Known space (say, { 0 , 1 } N ) Property P ⊆ { 0 , 1 } N ) Query (oracle) access to unknown x ∈ { 0 , 1 } N Proximity parameter ε ∈ ( 0 , 1 ] Must decide: x ∈ P 4
(and be correct on any x with probability at least 2 3) how? Known space (say, { 0 , 1 } N ) Property P ⊆ { 0 , 1 } N ) Query (oracle) access to unknown x ∈ { 0 , 1 } N Proximity parameter ε ∈ ( 0 , 1 ] Must decide: x ∈ P , or d ( x , P ) > ε ? 4
how? Known space (say, { 0 , 1 } N ) Property P ⊆ { 0 , 1 } N ) Query (oracle) access to unknown x ∈ { 0 , 1 } N Proximity parameter ε ∈ ( 0 , 1 ] Must decide: x ∈ P , or d ( x , P ) > ε ? (and be correct on any x with probability at least 2 / 3) 4
in an (egg)shell. how? Property Testing: 5
in an (egg)shell. how? Property Testing: 5
how? Property Testing: in an (egg)shell. 5
query-based vs. sample-based, uniform vs. distribution-free, adaptive vs. non-adaptive it’s complicated Many flavors… … one-sided vs. two-sided, 6
uniform vs. distribution-free, adaptive vs. non-adaptive it’s complicated Many flavors… … one-sided vs. two-sided, query-based vs. sample-based, 6
adaptive vs. non-adaptive it’s complicated Many flavors… … one-sided vs. two-sided, query-based vs. sample-based, uniform vs. distribution-free, 6
it’s complicated Many flavors… … one-sided vs. two-sided, query-based vs. sample-based, uniform vs. distribution-free, adaptive vs. non-adaptive 6
adaptivity
our focus: adaptivity Non-adaptive algorithm Makes all its queries upfront: Q ⊆ [ N ] = Q ( ε, r ) = { i 1 , . . . , i q } Adaptive algorithm Each query can depend arbitrarily on the previous answers: 8
some observations Dense graph model At most a quadratic gap between adaptive and non-adaptive algorithms: q vs. 2q 2 [AFKS00, GT03],[GR11] Boolean functions At most an exponential gap between adaptive and non-adaptive algorithms: q vs. 2 q Bounded-degree graph model Everything is possible: O ( 1 ) vs. Ω( √ n ) . [RS06] 9
But Many parallel queries can beat few sequential ones. Understanding the benefits and tradeoffs of adaptivity is crucial. why should we care? Of course Fewer queries is always better. 10
Understanding the benefits and tradeoffs of adaptivity is crucial. why should we care? Of course Fewer queries is always better. But Many parallel queries can beat few sequential ones. 10
why should we care? Of course Fewer queries is always better. But Many parallel queries can beat few sequential ones. Understanding the benefits and tradeoffs of adaptivity is crucial. 10
(and what does “amount of adaptivity” even mean?) this work A closer look Does the power of testing algorithms smoothly grow with the “amount of adaptivity?” 11
this work A closer look Does the power of testing algorithms smoothly grow with the “amount of adaptivity?” (and what does “amount of adaptivity” even mean?) 11
(i) Query Generation: The algorithm proceeds in k 1 rounds, such that at round 0, it produces a set of queries Q x 1 x Q , based on its own internal randomness and the answers to the previous sets of queries Q 0 Q 1 , and receives f Q f x 1 f x Q ; (ii) Completeness: If f , then it outputs accept with probability 2 3; (iii) Soundness: If dist f , then it outputs reject with probability 2 3. The query complexity q of the tester is the total number of queries made to k f, i.e., q 0 Q . coming up with a definition Definition (Round-Adaptive Testing Algorithms) Let Ω be a domain of size n, and k , q ≤ n. A randomized algorithm is said to be a ( k , q ) -round-adaptive tester for P ⊆ 2 Ω , if, on input ε ∈ ( 0 , 1 ] and granted query access to f : Ω → { 0 , 1 } : 12
(ii) Completeness: If f , then it outputs accept with probability 2 3; (iii) Soundness: If dist f , then it outputs reject with probability 2 3. The query complexity q of the tester is the total number of queries made to k f, i.e., q 0 Q . coming up with a definition Definition (Round-Adaptive Testing Algorithms) Let Ω be a domain of size n, and k , q ≤ n. A randomized algorithm is said to be a ( k , q ) -round-adaptive tester for P ⊆ 2 Ω , if, on input ε ∈ ( 0 , 1 ] and granted query access to f : Ω → { 0 , 1 } : (i) Query Generation: The algorithm proceeds in k + 1 rounds, such that at round ℓ ≥ 0, it produces a set of queries Q ℓ := { x ( ℓ ) , 1 , . . . , x ( ℓ ) , | Q ℓ | } ⊆ Ω , based on its own internal randomness and the answers to the previous sets of queries Q 0 , . . . , Q ℓ − 1 , and receives f ( Q ℓ ) = { f ( x ( ℓ ) , 1 ) , . . . , f ( x ( ℓ ) , | Q ℓ | ) } ; 12
The query complexity q of the tester is the total number of queries made to k f, i.e., q 0 Q . coming up with a definition Definition (Round-Adaptive Testing Algorithms) Let Ω be a domain of size n, and k , q ≤ n. A randomized algorithm is said to be a ( k , q ) -round-adaptive tester for P ⊆ 2 Ω , if, on input ε ∈ ( 0 , 1 ] and granted query access to f : Ω → { 0 , 1 } : (i) Query Generation: The algorithm proceeds in k + 1 rounds, such that at round ℓ ≥ 0, it produces a set of queries Q ℓ := { x ( ℓ ) , 1 , . . . , x ( ℓ ) , | Q ℓ | } ⊆ Ω , based on its own internal randomness and the answers to the previous sets of queries Q 0 , . . . , Q ℓ − 1 , and receives f ( Q ℓ ) = { f ( x ( ℓ ) , 1 ) , . . . , f ( x ( ℓ ) , | Q ℓ | ) } ; (ii) Completeness: If f ∈ P , then it outputs accept with probability 2 / 3; (iii) Soundness: If dist ( f , P ) > ε , then it outputs reject with probability 2 / 3. 12
Recommend
More recommend