community detection with secondary latent variables
play

Community Detection with Secondary Latent Variables Mohammad - PowerPoint PPT Presentation

Community Detection with Secondary Latent Variables Mohammad Esmaeili and Aria Nosratinia The University of Texas at Dallas { esmaeili, aria } @utdallas.edu 21-26 June 2020 1 / 18 Problem and Motivation Current Models: Edges independent


  1. Community Detection with Secondary Latent Variables Mohammad Esmaeili and Aria Nosratinia The University of Texas at Dallas { esmaeili, aria } @utdallas.edu 21-26 June 2020 1 / 18

  2. Problem and Motivation Current Models: Edges independent conditioned on communities. Reality: Communities don’t completely explain edge dependencies Our work: Brings modeling and analysis closer to reality. 2 / 18

  3. Examples Graphs and their Latent Variables: Social networks: republican/democrat communities leave dependence according to localities Product co-purchasing networks: products/buyer(men and women)/age Movie networks: type of movies (action, comedy, and romance)/ audiences (men and women)/age ratings 3 / 18

  4. Introduction We consider: Known second latent variable as an auxiliary latent variable Unknown second latent variable as a nuisance latent variable Related Models in the literature: Overlapping communities Latent space models Graphs with additional non-graph observations 4 / 18

  5. Our Contributions Community detection for the stochastic block model with secondary latent variable Applying semidefinite programming to community detection with secondary latent variable Calculating the exact recovery thresholds when the secondary latent variable is either known or unknown Showing that semidefinite programming is asymptotically optimal 5 / 18

  6. System Model Primary and secondary latent variables (binary): x , y Adjacency matrix: A Edges are drawn from a Bernoulli i.i.d. distribution, conditioned on both x and y Assumption for estimator: x T 1 = 0 (can be relaxed) Goals: Recovering x when y is known y is unknown 6 / 18

  7. Maximum Likelihood Detectors When y is known: T 1 x T ( A ∗ yy T ) x + T 2 x T Ax ˆ x =arg max x (1) subject to x i ∈ {± 1 } , i ∈ [ n ] x T 1 = 0 , where T 1 and T 2 are constants. When y is unknown: x T Ax x =arg max ˆ x (2) subject to x i ∈ {± 1 } , i ∈ [ n ] x T 1 = 0 , 7 / 18

  8. Semidefinite Programming Relaxation Define Z � xx T and B � A ∗ yy T . Semidefinite relaxation arises from: x T Ax = Tr ( xx T A ) and x T Bx = Tr ( xx T B ) substituting xx T → Z relaxing the rank-1 constraint on Z to a positivity constraint 8 / 18

  9. Semidefinite Programming Relaxation When y is known: ˆ Z =arg max � Z , T 1 B + T 2 A � Z subject to Z � 0 (3) Z ii = 1 , i ∈ [ n ] � Z , J � = 0 . When y is unknown: x =arg max ˆ � Z , A � x subject to Z � 0 (4) Z ii = 1 , i ∈ [ n ] � Z , J � = 0 . where J is all-one matrix. 9 / 18

  10. Exact Recovery Conditions Exact recovery metric: n →∞ P ( e = 0) = 1 lim (5) Lemma: Consider the Lagrange multipliers D ∗ = diag ( d ∗ λ ∗ , S ∗ . i ) , If we have � D ∗ + λ ∗ J − T 1 B − T 2 A when y is known S ∗ = when y is unknown , D ∗ + λ ∗ J − A S ∗ � 0 , λ 2 ( S ∗ ) > 0 , S ∗ x ∗ = 0 then ( λ ∗ , D ∗ , S ∗ ) is the dual optimal solution and Z SDP = x ∗ x ∗ T is the unique primal optimal solution of (3) and (4). 10 / 18

  11. Outline of Proof Optimality: via properties of Lagrangian and conditions of the Lemma Uniqueness: Simple contraposition exercise Showing S ∗ � 0 and λ 2 ( S ∗ ) > 0 with probability at least 1 − o (1). In other words, we show that � � V ⊥ x ∗ , � V � =1 V T S ∗ V > 0 P inf ≥ 1 − o (1) . 11 / 18

  12. Parameters and Quality Metrics ρ y is the empirical fraction of nodes with y i = 1. a , b , c , d define the distribution of graph edges conditioned on latent variables  Bern( a log n  n ) if x i = x j , y i = y j     Bern( b log n n ) if x i = x j , y i � = y j A ij ∼ Bern( c log n  n ) if x i � = x j , y i = y j     Bern( d log n n ) if x i � = x j , y i � = y j Quality Metrics η 1 , η 2 reflect the likelihood of graph edge w.r.t. x Quality Metrics η 3 , η 4 reflect the likelihood of graph edge w.r.t. x , averaged over y . 12 / 18

  13. Achievability Theorem When y is known, if min { η 1 , η 2 } > 1 and when y is unknown, if min { η 3 , η 4 } > 1 then the semidefinite programming estimator is asymptotically optimal, i.e., P ( Z SDP = Z ∗ ) ≥ 1 − o (1) . 13 / 18

  14. Converses Theorem When y is known, if min { η 1 , η 2 } < 1 and when y is unknown, if min { η 3 , η 4 } < 1 then for any sequence of estimators � Z n , P ( � Z n = Z ∗ ) → 0 as n → ∞ . The converse is obtained via failure of Maximum Likelihood. 14 / 18

  15. Results & Discussion Let γ 1 � min { η 1 , η 2 } and γ 2 � min { η 3 , η 4 } 2.5 y = 0.3, 1 y = 0.3, 2 2 y = 0.4, 1 y = 0.4, 2 y = 0.5, 1 1.5 y = 0.5, 2 1 , 2 Exact recovery region 1 0.5 0 4 6 8 10 12 14 a Figure: Exact recovery region of x with b = 3 , c = d = 1. 15 / 18

  16. Results & Discussion 2.5 y = 0.3, 1 y = 0.3, 2 2 y = 0.4, 1 y = 0.4, 2 y = 0.5, 1 1.5 y = 0.5, 2 1 , 2 Exact recovery region 1 0.5 0 4 6 8 10 12 14 a Figure: Exact recovery region of x with b = c = d = 1. 16 / 18

  17. Results & Discussion We introduced a generalization for stochastic block models with a secondary latent variable. Semidefinite programming relaxation of ML detector achieves exact recovery down to the optimal threshold 17 / 18

  18. Thank You 18 / 18

Recommend


More recommend