a common neighbors based random graph model for community
play

A common-neighbors-based random graph model for community structure - PowerPoint PPT Presentation

A common-neighbors-based random graph model for community structure Emily Fischer Cornell University May 12, 2017 A Exam Emily Fischer Outline 1. Introduction Preferential Attachment (PA) 2. Common Neighbors Model (CN) Degree


  1. A common-neighbors-based random graph model for community structure Emily Fischer Cornell University May 12, 2017

  2. A Exam Emily Fischer Outline 1. Introduction • Preferential Attachment (PA) 2. Common Neighbors Model (CN) • Degree distribution • Community structure

  3. A Exam Emily Fischer Preferential Attachment • Users prefer to connect to nodes of high degree

  4. A Exam Emily Fischer Preferential Attachment • Users prefer to connect to nodes of high degree • Results in heavy-tailed degree distribution

  5. A Exam Emily Fischer Issues with Preferential Attachmment The LinkedIn graph 1. does NOT have a power law degree distribution 2. has “community structure”

  6. A Exam Emily Fischer Log-log plots of degree distribution

  7. A Exam Emily Fischer Issues with Preferential Attachmment The LinkedIn graph 1. does NOT have a power law degree distribution 2. has “community structure”

  8. A Exam Emily Fischer What is “community structure”? • Strong community structure • More edges within community than between communities

  9. A Exam Emily Fischer What is “community structure"? • Preferential attachment • One central hub around high-degree node

  10. A Exam Emily Fischer Common Neighbors Model

  11. A Exam Emily Fischer Common Neighbors Model Users prefer to connect to nodes with whom they share many mutual friends

  12. A Exam Emily Fischer Common Neighbors Model Users prefer to connect to nodes with whom they share many mutual friends

  13. A Exam Emily Fischer Common Neighbors Model Sequence of graphs ( G t ) t ≥ 0 . • Given graph G t with n ( t ) nodes and m ( t ) edges

  14. A Exam Emily Fischer Common Neighbors Model Sequence of graphs ( G t ) t ≥ 0 . • Given graph G t with n ( t ) nodes and m ( t ) edges • At time t + 1, a new node v arrives with probability α • If no new arrival, select v uniformly among existing nodes

  15. A Exam Emily Fischer Common Neighbors Model Sequence of graphs ( G t ) t ≥ 0 . • Given graph G t with n ( t ) nodes and m ( t ) edges • At time t + 1, a new node v arrives with probability α • If no new arrival, select v uniformly among existing nodes • Select receiving node w with probability proportional to number of common neighbors between v and w • Γ v ( t ) is the neighborhood of v at time t • K vw ( t ) = | Γ v ( t ) ∩ Γ w ( t ) | K vw ( t ) + δ P (select w | sender = v ) = � u K vu ( t ) + δ n ( t )

  16. A Exam Emily Fischer Common Neighbors Model Sequence of graphs ( G t ) t ≥ 0 . • Given graph G t with n ( t ) nodes and m ( t ) edges • At time t + 1, a new node v arrives with probability α • If no new arrival, select v uniformly among existing nodes • Select receiving node w with probability proportional to number of common neighbors between v and w • Γ v ( t ) is the neighborhood of v at time t • K vw ( t ) = | Γ v ( t ) ∩ Γ w ( t ) | K vw ( t ) + δ P (select w | sender = v ) = � u K vu ( t ) + δ n ( t ) • Form directed edge ( v , w ).

  17. A Exam Emily Fischer Common Neighbors Model What does K vw ( t ) look like? Hard to analyze - feedback

  18. A Exam Emily Fischer Common Neighbors Model What does K vw ( t ) look like?

  19. A Exam Emily Fischer Common Neighbors Model What does K vw ( t ) look like? Hard to analyze - feedback

  20. A Exam Emily Fischer Common Neighbor Process • Want to model evolution of K ij ( t ) on its own. • Start at ˜ K ij (0) = 0 for all pairs i , j .

  21. A Exam Emily Fischer Common Neighbor Process • Want to model evolution of K ij ( t ) on its own. • Start at ˜ K ij (0) = 0 for all pairs i , j . • Given ( ˜ K ij ( t )) i , j ≥ 0 , at t + 1, • Select i uniformly from existing nodes • Choose η = c ( n ( t )) θ nodes, j 1 , j 2 , . . . , j η , preferentially with K ij ℓ ( t ) , and increase K ij ℓ ( t + 1) = K ij ℓ ( t ) + 1 .

  22. A Exam Emily Fischer Common Neighbor Process • Want to model evolution of K ij ( t ) on its own. • Start at ˜ K ij (0) = 0 for all pairs i , j . • Given ( ˜ K ij ( t )) i , j ≥ 0 , at t + 1, • Select i uniformly from existing nodes • Choose η = c ( n ( t )) θ nodes, j 1 , j 2 , . . . , j η , preferentially with ˜ K ij ℓ ( t ), and increase K ij ℓ ( t + 1) = ˜ ˜ K ij ℓ ( t ) + 1 .

  23. A Exam Emily Fischer Common Neighbor Process Let ˜ � N i ( t ) = K ij ( t ) j What is the distribution of N i ( t )?

  24. A Exam Emily Fischer Common Neighbor Process Theorem j ˜ Let N i ( t ) = � K ij ( t ) . Then there exists a random variable Z i such that N i ( t ) → Z i t θ in probability, where Z i has characteristic function � α θ � � 1 − α 1 t ( e itz − 1) dt φ Z ( z ) = exp . αθ 0

  25. A Exam Emily Fischer Common Neighbor Process

  26. A Exam Emily Fischer Common Neighbor Process Result • The “total common neighbors” N i ( t ) converges when scaled by t θ . In progress/Future • Limiting distribution for ˜ K ij ( t ). • Use these distributions to analyze degree distribution of the graph

  27. A Exam Emily Fischer Community Structure • How to quantify “strong community structure” • Compare community structure of CN and PA.

  28. A Exam Emily Fischer Community Structure CN vs. PA

  29. A Exam Emily Fischer Modularity Definition Given a graph partitioned into c communities, the modularity is c � ( e ii − a 2 Q = i ) i =1 where e ii is the fraction of edges with both end vertices in community i , and a i is the fraction of ends of edges with vertices in community i .

  30. A Exam Emily Fischer Community Detection • Community detection algorithms aim to assign nodes to communities in a way that is reasonable • Some algorithms maximize modularity: Fast-greedy (FG), Largest-eigenvector (LE) • But there are other methods as well: Edge-betweenness (EB), Walktrap (WC).

  31. A Exam Emily Fischer Modularity Averages of modularity over 100 trials ( α = . 2 , δ = . 5) Graph EB FG LE WC CN 500 .450 .472 .423 .401 PA 500 .276 .379 .333 .251 CN 1000 .310 .402 .350 .301 PA 1000 .103 .328 .279 .190 CN 5000 .145 .320 .176 PA 5000 .039 .277 .120

  32. A Exam Emily Fischer Conclusion 1. PA mode lacks characteristics of LinkedIn network: • Power-law degree distribution • Lack of community structure 2. Common Neighbors Model • Limiting distribution of N i ( t ) in the common neighbors process • Better community structure than PA

  33. A Exam Emily Fischer Edge Acceptance/Rejection Node v sends an invitation to a node w .

  34. A Exam Emily Fischer Model 1: Edge Acceptance/Rejection w accepts the invitation with probability p vw ( t ).

  35. A Exam Emily Fischer Edge Acceptance/Rejection How can acceptance probability achieve goals of (1) non-power law degree distribution and (2) community structure? • Rich may choose not to get richer • Probability of acceptance based on communities � p C v = C w p vw ( t ) = q C v � = C w .

  36. A Exam Emily Fischer Edge Acceptance/Rejection How can acceptance probability achieve goals of (1) non-power law degree distribution and (2) community structure? • Rich may choose not to get richer: p vw ( t ) ↓ 0 • Probability of acceptance based on communities � p C v = C w p vw ( t ) = q C v � = C w .

  37. A Exam Emily Fischer Edge Acceptance/Rejection How can acceptance probability achieve goals of (1) non-power law degree distribution and (2) community structure? • Rich may choose not to get richer: p vw ( t ) ↓ 0 • Probability of acceptance based on communities: � p C v = C w p vw ( t ) = C v � = C w . q

  38. A Exam Emily Fischer Edge Acceptance/Rejection For now, constant acceptance probability p vw ( t ) = p for all v , w and t ≥ 0 .

Recommend


More recommend