the densest subgraph of sparse random graphs
play

The densest subgraph of sparse random graphs Justin Salez (Universit - PowerPoint PPT Presentation

The densest subgraph of sparse random graphs Justin Salez (Universit e Paris 7) with Venkat Anantharam (UC Berkeley) The objective method (Aldous-Steele, 2004) The objective method (Aldous-Steele, 2004) Context: given a large interacting


  1. Examples of local weak limits Note: graphs must be sparse , i.e. | E | ≍ | V | ◮ G n = box of size n × . . . × n in Z d L = dirac at ( Z d , 0) ◮ G n = random d − regular graph on n nodes L = dirac at the d − regular infinite rooted tree enyi graph with p n = c ◮ G n = Erd˝ os-R´ n on n nodes L = law of a Galton-Watson tree with degree Poisson(c) ◮ G n = random graph with degree distribution π on n nodes L = law of a Galton-Watson tree with degree distribution π ◮ G n = uniform random tree on n nodes

  2. Examples of local weak limits Note: graphs must be sparse , i.e. | E | ≍ | V | ◮ G n = box of size n × . . . × n in Z d L = dirac at ( Z d , 0) ◮ G n = random d − regular graph on n nodes L = dirac at the d − regular infinite rooted tree enyi graph with p n = c ◮ G n = Erd˝ os-R´ n on n nodes L = law of a Galton-Watson tree with degree Poisson(c) ◮ G n = random graph with degree distribution π on n nodes L = law of a Galton-Watson tree with degree distribution π ◮ G n = uniform random tree on n nodes L = Infinite Skeleton Tree (Grimmett, 1980)

  3. Examples of local weak limits Note: graphs must be sparse , i.e. | E | ≍ | V | ◮ G n = box of size n × . . . × n in Z d L = dirac at ( Z d , 0) ◮ G n = random d − regular graph on n nodes L = dirac at the d − regular infinite rooted tree enyi graph with p n = c ◮ G n = Erd˝ os-R´ n on n nodes L = law of a Galton-Watson tree with degree Poisson(c) ◮ G n = random graph with degree distribution π on n nodes L = law of a Galton-Watson tree with degree distribution π ◮ G n = uniform random tree on n nodes L = Infinite Skeleton Tree (Grimmett, 1980) ◮ G n = preferential attachment graph on n nodes

  4. Examples of local weak limits Note: graphs must be sparse , i.e. | E | ≍ | V | ◮ G n = box of size n × . . . × n in Z d L = dirac at ( Z d , 0) ◮ G n = random d − regular graph on n nodes L = dirac at the d − regular infinite rooted tree enyi graph with p n = c ◮ G n = Erd˝ os-R´ n on n nodes L = law of a Galton-Watson tree with degree Poisson(c) ◮ G n = random graph with degree distribution π on n nodes L = law of a Galton-Watson tree with degree distribution π ◮ G n = uniform random tree on n nodes L = Infinite Skeleton Tree (Grimmett, 1980) ◮ G n = preferential attachment graph on n nodes L = Polya-point graph (Berger-Borgs-Chayes-Sabery, 2009)

  5. An illustration: the nullity of large graphs

  6. An illustration: the nullity of large graphs µ G ( { 0 } ) = dim ker ( A G ) . | V |

  7. An illustration: the nullity of large graphs µ G ( { 0 } ) = dim ker ( A G ) . Asymptotics when G is large ? | V |

  8. An illustration: the nullity of large graphs µ G ( { 0 } ) = dim ker ( A G ) . Asymptotics when G is large ? | V | n , c � � Conjecture (Bauer-Golinelli 2001). For G n : Erd˝ os-R´ enyi , n λ ∗ + e − c λ ∗ + c λ ∗ e − c λ ∗ − 1 , µ G n ( { 0 } ) − − − → n →∞ where λ ∗ ∈ [0 , 1] is the smallest root of λ = e − ce − c λ .

  9. An illustration: the nullity of large graphs µ G ( { 0 } ) = dim ker ( A G ) . Asymptotics when G is large ? | V | n , c � � Conjecture (Bauer-Golinelli 2001). For G n : Erd˝ os-R´ enyi , n λ ∗ + e − c λ ∗ + c λ ∗ e − c λ ∗ − 1 , µ G n ( { 0 } ) − − − → n →∞ where λ ∗ ∈ [0 , 1] is the smallest root of λ = e − ce − c λ . Theorem (Bordenave-Lelarge-S., 2011)

  10. An illustration: the nullity of large graphs µ G ( { 0 } ) = dim ker ( A G ) . Asymptotics when G is large ? | V | n , c � � Conjecture (Bauer-Golinelli 2001). For G n : Erd˝ os-R´ enyi , n λ ∗ + e − c λ ∗ + c λ ∗ e − c λ ∗ − 1 , µ G n ( { 0 } ) − − − → n →∞ where λ ∗ ∈ [0 , 1] is the smallest root of λ = e − ce − c λ . Theorem (Bordenave-Lelarge-S., 2011) 1. G n → L = ⇒ µ G n ( { 0 } ) → µ L ( { 0 } ).

  11. An illustration: the nullity of large graphs µ G ( { 0 } ) = dim ker ( A G ) . Asymptotics when G is large ? | V | n , c � � Conjecture (Bauer-Golinelli 2001). For G n : Erd˝ os-R´ enyi , n λ ∗ + e − c λ ∗ + c λ ∗ e − c λ ∗ − 1 , µ G n ( { 0 } ) − − − → n →∞ where λ ∗ ∈ [0 , 1] is the smallest root of λ = e − ce − c λ . Theorem (Bordenave-Lelarge-S., 2011) 1. G n → L = ⇒ µ G n ( { 0 } ) → µ L ( { 0 } ). 2. When L = Galton-Watson ( π ), f ′ (1) λλ ∗ + f (1 − λ ) + f (1 − λ ∗ ) − 1 � � µ L ( { 0 } ) = min , λ = λ ∗∗ n π n z n and λ ∗ = f ′ (1 − λ ) / f ′ (1). where f ( z ) = �

  12. Continuity with respect to local weak convergence

  13. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only.

  14. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − →

  15. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size.

  16. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size. ◮ Analytic implication: Φ admits a limit along most sparse graph sequences. The distributional self-similarity of L may even allow for an explicit determination of Φ( L ).

  17. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size. ◮ Analytic implication: Φ admits a limit along most sparse graph sequences. The distributional self-similarity of L may even allow for an explicit determination of Φ( L ). ◮ Examples:

  18. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size. ◮ Analytic implication: Φ admits a limit along most sparse graph sequences. The distributional self-similarity of L may even allow for an explicit determination of Φ( L ). ◮ Examples: number of spanning trees (Lyons, 2005),

  19. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size. ◮ Analytic implication: Φ admits a limit along most sparse graph sequences. The distributional self-similarity of L may even allow for an explicit determination of Φ( L ). ◮ Examples: number of spanning trees (Lyons, 2005), spectrum and rank (Bordenave-Lelarge-S, 2011),

  20. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size. ◮ Analytic implication: Φ admits a limit along most sparse graph sequences. The distributional self-similarity of L may even allow for an explicit determination of Φ( L ). ◮ Examples: number of spanning trees (Lyons, 2005), spectrum and rank (Bordenave-Lelarge-S, 2011), matching polynomial (idem, 2013),

  21. Continuity with respect to local weak convergence ◮ In the sparse regime, many important graph parameters Φ are essentially determined by the local geometry only. ◮ This can be rigorously formalized by a continuity theorem: loc . G n − n →∞ L − − → = ⇒ Φ( G n ) − n →∞ Φ( L ) − − → ◮ Algorithmic implication: Φ is efficiently approximable via local distributed algorithms, independently of network size. ◮ Analytic implication: Φ admits a limit along most sparse graph sequences. The distributional self-similarity of L may even allow for an explicit determination of Φ( L ). ◮ Examples: number of spanning trees (Lyons, 2005), spectrum and rank (Bordenave-Lelarge-S, 2011), matching polynomial (idem, 2013), Ising models (Dembo-Montanari-Sun, 2013)...

  22. The densest subgraph problem

  23. The densest subgraph problem Fix a finite graph G = ( V , E )

  24. The densest subgraph problem Fix a finite graph G = ( V , E ) Densest subgraph : H ⋆ = argmax � � | E ( H ) | | H | : H ⊆ V

  25. The densest subgraph problem Fix a finite graph G = ( V , E ) Densest subgraph : H ⋆ = argmax � � | E ( H ) | | H | : H ⊆ V � � Maximum subgraph density : ̺ ⋆ = max | E ( H ) | | H | : H ⊆ V

  26. The densest subgraph problem Fix a finite graph G = ( V , E ) Densest subgraph : H ⋆ = argmax � � | E ( H ) | | H | : H ⊆ V � � Maximum subgraph density : ̺ ⋆ = max | E ( H ) | = 17 | H | : H ⊆ V 10

  27. The densest subgraph problem on large sparse graphs

  28. The densest subgraph problem on large sparse graphs

  29. The densest subgraph problem on large sparse graphs

  30. Load balancing

  31. Load balancing An allocation on G is a function θ : � E → [0 , 1] satisfying θ ( i , j ) + θ ( j , i ) = 1

  32. Load balancing An allocation on G is a function θ : � E → [0 , 1] satisfying θ ( i , j ) + θ ( j , i ) = 1 The induced load at i ∈ V is � ∂θ ( i ) = θ ( j , i ) j ∼ i

  33. Load balancing An allocation on G is a function θ : � E → [0 , 1] satisfying θ ( i , j ) + θ ( j , i ) = 1 The induced load at i ∈ V is � ∂θ ( i ) = θ ( j , i ) j ∼ i The allocation is balanced if for each ( i , j ) ∈ � E ∂θ ( i ) < ∂θ ( j ) = ⇒ θ ( i , j ) = 0

  34. From local to global optimality

  35. From local to global optimality Claim. For an allocation θ , the following are equivalent:

  36. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced

  37. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced i ( ∂θ ( i )) 2 . 2. θ minimizes �

  38. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced i ( ∂θ ( i )) 2 . 2. θ minimizes � 3. θ minimizes � i f ( ∂θ ( i )) for any convex function f : R → R .

  39. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced i ( ∂θ ( i )) 2 . 2. θ minimizes � 3. θ minimizes � i f ( ∂θ ( i )) for any convex function f : R → R . Corollary 1. Balanced allocations always exist.

  40. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced i ( ∂θ ( i )) 2 . 2. θ minimizes � 3. θ minimizes � i f ( ∂θ ( i )) for any convex function f : R → R . Corollary 1. Balanced allocations always exist. Corollary 2. They all induce the same loads ∂θ : V → [0 , ∞ ).

  41. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced i ( ∂θ ( i )) 2 . 2. θ minimizes � 3. θ minimizes � i f ( ∂θ ( i )) for any convex function f : R → R . Corollary 1. Balanced allocations always exist. Corollary 2. They all induce the same loads ∂θ : V → [0 , ∞ ). Corollary 3. Balanced loads solve the densest subgraph problem:

  42. From local to global optimality Claim. For an allocation θ , the following are equivalent: 1. θ is balanced i ( ∂θ ( i )) 2 . 2. θ minimizes � 3. θ minimizes � i f ( ∂θ ( i )) for any convex function f : R → R . Corollary 1. Balanced allocations always exist. Corollary 2. They all induce the same loads ∂θ : V → [0 , ∞ ). Corollary 3. Balanced loads solve the densest subgraph problem: i ∈ V ∂θ ( i ) = ̺ ⋆ ∂θ ( i ) = H ⋆ max and argmax i ∈ V

  43. Example

  44. Example 5 10 5 10 5 1 10 5 2 9 8 3 10 8 5 6 8 5 8 9 4 4 7 2 5 2 7 5 8 5 3 5 5 2 2 5 10 3 7 5 5 7 3 10 1 5 5 7 10 1 6 10 2 3 2 8 5 10 9 10 10 8 5 10 5 10 10 5 10 10 5 5

  45. Example 1.5 1.6 1.6 1 1.5 1.6 1.7 1.7 1.7 1.7 1.6 1.7 1.5 1.5 1 1.7 1.7 1.7 1.6 1.5 1.7 1.5 1 1.7 1.5 1.5 1 1.5 1.5

  46. Example 1.5 1.6 1.6 1 1.5 1.6 1.7 1.7 1.7 1.7 1.6 1.7 1.5 1.5 1 1.7 1.7 1.7 1.6 1.5 1.7 1.5 1 1.7 1.5 1.5 1 1.5 1.5

  47. How do those ‘densities” look on a large sparse graph ?

  48. How do those ‘densities” look on a large sparse graph ?

  49. How do those ‘densities” look on a large sparse graph ?

  50. Density profile of a random graph with average degree 3

  51. Density profile of a random graph with average degree 3 | V | = 100

  52. Density profile of a random graph with average degree 3 | V | = 100 | V | = 10000

  53. Density profile of a random graph with average degree 4

  54. Density profile of a random graph with average degree 4 | V | = 100

  55. Density profile of a random graph with average degree 4 | V | = 100 | V | = 10000

  56. The conjecture (Hajek, 1990)

  57. The conjecture (Hajek, 1990) ∂ Θ( G , o ) : load induced at o by any balanced allocation on G .

  58. The conjecture (Hajek, 1990) ∂ Θ( G , o ) : load induced at o by any balanced allocation on G . Define the density profile of G = ( V , E ) as 1 � Λ G = δ ∂ Θ( G , o ) ∈ P ( R ) . | V | o ∈ V

  59. The conjecture (Hajek, 1990) ∂ Θ( G , o ) : load induced at o by any balanced allocation on G . Define the density profile of G = ( V , E ) as 1 � Λ G = δ ∂ Θ( G , o ) ∈ P ( R ) . | V | o ∈ V n , c � � Conjecture: G n Erd˝ os-R´ enyi ; c fixed, n → ∞ n

  60. The conjecture (Hajek, 1990) ∂ Θ( G , o ) : load induced at o by any balanced allocation on G . Define the density profile of G = ( V , E ) as 1 � Λ G = δ ∂ Θ( G , o ) ∈ P ( R ) . | V | o ∈ V n , c � � Conjecture: G n Erd˝ os-R´ enyi ; c fixed, n → ∞ n 1. Λ G n concentrates around a deterministic Λ ∈ P ( R )

  61. The conjecture (Hajek, 1990) ∂ Θ( G , o ) : load induced at o by any balanced allocation on G . Define the density profile of G = ( V , E ) as 1 � Λ G = δ ∂ Θ( G , o ) ∈ P ( R ) . | V | o ∈ V n , c � � Conjecture: G n Erd˝ os-R´ enyi ; c fixed, n → ∞ n 1. Λ G n concentrates around a deterministic Λ ∈ P ( R ) P 2. ̺ ⋆ ( G n ) − n →∞ sup { t ∈ R : Λ( t , + ∞ ) > 0 } − − →

  62. Result 1 : the density profile of sparse graphs

  63. Result 1 : the density profile of sparse graphs Theorem. Assume that L [deg( G , o )] < ∞ .

  64. Result 1 : the density profile of sparse graphs Theorem. Assume that L [deg( G , o )] < ∞ . Then, P ( R ) loc . G n − n →∞ L − − → = ⇒ Λ G n − n →∞ Λ L − − →

  65. Result 1 : the density profile of sparse graphs Theorem. Assume that L [deg( G , o )] < ∞ . Then, P ( R ) loc . G n − n →∞ L − − → = ⇒ Λ G n − n →∞ Λ L − − → where Λ L is the solution to a certain optimization problem on L .

  66. Result 1 : the density profile of sparse graphs Theorem. Assume that L [deg( G , o )] < ∞ . Then, P ( R ) loc . G n − n →∞ L − − → = ⇒ Λ G n − n →∞ Λ L − − → where Λ L is the solution to a certain optimization problem on L . � R ( x − t ) + Λ L ( dx ) Specifically, the excess function Φ: t �→

  67. Result 1 : the density profile of sparse graphs Theorem. Assume that L [deg( G , o )] < ∞ . Then, P ( R ) loc . G n − n →∞ L − − → = ⇒ Λ G n − n →∞ Λ L − − → where Λ L is the solution to a certain optimization problem on L . � R ( x − t ) + Λ L ( dx ) solves Specifically, the excess function Φ: t �→ � �� � � 1 Φ( t ) = max 2 L f ( G , i ) ∧ f ( G , o ) − t L [ f ( G , o )] f : G ⋆ → [0 , 1] i ∼ o

  68. Result 2 : maximum subgraph density of sparse graphs

Recommend


More recommend