detecting network effects
play

Detecting Network Effects Randomizing Over Randomized Experiments - PowerPoint PPT Presentation

Detecting Network Effects Randomizing Over Randomized Experiments Martin Saveski (@msaveski) MIT Detecting Network Effects Randomizing Over Randomized Experiments Guillaume Saint Jacques Martin Saveski Jean Pouget-Abadie MIT MIT


  1. Detecting Network Effects Randomizing Over Randomized Experiments Martin Saveski (@msaveski) MIT

  2. Detecting Network Effects Randomizing Over Randomized Experiments Guillaume Saint ‑ Jacques Martin Saveski Jean Pouget-Abadie MIT MIT Harvard Weitao Duan Souvik Ghosh Ya Xu Edo Airoldi LinkedIn LinkedIn LinkedIn Harvard

  3. Treatment Z i = 1 New Feed Ranking Algorithm

  4. Treatment Control Z i = 1 Z j = 0 New Feed Old Feed Ranking Algorithm Ranking Algorithm

  5. Treatment Control Z i = 1 Z j = 0 New Feed Old Feed Ranking Algorithm Ranking Algorithm

  6. Treatment Control Z i = 1 Z j = 0 New Feed Old Feed Ranking Algorithm Ranking Algorithm Y i Engagement

  7. Treatment Control Z i = 1 Z j = 0 New Feed Old Feed Ranking Algorithm Ranking Algorithm Y i Engagement

  8. Treatment Control Z i = 1 Z j = 0 New Feed Old Feed Ranking Algorithm Ranking Algorithm Y j Y i Engagement Engagement

  9. Completely-randomized Experiment

  10. Treatment (B) Completely-randomized Experiment

  11. Control (A) Treatment (B) Completely-randomized Experiment

  12. Control (A) Treatment (B) Σ Y Σ Y ( ) ( ) - μ = | | | | completely-randomized Completely-randomized Experiment

  13. Control (A) Treatment (B) Σ Y Σ Y ( ) ( ) - μ = | | | | completely-randomized SUTVA : Stable Unit Treatment Value Assumption Every user’s behavior is affected only by their treatment and NOT by the treatment of any other user Completely-randomized Experiment

  14. Cluster-based Randomized Experiment

  15. Cluster-based Randomized Experiment

  16. Cluster-based Randomized Experiment

  17. Cluster-based Randomized Experiment

  18. Control (A) Treatment (B) Cluster-based Randomized Experiment

  19. Completely-randomized Experiment Cluster-based Randomized Experiment OR

  20. Completely-randomized Experiment Cluster-based Randomized Experiment OR More Spillovers Less Spillovers Lower Variance Higher Variance

  21. Design for Detecting Network Effects

  22. Completely Randomized Experiment

  23. Completely Randomized Cluster-based Randomized Experiment Experiment

  24. Completely Randomized Cluster-based Randomized Experiment Experiment ? μ completely-randomized μ cluster-based =

  25. Hypothesis Test

  26. Hypothesis Test H 0 : SUTVA Holds

  27. Hypothesis Test H 0 : SUTVA Holds E W , Z [ˆ µ cbr − ˆ µ cr ] = 0

  28. Hypothesis Test H 0 : SUTVA Holds E W , Z [ˆ µ cbr − ˆ µ cr ] = 0 σ 2 ] var W , Z [ˆ µ cr − ˆ µ cbr ] ≤ E W , Z [ˆ

  29. Hypothesis Test H 0 : SUTVA Holds E W , Z [ˆ µ cbr − ˆ µ cr ] = 0 σ 2 ] var W , Z [ˆ µ cr − ˆ µ cbr ] ≤ E W , Z [ˆ Reject the null when:

  30. Hypothesis Test H 0 : SUTVA Holds E W , Z [ˆ µ cbr − ˆ µ cr ] = 0 σ 2 ] var W , Z [ˆ µ cr − ˆ µ cbr ] ≤ E W , Z [ˆ Reject the null when: | ˆ µ cbr | µ cr − ˆ 1 √ √ α ≥ σ 2 ˆ

  31. Hypothesis Test H 0 : SUTVA Holds E W , Z [ˆ µ cbr − ˆ µ cr ] = 0 σ 2 ] var W , Z [ˆ µ cr − ˆ µ cbr ] ≤ E W , Z [ˆ Reject the null when: | ˆ µ cbr | µ cr − ˆ 1 √ √ α ≥ σ 2 ˆ Type I error is no greater than α

  32. Nuts and Bolts of Running Cluster-based Randomized Experiments

  33. Why Balanced Clustering?

  34. Why Balanced Clustering? • Theoretical Motivation – Constants VS random variables

  35. Why Balanced Clustering? • Theoretical Motivation – Constants VS random variables • Practical Motivations

  36. Why Balanced Clustering? • Theoretical Motivation – Constants VS random variables • Practical Motivations – Variance reduction

  37. Why Balanced Clustering? • Theoretical Motivation – Constants VS random variables • Practical Motivations – Variance reduction – Balance on pre-treatment covariates (homophily => large homogenous clusters)

  38. Algorithms for Balanced Clustering

  39. Algorithms for Balanced Clustering Most clustering methods find skewed distributions of cluster sizes (Leskovec, 2009; Fortunato, 2010)

  40. Algorithms for Balanced Clustering Most clustering methods find skewed distributions of cluster sizes (Leskovec, 2009; Fortunato, 2010) => Algorithms that enforce equal cluster sizes

  41. Algorithms for Balanced Clustering Most clustering methods find skewed distributions of cluster sizes (Leskovec, 2009; Fortunato, 2010) => Algorithms that enforce equal cluster sizes Restreaming Linear Deterministic Greedy (Nishimura & Ugander, 2013)

  42. Algorithms for Balanced Clustering Most clustering methods find skewed distributions of cluster sizes (Leskovec, 2009; Fortunato, 2010) => Algorithms that enforce equal cluster sizes Restreaming Linear Deterministic Greedy (Nishimura & Ugander, 2013) – Streaming – Parallelizable – Stable

  43. Clustering the LinkedIn Graph – Graph: >100M nodes, >10B edges – 350 Hadoop nodes – 1% leniency

  44. Clustering the LinkedIn Graph – Graph: >100M nodes, >10B edges – 350 Hadoop nodes – 1% leniency 40% 35.6% % edges within clusters 28.5% 30% 26.2% 22.8% 21.1% 20% 10% 0% k = 1000 k = 3000 k = 5000 k = 7000 k = 10000

  45. Clustering the LinkedIn Graph – Graph: >100M nodes, >10B edges – 350 Hadoop nodes – 1% leniency 40% 35.6% % edges within clusters 30% 20% 10% 0% k = 1000 k = 3000 k = 5000 k = 7000 k = 10000

  46. Clustering the LinkedIn Graph – Graph: >100M nodes, >10B edges – 350 Hadoop nodes – 1% leniency 40% 35.6% % edges within clusters 28.5% 30% 26.2% 22.8% 21.1% 20% 10% 0% k = 1000 k = 3000 k = 5000 k = 7000 k = 10000

  47. Choosing the Number of Clusters

  48. Choosing the Number of Clusters small k large k

  49. Choosing the Number of Clusters small k large k small clusters large clusters

  50. Choosing the Number of Clusters small k large k small clusters large clusters small network effect large network effect small variance large variance

  51. Choosing the Number of Clusters Understanding the Type II error

  52. Choosing the Number of Clusters Understanding the Type II error Assuming an interference model

  53. Choosing the Number of Clusters Understanding the Type II error Assuming an interference model Y i = � 0 + � 1 Z i + � 2 ⇢ i + ✏ i ρ i : fraction of treated friends

  54. Choosing the Number of Clusters Understanding the Type II error Assuming an interference model Y i = � 0 + � 1 Z i + � 2 ⇢ i + ✏ i ρ i : fraction of treated friends E [ˆ µ cbr − ˆ µ cr ] ≈ ρ · β 2 : average fraction of a unit's neighbors contained in the cluster ρ i

  55. Choosing the Number of Clusters Understanding the Type II error Assuming an interference model Y i = � 0 + � 1 Z i + � 2 ⇢ i + ✏ i ρ i : fraction of treated friends E [ˆ µ cbr − ˆ µ cr ] ≈ ρ · β 2 : average fraction of a unit's neighbors contained in the cluster ρ i Choose number of clusters M and clustering C such that ρ max p σ 2 M,C ˆ C

  56. Experiments on LinkedIn

  57. Bernoulli Completely Randomized Cluster-based Randomized Randomized Experiment Experiment Experiment ? ( μ bernoulli ) μ completely-randomized μ cluster-based =

  58. Experiment 1

  59. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%]

  60. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%] Time period: 2 weeks –

  61. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%] Time period: 2 weeks – Number of clusters: k = 3,000 –

  62. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%] Time period: 2 weeks – Number of clusters: k = 3,000 – Outcome: feed engagement –

  63. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%] Time period: 2 weeks – Number of clusters: k = 3,000 – Outcome: feed engagement – Treatment effect Standard Deviation Bernoulli Randomization (BR) 0.0559 0.0050 Cluster-based Randomization (CBR) 0.0771 0.0260 Delta (CBR – BR) -0.0211 0.0265 p-value: 0.4246

  64. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%] Time period: 2 weeks – Number of clusters: k = 3,000 – Outcome: feed engagement – Treatment effect Standard Deviation Bernoulli Randomization (BR) 0.0559 0.0050 Cluster-based Randomization (CBR) 0.0771 0.0260 Delta (CBR – BR) -0.0211 0.0265 p-value: 0.4246

  65. Experiment 1 – Population: 20% of all LinkedIn users [Bernoulli: 10%, Cluster-based: 10%] Time period: 2 weeks – Number of clusters: k = 3,000 – Outcome: feed engagement – Treatment effect Standard Deviation Bernoulli Randomization (BR) 0.0559 0.0050 Cluster-based Randomization (CBR) 0.0771 0.0260 Delta (CBR – BR) -0.0211 0.0265 p-value: 0.4246

Recommend


More recommend