the computer network behind the social network
play

The Computer Network behind the Social Network Jam James s Hon - PowerPoint PPT Presentation

The Computer Network behind the Social Network Jam James s Hon Hongyi yi Ze Zeng ng Engineering Manager, Network Infra APNet 2019, Beijing, China Facebook Family 2.7B people every month 2.1B people every day (Q2, 2019) About Me


  1. The Computer Network behind the Social Network Jam James s Hon Hongyi yi Ze Zeng ng Engineering Manager, Network Infra APNet 2019, Beijing, China

  2. Facebook Family 2.7B people every month 2.1B people every day (Q2, 2019)

  3. About Me • Joined Facebook networking in 2014 • Supporting Routing and UI team • https://research.fb.com/category/systems-and-networking/

  4. Lu Luleå, , Sweden Alt Altoona, IA Clonee, Cl , Ireland Odense, Od , Denmark Pr Prineville, OR Papillion, NE Pa Forest City, Fo , NC Los Lu Lo Lunas, NM Fort Worth, Fo , TX

  5. How Users Reach Facebook Datacenter Network Backbone Internet Edge Network Backbone Network

  6. • Edge Network Agenda • Backbone Network • Datacenter Network

  7. • Edge Network Agenda • Backbone Network • Datacenter Network

  8. Edge Network • Goal: Delivers the traffic to ISP and ultimately to users • Majority of users are on mobile • Majority of users are on IPv6 • IPv6 penetration rate is at 56% in the United States • https://www.facebook.com/ipv6/

  9. Facebook’s Traffic Dynamic Requests Static Requests (not Cachable) (Cachable) News Feed Photos Likes Videos Messaging JavaScript Status Updates

  10. DNS Based Load Balancing US-EAST Web Web L7LB L4LB Web L7LB Server L4LB Web L7LB Server L4LB L7LB Server Server Internet US-WEST Web Web L7LB L4LB Web L7LB Server L4LB Web L7LB Server L4LB L7LB Server Server www? DNS LB US-east .facebook.com

  11. POP + DC L7LB L4LB L7LB L4LB L7LB L4LB L7LB Internet Backbone Web Web L7LB L4LB Web L7LB Server L4LB Web L7LB Server L4LB L7LB Server Server

  12. How about static content? JS L7LB L4LB Photo L7LB L4LB L7LB L4LB L7LB Video Internet Backbone Web Web L7LB L4LB Web L7LB Server L4LB Web L7LB Server L4LB L7LB Server Server

  13. Edge Network Summary • Software Hierarchy to scale • DNS Load Balancer (to Datacenter/POP) • Router + Anycast BGP, Layer 3 Load balancer (to Layer 4 Load Balancer) • Layer 4 Load Balancer (to Layer 7 Load Balancer) • Layer 7 Load Balancer (to Web Server) • POP + DC to scale • Reduce RTT for initial setup • Cache content closer to users

  14. • Edge Network Agenda • Backbone Network • Datacenter Network

  15. Backbones at Facebook • Classic Backbone (CBB) • Connects POP and DCs • RSVP-TE, Vendor software solution • Express Backbone (EBB) • Connects DC and DC • Centralized control

  16. Three Datacenters A B C

  17. Add Planes A B C

  18. N-way Active-active Redundancy A B C

  19. Incremental changes and canary A B C

  20. A/B Testing Algorithm 1 Algorithm 2

  21. Open/R • Routing Protocol supports EBB • Establish basic reachability among routers (OSPF, IS-IS) • Extensible (e.g., key-value store) • In-house software • Run as agent on EBB routers • EBB is first production network where Open/R is the sole IGP

  22. Typical IGP metric configuration Type Link Metric Trans Atlantic 100 Trans Pacific 150 US-West to US- 50 East

  23. Open/R: Calculate link metric with RTT RTT = 200ms Open/R Open/R metric = 200

  24. Backbone Network Summary • Two backbones • CBB: Connects POPs and DCs • EBB: Inter-DC backbone • Plane architecture • Reliability, maintenance, experiment • Software • Centralized control • Innovative distributed routing protocols to minimize configuration

  25. • Edge Network Agenda • Backbone Network • Datacenter Network

  26. Classic Facebook Fabric 48 48 3 48 3 2 3 48 2 1 3 2 1 1 2 1 4 4 4 1 4 1 2 4 1 1 2 4 3 2 3 4 1 3 4 4 2 1 4 3 2 4 48 ports in 3 Pod Y 4 Pod X Pod 4 48 Pod 3 Pod 2 48 Pod 1 1 2 3 4 5 48 6 7 8 48 9 10 48 11 48

  27. Growing Pressure Mega Regions . 48 Expanding Mega Regions . . . . . 48 3 . . . 48 3 2 (5-6 buildings) = accelerated 3 . . . 48 2 1 3 2 1 fabric-to-fabric East-West demand 1 2 1 4 4 . . . Compute-Storage and AI disaggregation . . . 4 1 4 1 2 4 1 1 . . . 2 3 4 requires Terabit capacity per Rack . 2 . . 3 4 1 3 4 4 2 1 4 3 2 4 3 4 . . . . . . 48 Both require larger fabric Spine . . . 48 1 2 . 3 . 4 . 5 48 6 capacity (by 2-4x ) ... 7 8 48 9 10 48 11 . . . 48 Disaggregated Services

  28. F16 – Facebook’s new topology 16-plane architecture 6-16x spine capacity on day 1 1.6T raw capacity per rack Fewer chips* = better power & space

  29. Mega Region

  30. Mega Region F16 Fabric Aggregator

  31. Minipack – 128 x 100G Switch • Single 12.8T ASIC • Modular design • Mature optics • Lower power/smaller size

  32. Fabric Aggregator • Disaggregated design for scale • Built upon smaller commodity switches

  33. White Box Switch Customizable switch hardware and software Power Supply • Customized hardware Fan Temperature • Pick the minimal Sensor x86 CPU software needed for the SSD specific network BMC • Powerful CPU to run Switch ASIC CPLD more complex software QSFP Ports

  34. FBOSS Overview External Software Protocols Monitoring Network (BGP, ECMP) Service Configurator Switch Software FBOSS Switch Hardware Switch ASIC

  35. FBOSS Design Principles • Switch-as-a-Server • Continuous integration and staged deployment • Integrate closely with existing software services • Open-source software • Deploy-Early-and-Iterate • Focus on developing and deploying minimal set of features • Quickly iterate with smaller “ diffs”

  36. FBOSS Testing and Deployment 3 Stage Deployment via fbossdeploy • Continuous Canary • Deploy all commits continuously to 1~2 switches for each type • Daily Canary • Deploy all of single day’s commits to 10~20 switches for each type • Staged Deployment • Final stage to push all the commits to all the switches in the DC • Performed once every two weeks for reliability

  37. Datacenter Network Summary • Datacenters are huge • Internally: Clos topology • Intra-region connectivity is challenging too • In-house Hardware and Software • Minipack, Fabric Aggregator • FBOSS

  38. Summary Datacenter Network Backbone Internet Edge Network Backbone Network

  39. Extended Reading • Inside the Social Network’s (Datacenter) Network, SIGCOMM 2015 • Robotron: Top-down Network Management at Facebook Scale, SIGCOMM 2016 • Engineering Egress with Edge Fabric: Steering Oceans of Content to the World, SIGCOMM 2017 • FBOSS: Building Switch Software at Scale, SIGCOMM 2018

Recommend


More recommend