devoflow scaling flow management for high performance
play

DevoFlow: Scaling Flow Management for High-Performance Networks - PowerPoint PPT Presentation

DevoFlow: Scaling Flow Management for High-Performance Networks Andy Curtis Je ff Mogul Jean Tourrilhes Praveen Yalagandula Puneet Sharma Sujata Banerjee Wednesday, August 17, 11 Software-de fi ned networking Wednesday, August 17, 11


  1. DevoFlow: Scaling Flow Management for High-Performance Networks Andy Curtis Je ff Mogul Jean Tourrilhes Praveen Yalagandula Puneet Sharma Sujata Banerjee Wednesday, August 17, 11

  2. Software-de fi ned networking Wednesday, August 17, 11

  3. Software-de fi ned networking • Enables programmable networks Wednesday, August 17, 11

  4. Software-de fi ned networking • Enables programmable networks • Implemented by OpenFlow Wednesday, August 17, 11

  5. Software-de fi ned networking • Enables programmable networks • Implemented by OpenFlow • OpenFlow is a great concept, but... - its original design imposes excessive overheads Wednesday, August 17, 11

  6. Traditional switch Control-plane Data-plane Wednesday, August 17, 11

  7. Traditional switch Control-plane Routed packets Inbound Data-plane packets Wednesday, August 17, 11

  8. Traditional switch Reachability Reachability Control-plane Routed packets Inbound Data-plane packets Wednesday, August 17, 11

  9. Control-plane Centralized controller OpenFlow switch Inbound packets Routed packets Data-plane Wednesday, August 17, 11

  10. Control-plane Centralized controller Flow setups Forwarding table entries Link state Statistics requests Forwarding rule stats Inbound packets Routed packets Data-plane Wednesday, August 17, 11

  11. OpenFlow enables innovative management solutions Wednesday, August 17, 11

  12. OpenFlow enables innovative management solutions • Consistent routing and security policy enforcement [Ethane, SIGCOMM 2007] • Data center network architectures like VL2 and PortLand [Tavakoli et al. Hotnets 2009] • Client load-balancing with commodity switches [Aster*x, ACLD demo 2010; Wang et al., HotICE 2011] • Flow scheduling [Hedera, NSDI 2010] • Energy-proportional networking [ElasticTree, NSDI 2010] • Automated data center QoS [Kim et al., INM/WREN 2010] Wednesday, August 17, 11

  13. But OpenFlow is not perfect... • Scaling these solutions to data center- sized networks is challenging Wednesday, August 17, 11

  14. Contributions • Characterize overheads of implementing OpenFlow in hardware • Propose DevoFlow to enable cost- e ff ective, scalable fl ow management • Evaluate DevoFlow by applying it to data center fl ow scheduling Wednesday, August 17, 11

  15. Contributions • Characterize overheads of implementing OpenFlow in hardware Experience drawn from implementing OpenFlow on HP ProCurve switches Wednesday, August 17, 11

  16. OpenFlow couples fl ow setup with visibility Central controller . . . Core Aggregation . . . Edge switches Wednesday, August 17, 11

  17. Central controller . . . Core Aggregation . . . Edge switches Flow arrival Wednesday, August 17, 11

  18. If no forwarding table rule at switch (exact-match or wildcard) Central controller . . . Core Aggregation . . . Edge switches Wednesday, August 17, 11

  19. Two problems arise... Central controller . . . Core Aggregation . . . Edge switches Wednesday, August 17, 11

  20. problem 1: bottleneck at controller Central controller . . . . . . Wednesday, August 17, 11

  21. problem 1: bottleneck at controller Up to 10 million new fl ows per second in data center with 100 edge switches [Benson et al. IMC 2010] Central controller . . . . . . Wednesday, August 17, 11

  22. problem 1: bottleneck at controller Up to 10 million new fl ows per second in data center with 100 edge switches [Benson et al. IMC 2010] If controller can handle 30K fl ow setups/ sec. then, we need at least 333 controllers! Central controller . . . . . . Wednesday, August 17, 11

  23. Onix [Koponen et al. OSDI 2010] Maestro [Cai et al. Tech Report 2010] HyperFlow [Tootoonchian and Ganjali, WREN 2010] Devolved controller [Tam et al. WCC 2011] Central controller . . . . . . Wednesday, August 17, 11

  24. problem 2: stress on switch control-plane Central controller . . . . . . Wednesday, August 17, 11

  25. Control-plane Inbound packets Routed packets Data-plane Wednesday, August 17, 11

  26. Control-plane Switch Switch CPU control-plane Inbound packets Routed packets ASIC Data-plane Wednesday, August 17, 11

  27. Control-plane Switch Switch CPU control-plane Inbound packets Routed packets ASIC Data-plane Wednesday, August 17, 11

  28. Scaling problem: switches • Inherent overheads • Implementation-imposed overheads Wednesday, August 17, 11

  29. Scaling problem: switches • Inherent overheads - Bandwidth • OpenFlow creates too much control tra ffi c ~1 control packet for every 2–3 data packets - Latency • Implementation-imposed overheads Wednesday, August 17, 11

  30. Scaling problem: switches • Inherent overheads • Implementation-imposed overheads - Flow setup - Statistics gathering - State size (see paper) Wednesday, August 17, 11

  31. Flow setup OpenFlow controller Client A Client B ProCurve 5406 zl switch Wednesday, August 17, 11

  32. Flow setup OpenFlow controller Client A Client B ProCurve 5406 zl switch We believe our measurement numbers are representative of the current generation of OpenFlow switches Wednesday, August 17, 11

  33. Flow setup 300 275 250 Flow setups per sec. 200 150 100 50 0 5406 zl Wednesday, August 17, 11

  34. Flow setup We can expect up 10,000 10000 to 10K fl ow 9000 arrivals / sec. 8000 [Benson et al. IMC 2010] Flow setups per sec. 7000 40x di ff erence! 6000 5000 4000 3000 2000 1000 0 5406 zl Expected Wednesday, August 17, 11

  35. Flow setup 10,000 10000 9000 Too much latency: 8000 Flow setups per sec. adds 2ms to fl ow 7000 setup 6000 5000 4000 3000 2000 1000 0 5406 zl Expected Wednesday, August 17, 11

  36. Control-plane 300 Gbps Switch CPU Inbound Routed packets packets Data-plane Wednesday, August 17, 11

  37. Control-plane 80 Mbps Switch CPU Inbound Routed packets packets Data-plane Wednesday, August 17, 11

  38. Control-plane 17 Mbps Switch CPU Inbound Routed packets packets Data-plane Wednesday, August 17, 11

  39. Stats-gathering • Flow setups and stat-pulling compete for this bandwidth Wednesday, August 17, 11

  40. Stats-gathering • Flow setups and stat-pulling compete for this bandwidth 300 Flow setups per sec. 250 200 150 100 50 0 Stat-pull frequency: Never 1 s 500 ms Wednesday, August 17, 11

  41. Stats-gathering • Flow setups and stat-pulling compete for this bandwidth - 2.5 sec. to collect stats from the average data center edge switch Wednesday, August 17, 11

  42. Can we solve the problem with more hardware? • Faster CPU may help, but won’t be enough - Control-plane datapath needs at least two orders of magnitude more bandwidth • Ethernet speeds accelerating faster than CPU speeds • OpenFlow won’t drive chip-area budgets for several generations Wednesday, August 17, 11

  43. Contributions • Characterize overheads of implementing OpenFlow in hardware • Propose DevoFlow to enable cost- e ff ective, scalable fl ow management • Evaluate DevoFlow by applying it to data center fl ow scheduling Wednesday, August 17, 11

  44. Devolved OpenFlow We devolve control over most fl ows back to the switches Wednesday, August 17, 11

  45. DevoFlow design • Keep fl ows in the data-plane • Maintain just enough visibility for e ff ective fl ow management • Simplify the design and implementation of high-performance switches Wednesday, August 17, 11

  46. DevoFlow mechanisms • Control mechanisms • Statistics-gathering mechanisms Wednesday, August 17, 11

  47. DevoFlow mechanisms • Control mechanisms - Rule cloning • ASIC clones a wildcard rule as an exact match rule for new micro fl ows Wednesday, August 17, 11

  48. DevoFlow mechanisms • Control mechanisms - Rule cloning src dst src port dst Port wildcard rules * 129.100.1.5 * * src dst src port dst Port exact-match rules Wednesday, August 17, 11

  49. DevoFlow mechanisms • Control mechanisms - Rule cloning src dst src port dst Port wildcard rules * 129.100.1.5 * * src dst src port dst Port exact-match rules Wednesday, August 17, 11

  50. DevoFlow mechanisms • Control mechanisms - Rule cloning src dst src port dst Port wildcard rules * 129.100.1.5 * * src dst src port dst Port exact-match rules Wednesday, August 17, 11

  51. DevoFlow mechanisms • Control mechanisms - Rule cloning src dst src port dst Port wildcard rules * 129.100.1.5 * * src dst src port dst Port exact-match 129.200.1.1 129.100.1.5 4832 80 rules Wednesday, August 17, 11

  52. DevoFlow mechanisms • Control mechanisms - Rule cloning • ASIC clones a wildcard rule as an exact match rule for new micro fl ows - Local actions • Rapid re-routing • Gives fallback paths for when a port fails • Multipath support Wednesday, August 17, 11

  53. Control-mechanisms • Control mechanisms - Rule cloning • ASIC clones a wildcard rule as an exact match rule for new micro fl ows - Local actions • 1/3 1/6 1/2 Rapid re-routing • Gives fallback paths for when a port fails • Multipath support Wednesday, August 17, 11

  54. Statistics-gathering mechanisms Wednesday, August 17, 11

  55. Statistics-gathering mechanisms • Sampling - Packet header is sent to controller with 1/1000 probability Wednesday, August 17, 11

Recommend


More recommend