Facebook and the Open Compute Project Charlie Manese Infrastructure NSF SDC - June 22, 2015
January 2015
Data Triplet Mezzanine Center Rack Card v2 Battery Freedom Windmill Open Group Hug Cold Open Cabinet Servers (Intel) Rack v1 Storage Rack v2 Spitfire Power Watermark Mezzanine Winterfell Knox Micro Server Honey Supply Server (AMD) (AMD) Card v1 (Panther) Badger 2011 2012 2013 2014
Open data center stack Cold Storage Cooling Leopard Knox Open Rack Wedge Battery Power 6-Pack
Software Network HipHop Virtual Machine 5 x ¡– ¡6 x faster ¡than ¡ Servers & Storage Zend Data Center
Software Network Servers & Storage Data Center
Original OCP designs Software Cost Network 38 % 24 Servers & Storage % Energy Efficiency Data Center
Software Network Servers & Storage Data Center
Efficiency gains with OCP $2 Billion
Efficiency gains with OCP 95,000 80,000 Cars Homes Annual Energy Savings Annual Carbon Savings
Design principles ▪ Efficiency ▪ Scale ▪ Simplicity ▪ Vanity Free ▪ Easy to Operate
DATA CENTER
Facebook greenfield datacenter Goal ▪ Design and build the most efficient datacenter eco-system possible Control ▪ Application ▪ Server configuration ▪ Datacenter design
Prineville , OR Forest City, NC Luleå, Sweden
Electrical overview ▪ Eliminate 480V to 208V transformation ▪ Used 480/277VAC distribution to IT equipment ▪ Remove centralized UPS ▪ Implemented 48VDC UPS System ▪ Result a highly efficient electrical system and small failure domain
Typical Power Prineville Power Utility Transformer Utility Transformer Standby Standby 480/277 VAC 480/277 VAC Generator Generator 2% loss 2% loss AC/DC UPS 480VAC DC/AC 6% - 12% loss ASTS/PDU 208/120VAC 99.9999% 99.999% Availability Availability 3% loss 480/277VAC FB SERVER SERVER PS 10% loss 48VDC DC UPS (assuming 90% plus PS (Stand-by) PS) 5.5% loss Total loss up to server: Total loss up to server: 21% to 27% 7.5%
Reactor power panel ▪ Custom Fabricated RPP ▪ Delivers 165kW, 480/277V, 3-phase to CAB level ▪ Contains Cam-Lock connector for maintenance wrap around ▪ Line Reactor ▪ Reduces short circuit current < 10kA ▪ Corrects leading power factor towards unity (3% improvement) ▪ Reduces THD for improved electrical system performance (iTHD 2% improvement) ▪ Power consumption = 360 Watt
Battery cabinet ▪ Custom DC UPS ▪ 56kW or 85kW ▪ 480VAC, 3-phase input ▪ 45 second back-up ▪ 20 sealed VRLA batteries ▪ Battery Validation System ▪ Six 48VDC Output ▪ Two 50A 48VDC aux outputs
Mechanical overview ▪ Removed ▪ Centralized chiller plant ▪ HVAC ductwork ▪ System Basis of Design ▪ ASHRAE Weather Data: N=50 years ▪ TC9.9 2008: Recommended Envelopes ▪ Built-up penthouse air handling system ▪ Server waste heat is used for office space heating
Typical datacenter cooling Return Ductwork CT AHU CHILLER SUPPLY DUCT DATA CENTER Prineville datacenter cooling Relief Air Return Air Ductless Return Air Ductless Relief Air Return 100% Outside Air Air Intake Plenum DUCTLESS DATA CENTER SUPPLY Filter Evap Fan Wall System Wall
PRN datacenter cooling Evap Filter Fan System Wall Wall Mixed Air Supply Air Relief Air 100% Corridor Corridor Corridor Outside Air Intake Relief Air Fan Common Hot Aisle Hot Aisle Return Air Relief Air Return Air Plenum Common Hot Aisle Hot Aisle Cold Aisle Data Center Server Server Cabinets Cabinets
Cold aisle pressurization – ductless supply
Basis of design comparison 80ºF/27ºC inlet 85ºF/30ºC inlet 85ºF/30ºC inlet 85ºF/30ºC inlet 65% humidity 80% humidity 90% humidity 80% humidity 20ºF/11ºC Δ T 22ºF/11ºC Δ T 22ºF/11ºC Δ T 22ºF/11ºC Δ T PRN1A1B PRN1C1D FRC1A1B LLA1A1B
RACK, SERVERS, AND STORAGE
Open Compute Rack: Open Rack • Well-defined “Mechanical API” between the server and the rack • Accepts any size equipment 1U – 10U • Wide 21” equipment bay for maximum space efficiency • Shared 12v DC power system
Open Compute Server v2 • First step with shared components by reusing PSU and fans between two servers • Increased rack density without sacrificing efficiency or cost • All new Facebook deployments in 2012 were “v2” servers
Open Compute Server v3 • Reuses the “v2” half-width motherboards • Self-contained sled for Open Rack • 3-across 2U form factor enables 80mm fans with 45 servers per rack
Open Vault • Storage JBOD for Open Rack • Fills the volume of the rack without sacrificing hot-swap
NETWORK
Traffic growth
Fabric
Wedge
FBOSS
6-Pack
SERVICEABILITY
Complex designs Typical large datacenter: 1000 Servers per Technician
Complex Simple designs Typical large datacenter: 1000 Servers per Technician Facebook datacenter: 25,000 Servers per Technician
Efficiency through serviceability Standing ¡at ¡Machine OEM ¡REPAIRS Pre-‑Repair ¡Activities ¡ Part ¡Swap ¡Duration ¡ Additional ¡ Post-‑Repair ¡ Total ¡Repair ¡Time ¡ Min Min Steps ¡Min Activities ¡Min Min Hard ¡Drive ¡(Non-‑raid) 2 3 0 2 7 DIMM ¡(Offline) 2 3 0 2 7 Motherboard 2 20 20 2 44 PSU ¡(Hot ¡Swap) 2 5 0 2 9 OCP#1 ¡REPAIRS Pre-‑Repair ¡Activities ¡ Part ¡Swap ¡Duration ¡ Additional ¡ Post-‑Repair ¡ Total ¡Repair ¡Time ¡ Min Min Steps ¡Min Activities ¡Min Min Hard ¡Drive ¡(Non-‑raid) 0 0.98 0 0 0.98 DIMM ¡(Offline) 0 0.82 0 0 0.82 Motherboard 2.5 10.41 2.5 0 15.41 PSU ¡(Hot ¡Swap) 0 0.65 0 0 0.65
First-time-fix repair rates 100% 99% 98% 97% 96% 95% Target 94% 93% 92% 91% 90% 89% 88% 87% 86% 85% Jul 12 Aug 12 Sep 12 Oct 12 Nov 12 Dec 12
Let’s engage
KEYNOTE
Recommend
More recommend