lect ure 24 advanced database
play

Lect ure # 24 ADVANCED DATABASE SYSTEMS Databases on New Hardware - PowerPoint PPT Presentation

Lect ure # 24 ADVANCED DATABASE SYSTEMS Databases on New Hardware @ Andy_Pavlo // 15- 721 // Spring 2018 2 ADM IN ISTRIVIA Snowflake Guest: May 2 th @ 3:00pm Final Exam Handout: May 2 nd Code Review #2: May 2 nd @ 11:59pm We will use the


  1. Lect ure # 24 ADVANCED DATABASE SYSTEMS Databases on New Hardware @ Andy_Pavlo // 15- 721 // Spring 2018

  2. 2 ADM IN ISTRIVIA Snowflake Guest: May 2 th @ 3:00pm Final Exam Handout: May 2 nd Code Review #2: May 2 nd @ 11:59pm → We will use the same group pairings as before. Final Presentations: May 14 th @ 8:30am → GHC 4303 (ignore schedule!) → 12 minutes per group → Food and prizes for everyone! CMU 15-721 (Spring 2018)

  3. 3 ADM IN ISTRIVIA Course Evaluation → Please tell me what you really think of me. → I actually take your feedback in consideration. → Take revenge on next year's students. https://cmu.smartevals.com/ CMU 15-721 (Spring 2018)

  4. 3 ADM IN ISTRIVIA Course Evaluation → Please tell me what you really think of me. → I actually take your feedback in consideration. → Take revenge on next year's students. https://cmu.smartevals.com/ CMU 15-721 (Spring 2018)

  5. 4 DATABASE H ARDWARE People have been thinking about using hardware to accelerate DBMSs for decades. 1980s: Database Machines 2000s: FPGAs + Appliances 2010s: FPGAs + GPUs DATABASE MACHINES: AN IDEA WHOSE TIME HAS PASSED? A CRITIQUE O OF THE FUTURE OF DATABASE M MACHINES University of Wisconsin 1983 CMU 15-721 (Spring 2018)

  6. 5 Non-Volatile Memory GPU Acceleration Hardware Transactional Memory CMU 15-721 (Spring 2018)

  7. 6 N O N- VO LATILE M EM O RY Emerging storage technology that provide low latency read/writes like DRAM, but with persistent writes and large capacities like SSDs. → aka Storage-class Memory, Persistent Memory First devices will be block-addressable (NVMe) Later devices will be byte-addressable. CMU 15-721 (Spring 2018)

  8. 7 FUN DAM EN TAL ELEM EN TS O F CIRCUITS Capacitor Resistor Inductor (ca. 1745) (ca. 1827) (ca. 1831) CMU 15-721 (Spring 2018)

  9. 8 FUN DAM EN TAL ELEM EN TS O F CIRCUITS In 1971, Leon Chua at Berkeley predicted the existence of a fourth fundamental element. A two-terminal device whose resistance depends on the voltage applied to it, but when that voltage is turned off it permanently remembers its last resistive state. TWO CENTURIES OF MEMRISTORS Nature Materials 2012 CMU 15-721 (Spring 2018)

  10. 9 FUN DAM EN TAL ELEM EN TS O F CIRCUITS Capacitor Resistor Inductor Memristor (ca. 1745) (ca. 1827) (ca. 1831) (ca. 1971) CMU 15-721 (Spring 2018)

  11. 10 M ERISTO RS A team at HP Labs led by Stanley Williams stumbled upon a nano-device that had weird properties that they could not understand. It wasn’t until they found Chua’s 1971 paper that they realized what they had invented. HOW WE FOUND THE MISSING MEMRISTOR IEEE Spectrum 2008 CMU 15-721 (Spring 2018)

  12. 11 TECH N O LO GIES Phase-Change Memory (PRAM) Resistive RAM (ReRAM) Magnetoresistive RAM (MRAM) CMU 15-721 (Spring 2018)

  13. 12 PH ASE- CH AN GE M EM O RY Storage cell is comprised of two metal electrodes separated by a resistive heater and the phase change material (chalcogenide). The value of the cell is changed based on Bitline how the material is heated. chalcogenide → A short pulse changes the cell to a ‘0’. → A long, gradual pulse changes the cell to a ‘1’. Heater Access PHASE CHANGE MEMORY ARCHITECTURE AND THE QUEST FOR SCALABILITY Communications of the ACM 2010 CMU 15-721 (Spring 2018)

  14. 13 RESISTIVE RAM Two metal layers with two TiO 2 layers in between. Running a current one direction moves electrons from the top TiO 2 layer to the bottom, thereby changing the resistance. May be programmable storage fabric… Platinum → Bertrand Russell’s Material Implication Logic TiO 2-x Layer TiO 2 Layer Platinum HOW WE FOUND THE MISSING MEMRISTOR IEEE Spectrum 2008 CMU 15-721 (Spring 2018)

  15. 14 M AGN ETO RESISTIVE RAM Stores data using magnetic storage elements instead of electric charge or current flows. Spin-Transfer Torque (STT-MRAM) is the leading technology for this type of NVM. → Supposedly able to scale to very small Fixed FM Layer→ sizes (10nm) and have SRAM latencies. Oxide Layer Free FM Layer ↔ SPIN MEMORY SHOWS ITS M MIGHT IEEE Spectrum 2014 CMU 15-721 (Spring 2018)

  16. 15 WH Y TH IS IS FO R REAL TH IS TIM E Industry has agreed to standard technologies and form factors. Linux and Microsoft have added support for NVM in their kernels (DAX). Intel has added new instructions for flushing cache lines to NVM ( CLFLUSH , CLWB ). CMU 15-721 (Spring 2018)

  17. 16 N VM DIM M FO RM FACTO RS NVDIMM-F (2015) → Flash only. Has to be paired with DRAM DIMM. NVDIMM-N (2015) → Flash and DRAM together on the same DIMM. → Appears as volatile memory to the OS. NVDIMM-P (2018) → True persistent memory. No DRAM or flash. CMU 15-721 (Spring 2018)

  18. 17 N VM CO N FIGURATIO N S DRAM as Hardware- NVM Next to NVM as Persistent Managed Cache DRAM Memory DBMS DBMS DBMS DBMS Address Space DBMS Address Space DBMS Address Space Virtual Memory Subsystem Virtual Memory Subsystem Buffer Pool DRAM Disk NVM DRAM NVM Filesystem Filesystem NVM Source: Ismail Oukid CMU 15-721 (Spring 2018)

  19. 18 N VM FO R DATABASE SYSTEM S Block-addressable NVM is not that interesting. Byte-addressable NVM will be a game changer but will require some work to use correctly. → In-memory DBMSs will be better positioned to use byte- addressable NVM. → Disk-oriented DBMSs will initially treat NVM as just a faster SSD. CMU 15-721 (Spring 2018)

  20. 19 STO RAGE & RECOVERY M ETH O DS Understand how a DBMS will behave on a system that only has byte-addressable NVM. Develop NVM-optimized implementations of standard DBMS architectures. Based on the N-Store prototype DBMS. LET'S TALK ABOUT STORAGE & RECOVERY METHODS FOR NON- VOLATILE MEMORY DATABASE SYSTEMS SIGMOD 2015 CMU 15-721 (Spring 2018)

  21. 20 SYN CH RO N IZATIO N Existing programming models assume that any write to memory is non-volatile. → CPU decides when to move data from caches to DRAM. The DBMS needs a way to ensure that data is flushed from caches to NVM. STORE Memory L1 Cache Controller L2 Cache CMU 15-721 (Spring 2018)

  22. 20 SYN CH RO N IZATIO N Existing programming models assume that any write to memory is non-volatile. → CPU decides when to move data from caches to DRAM. The DBMS needs a way to ensure that data is flushed from caches to NVM. STORE CLWB Memory L1 Cache Controller L2 Cache CMU 15-721 (Spring 2018)

  23. 20 SYN CH RO N IZATIO N Existing programming models assume that any write to memory is non-volatile. → CPU decides when to move data from caches to DRAM. The DBMS needs a way to ensure that data is flushed from caches to NVM. STORE CLWB ADR Memory L1 Cache Controller L2 Cache CMU 15-721 (Spring 2018)

  24. 21 N AM IN G If the DBMS process restarts, we need to make sure that all of the pointers for in-memory data point to the same data. Index Table Heap Tuple #00 Tuple #01 Tuple #02 Tuple #00 ( v2 ) CMU 15-721 (Spring 2018)

  25. 21 N AM IN G If the DBMS process restarts, we need to make sure that all of the pointers for in-memory data X X point to the same data. Index Table Heap Tuple #00 Tuple #01 Tuple #02 Tuple #00 ( v2 ) CMU 15-721 (Spring 2018)

  26. 21 N AM IN G If the DBMS process restarts, we need to make sure that all of the pointers for in-memory data point to the same data. Index Table Heap Tuple #00 Tuple #01 Tuple #02 Tuple #00 ( v2 ) CMU 15-721 (Spring 2018)

  27. 22 N VM - AWARE M EM O RY ALLO CATO R Feature #1: Synchronization → The allocator writes back CPU cache lines to NVM using the CLFLUSH instruction. → It then issues a SFENCE instruction to wait for the data to become durable on NVM. Feature #2: Naming → The allocator ensures that virtual memory addresses assigned to a memory-mapped region never change even after the OS or DBMS restarts. CMU 15-721 (Spring 2018)

  28. 23 DBM S EN GIN E ARCH ITECTURES Choice #1: In-place Updates → Table heap with a write-ahead log + snapshots. → Example: VoltDB Choice #2: Copy-on-Write → Create a shadow copy of the table when updated. → No write-ahead log. → Example: LMDB Choice #3: Log-structured → All writes are appended to log. No table heap. → Example: RocksDB CMU 15-721 (Spring 2018)

  29. 24 IN- PLACE UPDATES EN GIN E In-Memory In-Memory Durable Index Table Heap Storage Write-Ahead Log Tuple #00 Tuple #01 Tuple #02 Snapshots CMU 15-721 (Spring 2018)

  30. 24 IN- PLACE UPDATES EN GIN E In-Memory In-Memory Durable Index Table Heap Storage Write-Ahead Log Tuple #00 1 Tuple Delta Tuple #01 Tuple #02 Snapshots CMU 15-721 (Spring 2018)

  31. 24 IN- PLACE UPDATES EN GIN E In-Memory In-Memory Durable Index Table Heap Storage Write-Ahead Log Tuple #00 2 1 Tuple Delta Tuple #01 (!) Tuple #01 Tuple #02 Snapshots CMU 15-721 (Spring 2018)

  32. 24 IN- PLACE UPDATES EN GIN E In-Memory In-Memory Durable Index Table Heap Storage Write-Ahead Log Tuple #00 2 1 Tuple Delta Tuple #01 (!) Tuple #01 Tuple #02 Snapshots 3 Tuple #01 (!) CMU 15-721 (Spring 2018)

  33. 24 IN- PLACE UPDATES EN GIN E In-Memory In-Memory Durable Index Table Heap Storage Duplicate Data Write-Ahead Log Tuple #00 2 1 Tuple Delta Tuple #01 (!) Tuple #01 Tuple #02 Recovery Latency Snapshots 3 Tuple #01 (!) CMU 15-721 (Spring 2018)

Recommend


More recommend