reduction of operating system jitter
play

Reduction of Operating System Jitter Caused by Page Reclaim - PowerPoint PPT Presentation

Reduction of Operating System Jitter Caused by Page Reclaim Yoshihiro Oyama 1,3 Shun Ishiguro 1 Jun Murakami 1 Shin Sasaki 1 Ryo Matsumiya 1 Osamu Tatebe 2,3 1. The University of Electro-Communications 2. University of Tsukuba 3. Japan Science


  1. Reduction of Operating System Jitter Caused by Page Reclaim Yoshihiro Oyama 1,3 Shun Ishiguro 1 Jun Murakami 1 Shin Sasaki 1 Ryo Matsumiya 1 Osamu Tatebe 2,3 1. The University of Electro-Communications 2. University of Tsukuba 3. Japan Science and Technology Agency

  2. Background • OS jitter: interference into applications by OS – Services by OS kernel • E.g., interrupt handling and tasklets – Daemon processes developed to provide OS services • E.g., memory management daemons • Jitter degrades application performance – It deprives applications from computing resources such as CPU and memory • Minimizing the impact of jitter is critical in HPC

  3. Jitter Focused in This Study • We focus on jitter observed when application frequently executes disk I/O of large data – Footprint of file data exceeds the physical memory size – Kernel must discard page cache or swap out processes to obtain free memory – Overhead is imposed on memory allocation operations • This jitter has not attracted much attention, but HPC people should be aware of its potential impact

  4. Overview of This Study 1. We clarify the impact of the jitter caused by page reclaim – Target OS is Linux 2. We propose a mechanism for minimizing the impact – It increases the amount of page cache released at one time – It reduces the number of page reclaim operations

  5. Page Cache application application application access memory read disk blocks disk

  6. Memory Pressure by Page Cache application application application memory disk

  7. Memory Pressure by Page Cache application application application memory Page reclaim frequently occurs when: • Pages are consumed fast • Only a small number of pages are released at one time disk

  8. Memory Pressure by Page Cache application application application memory Page reclaim frequently occurs when: • Pages are consumed fast • Only a small number of pages are released at one time disk

  9. Page Reclaim in Linux • Memory pages are running short -> it immediately reclaims memory (direct reclaim) -> it awakens a kernel thread kswapd • kswapd reclaims pages by flushing page cache or swapping process memory • Two values inside kswapd are particularly important – Freemem threshold • Kswapd is awakened if the amount of free memory falls below this threshold – Watermark • Kswapd continues to reclaim pages until the amount of free pages exceeds this value Modifiable indirectly through /proc/sys/vm/min_free_kbytes Unfortunately, it is the only parameter effective in minimizing page reclaim jitter

  10. Page Reclaim by kswapd memory objects page cache freemem free threshold watermark normal state page reclaim started page reclaim finished reclaimed pages

  11. Proposed Mechanism (1) • Introduces new kernel module and kernel thread – Starts page reclaim before kswapd – Reclaims larger # of pages at once our kswapd our kswapd freemem freemem watermark watermark threshold threshold normal state page reclaim started page reclaim finished reclaimed pages

  12. System Structure Specifies # of pages Specifies threshold reclaimed at once of free pages /proc/min_free_pages /proc/reclaimed_pages Invokes page kernel kernel thread reclaim function Read Monitors # of Monitors disk-read requests free pages memory disk

  13. Proposed Mechanism (2) • It starts when both conditions are satisfied: – Cond. 1: # of free pages < our freemem threshold – Cond. 2: our mechanism determines that memory shortage is caused by frequent I/O • Otherwise, our kernel thread does not start – And eventually kswapd will be awakened • We expect kswapd will do a good job in minimizing page-outs of memory objects

  14. Discussion • Q: Why introducing a kernel thread, instead of customizing kswapd? – Tuning kswapd parameters – Modifying kswapd code • A: Kswapd provides only a few parameters – For example, kswapd users cannot directly specify the amount of reclaimed memory – But, we would like to investigate a vast space of parameters and algorithms – This inconvenience is also pointed out by another Linux engineer: https://lwn.net/Articles/422291/

  15. Experiments • We measured the impact of jitter on the performance of a scientific application • Application: WRF (weather forecasting software) – Simulated the weather around Japan in one hour (6 s x 600 steps) • Jitter generator – Program that repeatedly reads a 100-GB file sequentially • Although it represents an extreme case, we believe that a similar case can possibly occur in some configurations and job sets

  16. Condition InfiniBand QDR 4X, MPI Node 2 Node 1 Node 3 Node 4 ... ... ... ... 11 threads 11 threads 11 threads 11 threads Jitter generator Machine specification: CPU: Intel Xeon E5645 2.4 GHz (6 cores) x 2 Memory: 48 GB HDD: SAS 15,000 rpm

  17. Experiment 1 • We compared WRF performance in 3 cases – Original – With jitter – With jitter and proposed mechanism (Jitter+Proposed)

  18. Result (Not Using Proposed Mechanism) 20 Original Computation time of each step (s) 18 16 14 12 10 8 6 4 Original 2 0 50 100 150 200 250 Step

  19. Result (Not Using Proposed Mechanism) 20 Computation time of each step (s) 18 16 14 12 10 8 6 4 Original Jitter 2 0 50 100 150 200 250 Step

  20. Accumulated Computation Time (Not Using Proposed Mechanism) 7000 6000 26.6% slowdown because of jitter! 5000 Computation time (s) 4000 3000 2000 1000 0 Original Jitter Jitter+Proposed Jitter+Proposed (4 GiB reclaim) (48 GiB reclaim)

  21. Result (Using Proposed Mechanism) 20 Computation time of each step (s) 18 16 14 12 10 8 6 4 Original Jitter 2 0 50 100 150 200 250 Step

  22. Result (Using Proposed Mechanism) 20 Computation time of each step (s) 18 16 14 12 10 8 6 Original 4 Jitter 2 Jitter+Proposed (4 GiB reclaim) 0 50 100 150 200 250 Step

  23. Result (Using Proposed Mechanism) 20 Computation time of each step (s) 18 16 14 12 10 8 6 Original 4 Jitter Jitter+Proposed (4 GiB reclaim) 2 Jitter+Proposed (48 GiB reclaim) 0 50 100 150 200 250 Step

  24. Accumulated Computation Time (Using Proposed Mechanism) 7000 Only 1.9% slowdown 26.6% 6000 slowdown 5000 Computation time (s) 4000 3000 2000 1000 0 Original Jitter Jitter+Proposed Jitter+Proposed (4 GiB reclaim) (48 GiB reclaim)

  25. Experiment 2 • In addition, we must answer – “How good performance can we get by changing parameters of kswapd ?” – “Is kswapd parameter tuning sufficient to obtain comparative performance?” • We measured WRF performance in Jitter case with various kswapd parameters

  26. Effect of kswapd Parameter Changes 20 Computation time of each step (s) 18 16 14 12 10 8 6 Original 4 Jitter (kswapd threshold: 88 MiB) 2 0 50 100 150 200 250 Step

  27. Effect of kswapd Parameter Changes 20 Computation time of each step (s) 18 16 14 12 10 8 6 Original Jitter (kswapd threshold: 88 MiB) 4 Jitter (kswapd threshold: 2 GiB) 2 Jitter (kswapd threshold: 4 GiB) 0 50 100 150 200 250 Step

  28. Effect of kswapd Parameter Changes 7000 +12.8% +14.4% 6000 Computation time (s) 5000 4000 3000 2000 1000 0 Original Jitter Jitter+Proposed Jitter+Proposed Jitter Jitter (4 GiB reclaim) (48 GiB reclaim) (2 GiB threshold) (4 GiB threshold)

  29. Related Work • “Core separation” approaches – [De et al. IPDPS 2009], [Oral et al. 2010], [Rosenthal et al. 2013], [Seelam et al. IPDPS 2011] – Executes the kernel and daemons on dedicated CPU cores – Executes applications on remaining CPU cores – Prevents the kernel and daemons from depriving applications of CPU resources It is unclear how many CPU cores are sufficient for hosting kswapd threads and other system tasks – Their approach should be combined with another approach for reducing the impact of jitter

  30. Summary and Future Work • Summary – We proposed a mechanism for reducing the impact of jitter caused by page reclaim – Jitter caused by an I/O-intensive process increased the execution time of WRF by 26.6% – The mechanism lowered the increase to 1.9% • Future Work – Understanding jitter caused by reading many small files or by writing to a file – Improving the proposed mechanism in order to monitor accesses to files on remote I/O nodes – Analyzing the experimental results in more detail

Recommend


More recommend