preparing the lhc computing grid for mpi applications
play

Preparing the LHC Computing Grid for MPI Applications Richard de - PowerPoint PPT Presentation

Enabling Grids for E-sciencE Preparing the LHC Computing Grid for MPI Applications Richard de Jong & Matthijs Koot www.eu-egee.org INFSO-RI-508833 Outline Enabling Grids for E-sciencE Context CERN Grid LCG YAIM


  1. Enabling Grids for E-sciencE Preparing the LHC Computing Grid for MPI Applications Richard de Jong & Matthijs Koot www.eu-egee.org INFSO-RI-508833

  2. Outline Enabling Grids for E-sciencE • Context – CERN – Grid – LCG – YAIM • Problems (and solutions) – YAIM • Context (ct’d) – Parallel programming • Problems (and solutions) (ct’d) – Single-site MPI – Cross-site MPI • Conclusion INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 2

  3. Context: CERN & The LHC Project Enabling Grids for E-sciencE • A new particle accelerator is being built at CERN • 10 PB/year • Save (backup) & Analyze all data • By physicists around the world • O(8K) nodes at CERN • But that’s not enough… • Solution: Tier 0, 1, 2 INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 3

  4. Context: Data flow Enabling Grids for E-sciencE INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 4

  5. Context: Grid Enabling Grids for E-sciencE • A grid is a system that (Foster): – coordinates resources that are not subject to centralized control – using standard, open, general-purpose protocols and interfaces – to deliver nontrivial qualities of service. • Keyword: Middleware INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 5

  6. Context: LHC Computing Grid Enabling Grids for E-sciencE • LCG middleware: – LCG – gLite • Both based on existing technology: – Globus Toolkit – VDT – GridFTP – Condor • Components in the LHC architecture: – User Interface (UI) – To submit a job, retrieve output – Resource Broker (RB) – To find a suitable CE – Compute Element (CE) – To schedule the job – LRMS – Worker Node (WN) – To execute the job – Many more… INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 6

  7. Context: YAIM Enabling Grids for E-sciencE • YAIM = Yet Another Installation Mechanism • A tool for Grid middleware deployment – Scope – Structure – Evaluation INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 7

  8. YAIM: Scope Enabling Grids for E-sciencE • EDG - WP4: – "provide means to install OS and applications over the network according to policies, to bring a machine in a desired state" • Configuration by LCFG(ng), but inflexible and error- prone • So: QUATTOR – Scope on OS and applications – But: not (so much) on Grid middleware INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 8

  9. YAIM: Scope Enabling Grids for E-sciencE • CERN IT-GD: "should not be so hard"  YAIM – Configuration template – For small and/or simple sites • But for larger sites? INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 9

  10. YAIM: Structure Enabling Grids for E-sciencE • Configuration files – site-info.def - Site wide configuration – node-info.def - What to configure for which role • Scripts – install_node - Install packages – configure_node - Configure services • Functions – One function covers an atomic piece of configuration – One file per function, one function per file • Utilities – Helper routines INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 10

  11. YAIM: Evaluation Enabling Grids for E-sciencE • Pro: – Straightforward – Modular – Simple – Easy to use • Con: – Not atomic – Unstructured output – Reconfiguring – Unconfiguring – Users learn to ignore the errors • Conclusion: fine for small and/or simple sites – Meanwhile: QWG to integrate YAIM functionality for large sites INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 11

  12. Context: Parallel Programming Enabling Grids for E-sciencE - Traditional: - Use a single processor to solve multiple problems = SISD - Parallel programming: - Using multiple processors to solve a single problem - Divide the work: - Functional decomposition - Each processor a different role = MISD - Domain decomposition - Each processor different data = SIMD - Both - Each processor it’s own program and data = MIMD INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 12

  13. Context: Parallel Programming Enabling Grids for E-sciencE - SISD - SIMD INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 13

  14. Context: Parallel Programming Enabling Grids for E-sciencE - MISD - MIMD INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 14

  15. Context: MPI Enabling Grids for E-sciencE • MPI = Message Passing Interface • The MPI library is a widely known and used standard for parallel programming • Multiple implementations available – MPICH – LAM/MPI – OpenMPI – MS MPI – etc • Traffic profile depends on parallel algorithm INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 15

  16. Enabling Grids for E-sciencE • Next in line: mrkoot@os3.nl INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 16

  17. Outline Enabling Grids for E-sciencE • Context – CERN – Grid – LCG – YAIM • Problems (and solutions) – YAIM • Context (ct’d) – Parallel programming • Problems (and solutions) (ct’d) – Single-site MPI – Cross-site MPI • Conclusion INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 17

  18. Problems: Single-Site MPI Enabling Grids for E-sciencE • MPI within a single grid-connected cluster • Integration between MPI and LRMS INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 18

  19. Problems: Cross-Site MPI Enabling Grids for E-sciencE INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 19

  20. Problems: Cross-Site MPI Enabling Grids for E-sciencE INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 20

  21. Problems: Cross-Site MPI Enabling Grids for E-sciencE • Job decomposition – Functional ~, domain ~ – Is it possible to automate? INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 21

  22. Problems: Cross-Site MPI Enabling Grids for E-sciencE • Co-allocation – Resource scheduling is tough! – Learn from Koala in DAS-2? – LRMS interoperability? INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 22

  23. Problems: Cross-Site MPI Enabling Grids for E-sciencE INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 23

  24. Problems: Cross-Site MPI Enabling Grids for E-sciencE • MPI interoperability – MPI is portable , not interoperable – Sites should provide common MPI INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 24

  25. Problems: Cross-Site MPI Enabling Grids for E-sciencE • Networking – Latency(-tolerance?) – Protocols (public IP addresses?) – Security (dynamic firewall with VO-EAP?) INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 25

  26. Problems: Cross-Site MPI Enabling Grids for E-sciencE • Networking (2) – Collective ops? INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 26

  27. Problems: Cross-Site MPI Enabling Grids for E-sciencE • Integration is needed between MPI and grid middleware (AuthN/Z, monitoring) • MPICH-G2 demonstrates integration between MPICH and Globus • Might (still?) be usable on LCG INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 27

  28. Problems: Cross-Site MPI Enabling Grids for E-sciencE INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 28

  29. Problems: Cross-Site MPI Enabling Grids for E-sciencE • It has been demonstrated – Legion (2001) – TeraGrid (2003) – K*Grid in Korea (2005) • …but not yet over a high-latency, heterogeneous, Internet-based grid INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 29

  30. Conclusion(s) Enabling Grids for E-sciencE • Single-Site MPI is probably OK • Cross-Site MPI is tough, but possible – Good software engineering required • YAIM needs improvement INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 30

  31. Questions Enabling Grids for E-sciencE • Questions? INFSO-RI-508833 Preparing the LHC Computing Grid for MPI - Richard de Jong & Matthijs Koot 31

Recommend


More recommend