6. Programming HPC systems with OpenMP and MPI
Part1: The Message Passing Interface (MPI) is the de-facto standard for programming scalable compute architectures with distributed memory architectures like, for example, supercomputers or clusters or workstations.In MPI instances of the same program run Independently on each compute node, but thesis instances exchange data via explicit messages.The course gives an introduction to the message-passing paradigm of parallel programming and illustrates the most relevant features of MPI-including collective communication operations.
Part2: OpenMP is the most wide-spread standard for programming shared memory parallel computers, ie the majority of today's multi-core processor based desktop and server systems.The approach taken by OpenMP is to augment (mostly) ordinary C or Fortran programs with compiler directives, so-called pragmas.These directives instruct an OpenMP-aware compiler where to safely generate parallel executable code from otherwiseArrival arrival arrival sequential programs.
- Duration: 4 hours.
- Date: 27/01/2017.
- Time: 13:00-17:00.
- Location: Science Park 904, Room: SP-A1.04.
- Target group: Researchers wishing to build or modify applications and who are interested in using the possibilities offered by modern hardware.
- Prerequisites: Introduction Unix;Lisa Using or Using Huygens;Knowledge of the C or Fortran languages.
- Course leader: Clemens Grelck (Informatics Institute).