MPI

From HPC Wiki
Revision as of 12:44, 9 April 2018 by 134.130.1.101 (talk)
Jump to navigation Jump to search

MPI is an implementation of Distributed Memory parallelization. Information on how to run an existing MPI program can be found in the How_to_Use_MPI Section.

General

In MPI the basic operations are the send

int MPI_Send (void* buf, int count, MPI_Datatype datatype, int dest, int tag, MPI_Comm comm)

and receive

int MPI_Recv (void* buf, int count, MPI_Datatype datatype, int source, int tag, MPI_Comm comm, MPI_Status* status)

Next to those, a lot of other different functions exist for non-blocking or all-to-all communication. Programs written with these functions have to be compiled with a specific compiler (options) and executed with a special startup program like detailed here.

Please check the more detailed tutorials in the References.

References

Introduction to MPI from PPCES (@RWTH Aachen) Part 1

Introduction to MPI from PPCES (@RWTH Aachen) Part 2

Introduction to MPI from PPCES (@RWTH Aachen) Part 3