You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This exercise builds on #54. It is part of a series that looks at execution time of different ways to calculate π using the same Monte Carlo approach.
This exercise uses the Message Passing Interface (MPI) to accomplish this approximation of π. The code is already written, and you can find it in calc_pi_mpi.py in the pi_calculation repository, on the class branch. Your job is to install MPI, and measure how much time it takes to complete in comparison to #54.
MPI
MPI allows parallelisation of computation. An MPI program consists of multiple processes, existing within a group called a communicator. The default communicator contains all available processes and is called MPI_COMM_WORLD.
Each process has its own rank and can execute different code. A typical way of using MPI is to divide the computation into smaller chunks, have each process deal with a chunk, and have one "main" process to coordinate this and gather all the results. The processes can communicate with each other in pre-determined ways as specified by the MPI protocol -- for example, sending and receiving data to a particular process, or broadcasting a message to all processes.
Preparation
We are going to run the original (non-numpy) version in parallel, and compare it to the non-parallel version.
We will be using mpi4py, a Python library that gives us access to MPI functionality.
The MPI version of the code is available at calc_pi_mpi.py. Look at the file and try to identify what it is doing -- it's fine if you don't understand all the details! Can you see how the concepts in the brief description of MPI above are reflected in the code?
Execution
Run the MPI version as:
mpiexec -n 4 python calc_pi_mpi.py
The -n argument controls how many processes you start.
Increase the number of points and proceses, and compare the time it takes against the normal version. Note that to pass arguments to the python file (like -np below), we have to give those after the file name.
This exercise builds on #54. It is part of a series that looks at execution time of different ways to calculate π using the same Monte Carlo approach.
This exercise uses the Message Passing Interface (MPI) to accomplish this approximation of π. The code is already written, and you can find it in
calc_pi_mpi.py
in the pi_calculation repository, on theclass
branch. Your job is to install MPI, and measure how much time it takes to complete in comparison to #54.MPI
MPI allows parallelisation of computation. An MPI program consists of multiple processes, existing within a group called a communicator. The default communicator contains all available processes and is called
MPI_COMM_WORLD
.Each process has its own rank and can execute different code. A typical way of using MPI is to divide the computation into smaller chunks, have each process deal with a chunk, and have one "main" process to coordinate this and gather all the results. The processes can communicate with each other in pre-determined ways as specified by the MPI protocol -- for example, sending and receiving data to a particular process, or broadcasting a message to all processes.
Preparation
We are going to run the original (non-numpy) version in parallel, and compare it to the non-parallel version.
We will be using
mpi4py
, a Python library that gives us access to MPI functionality.Install
mpi4py
using conda:or pip:
On windows you will also need to install MS MPI.
The MPI version of the code is available at
calc_pi_mpi.py
. Look at the file and try to identify what it is doing -- it's fine if you don't understand all the details! Can you see how the concepts in the brief description of MPI above are reflected in the code?Execution
-n
argument controls how many processes you start.-np
below), we have to give those after the file name.timeit
(1 and 1 in this example)numpy
-based implementation?The text was updated successfully, but these errors were encountered: