
Security News
Open Source CAI Framework Handles Pen Testing Tasks up to 3,600× Faster Than Humans
CAI is a new open source AI framework that automates penetration testing tasks like scanning and exploitation up to 3,600× faster than humans.
This package provides Python bindings for the Message Passing Interface (MPI_) standard. It is implemented on top of the MPI specification and exposes an API which grounds on the standard MPI-2 C++ bindings.
.. _MPI: https://www.mpi-forum.org
This package supports:
Convenient communication of any picklable Python object
Fast communication of Python object exposing the Python buffer interface (NumPy arrays, builtin bytes/string/array objects)
Process groups and communication domains
Parallel input/output:
Dynamic process management
One-sided operations
The mpi4py project builds and publishes binary wheels able to run in a variety of:
.. _MPICH: https://mpich.org .. _Open MPI: https://open-mpi.org .. _MVAPICH: https://mvapich.cse.ohio-state.edu .. _HPE Cray MPICH: https://cpe.ext.hpe.com/docs/latest/mpt/mpich/ .. _NVIDIA HPC-X: https://developer.nvidia.com/networking/hpc-x .. _Intel MPI: https://software.intel.com/intel-mpi-library .. _Microsoft MPI: https://learn.microsoft.com/message-passing-interface/microsoft-mpi
These mpi4py wheels are distributed via the Python Package Index
(PyPI <https://pypi.org/project/mpi4py/>
) and can be installed
with Python package managers like pip
:
.. code:: sh
python -m pip install mpi4py
.. _pip: https://pip.pypa.io
The mpi4py wheels can be installed in standard Python virtual environments. The MPI runtime can be provided by other wheels installed in the same virtual environment.
.. tip::
Intel publishes production-grade Intel MPI wheels <impi-rt-wheels_>
_ for Linux (x86_64) and Windows (AMD64).
mpi4py and MPI wheels can be installed side by side to get a
ready-to-use Python+MPI environment:
.. code:: sh
python -m pip install mpi4py impi-rt
.. _impi-rt-wheels: https://pypi.org/project/impi-rt/#files
.. tip::
The mpi4py project publishes MPICH wheels <mpich-wheels_>
_ and
Open MPI wheels <openmpi-wheels_>
_ for Linux
(x86_64/aarch64) and macOS (arm64/x86_64).
mpi4py and MPI wheels can be installed side by side to get a
ready-to-use Python+MPI environment:
.. code:: sh
python -m pip install mpi4py mpich # for MPICH
python -m pip install mpi4py openmpi # for Open MPI
.. _mpich-wheels: https://pypi.org/project/mpich/#files .. _openmpi-wheels: https://pypi.org/project/openmpi/#files
.. warning::
The MPI wheels are distributed with special focus on ease of
use, convenience, compatibility, and interoperability. The Linux
wheels are built in somewhat constrained environments with
relatively dated Linux distributions (`manylinux`_ container
images). Therefore, they may lack support for features like GPU
awareness (CUDA/ROCm) and C++/Fortran bindings. In production
scenarios, it is recommended to use external (either
custom-built or system-provided) MPI installations.
.. _manylinux: https://github.com/pypa/manylinux
The mpi4py wheels can also be installed (with pip
) in conda
environments and they should work out of the box, without any special
tweak to environment variables, for any of the MPI packages provided
by conda-forge
_.
Externally-provided MPI implementations may come from a system package manager, sysadmin-maintained builds accessible via module files, or customized user builds. Such usage is supported and encouraged. However, there are a few platform-specific considerations to take into account.
Linux ^^^^^
The Linux (x86_64/aarch64) wheels require one of
MPICH
_ or any other ABI-compatible derivative,
like MVAPICH
, Intel MPI
, HPE Cray MPICH
_
Open MPI
_ or any other ABI-compatible derivative,
like NVIDIA HPC-X
_
Users may need to set the LD_LIBRARY_PATH
environment variable
such that the dynamic linker is able to find at runtime the MPI shared
library file (libmpi.so.*
).
Fedora/RHEL
On Fedora/RHEL systems, both MPICH and Open MPI are available for
installation. There is no default or preferred MPI implementation.
Instead, users must select their favorite MPI implementation by
loading the proper MPI module.
.. code:: sh
module load mpi/mpich-$(arch) # for MPICH
module load mpi/openmpi-$(arch) # for Open MPI
After loading the requested MPI module, the ``LD_LIBRARY_PATH``
environment variable should be properly setup.
Debian/Ubuntu
On Debian/Ubuntu systems, Open MPI is the default MPI implementation and most of the MPI-based applications and libraries provided by the distribution depend on Open MPI. Nonetheless, MPICH is also available to users for installation.
In Ubuntu 22.04 and older, due to legacy reasons, the MPICH ABI is
slightly broken: the MPI shared library file is named
libmpich.so.12
instead of libmpi.so.12
as required by the
MPICH ABI Compatibility Initiative <https://www.mpich.org/abi/>
_.
Users without sudo
access can workaround this issue creating a
symbolic link anywhere in their home directory and appending to
LD_LIBRARY_PATH
.
.. code:: sh
mkdir -p ~/.local/lib
libdir=/usr/lib/$(arch)-linux-gnu
ln -s $libdir/libmpich.so.12 /.local/lib/libmpi.so.12
export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:/.local/lib
A system-wide fix for all users requires sudo
access:
.. code:: sh
libdir=/usr/lib/$(arch)-linux-gnu sudo ln -sr $libdir/libmpi{ch,}.so.12
HPE Cray OS
On HPE Cray systems, users must load the ``cray-mpich-abi`` module.
For further details, refer to `man intro_mpi <cray-mpt-mpichabi_>`_.
.. _cray-mpt-mpichabi: https://cpe.ext.hpe.com/docs/latest/mpt/mpich/intro_mpi.html#using-mpich-abi-compatibility
macOS
^^^^^
The macOS (arm64/x86_64) wheels require
* `MPICH`_ or `Open MPI`_ installed (either manually or via a package
manager) in the standard system prefix ``/usr/local``
* `MPICH`_ or `Open MPI`_ installed via `Homebrew`_ in the default
prefix ``/opt/homebrew``
* `MPICH`_ or `Open MPI`_ installed via `MacPorts`_ in the default
prefix ``/opt/local``
.. _Homebrew: https://brew.sh/
.. _MacPorts: https://www.macports.org/
Windows
^^^^^^^
The Windows (AMD64) wheels require one of
* `Intel MPI`_
* `Microsoft MPI`_
User may need to set the ``I_MPI_ROOT`` or ``MSMPI_BIN`` environment
variables such that the MPI dynamic link library (DLL) (``impi.dll``
or ``msmpi.dll``) can be found at runtime.
Intel MPI is under active development and supports recent versions of
the MPI standard. Intel MPI can be installed with ``pip`` (see the
`impi-rt`_ package on PyPI), being therefore straightforward to get it
up and running within a Python environment. Intel MPI can also be
installed system-wide as part of the Intel oneAPI HPC Toolkit for
Windows or via standalone online/offline installers.
.. _impi-rt: https://pypi.org/project/impi-rt/
**Conda** packages
------------------
The `conda-forge`_ community provides ready-to-use binary packages
from an ever growing collection of software libraries built around the
multi-platform *conda* package manager. Four MPI implementations are
available on conda-forge: Open MPI (Linux and macOS), MPICH (Linux and
macOS), Intel MPI (Linux and Windows), and Microsoft MPI (Windows).
You can install mpi4py and your preferred MPI implementation using the
`conda`_ package manager:
* to use MPICH do:
.. code:: sh
conda install -c conda-forge mpi4py mpich
* to use Open MPI do:
.. code:: sh
conda install -c conda-forge mpi4py openmpi
* to use Intel MPI do:
.. code:: sh
conda install -c conda-forge mpi4py impi_rt
* to use Microsoft MPI do:
.. code:: sh
conda install -c conda-forge mpi4py msmpi
MPICH and many of its derivatives are ABI-compatible. You can provide
the package specification ``mpich=X.Y.*=external_*`` (where ``X`` and
``Y`` are the major and minor version numbers) to request the conda
package manager to use system-provided MPICH (or derivative)
libraries. Similarly, you can provide the package specification
``openmpi=X.Y.*=external_*`` to use system-provided Open MPI
libraries.
The ``openmpi`` package on conda-forge has built-in CUDA support, but
it is disabled by default. To enable it, follow the instruction
outlined during ``conda install``. Additionally, UCX support is also
available once the ``ucx`` package is installed.
.. warning::
The MPI conda-forge packages are built with special focus on
compatibility. The MPICH and Open MPI packages are built in a
constrained environment with relatively dated OS images. Therefore,
they may lack support for high-performance features like
cross-memory attach (XPMEM/CMA). In production scenarios, it is
recommended to use external (either custom-built or system-provided)
MPI installations. See the relevant conda-forge documentation about
`using external MPI libraries <cf-mpi-docs_>`_ .
.. _conda: https://docs.conda.io
.. _conda-forge: https://conda-forge.org/
.. _cf-mpi-docs: https://conda-forge.org/docs/user/tipsandtricks/#using-external-message-passing-interface-mpi-libraries
System packages
---------------
mpi4py is readily available through system package managers of most
Linux distributions and the most popular community package managers
for macOS.
.. _sys-pkg-linux:
Linux
^^^^^
On **Fedora Linux** systems (as well as **RHEL** and their derivatives
using the EPEL software repository), you can install binary packages
with the system package manager:
* using ``dnf`` and the ``mpich`` package:
.. code:: sh
sudo dnf install python3-mpi4py-mpich
* using ``dnf`` and the ``openmpi`` package:
.. code:: sh
sudo dnf install python3-mpi4py-openmpi
Please remember to load the correct MPI module for your chosen MPI
implementation:
* for the ``mpich`` package do:
.. code:: sh
module load mpi/mpich-$(arch)
python -c "from mpi4py import MPI"
* for the ``openmpi`` package do:
.. code:: sh
module load mpi/openmpi-$(arch)
python -c "from mpi4py import MPI"
On **Ubuntu Linux** and **Debian Linux** systems, binary packages are
available for installation using the system package manager:
.. code:: sh
sudo apt install python3-mpi4py
On **Arch Linux** systems, binary packages are available for
installation using the system package manager:
.. code:: sh
sudo pacman -S python-mpi4py
.. _sys-pkg-macos:
macOS
^^^^^
macOS users can install mpi4py using the `Homebrew`_ package
manager:
.. code:: sh
brew install mpi4py
Note that the Homebrew mpi4py package uses Open MPI. Alternatively,
install the ``mpich`` package and next install mpi4py from sources
using ``pip``.
Alternatively, mpi4py can be installed from `MacPorts`_:
.. code:: sh
sudo port install py-mpi4py
Building from sources
---------------------
Installing mpi4py from pre-built binary wheels, conda packages, or
system packages is not always desired or appropriate. For example, the
mpi4py wheels published on PyPI may not be interoperable with
non-mainstream, vendor-specific MPI implementations; or a system
mpi4py package may be built with a alternative, non-default MPI
implementation. In such scenarios, mpi4py can still be installed from
its source distribution (sdist) using ``pip``:
.. code:: sh
python -m pip install --no-binary=mpi4py mpi4py
You can also install the in-development version with:
.. code:: sh
python -m pip install git+https://github.com/mpi4py/mpi4py
or:
.. code:: sh
python -m pip install https://github.com/mpi4py/mpi4py/tarball/master
.. note::
Installing mpi4py from its source distribution (available on PyPI)
or Git source code repository (available on GitHub) requires a C
compiler and a working MPI implementation with development headers
and libraries.
.. warning::
``pip`` keeps previously built wheel files in its cache for future
reuse. If you want to reinstall the ``mpi4py`` package from its source
distribution using a different or updated MPI implementation, you have
to either first remove the cached wheel file:
.. code:: sh
python -m pip cache remove mpi4py
python -m pip install --no-binary=mpi4py mpi4py
or ask ``pip`` to disable the cache:
.. code:: sh
python -m pip install --no-cache-dir --no-binary=mpi4py mpi4py
Citation
========
If MPI for Python been significant to a project that leads to an
academic publication, please acknowledge that fact by citing the
project.
* M. Rogowski, S. Aseeri, D. Keyes, and L. Dalcin,
*mpi4py.futures: MPI-Based Asynchronous Task Execution for Python*,
IEEE Transactions on Parallel and Distributed Systems, 34(2):611-622, 2023.
https://doi.org/10.1109/TPDS.2022.3225481
* L. Dalcin and Y.-L. L. Fang,
*mpi4py: Status Update After 12 Years of Development*,
Computing in Science & Engineering, 23(4):47-54, 2021.
https://doi.org/10.1109/MCSE.2021.3083216
* L. Dalcin, P. Kler, R. Paz, and A. Cosimo,
*Parallel Distributed Computing using Python*,
Advances in Water Resources, 34(9):1124-1139, 2011.
https://doi.org/10.1016/j.advwatres.2011.04.013
* L. Dalcin, R. Paz, M. Storti, and J. D'Elia,
*MPI for Python: performance improvements and MPI-2 extensions*,
Journal of Parallel and Distributed Computing, 68(5):655-662, 2008.
https://doi.org/10.1016/j.jpdc.2007.09.005
* L. Dalcin, R. Paz, and M. Storti,
*MPI for Python*,
Journal of Parallel and Distributed Computing, 65(9):1108-1115, 2005.
https://doi.org/10.1016/j.jpdc.2005.03.010
FAQs
Python bindings for MPI
We found that mpi4py demonstrated a healthy version release cadence and project activity because the last version was released less than a year ago. It has 2 open source maintainers collaborating on the project.
Did you know?
Socket for GitHub automatically highlights issues in each pull request and monitors the health of all your open source dependencies. Discover the contents of your packages and block harmful activity before you install or update your dependencies.
Security News
CAI is a new open source AI framework that automates penetration testing tasks like scanning and exploitation up to 3,600× faster than humans.
Security News
Deno 2.4 brings back bundling, improves dependency updates and telemetry, and makes the runtime more practical for real-world JavaScript projects.
Security News
CVEForecast.org uses machine learning to project a record-breaking surge in vulnerability disclosures in 2025.