No Travel? No Problem.

Remote Participation
Performance Evaluation of Python Parallel Programming Models: Charm4Py and Mpi4py
Author/Presenter
Event Type
Workshop
Tags
Architectures
Big Data
Cloud and Distributed Computing
Extreme Scale Comptuing
Parallel Programming Languages and Models
Parallel Programming Systems
Quantum Computing
Scientific Computing
System Software and Runtime Systems
Registration Categories
W
TimeMonday, 15 November 20212pm - 2:30pm CST
Location242
DescriptionPython is rapidly becoming the lingua franca of machine learning and scientific computing. With the broad use of frameworks such as Numpy, SciPy, and TensorFlow, scientific computing and machine learning are seeing a productivity boost on systems without a requisite loss in performance. While high-performance libraries often provide adequate performance within a node, distributed computing is required to scale Python across nodes and make it truly competitive in large-scale high-performance computing. Many frameworks, such as Charm4Py, DaCe, Dask, Legate Numpy, mpi4py, and Ray, scale Python across nodes. However, little is known about these frameworks' relative strengths and weaknesses, leaving practitioners and scientists without enough information about which frameworks are suitable for their requirements. In this paper, we seek to narrow this knowledge gap by studying the relative performance of two such frameworks: Charm4Py and mpi4py.

We perform a comparative performance analysis of Charm4Py and mpi4py using CPU and GPU-based microbenchmarks, including TaskBench and other representative mini-apps for scientific computing.
Back To Top Button