Frequently Asked Questions
|
Home
|
Support
|
FAQ
|
all
just the FAQ
About
Presentations
Open MPI Team
FAQ
Rollup/ALL
General information
General information
Supported systems
Contributing
Developer information
Sysadmin information
Fault Tolerance
Building
Building Open MPI
Removed MPI constructs
Compiling MPI apps
Running Jobs
Running MPI jobs
Troubleshooting
Parallel debugging
rsh/ssh
BProc
Torque / PBS Pro
Slurm
SGE
Large clusters
Tuning
General tuning
Shared memory (Vader)
TCP
IB, RoCE, iWARP
Omni-Path
Performance tools
OMPIO
UDAPL
Myrinet
Platform
OS X
AIX (unsupported)
Contrib
VampirTrace
Languages
Java
CUDA-aware
Building CUDA-aware
Running CUDA-aware
Videos
Performance
Open MPI Software
Download
Documentation
Source Code Access
Bug Tracking
Regression Testing
Version Information
Sub-Projects
Hardware Locality
Network Locality
MPI Testing Tool
Open Tool for Parameter Optimization
Community
Mailing Lists
Getting Help/Support
Contribute
Contact
License
This FAQ is for Open MPI v4.x and earlier.
If you are looking for documentation for Open MPI v5.x and later, please visit
docs.open-mpi.org
.
FAQ categories:
Rollup of ALL FAQ categories and questions
General information about the Open MPI Project
General information about the Open MPI Project
What kinds of systems / networks / run-time environments does Open MPI support?
Contributing to the Open MPI project
Developer-level technical information on the internals of Open MPI
System administrator-level technical information about Open MPI
Fault tolerance for parallel MPI jobs
Building
Building Open MPI
Removed MPI constructs
Compiling MPI applications
Running Jobs with Open MPI
Running MPI jobs
Troubleshooting building and running MPI jobs
Debugging applications in parallel
Running jobs under rsh/ssh
Running jobs under BProc
Running jobs under Torque / PBS Pro
Running jobs under Slurm
Running jobs under SGE
Running on large clusters
Tuning
General run-time tuning
Tuning the run-time characteristics of MPI shared memory communications
Tuning the run-time characteristics of MPI TCP communications
Tuning the run-time characteristics of MPI InfiniBand, RoCE, and iWARP communications
Tuning the run-time characteristics of MPI Omni-Path communications
Performance analysis tools
Tuning the OMPIO parallel I/O component
Tuning the run-time characteristics of MPI UDAPL communications
Tuning the run-time characteristics of MPI Myrinet communications
Platform Specific Questions
OS X
AIX (unsupported)
Contributed Software
VampirTrace Integration
Languages
Java
CUDA-aware Support
Building CUDA-aware Open MPI
Running CUDA-aware Open MPI
Open MPI is an Associated Project
of the
Software in the Public Interest
non-profit organization
Contact the Open MPI webmaster
Page last modified: 20-May-2019
©2004-2025 The Open MPI Project