-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Install Request: OpenMPI 4.0.2 or later (stale shared memory segments bug) #337
Comments
Building on:
|
Installed, need to test if working before doing the modulefile. |
mpi_pi 2-node job worked on Kathleen. |
Also working 2-node on Grace, single node on Myriad and single node on the build node on Legion. (Don't think I'm currently in anything that can submit multi-node jobs on Legion remains). Need to check multi-node on the Economics part of Myriad. |
Not working across 2 nodes on Myriad-Economics.
|
We don't have openucx installed, which is the recommended replacement for openib and probably something we should get installed.
The mpirun wrapper is doing |
I've been adding some of the verbose options.
If we just specify ofi then ofi is picking verbs as the provider (and I left in the
It is possible to set open-mpi/ompi#6570 (comment) says that "The default settings are indeed based on an assumption that ofi will not be the transport of choice for Mellanox IB cards as Mellanox directs their users to install UCX and go that route." |
So, back to installing UCX? |
Ok, OpenUCX is BSD-licensed and looks straightforward to build: https://openucx.readthedocs.io/en/master/ I will install it centrally and test with a test module first.
(On the OmniPath systems it ought to work but be slower than psm2 - don't want to use it there but could install it).
|
A default config with these modules
gets this:
|
OpenMPI 4.0.3 on Myriad now built with ucx.
Trying
|
That ran, but I think it used tcp entirely.
The line that is too long to read is |
Related to IN:04155266 on Legion but is a general issue, as our most recent OpenMPIs are 3.1.4 and 3.1.5 (beta module). OpenMPI 3 > 3.1.1 has a bug where
vader_segment.x
shared memory files are left behind (only/mostly on an aborted run?). If they exist, then a new run on those nodes will fail with this:Note that
/dev/shm
is not full in this case.OpenMPI 4.0.2 and later have fixed a bunch of vader issues, and are using PMIx 3 rather than 2, which has better hooks for doing job shutdown cleanup.
Note: 4.0.x deprecates the openib BTL in favour of UCX.
https://www.open-mpi.org/software/ompi/major-changes.php
https://www.open-mpi.org/faq/?category=openfabrics#run-ucx
https://www.open-mpi.org/faq/?category=building#build-p2p
It also suggests to build
--without-verbs
when using UCX.See open-mpi/ompi#6322 and open-mpi/ompi#7220 for bug.
The text was updated successfully, but these errors were encountered: