From ryandeng at mit.edu Sat Jul 5 01:48:00 2025 From: ryandeng at mit.edu (Ryan Deng) Date: Sat, 5 Jul 2025 06:48:00 +0000 Subject: [mpich-discuss] MPIX Streams Usage Message-ID: Hi, I had some questions about how MPIX_Streams work under the hood. I had posted a couple of times on the github issues page before about it but thought this might be a more appropriate place for it. Right now, my code involves many MPIX_Stream_isends/MPIX_Stream_irecvs as processes asynchronously communicate with one another, and I have a separate thread in the background constantly advancing progress on streams. It's my understanding that all operations on a stream must occur serially. I assume this means that things operations like sending and receiving, calling MPIX_Stream_progress/MPI_Wait cannot occur at the same time without locks? Are there any other operations on streams that cannot occur concurrently? In addition, I had also triggered the assertion here. [https://urldefense.us/v3/__https://opengraph.githubassets.com/a68c91e94d3a9dcfc8cbfd9fbffd5b2cc6514180d7dd8cffd36c9ccf563b3fa9/pmodels/mpich__;!!G_uCfscf7eWS!cK1AJPZW8zQ8gnmJ2XatdQNoOydpD7-3SuzANwgQMZT68UGmFP9sXZO2GMboYtXFR9JgbT217u1Y9sI$ ] mpich/src/mpid/ch4/shm/posix/posix_progress.h at 837914f6e810bbbd5e7ad391bbad9f546ffe3dab ? pmodels/mpich Official MPICH Repository. Contribute to pmodels/mpich development by creating an account on GitHub. github.com Does this mean that I cannot send multiple messages at the same time if it's between two different streams? On the github issue here, Hui Zhou mentioned that it is due to messaging protocol consistency. I'm not quite sure what he means by it and what that means in terms of what I can and cannot do in my application involving MPIX_Streams. Thanks, Ryan -------------- next part -------------- An HTML attachment was scrubbed... URL: From kurt.e.mccall at nasa.gov Mon Jul 21 16:19:28 2025 From: kurt.e.mccall at nasa.gov (Mccall, Kurt E. (MSFC-EV41)) Date: Mon, 21 Jul 2025 21:19:28 +0000 Subject: [mpich-discuss] Configuration failure for 4.3.1. Message-ID: I'm having some trouble configuring 4.3.1. The following command finishes successfully: $ ../mpich-4.3.1/configure --prefix=/opt/mpich --with-device=ch4:ofi --with-slurm -enable-debuginfo --enable-g=debug 2>&1 | tee c.txt But this one fails with the error below: $ ../mpich-4.3.1/configure --prefix=/opt/mpich --with-device=ch4:ucx --with-ucx=embedded --with-slurm -enable-debuginfo --enable-g=debug 2>&1 | tee c.txt config.status: creating src/tools/perf/rocm/Makefile mv: cannot move './confYHXukm/out' to 'src/tools/perf/rocm/Makefile': Device or resource busy config.status: error: could not create src/tools/perf/rocm/Makefile Are my arguments incorrect? I can send c.txt in a separate message. Redhat 4.18.0, gcc 8.5.0. Thanks, Kurt -------------- next part -------------- An HTML attachment was scrubbed... URL: From zhouh at anl.gov Tue Jul 22 09:16:26 2025 From: zhouh at anl.gov (Zhou, Hui) Date: Tue, 22 Jul 2025 14:16:26 +0000 Subject: [mpich-discuss] Configuration failure for 4.3.1. In-Reply-To: References: Message-ID: Hi Kurt, The error is complaining the file `src/tools/perf/rocm/Makefile` is in use while config.status is trying to recreate the file. I am not sure what is causing it. Did you try again and does the same error occur? Hui ________________________________ From: Mccall, Kurt E. (MSFC-EV41) via discuss Sent: Monday, July 21, 2025 4:19 PM To: Zhou, Hui via discuss Cc: Mccall, Kurt E. (MSFC-EV41) Subject: [mpich-discuss] Configuration failure for 4.3.1. I?m having some trouble configuring 4.?3.?1. The following command finishes successfully: $ ..?/mpich-4.?3.?1/configure --prefix=/opt/mpich --with-device=ch4:?ofi --with-slurm -enable-debuginfo --enable-g=debug 2>&1 | tee c.?txt But this ZjQcmQRYFpfptBannerStart This Message Is From an External Sender This message came from outside your organization. ZjQcmQRYFpfptBannerEnd I?m having some trouble configuring 4.3.1. The following command finishes successfully: $ ../mpich-4.3.1/configure --prefix=/opt/mpich --with-device=ch4:ofi --with-slurm -enable-debuginfo --enable-g=debug 2>&1 | tee c.txt But this one fails with the error below: $ ../mpich-4.3.1/configure --prefix=/opt/mpich --with-device=ch4:ucx --with-ucx=embedded --with-slurm -enable-debuginfo --enable-g=debug 2>&1 | tee c.txt config.status: creating src/tools/perf/rocm/Makefile mv: cannot move './confYHXukm/out' to 'src/tools/perf/rocm/Makefile': Device or resource busy config.status: error: could not create src/tools/perf/rocm/Makefile Are my arguments incorrect? I can send c.txt in a separate message. Redhat 4.18.0, gcc 8.5.0. Thanks, Kurt -------------- next part -------------- An HTML attachment was scrubbed... URL: From hppritcha at gmail.com Mon Jul 28 09:41:04 2025 From: hppritcha at gmail.com (Howard Pritchard) Date: Mon, 28 Jul 2025 08:41:04 -0600 Subject: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem Message-ID: Hi Folks, We are seeing a strange performance issue on our HPE SS11 system when testing osu_latency inter-node with MPICH. First the info: system using libfabric 1.22.0 slurm - 24.11.5 Here's my mpichversion output: MPICH Version: 5.0.0a1 MPICH Release date: unreleased development copy MPICH ABI: 0:0:0 MPICH Device: ch4:ofi MPICH configure: --prefix=/XXXX/mpich_again/install --enable-g=no --enable-error-checking=no --with-device=ch4:ofi --enable-threads=multiple --with-ch4-shmmods=posix,xpmem --enable-thread-cs=per-vci --with-libfabric=/opt/cray/libfabric/1.22.0 --with-xpmem=/opt/cray/xpmem/default --with-pmix=/opt/pmix/gcc4x/5.0.8 --enable-fast=O3 MPICH CC: gcc -O3 MPICH CXX: g++ -O3 MPICH F77: gfortran -O3 MPICH FC: gfortran -O3 MPICH features: threadcomm And here's the OSU latency results: slurmstepd: error: mpi/pmix_v4: pmixp_coll_belong_chk: nid001439 [1]: pmixp_coll.c:280: No process controlled by this slurmstepd is involved in this collective. slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001439 [1]: pmixp_server.c:923: Unable to pmixp_state_coll_get() slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_check: nid001438 [0]: pmixp_coll_ring.c:614: 0x15005c005dc0: unexpected contrib from nid001439:1, expected is 0 slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001438 [0]: pmixp_server.c:937: 0x15005c005dc0: unexpected contrib from nid001439:1, coll->seq=0, seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001438 [0]: pmixp_coll_ring.c:738: 0x1500580532f0: collective timeout seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001438 [0]: pmixp_coll.c:286: Dumping collective state slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:756: 0x1500580532f0: COLL_FENCE_RING state seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:758: my peerid: 0:nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:765: neighbor id: next 1:nid001439, prev 1:nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x150058053368, #0, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x1500580533a0, #1, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x1500580533d8, #2, in-use=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:788: neighbor contribs [2]: slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:821: done contrib: - slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:823: wait contrib: nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:829: buf (offset/size): 36/16384 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001439 [1]: pmixp_coll_ring.c:738: 0x151d0c053400: collective timeout seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001439 [1]: pmixp_coll.c:286: Dumping collective state slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:756: 0x151d0c053400: COLL_FENCE_RING state seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:758: my peerid: 1:nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:765: neighbor id: next 0:nid001438, prev 0:nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c053478, #0, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c0534b0, #1, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c0534e8, #2, in-use=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:788: neighbor contribs [2]: slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:821: done contrib: - slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:823: wait contrib: nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:829: buf (offset/size): 36/16384 # OSU MPI Latency Test v5.8 # Size Latency (us) 0 1.66 1 9.29 2 9.57 4 9.69 8 9.76 16 9.77 32 9.76 64 9.77 128 10.32 256 7.54 512 7.45 1024 7.38 2048 7.37 4096 7.45 8192 9.21 16384 9.70 32768 10.63 65536 13.15 131072 16.96 262144 23.84 524288 36.16 1048576 60.36 2097152 108.43 4194304 228.31 Note the slurm behavior is - I launch the job. Go get coffee, do some duo-lingo, read some emails, then after about 10 minutes the osu latency runs. I did not get the slurm problems using an older mpich 4.3.1 but did get the same performance issue. 9 usecs doesn't seem right for an 8-byte pingpong over libfabric S11. I was expecting more like 1.6 or so. I am confident the slurm issue is unrelated to the latency issue. Thanks for any suggestions on how to address either issue however. -------------- next part -------------- An HTML attachment was scrubbed... URL: From zhouh at anl.gov Mon Jul 28 17:15:56 2025 From: zhouh at anl.gov (Zhou, Hui) Date: Mon, 28 Jul 2025 22:15:56 +0000 Subject: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem In-Reply-To: References: Message-ID: Hi Howard, I wonder whether it is due to the overhead of querying pointer attributes. Could you try disable GPU support with `MPIR_CVAR_ENABLE_GPU=0` and see if the latency improves? Hui ________________________________ From: Howard Pritchard via discuss Sent: Monday, July 28, 2025 9:41 AM To: discuss at mpich.org Cc: Howard Pritchard Subject: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem Hi Folks, We are seeing a strange performance issue on our HPE SS11 system when testing osu_latency inter-node with MPICH. First the info: system using libfabric 1.?22.?0 slurm - 24.?11.?5 Here's my mpichversion output: MPICH Version:? 5.?0.?0a1 ZjQcmQRYFpfptBannerStart This Message Is From an External Sender This message came from outside your organization. ZjQcmQRYFpfptBannerEnd Hi Folks, We are seeing a strange performance issue on our HPE SS11 system when testing osu_latency inter-node with MPICH. First the info: system using libfabric 1.22.0 slurm - 24.11.5 Here's my mpichversion output: MPICH Version: 5.0.0a1 MPICH Release date: unreleased development copy MPICH ABI: 0:0:0 MPICH Device: ch4:ofi MPICH configure: --prefix=/XXXX/mpich_again/install --enable-g=no --enable-error-checking=no --with-device=ch4:ofi --enable-threads=multiple --with-ch4-shmmods=posix,xpmem --enable-thread-cs=per-vci --with-libfabric=/opt/cray/libfabric/1.22.0 --with-xpmem=/opt/cray/xpmem/default --with-pmix=/opt/pmix/gcc4x/5.0.8 --enable-fast=O3 MPICH CC: gcc -O3 MPICH CXX: g++ -O3 MPICH F77: gfortran -O3 MPICH FC: gfortran -O3 MPICH features: threadcomm And here's the OSU latency results: slurmstepd: error: mpi/pmix_v4: pmixp_coll_belong_chk: nid001439 [1]: pmixp_coll.c:280: No process controlled by this slurmstepd is involved in this collective. slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001439 [1]: pmixp_server.c:923: Unable to pmixp_state_coll_get() slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_check: nid001438 [0]: pmixp_coll_ring.c:614: 0x15005c005dc0: unexpected contrib from nid001439:1, expected is 0 slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001438 [0]: pmixp_server.c:937: 0x15005c005dc0: unexpected contrib from nid001439:1, coll->seq=0, seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001438 [0]: pmixp_coll_ring.c:738: 0x1500580532f0: collective timeout seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001438 [0]: pmixp_coll.c:286: Dumping collective state slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:756: 0x1500580532f0: COLL_FENCE_RING state seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:758: my peerid: 0:nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:765: neighbor id: next 1:nid001439, prev 1:nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x150058053368, #0, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x1500580533a0, #1, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x1500580533d8, #2, in-use=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:788: neighbor contribs [2]: slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:821: done contrib: - slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:823: wait contrib: nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:829: buf (offset/size): 36/16384 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001439 [1]: pmixp_coll_ring.c:738: 0x151d0c053400: collective timeout seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001439 [1]: pmixp_coll.c:286: Dumping collective state slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:756: 0x151d0c053400: COLL_FENCE_RING state seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:758: my peerid: 1:nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:765: neighbor id: next 0:nid001438, prev 0:nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c053478, #0, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c0534b0, #1, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c0534e8, #2, in-use=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:788: neighbor contribs [2]: slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:821: done contrib: - slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:823: wait contrib: nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:829: buf (offset/size): 36/16384 # OSU MPI Latency Test v5.8 # Size Latency (us) 0 1.66 1 9.29 2 9.57 4 9.69 8 9.76 16 9.77 32 9.76 64 9.77 128 10.32 256 7.54 512 7.45 1024 7.38 2048 7.37 4096 7.45 8192 9.21 16384 9.70 32768 10.63 65536 13.15 131072 16.96 262144 23.84 524288 36.16 1048576 60.36 2097152 108.43 4194304 228.31 Note the slurm behavior is - I launch the job. Go get coffee, do some duo-lingo, read some emails, then after about 10 minutes the osu latency runs. I did not get the slurm problems using an older mpich 4.3.1 but did get the same performance issue. 9 usecs doesn't seem right for an 8-byte pingpong over libfabric S11. I was expecting more like 1.6 or so. I am confident the slurm issue is unrelated to the latency issue. Thanks for any suggestions on how to address either issue however. -------------- next part -------------- An HTML attachment was scrubbed... URL: From hppritcha at gmail.com Wed Jul 30 13:42:41 2025 From: hppritcha at gmail.com (Howard Pritchard) Date: Wed, 30 Jul 2025 12:42:41 -0600 Subject: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem In-Reply-To: References: Message-ID: Hi Hui That didn?t help. I am not surprised though as our cluster is an NVIDIA free zone. What did help is to switch to the mpich 4.3.x branch and latency results are nominal and the slurm problem went away too. So we will stick with that branch. Howard On Mon, Jul 28, 2025 at 4:15?PM Zhou, Hui wrote: > Hi Howard, > > I wonder whether it is due to the overhead of querying pointer > attributes. Could you try disable GPU support with `MPIR_CVAR_ENABLE_GPU=0` > and see if the latency improves? > > Hui > ------------------------------ > *From:* Howard Pritchard via discuss > *Sent:* Monday, July 28, 2025 9:41 AM > *To:* discuss at mpich.org > *Cc:* Howard Pritchard > *Subject:* [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more > - a slurm problem > > Hi Folks, We are seeing a strange performance issue on our HPE SS11 system > when testing osu_latency inter-node with MPICH. First the info: system > using libfabric 1. 22. 0 slurm - 24. 11. 5 Here's my mpichversion output: > MPICH Version: 5. 0. 0a1 > ZjQcmQRYFpfptBannerStart > This Message Is From an External Sender > This message came from outside your organization. > > ZjQcmQRYFpfptBannerEnd > Hi Folks, > > We are seeing a strange performance issue on our HPE SS11 system when > testing osu_latency inter-node with MPICH. > > First the info: > system using libfabric 1.22.0 > slurm - 24.11.5 > > Here's my mpichversion output: > > MPICH Version: 5.0.0a1 > > MPICH Release date: unreleased development copy > > MPICH ABI: 0:0:0 > > MPICH Device: ch4:ofi > > MPICH configure: --prefix=/XXXX/mpich_again/install --enable-g=no > --enable-error-checking=no --with-device=ch4:ofi --enable-threads=multiple > --with-ch4-shmmods=posix,xpmem --enable-thread-cs=per-vci > --with-libfabric=/opt/cray/libfabric/1.22.0 > --with-xpmem=/opt/cray/xpmem/default --with-pmix=/opt/pmix/gcc4x/5.0.8 > --enable-fast=O3 > > MPICH CC: gcc -O3 > > MPICH CXX: g++ -O3 > > MPICH F77: gfortran -O3 > > MPICH FC: gfortran -O3 > > MPICH features: threadcomm > > > > And here's the OSU latency results: > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_belong_chk: nid001439 [1]: > pmixp_coll.c:280: No process controlled by this slurmstepd is involved in > this collective. > > slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001439 [1]: > pmixp_server.c:923: Unable to pmixp_state_coll_get() > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_check: nid001438 [0]: > pmixp_coll_ring.c:614: 0x15005c005dc0: unexpected contrib from nid001439:1, > expected is 0 > > slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001438 [0]: > pmixp_server.c:937: 0x15005c005dc0: unexpected contrib from nid001439:1, > coll->seq=0, seq=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001438 > [0]: pmixp_coll_ring.c:738: 0x1500580532f0: collective timeout seq=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001438 [0]: > pmixp_coll.c:286: Dumping collective state > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:756: 0x1500580532f0: COLL_FENCE_RING state seq=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:758: my peerid: 0:nid001438 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:765: neighbor id: next 1:nid001439, prev 1:nid001439 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:775: Context ptr=0x150058053368, #0, in-use=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:775: Context ptr=0x1500580533a0, #1, in-use=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:775: Context ptr=0x1500580533d8, #2, in-use=1 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:788: neighbor contribs [2]: > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:821: done contrib: - > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:823: wait contrib: nid001439 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: > pmixp_coll_ring.c:829: buf (offset/size): 36/16384 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001439 > [1]: pmixp_coll_ring.c:738: 0x151d0c053400: collective timeout seq=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001439 [1]: > pmixp_coll.c:286: Dumping collective state > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:756: 0x151d0c053400: COLL_FENCE_RING state seq=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:758: my peerid: 1:nid001439 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:765: neighbor id: next 0:nid001438, prev 0:nid001438 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:775: Context ptr=0x151d0c053478, #0, in-use=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:775: Context ptr=0x151d0c0534b0, #1, in-use=0 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:775: Context ptr=0x151d0c0534e8, #2, in-use=1 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:788: neighbor contribs [2]: > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:821: done contrib: - > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:823: wait contrib: nid001438 > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS > > slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: > pmixp_coll_ring.c:829: buf (offset/size): 36/16384 > > # OSU MPI Latency Test v5.8 > > # Size Latency (us) > > 0 1.66 > > 1 9.29 > > 2 9.57 > > 4 9.69 > > 8 9.76 > > 16 9.77 > > 32 9.76 > > 64 9.77 > > 128 10.32 > > 256 7.54 > > 512 7.45 > > 1024 7.38 > > 2048 7.37 > > 4096 7.45 > > 8192 9.21 > > 16384 9.70 > > 32768 10.63 > > 65536 13.15 > > 131072 16.96 > > 262144 23.84 > > 524288 36.16 > > 1048576 60.36 > > 2097152 108.43 > > 4194304 228.31 > > > Note the slurm behavior is - I launch the job. Go get coffee, do some > duo-lingo, read some emails, then after about 10 minutes the osu latency > runs. > > > I did not get the slurm problems using an older mpich 4.3.1 but did get > the same performance issue. 9 usecs doesn't seem right for an 8-byte > pingpong over libfabric S11. I was expecting more like 1.6 or so. > > > I am confident the slurm issue is unrelated to the latency issue. > > Thanks for any suggestions on how to address either issue however. > > > > -------------- next part -------------- An HTML attachment was scrubbed... URL: From thakur at anl.gov Wed Jul 30 13:48:45 2025 From: thakur at anl.gov (Thakur, Rajeev) Date: Wed, 30 Jul 2025 18:48:45 +0000 Subject: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem In-Reply-To: References: Message-ID: <0F678B6E-4B94-4F1E-A440-AA58A7FEF241@anl.gov> Hi Howard, What was the latency with the 4.3.x branch? Rajeev From: Howard Pritchard via discuss Reply-To: "discuss at mpich.org" Date: Wednesday, July 30, 2025 at 1:43?PM To: "Zhou, Hui" Cc: Howard Pritchard , "discuss at mpich.org" Subject: Re: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem Hi Hui That didn?t help. I am not surprised though as our cluster is an NVIDIA free zone. What did help is to switch to the mpich 4.?3.?x branch and latency results are nominal and the slurm problem went away too. So we will stick with that branch.? ZjQcmQRYFpfptBannerStart This Message Is From an External Sender This message came from outside your organization. ZjQcmQRYFpfptBannerEnd Hi Hui That didn?t help. I am not surprised though as our cluster is an NVIDIA free zone. What did help is to switch to the mpich 4.3.x branch and latency results are nominal and the slurm problem went away too. So we will stick with that branch. Howard On Mon, Jul 28, 2025 at 4:15?PM Zhou, Hui > wrote: Hi Howard, I wonder whether it is due to the overhead of querying pointer attributes. Could you try disable GPU support with `MPIR_CVAR_ENABLE_GPU=0` and see if the latency improves? Hui ________________________________ From: Howard Pritchard via discuss > Sent: Monday, July 28, 2025 9:41 AM To: discuss at mpich.org > Cc: Howard Pritchard > Subject: [mpich-discuss] MPICH 5.0.1 performance on HPE SS11 plus more - a slurm problem Hi Folks, We are seeing a strange performance issue on our HPE SS11 system when testing osu_latency inter-node with MPICH. First the info: system using libfabric 1.?22.?0 slurm - 24.?11.?5 Here's my mpichversion output: MPICH Version:? 5.?0.?0a1 ZjQcmQRYFpfptBannerStart This Message Is From an External Sender This message came from outside your organization. ZjQcmQRYFpfptBannerEnd Hi Folks, We are seeing a strange performance issue on our HPE SS11 system when testing osu_latency inter-node with MPICH. First the info: system using libfabric 1.22.0 slurm - 24.11.5 Here's my mpichversion output: MPICH Version: 5.0.0a1 MPICH Release date: unreleased development copy MPICH ABI: 0:0:0 MPICH Device: ch4:ofi MPICH configure: --prefix=/XXXX/mpich_again/install --enable-g=no --enable-error-checking=no --with-device=ch4:ofi --enable-threads=multiple --with-ch4-shmmods=posix,xpmem --enable-thread-cs=per-vci --with-libfabric=/opt/cray/libfabric/1.22.0 --with-xpmem=/opt/cray/xpmem/default --with-pmix=/opt/pmix/gcc4x/5.0.8 --enable-fast=O3 MPICH CC: gcc -O3 MPICH CXX: g++ -O3 MPICH F77: gfortran -O3 MPICH FC: gfortran -O3 MPICH features: threadcomm And here's the OSU latency results: slurmstepd: error: mpi/pmix_v4: pmixp_coll_belong_chk: nid001439 [1]: pmixp_coll.c:280: No process controlled by this slurmstepd is involved in this collective. slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001439 [1]: pmixp_server.c:923: Unable to pmixp_state_coll_get() slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_check: nid001438 [0]: pmixp_coll_ring.c:614: 0x15005c005dc0: unexpected contrib from nid001439:1, expected is 0 slurmstepd: error: mpi/pmix_v4: _process_server_request: nid001438 [0]: pmixp_server.c:937: 0x15005c005dc0: unexpected contrib from nid001439:1, coll->seq=0, seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001438 [0]: pmixp_coll_ring.c:738: 0x1500580532f0: collective timeout seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001438 [0]: pmixp_coll.c:286: Dumping collective state slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:756: 0x1500580532f0: COLL_FENCE_RING state seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:758: my peerid: 0:nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:765: neighbor id: next 1:nid001439, prev 1:nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x150058053368, #0, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x1500580533a0, #1, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:775: Context ptr=0x1500580533d8, #2, in-use=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:788: neighbor contribs [2]: slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:821: done contrib: - slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:823: wait contrib: nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001438 [0]: pmixp_coll_ring.c:829: buf (offset/size): 36/16384 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_reset_if_to: nid001439 [1]: pmixp_coll_ring.c:738: 0x151d0c053400: collective timeout seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_log: nid001439 [1]: pmixp_coll.c:286: Dumping collective state slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:756: 0x151d0c053400: COLL_FENCE_RING state seq=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:758: my peerid: 1:nid001439 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:765: neighbor id: next 0:nid001438, prev 0:nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c053478, #0, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c0534b0, #1, in-use=0 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:775: Context ptr=0x151d0c0534e8, #2, in-use=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:786: seq=0 contribs: loc=1/prev=0/fwd=1 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:788: neighbor contribs [2]: slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:821: done contrib: - slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:823: wait contrib: nid001438 slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:825: status=PMIXP_COLL_RING_PROGRESS slurmstepd: error: mpi/pmix_v4: pmixp_coll_ring_log: nid001439 [1]: pmixp_coll_ring.c:829: buf (offset/size): 36/16384 # OSU MPI Latency Test v5.8 # Size Latency (us) 0 1.66 1 9.29 2 9.57 4 9.69 8 9.76 16 9.77 32 9.76 64 9.77 128 10.32 256 7.54 512 7.45 1024 7.38 2048 7.37 4096 7.45 8192 9.21 16384 9.70 32768 10.63 65536 13.15 131072 16.96 262144 23.84 524288 36.16 1048576 60.36 2097152 108.43 4194304 228.31 Note the slurm behavior is - I launch the job. Go get coffee, do some duo-lingo, read some emails, then after about 10 minutes the osu latency runs. I did not get the slurm problems using an older mpich 4.3.1 but did get the same performance issue. 9 usecs doesn't seem right for an 8-byte pingpong over libfabric S11. I was expecting more like 1.6 or so. I am confident the slurm issue is unrelated to the latency issue. Thanks for any suggestions on how to address either issue however. -------------- next part -------------- An HTML attachment was scrubbed... URL: