[mpich-discuss] [EXTERNAL] Re: Spawned process hanging in MPI_Finalize

Mccall, Kurt E. (MSFC-EV41) kurt.e.mccall at nasa.gov
Wed Mar 3 18:55:52 CST 2021


Thanks, that helped!   Is it necessary to call MPI_Comm_disconnect on inter-communicators that are created by MPI_Intercomm_create?


-----Original Message-----
From: Joachim Protze <protze at itc.rwth-aachen.de> 
Sent: Wednesday, March 3, 2021 5:52 AM
To: discuss at mpich.org
Cc: Mccall, Kurt E. (MSFC-EV41) <kurt.e.mccall at nasa.gov>
Subject: [EXTERNAL] Re: [mpich-discuss] Spawned process hanging in MPI_Finalize

Hi Kurt,

did you call MPI_Comm_disconnect on all processes connected with the
inter-communicator? Also the parent process needs to disconnect from the
inter-comm before the MPI_Comm_disconnect can return.

- Joachim

Am 03.03.21 um 01:54 schrieb Mccall, Kurt E. (MSFC-EV41) via discuss:
> I have a parent process that creates a child via MPI_Comm_spawn().     When the child decides it has to exit, it is hanging in MPI_Finalize().   It does the same if it calls MPI_Comm_disconnect() before MPI_Finalize.
> Here is the stack trace in the child:
> (gdb) where
> #0  0x00007fc6f2fedde0 in __poll_nocancel () from /usr/lib/gcc/x86_64-redhat-linux/4.8.5/../../../../lib64/libc.so.6
> #1  0x00007fc6f4dc840e in MPID_nem_tcp_connpoll () at src/mpid/ch3/channels/nemesis/netmod/tcp/socksm.c:1819
> #2  0x00007fc6f4db857e in MPID_nem_network_poll () at src/mpid/ch3/channels/nemesis/src/mpid_nem_network_poll.c:16
> #3  0x00007fc6f4dafc43 in MPIDI_CH3I_Progress () at src/mpid/ch3/channels/nemesis/src/ch3_progress.c:1019
> #4  0x00007fc6f4d5094d in MPIDI_CH3U_VC_WaitForClose () at src/mpid/ch3/src/ch3u_handle_connection.c:383
> #5  0x00007fc6f4d94efa in MPID_Finalize () at src/mpid/ch3/src/mpid_finalize.c:110
> #6  0x00007fc6f4c432ca in PMPI_Finalize () at src/mpi/init/finalize.c:260
> #7  0x0000000000408a85 in needles::MpiWorker::finalize () at src/MpiWorker.cpp:470
> Maybe I have a communication that hasn't completed, or the child is waiting for the parent to call MPI_Finalize.   I believe that you (Ken, Hui) told me that it shouldn't do the latter.
> Is there a way for the child to cleanly exit without hanging in MPI_Finalize?   I tried calling MPI_Cancel() in the child on the only possible communication request that I knew of, but it didn't help.
> It just occurred to me that I haven't tried calling MPI_Cancel on the requests in the parent...
> Thanks,
> Kurt
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss

Dipl.-Inf. Joachim Protze

IT Center
Group: High Performance Computing
Division: Computational Science and Engineering
RWTH Aachen University
Seffenter Weg 23
D 52074  Aachen (Germany)
Tel: +49 241 80- 24765
Fax: +49 241 80-624765
protze at itc.rwth-aachen.de
-------------- next part --------------
A non-text attachment was scrubbed...
Name: smime.p7s
Type: application/pkcs7-signature
Size: 10416 bytes
Desc: not available
URL: <http://lists.mpich.org/pipermail/discuss/attachments/20210304/745e2a07/attachment.p7s>

More information about the discuss mailing list