[mpich-discuss] error spawning processes in mpich-3.2rc1

Min Si msi at il.is.s.u-tokyo.ac.jp
Wed Oct 7 16:48:06 CDT 2015


Hi Siegmar,

We cannot reproduce this error on our test machines (Solaris i386, 
Ubuntu x86_64) by using your programs. And unfortunately we do not have 
Solaris Sparc machine thus could not verify it.

Sometime, it can happen that you need to add "./" in front of the 
program path, could you try it ?

For example, in spawn_master.c MPI: A Message-Passing Interface Standard
> #define SLAVE_PROG      "./spawn_slave"

Min

On 10/7/15 5:03 AM, Siegmar Gross wrote:
> Hi,
>
> today I've built mpich-3.2rc1 on my machines (Solaris 10 Sparc,
> Solaris 10 x86_64, and openSUSE Linux 12.1 x86_64) with gcc-5.1.0
> and Sun C 5.13. I still get the following errors on my Sparc machine
> which I'd already reported September 8th. "mpiexec" is aliased to
> 'mpiexec -genvnone'. It still doesn't matter if I use my cc- or
> gcc-version of MPICH.
>
>
> tyr spawn 119 mpichversion
> MPICH Version:          3.2rc1
> MPICH Release date:     Wed Oct  7 00:00:33 CDT 2015
> MPICH Device:           ch3:nemesis
> MPICH configure:        --prefix=/usr/local/mpich-3.2_64_cc 
> --libdir=/usr/local/mpich-3.2_64_cc/lib64 
> --includedir=/usr/local/mpich-3.2_64_cc/include64 CC=cc CXX=CC F77=f77 
> FC=f95 CFLAGS=-m64 CXXFLAGS=-m64 FFLAGS=-m64 FCFLAGS=-m64 LDFLAGS=-m64 
> -L/usr/lib/sparcv9 -R/usr/lib/sparcv9 --enable-fortran=yes 
> --enable-cxx --enable-romio --enable-debuginfo --enable-smpcoll 
> --enable-threads=multiple --with-thread-package=posix --enable-shared
> MPICH CC:       cc -m64   -O2
> MPICH CXX:      CC -m64  -O2
> MPICH F77:      f77 -m64
> MPICH FC:       f95 -m64  -O2
> tyr spawn 120
>
>
>
> tyr spawn 111 mpiexec -np 1 spawn_master
>
> Parent process 0 running on tyr.informatik.hs-fulda.de
>   I create 4 slave processes
>
> Fatal error in MPI_Comm_spawn: Unknown error class, error stack:
> MPI_Comm_spawn(144)...........: MPI_Comm_spawn(cmd="spawn_slave", 
> argv=0, maxprocs=4, MPI_INFO_NULL, root=0, MPI_COMM_WORLD, 
> intercomm=ffffffff7fffde50, errors=0) failed
> MPIDI_Comm_spawn_multiple(274):
> MPID_Comm_accept(153).........:
> MPIDI_Comm_accept(1057).......:
> MPIR_Bcast_intra(1287)........:
> MPIR_Bcast_binomial(310)......: Failure during collective
>
>
>
>
> tyr spawn 112 mpiexec -np 1 spawn_multiple_master
>
> Parent process 0 running on tyr.informatik.hs-fulda.de
>   I create 3 slave processes.
>
> Fatal error in MPI_Comm_spawn_multiple: Unknown error class, error stack:
> MPI_Comm_spawn_multiple(162)..: MPI_Comm_spawn_multiple(count=2, 
> cmds=ffffffff7fffde08, argvs=ffffffff7fffddf8, 
> maxprocs=ffffffff7fffddf0, infos=ffffffff7fffdde8, root=0, 
> MPI_COMM_WORLD, intercomm=ffffffff7fffdde4, errors=0) failed
> MPIDI_Comm_spawn_multiple(274):
> MPID_Comm_accept(153).........:
> MPIDI_Comm_accept(1057).......:
> MPIR_Bcast_intra(1287)........:
> MPIR_Bcast_binomial(310)......: Failure during collective
>
>
>
>
> tyr spawn 113 mpiexec -np 1 spawn_intra_comm
> Parent process 0: I create 2 slave processes
> Fatal error in MPI_Comm_spawn: Unknown error class, error stack:
> MPI_Comm_spawn(144)...........: MPI_Comm_spawn(cmd="spawn_intra_comm", 
> argv=0, maxprocs=2, MPI_INFO_NULL, root=0, MPI_COMM_WORLD, 
> intercomm=ffffffff7fffded4, errors=0) failed
> MPIDI_Comm_spawn_multiple(274):
> MPID_Comm_accept(153).........:
> MPIDI_Comm_accept(1057).......:
> MPIR_Bcast_intra(1287)........:
> MPIR_Bcast_binomial(310)......: Failure during collective
> tyr spawn 114
>
>
> I would be grateful if somebody can fix the problem. Thank you very
> much for any help in advance. I've attached my programs. Please let
> me know if you need anything else.
>
>
> Kind regards
>
> Siegmar
>
>
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mpich.org/pipermail/discuss/attachments/20151007/9a0a8154/attachment.html>
-------------- next part --------------
_______________________________________________
discuss mailing list     discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss


More information about the discuss mailing list