[mpich-devel] Hydra fails to launch hello world on 1 proc
Dave Goodell
goodell at mcs.anl.gov
Wed Apr 10 22:24:09 CDT 2013
Does it run non-MPI jobs OK? ("mpiexec -n 1 hostname", for example)
Is this Linux or a Mac?
If you temporarily disable the firewall, does that make a difference?
-Dave
On Apr 10, 2013, at 6:34 PM CDT, Jeff Hammond <jhammond at alcf.anl.gov> wrote:
> Hi,
>
> I'm using the latest Git trunk build of MPICH with GCC and am unable
> to run a 'hello, world' program using mpiexec.
>
> Any clues what the problem is? I have not seen this problem before,
> but this is newly refreshed laptop. The firewall is active but I
> would not have expected Hydra to need to go through the firewall to
> launch a serial job.
>
> If there's something wrong with my setup, it would be nice if Hydra
> would issue a warning/error instead of handing.
>
> Thanks,
>
> Jeff
>
> I compiled MPICH like this:
> ../configure CC=gcc CXX=g++ FC=gfortran F77=gfortran --enable-threads
> --enable-f77 --enable-fc --enable-g --with-pm=hydra --enable-rpath
> --disable-static --enable-shared --with-device=ch3:nemesis
> --prefix=/home/jeff/eclipse/MPICH/git/install-gcc
>
> jeff at goldstone:~/eclipse/OSPRI/mcs.svn/trunk/tests/devices/mpi-pt> mpicc -show
> gcc -I/home/jeff/eclipse/MPICH/git/install-gcc/include
> -L/home/jeff/eclipse/MPICH/git/install-gcc/lib64 -Wl,-rpath
> -Wl,/home/jeff/eclipse/MPICH/git/install-gcc/lib64 -lmpich -lopa -lmpl
> -lrt -lpthread
>
> jeff at goldstone:~/eclipse/OSPRI/mcs.svn/trunk/tests/devices/mpi-pt> make
> mpicc -g -O0 -Wall -std=gnu99 -DDEBUG -c hello.c -o hello.o
> mpicc -g -O0 -Wall -std=gnu99 safemalloc.o hello.o -lm -o hello.x
> rm hello.o
>
> jeff at goldstone:~/eclipse/OSPRI/mcs.svn/trunk/tests/devices/mpi-pt>
> mpiexec -n 1 ./hello.x
> ^C[mpiexec at goldstone.mcs.anl.gov] Sending Ctrl-C to processes as requested
> [mpiexec at goldstone.mcs.anl.gov] Press Ctrl-C again to force abort
> [mpiexec at goldstone.mcs.anl.gov] HYDU_sock_write
> (../../../../src/pm/hydra/utils/sock/sock.c:291): write error (Bad
> file descriptor)
> [mpiexec at goldstone.mcs.anl.gov] HYD_pmcd_pmiserv_send_signal
> (../../../../src/pm/hydra/pm/pmiserv/pmiserv_cb.c:170): unable to
> write data to proxy
> [mpiexec at goldstone.mcs.anl.gov] ui_cmd_cb
> (../../../../src/pm/hydra/pm/pmiserv/pmiserv_pmci.c:79): unable to
> send signal downstream
> [mpiexec at goldstone.mcs.anl.gov] HYDT_dmxu_poll_wait_for_event
> (../../../../src/pm/hydra/tools/demux/demux_poll.c:77): callback
> returned error status
> [mpiexec at goldstone.mcs.anl.gov] HYD_pmci_wait_for_completion
> (../../../../src/pm/hydra/pm/pmiserv/pmiserv_pmci.c:197): error
> waiting for event
> [mpiexec at goldstone.mcs.anl.gov] main
> (../../../../src/pm/hydra/ui/mpich/mpiexec.c:331): process manager
> error waiting for completion
>
> jeff at goldstone:~/eclipse/OSPRI/mcs.svn/trunk/tests/devices/mpi-pt> ./hello.x
> <no errors>
>
> jeff at goldstone:~/eclipse/OSPRI/mcs.svn/trunk/tests/devices/mpi-pt> cat hello.c
> #include <stdio.h>
> #include <stdlib.h>
>
> #include <mpi.h>
>
> int main(int argc, char * argv[])
> {
> int provided;
>
> MPI_Init_thread(&argc, &argv, MPI_THREAD_MULTIPLE, &provided);
> if (provided!=MPI_THREAD_MULTIPLE)
> MPI_Abort(MPI_COMM_WORLD, 1);
>
> int rank, size;
> MPI_Comm_rank(MPI_COMM_WORLD, &rank);
> MPI_Comm_size(MPI_COMM_WORLD, &size);
>
> MPI_Finalize();
>
> return 0;
> }
>
>
> --
> Jeff Hammond
> Argonne Leadership Computing Facility
> University of Chicago Computation Institute
> jhammond at alcf.anl.gov / (630) 252-5381
> http://www.linkedin.com/in/jeffhammond
> https://wiki.alcf.anl.gov/parts/index.php/User:Jhammond
More information about the devel
mailing list