[mpich-discuss] having problem running MPICH on multiple nodes
Kenneth Raffenetti
raffenet at mcs.anl.gov
Wed Nov 26 09:25:38 CST 2014
The connection refused makes me think a firewall is getting in the way.
Is TCP communication limited to specific ports on the cluster? If so,
you can use this envvar to enforce a range of ports in MPICH.
MPIR_CVAR_CH3_PORT_RANGE
Description: The MPIR_CVAR_CH3_PORT_RANGE environment variable
allows you to specify the range of TCP ports to be used by the process
manager and the MPICH library. The format of this variable is
<low>:<high>. To specify any available port, use 0:0.
Default: {0,0}
On 11/25/2014 11:50 PM, Amin Hassani wrote:
> Tried with the new configure too. same problem :(
>
> $ mpirun -hostfile hosts-hydra -np 2 test_dup
> Fatal error in MPI_Send: Unknown error class, error stack:
> MPI_Send(174)..............: MPI_Send(buf=0x7fffd90c76c8, count=1,
> MPI_INT, dest=1, tag=0, MPI_COMM_WORLD) failed
> MPID_nem_tcp_connpoll(1832): Communication error with rank 1: Connection
> refused
>
> ===================================================================================
> = BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> = PID 5459 RUNNING AT oakmnt-0-a
> = EXIT CODE: 1
> = CLEANING UP REMAINING PROCESSES
> = YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> ===================================================================================
> [proxy:0:1 at oakmnt-0-b] HYD_pmcd_pmip_control_cmd_cb
> (../../../../src/pm/hydra/pm/pmiserv/pmip_cb.c:885): assert (!closed) failed
> [proxy:0:1 at oakmnt-0-b] HYDT_dmxu_poll_wait_for_event
> (../../../../src/pm/hydra/tools/demux/demux_poll.c:76): callback
> returned error status
> [proxy:0:1 at oakmnt-0-b] main
> (../../../../src/pm/hydra/pm/pmiserv/pmip.c:206): demux engine error
> waiting for event
> [mpiexec at oakmnt-0-a] HYDT_bscu_wait_for_completion
> (../../../../src/pm/hydra/tools/bootstrap/utils/bscu_wait.c:76): one of
> the processes terminated badly; aborting
> [mpiexec at oakmnt-0-a] HYDT_bsci_wait_for_completion
> (../../../../src/pm/hydra/tools/bootstrap/src/bsci_wait.c:23): launcher
> returned error waiting for completion
> [mpiexec at oakmnt-0-a] HYD_pmci_wait_for_completion
> (../../../../src/pm/hydra/pm/pmiserv/pmiserv_pmci.c:218): launcher
> returned error waiting for completion
> [mpiexec at oakmnt-0-a] main
> (../../../../src/pm/hydra/ui/mpich/mpiexec.c:344): process manager error
> waiting for completion
>
>
> Amin Hassani,
> CIS department at UAB,
> Birmingham, AL, USA.
>
> On Tue, Nov 25, 2014 at 11:44 PM, Lu, Huiwei <huiweilu at mcs.anl.gov
> <mailto:huiweilu at mcs.anl.gov>> wrote:
>
> So the error only happens when there is communication.
>
> It may be caused by IB as your guessed before. Could you try to
> reconfigure MPICH using "./configure --with-device=ch3:nemesis:tcp”
> and try again?
>
> —
> Huiwei
>
> > On Nov 25, 2014, at 11:23 PM, Amin Hassani <ahassani at cis.uab.edu
> <mailto:ahassani at cis.uab.edu>> wrote:
> >
> > Yes it works.
> > output:
> >
> > $ mpirun -hostfile hosts-hydra -np 2 test
> > rank 1
> > rank 0
> >
> >
> > Amin Hassani,
> > CIS department at UAB,
> > Birmingham, AL, USA.
> >
> > On Tue, Nov 25, 2014 at 11:20 PM, Lu, Huiwei
> <huiweilu at mcs.anl.gov <mailto:huiweilu at mcs.anl.gov>> wrote:
> > Could you try to run the following simple code to see if it works?
> >
> > #include <mpi.h>
> > #include <stdio.h>
> > int main(int argc, char** argv)
> > {
> > int rank, size;
> > MPI_Init(&argc, &argv);
> > MPI_Comm_rank(MPI_COMM_WORLD, &rank);
> > printf("rank %d\n", rank);
> > MPI_Finalize();
> > return 0;
> > }
> >
> > —
> > Huiwei
> >
> > > On Nov 25, 2014, at 11:11 PM, Amin Hassani
> <ahassani at cis.uab.edu <mailto:ahassani at cis.uab.edu>> wrote:
> > >
> > > No, I checked. Also I always install my MPI's in
> /nethome/students/ahassani/usr/mpi. I never install them in
> /nethome/students/ahassani/usr. So MPI files will never get there.
> Even if put the /usr/mpi/bin in front of /usr/bin, it won't affect
> anything. There has never been any mpi installed in /usr/bin.
> > >
> > > Thank you.
> > > _______________________________________________
> > > discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
> > > To manage subscription options or unsubscribe:
> > > https://lists.mpich.org/mailman/listinfo/discuss
> >
> > _______________________________________________
> > discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
> > To manage subscription options or unsubscribe:
> > https://lists.mpich.org/mailman/listinfo/discuss
> >
> > _______________________________________________
> > discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
> > To manage subscription options or unsubscribe:
> > https://lists.mpich.org/mailman/listinfo/discuss
>
> _______________________________________________
> discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
>
>
>
>
> _______________________________________________
> discuss mailing list discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
>
_______________________________________________
discuss mailing list discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss
More information about the discuss
mailing list