[mpich-discuss] discuss Digest, Vol 16, Issue 7
Kenneth Raffenetti
raffenet at mcs.anl.gov
Thu Feb 20 12:51:26 CST 2014
If you are replying to a digest email, it is hard to discern who/what
you are replying to without providing context. Now, as a reminder, this
is a list for discussing the usage of MPICH. A problem using or
understanding malloc() should be raised in a more appropriate forum.
Ken
On 02/20/2014 11:15 AM, Soheil Hooshdaran wrote:
> It causes no problem at all, since the categories/buckets are large
> enough. Only the malloc() function causes the problem. I read that it
> is not 'interrupt safe'. What does that mean and what should be done
> against it?
>
>
> On Wed, Feb 19, 2014 at 8:08 AM, <discuss-request at mpich.org
> <mailto:discuss-request at mpich.org>> wrote:
>
> Send discuss mailing list submissions to
> discuss at mpich.org <mailto:discuss at mpich.org>
>
> To subscribe or unsubscribe via the World Wide Web, visit
> https://lists.mpich.org/mailman/listinfo/discuss
> or, via email, send a message with subject or body 'help' to
> discuss-request at mpich.org <mailto:discuss-request at mpich.org>
>
> You can reach the person managing the list at
> discuss-owner at mpich.org <mailto:discuss-owner at mpich.org>
>
> When replying, please edit your Subject line so it is more specific
> than "Re: Contents of discuss digest..."
>
>
> Today's Topics:
>
> 1. Re: dynamic 2D array creation error (Kenneth Raffenetti)
> 2. Re: urgent-malloc problem (Gilles Gouaillardet)
> 3. Re: Communication Error when installing MPICH on multi
> HOSTS. (Balaji, Pavan)
>
>
> ----------------------------------------------------------------------
>
> Message: 1
> Date: Tue, 18 Feb 2014 07:48:39 -0600
> From: Kenneth Raffenetti <raffenet at mcs.anl.gov
> <mailto:raffenet at mcs.anl.gov>>
> To: <discuss at mpich.org <mailto:discuss at mpich.org>>
> Subject: Re: [mpich-discuss] dynamic 2D array creation error
> Message-ID: <530364B7.5040908 at mcs.anl.gov
> <mailto:530364B7.5040908 at mcs.anl.gov>>
> Content-Type: text/plain; charset="UTF-8"; format=flowed
>
> This list is for asking questions about the usage of mpich. General
> programming questions like the below are better suited for a forum like
> Stackoverflow.
>
> On 02/18/2014 12:44 AM, Soheil Hooshdaran wrote:
> > Hello. What's wrong with this code snippet?
> >
> > |1
> > 2
> > 3
> > 4
> > 5
> > 6
> > |
> >
> >
> >
> > | int **lBucket;
> >
> > lBucket =new int*[iWorldSize];//iWorldSize is the
> nummber of processors
> >
> > for(i=0;i<iWorldSize;++i)
> > lBucket[i] =new int[m];|
> >
> >
> >
> > Thanks in advance
> >
> >
> >
> >
> > _______________________________________________
> > discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
> > To manage subscription options or unsubscribe:
> > https://lists.mpich.org/mailman/listinfo/discuss
> >
>
>
> ------------------------------
>
> Message: 2
> Date: Wed, 19 Feb 2014 12:07:43 +0900
> From: Gilles Gouaillardet <gilles.gouaillardet at iferc.org
> <mailto:gilles.gouaillardet at iferc.org>>
> To: discuss at mpich.org <mailto:discuss at mpich.org>
> Subject: Re: [mpich-discuss] urgent-malloc problem
> Message-ID: <53041FFF.4040101 at iferc.org
> <mailto:53041FFF.4040101 at iferc.org>>
> Content-Type: text/plain; charset=ISO-8859-1
>
> Hello,
>
> how do you run your program ?
> e.g.
> ./a.out <m> <base> <r>
> which values for m, base and r ?
>
> there is no check on boundaries, and depending on the input parameters,
>
> cat[count][ cursor[count]++ ]
>
> indexes can be out of bound
>
> imho, this is not mpich related at all
>
> Best regards,
>
> Gilles
>
>
> On 2014/02/13 0:22, Soheil Hooshdaran wrote:
> > Hello.
> > I have a memory allocation problem (using malloc). I can't figure
> out its
> > cause. Could you help me please?
> >
>
>
>
> ------------------------------
>
> Message: 3
> Date: Wed, 19 Feb 2014 04:38:38 +0000
> From: "Balaji, Pavan" <balaji at anl.gov <mailto:balaji at anl.gov>>
> To: "discuss at mpich.org <mailto:discuss at mpich.org>"
> <discuss at mpich.org <mailto:discuss at mpich.org>>
> Subject: Re: [mpich-discuss] Communication Error when installing MPICH
> on multi HOSTS.
> Message-ID: <CF29912C.5996E%balaji at anl.gov
> <mailto:CF29912C.5996E%25balaji at anl.gov>>
> Content-Type: text/plain; charset="utf-8"
>
>
> It?s hard to tell, but this does indicate some problem with your
> communication setup. Did you verify your /etc/hosts like described
> on the FAQ page?
>
> ? Pavan
>
> From: ???? <wu_0317 at qq.com
> <mailto:wu_0317 at qq.com><mailto:wu_0317 at qq.com <mailto:wu_0317 at qq.com>>>
> Reply-To: "discuss at mpich.org
> <mailto:discuss at mpich.org><mailto:discuss at mpich.org
> <mailto:discuss at mpich.org>>" <discuss at mpich.org
> <mailto:discuss at mpich.org><mailto:discuss at mpich.org
> <mailto:discuss at mpich.org>>>
> Date: Tuesday, February 18, 2014 at 5:21 AM
> To: discuss <discuss at mpich.org
> <mailto:discuss at mpich.org><mailto:discuss at mpich.org
> <mailto:discuss at mpich.org>>>
> Subject: [mpich-discuss] Communication Error when installing MPICH
> on multi HOSTS.
>
> HI.
>
> My environment:
> Two Vmware VMs with ubuntu-server12.04 OS, called mpimaster,mpislaver1
> they both linked to a virtual network 10.0.0.1;
> they can ssh to each other without password;
> I have disabled the fire walls with "sudo ufw disable"
> I install mpich3.0.4 on a NFS servered by mpimaster.
>
> I installed mpich3.0.4 follow the "readme.txt", it has Communication
> problem when progresses from different host comunicate with each other.
> [cid:E1179051 at 3853243A.43420353.png]
>
> From picture above we can see it's ok to run "cpi" on both hosts
> separately.
>
> If you can't see the picture,plz see the shell's below.
>
> ailab at mpimaster:~/Downloads/mpich-3.0.4$ mpiexec -n 4 ./examples/cpi
> Process 0 of 4 is on mpimaster
> Process 1 of 4 is on mpimaster
> Process 2 of 4 is on mpimaster
> Process 3 of 4 is on mpimaster
> pi is approximately 3.1415926544231239, Error is 0.0000000008333307
> wall clock time = 0.028108
> ailab at mpimaster:~/Downloads/mpich-3.0.4$ mpiexec -hosts mpimaster -n
> 4 ./examples/cpi
> Process 2 of 4 is on mpimaster
> Process 0 of 4 is on mpimaster
> Process 1 of 4 is on mpimaster
> Process 3 of 4 is on mpimaster
> pi is approximately 3.1415926544231239, Error is 0.0000000008333307
> wall clock time = 0.027234
> ailab at mpimaster:~/Downloads/mpich-3.0.4$ mpiexec -hosts mpislaver1
> -n 4 ./examples/cpi
> Process 0 of 4 is on mpislaver1
> pi is approximately 3.1415926544231239, Error is 0.0000000008333307
> wall clock time = 0.000093
> Process 1 of 4 is on mpislaver1
> Process 2 of 4 is on mpislaver1
> Process 3 of 4 is on mpislaver1
> ailab at mpimaster:~/Downloads/mpich-3.0.4$ mpiexec -hosts
> mpimaster,mpislaver1 -n 4 ./examples/cpi
> Process 0 of 4 is on mpimaster
> Process 2 of 4 is on mpimaster
> Fatal error in PMPI_Reduce: A process has failed, error stack:
> PMPI_Reduce(1217)...............: MPI_Reduce(sbuf=0x7fff73a51ce8,
> rbuf=0x7fff73a51cf0, count=1, MPI_DOUBLE, MPI_SUM, root=0,
> MPI_COMM_WORLD) failed
> MPIR_Reduce_impl(1029)..........:
> MPIR_Reduce_intra(779)..........:
> MPIR_Reduce_impl(1029)..........:
> MPIR_Reduce_intra(835)..........:
> MPIR_Reduce_binomial(144).......:
> MPIDI_CH3U_Recvq_FDU_or_AEP(667): Communication error with rank 1
> MPIR_Reduce_intra(799)..........:
> MPIR_Reduce_impl(1029)..........:
> MPIR_Reduce_intra(835)..........:
> MPIR_Reduce_binomial(206).......: Failure during collective
>
> ================================================================================
> = BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> = EXIT CODE: 1
> = CLEANING UP REMAINING PROCESSES
> = YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> ================================================================================
> [proxy:0:1 at mpislaver1] HYD_pmcd_pmip_control_cmd_cb
> (./pm/pmiserv/pmip_cb.c:886)
> [proxy:0:1 at mpislaver1] HYDT_dmxu_poll_wait_for_event
> (./tools/demux/demux_poll.c
> [proxy:0:1 at mpislaver1] main (./pm/pmiserv/pmip.c:206): demux engine
> error waitin
> [mpiexec at mpimaster] HYDT_bscu_wait_for_completion
> (./tools/bootstrap/utils/bscu_
> [mpiexec at mpimaster] HYDT_bsci_wait_for_completion
> (./tools/bootstrap/src/bsci_wa
> [mpiexec at mpimaster] HYD_pmci_wait_for_completion
> (./pm/pmiserv/pmiserv_pmci.c:21
> [mpiexec at mpimaster] main (./ui/mpich/mpiexec.c:331): process manager
> error waiti
> ailab at mpimaster:~/Downloads/mpich-3.0.4$
>
> plz help,THX!
>
>
> ------------------
> Jie-Jun Wu
> Department of Computer Science,
> Sun Yat-sen University,
> Guangzhou,
> P.R. China
>
> -------------- next part --------------
> An HTML attachment was scrubbed...
> URL:
> <http://lists.mpich.org/pipermail/discuss/attachments/20140219/d9519600/attachment.html>
> -------------- next part --------------
> A non-text attachment was scrubbed...
> Name: E1179051 at 3853243A.43420353.png
> Type: application/octet-stream
> Size: 68330 bytes
> Desc: E1179051 at 3853243A.43420353.png
> URL:
> <http://lists.mpich.org/pipermail/discuss/attachments/20140219/d9519600/attachment.obj>
>
> ------------------------------
>
> _______________________________________________
> discuss mailing list
> discuss at mpich.org <mailto:discuss at mpich.org>
> https://lists.mpich.org/mailman/listinfo/discuss
>
> End of discuss Digest, Vol 16, Issue 7
> **************************************
>
>
>
>
> --
> قال المصطفی علیه السلام: «إرحموا ثلاثا عزیزاً، قوم الذل و غنی قوم افتقر
> و عالماً یلعب به الجُهّال»
>
>
>
> _______________________________________________
> discuss mailing list discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
>
More information about the discuss
mailing list