<html dir="ltr"><head>
<meta http-equiv="Content-Type" content="text/html; charset=Windows-1252">
<style id="owaParaStyle" type="text/css">P {margin-top:0;margin-bottom:0;}</style>
</head>
<body ocsi="0" fpstyle="1">
<div style="direction: ltr;font-family: Tahoma;color: #000000;font-size: 10pt;">Thanks for the "bash line", the output is:<br>
<br>
$ grep "define MPICH_VERSION" $CRAY_MPICH2_DIR/include/mpi.h<br>
#define MPICH_VERSION "3.1.2"<br>
<br>
Florian<br>
<br>
<div style="font-family: Times New Roman; color: #000000; font-size: 16px">
<hr tabindex="-1">
<div style="direction: ltr;" id="divRpF391591"><font face="Tahoma" size="2" color="#000000"><b>Von:</b> Jeff Hammond [jeff.science@gmail.com]<br>
<b>Gesendet:</b> Mittwoch, 2. Dezember 2015 14:58<br>
<b>An:</b> MPICH<br>
<b>Betreff:</b> Re: [mpich-discuss] MPI_Gather fails with 2048 processes and 4096 MB total<br>
</font><br>
</div>
<div></div>
<div>
<div dir="ltr">1) You should address this with Cray. Their implementation of MPI has closed-source modifications to MPICH that we cannot analyze for count-safety.
<div><br>
</div>
<div>2) You can discover the version of MPICH associated with Cray MPI like this:<br>
jhammond@cori11:~> grep "define MPICH_VERSION" $CRAY_MPICH2_DIR/include/mpi.h</div>
<div>#define MPICH_VERSION "3.1.2"<br>
</div>
<div><br>
</div>
<div>Best,</div>
<div><br>
</div>
<div>Jeff</div>
</div>
<div class="gmail_extra"><br>
<div class="gmail_quote">On Wed, Dec 2, 2015 at 12:15 AM, <span dir="ltr"><<a href="mailto:Florian.Willich@dlr.de" target="_blank">Florian.Willich@dlr.de</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex; border-left:1px #ccc solid; padding-left:1ex">
Hi Rob,<br>
<br>
well maybe I was addressing the wrong organisation... I am currently testing on the Cray swan super computer which provides the module cray-mpich/7.2.6 ("Cray Message Passing Toolkit 7.2.6").<br>
<br>
I can not determine whether the cray mpich version is mpich with additional implementations or if it is totally different from the official mpich releases. Additioanlly, I can not figure out on which mpich version this cray-mpich module is based on. I'll continue
investigation and keep you updated.<br>
<br>
Best Regards<br>
<br>
Florian<br>
________________________________________<br>
Von: Rob Latham [<a href="mailto:robl@mcs.anl.gov" target="_blank">robl@mcs.anl.gov</a>]<br>
Gesendet: Dienstag, 1. Dezember 2015 16:48<br>
An: <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
Betreff: Re: [mpich-discuss] MPI_Gather fails with 2048 processes and 4096 MB total<br>
<div class="HOEnZb">
<div class="h5"><br>
On 11/26/2015 12:38 PM, Archer, Charles J wrote:<br>
> FYI, we hit various flavors of this problem when I was still at IBM, I think mostly in weather codes.<br>
> Apparently Cray hit this too:<br>
><br>
> <a href="https://trac.mpich.org/projects/mpich/ticket/1767" rel="noreferrer" target="_blank">
https://trac.mpich.org/projects/mpich/ticket/1767</a><br>
><br>
> We pretty much told our customers back then that a fix was forthcoming (with no ETA :) )with the revamp of datatypes to use internal 64-bit counts.<br>
> We also provided workarounds.<br>
><br>
> In the case of this gather operation, we asked the customer to implement gather as a flat tree using point to point.<br>
> Root posts irecvs, then barrier, children send to root.<br>
><br>
> IRC, the giant gather we were debugging was at the very end of the application and used to gather some statistics for IO at the root, so it wasn’t critical to perform well.<br>
> I also attempted a workaround using some derived datatypes, but I hit another truncation in the datatype code itself :\<br>
> I should see if I can dig up that implementation and make sure it isn’t still broken for large counts.<br>
<br>
those are all fine approaches to work around the problem. the internals<br>
of MPICH, though, need to be 64 bit clean -- there are still 4500 places<br>
where clang warns of a 64 bit value being assigned to a 32 bit type.<br>
<br>
Florian Willich, what version of MPICH is this? The line numbers in<br>
the back trace don't match up with what I've got, and<br>
I really thought we fixed this class of bug with commits 31d95ed7b18c<br>
and 68f8c7aa7 over the summer.<br>
<br>
==rob<br>
<br>
<br>
><br>
><br>
><br>
><br>
><br>
> On Nov 26, 2015, at 12:14 PM, Balaji, Pavan <<a href="mailto:balaji@anl.gov" target="_blank">balaji@anl.gov</a><mailto:<a href="mailto:balaji@anl.gov" target="_blank">balaji@anl.gov</a>>> wrote:<br>
><br>
><br>
> Thanks for reporting. This looks like an integer-overflow issue, which fails when the summation of data elements from all processes is larger than INT_MAX (2 billion). We'll look into it. I've created a ticket for it, and added you as the reporter, so
you'll get notified as they are updates.<br>
><br>
> <a href="http://trac.mpich.org/projects/mpich/ticket/2317" rel="noreferrer" target="_blank">
http://trac.mpich.org/projects/mpich/ticket/2317</a><br>
><br>
> Rob: can you create a simple test program for this and add it to the test bucket, so it shows up on the nightlies?<br>
><br>
> Thanks,<br>
><br>
> -- Pavan<br>
><br>
> On Nov 26, 2015, at 10:18 AM, <a href="mailto:Florian.Willich@dlr.de" target="_blank">
Florian.Willich@dlr.de</a> wrote:<br>
><br>
> Dear mpich discussion group,<br>
><br>
> the following issue appeared when running some benchmarks with MPI Gather:<br>
><br>
> Gathering data (calling MPI_Gather(...) ) involing 2048 processes and 2 MB of data (4096 MB total) that each process sends fails with the following output:<br>
> ____________________________<br>
><br>
> Rank 1024 [Thu Nov 26 09:43:16 2015] [c1-0c1s12n3] Fatal error in PMPI_Gather: Invalid count, error stack:<br>
> PMPI_Gather(959)......: MPI_Gather(sbuf=0x2aaab826c010, scount=524288, MPI_INT, rbuf=(nil), rcount=524288, MPI_INT, root=0, MPI_COMM_WORLD) failed<br>
> MPIR_Gather_impl(775).:<br>
> MPIR_Gather(735)......:<br>
> MPIR_Gather_intra(347):<br>
> MPIC_Send(360)........: Negative count, value is -2147483648<br>
> _pmiu_daemon(SIGCHLD): [NID 00307] [c1-0c1s12n3] [Thu Nov 26 09:43:16 2015] PE RANK 1024 exit signal Aborted<br>
> [NID 00307] 2015-11-26 09:43:16 Apid 949450: initiated application termination<br>
> Application 949450 exit codes: 134<br>
> Application 949450 exit signals: Killed<br>
> Application 949450 resources: utime ~1s, stime ~137s, Rss ~2110448, inblocks ~617782, outblocks ~1659320<br>
> ____________________________<br>
><br>
> The following are some tests that I ran to better understand the problem:<br>
><br>
> 2047 processes - 2 MB (4094 MB total) -> works!<br>
><br>
> 2048 processes - 2047.5 KB (~1.999512 MB) (4095 MB total) -> works!<br>
><br>
> 2048 processes - 3 MB (6144 MB total) -> fails:<br>
> ____________________________<br>
><br>
> Rank 1024 [Thu Nov 26 09:41:15 2015] [c1-0c1s12n3] Fatal error in PMPI_Gather: Invalid count, error stack:<br>
> PMPI_Gather(959)......: MPI_Gather(sbuf=0x2aaab826c010, scount=786432, MPI_INT, rbuf=(nil), rcount=786432, MPI_INT, root=0, MPI_COMM_WORLD) failed<br>
> MPIR_Gather_impl(775).:<br>
> MPIR_Gather(735)......:<br>
> MPIR_Gather_intra(347):<br>
> MPIC_Send(360)........: Negative count, value is -1073741824<br>
> _pmiu_daemon(SIGCHLD): [NID 00307] [c1-0c1s12n3] [Thu Nov 26 09:41:15 2015] PE RANK 1024 exit signal Aborted<br>
> [NID 00307] 2015-11-26 09:41:15 Apid 949448: initiated application termination<br>
> Application 949448 exit codes: 134<br>
> Application 949448 exit signals: Killed<br>
> Application 949448 resources: utime ~1s, stime ~139s, Rss ~3159984, inblocks ~617782, outblocks ~1659351<br>
> ____________________________<br>
><br>
> 2047 processes - 3 MB (6141 MB total) -> fails:<br>
> ____________________________<br>
><br>
> Rank 1024 [Thu Nov 26 09:40:31 2015] [c1-0c1s12n3] Fatal error in PMPI_Gather: Invalid count, error stack:<br>
> PMPI_Gather(959)......: MPI_Gather(sbuf=0x2aaab826c010, scount=786432, MPI_INT, rbuf=(nil), rcount=786432, MPI_INT, root=0, MPI_COMM_WORLD) failed<br>
> MPIR_Gather_impl(775).:<br>
> MPIR_Gather(735)......:<br>
> MPIR_Gather_intra(347):<br>
> MPIC_Send(360)........: Negative count, value is -1076887552<br>
> _pmiu_daemon(SIGCHLD): [NID 00307] [c1-0c1s12n3] [Thu Nov 26 09:40:32 2015] PE RANK 1024 exit signal Aborted<br>
> [NID 00307] 2015-11-26 09:40:32 Apid 949446: initiated application termination<br>
> Application 949446 exit codes: 134<br>
> Application 949446 exit signals: Killed<br>
> Application 949446 resources: utime ~1s, stime ~134s, Rss ~3157072, inblocks ~617780, outblocks ~1659351<br>
> ____________________________<br>
><br>
> 8 processes - 625 MB (5000 MB total) -> works!<br>
><br>
> I can think of some pitfalls that might cause this issue but I do not have the knowledge of the internally called routines to check them. Is someone familier with the implementation of MPI_Gather(...) and willing to help me?<br>
><br>
> Best regards<br>
><br>
> Florian<br>
><br>
> Deutsches Zentrum für Luft- und Raumfahrt e. V. (DLR)<br>
> German Aerospace Center<br>
> Institute of Planetary Research | Planetary Physics | Rutherfordstraße 2 | 12489 Berlin<br>
><br>
> Florian Willich| Intern - Software Developer (Parallel Applications)<br>
> <a href="mailto:florian.willlich@dlr.de" target="_blank">florian.willlich@dlr.de</a><br>
> DLR.de<br>
> _______________________________________________<br>
> discuss mailing list <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
> To manage subscription options or unsubscribe:<br>
> <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">
https://lists.mpich.org/mailman/listinfo/discuss</a><br>
><br>
> _______________________________________________<br>
> discuss mailing list <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
> To manage subscription options or unsubscribe:<br>
> <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">
https://lists.mpich.org/mailman/listinfo/discuss</a><br>
><br>
> _______________________________________________<br>
> discuss mailing list <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
> To manage subscription options or unsubscribe:<br>
> <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">
https://lists.mpich.org/mailman/listinfo/discuss</a><br>
><br>
<br>
--<br>
Rob Latham<br>
Mathematics and Computer Science Division<br>
Argonne National Lab, IL USA<br>
_______________________________________________<br>
discuss mailing list <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
_______________________________________________<br>
discuss mailing list <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
</div>
</div>
</blockquote>
</div>
<br>
<br clear="all">
<div><br>
</div>
-- <br>
<div class="gmail_signature">Jeff Hammond<br>
<a href="mailto:jeff.science@gmail.com" target="_blank">jeff.science@gmail.com</a><br>
<a href="http://jeffhammond.github.io/" target="_blank">http://jeffhammond.github.io/</a></div>
</div>
</div>
</div>
</div>
</body>
</html>