<div dir="ltr">Thank you.<div>Anatoly.</div></div><div class="gmail_extra"><br><br><div class="gmail_quote">On Thu, May 8, 2014 at 3:57 PM, Balaji, Pavan <span dir="ltr"><<a href="mailto:balaji@anl.gov" target="_blank">balaji@anl.gov</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><br>
This is expected. Currently, the only way to not have MPICH poll is to configure with --with-device=ch3:sock. Please note that this can cause performance loss (the polling is helpful for performance in the common case).<br>
<br>
We are planning to allow this in the default build as well in the future.<br>
<br>
— Pavan<br>
<div><div class="h5"><br>
On May 8, 2014, at 7:54 AM, Anatoly G <<a href="mailto:anatolyrishon@gmail.com">anatolyrishon@gmail.com</a>> wrote:<br>
<br>
> Dear MPICH forum.<br>
> I created an endless MPI program.<br>
> In this program each process calls MPI_Recv from other process, w/o any MPI_Send.<br>
> When I execute this program I see each process takes ~ 100% CPU core.<br>
> Is this behavior (I suppose polling) is normal?<br>
> May I reduce MPI_Recv CPU penalty?<br>
><br>
> Regards,<br>
> Anatoly.<br>
</div></div>> <mpi_polling.cpp>_______________________________________________<br>
> discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
> To manage subscription options or unsubscribe:<br>
> <a href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
<br>
_______________________________________________<br>
discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
</blockquote></div><br></div>