<div dir="ltr"><div><div><div><br></div>I am running bandwidth test from OSU-Micro-Benchmarks and a simple Ping-Pong program implemented using MPI Send and Recv methods. I have also tested bandwidth with NetPipe MPI implementation as well but no improvements in results.<br>
<br></div>Interrupt coalescing was ON by default and disabling it makes latency consistent at 37 us through out messages of 1 byte to 2 MB data size. I have done following vendor recommended optimizations for network buffer. Latency is not increasing with the message size may be because of these settings.<br>
<pre>net.core.rmem_max = 16777216
net.core.wmem_max = 16777216
net.ipv4.tcp_rmem = 4096 87380 16777216
net.ipv4.tcp_wmem = 4096 65536 16777216
net.core.netdev_max_backlog = 250000</pre><br></div>I have not run streaming benchmark. I found a Streaming MPI benchmark (<a href="http://www.cs.virginia.edu/stream/">http://www.cs.virginia.edu/stream/</a>) but its C implementation is not available. Can you please suggest me any Streaming MPI-C benchmark?<br>
<div><div><br></div><div>Latency test with netperf TCP_RR gives 37 us for message size 1 Bytes. Moreover, according to Netperf docs, _RR test support only 1 bytes message size. Please correct me if I am wrong.<br><br></div>
<div>Thanks,<br></div><div><div><div> </div></div></div></div></div><div class="gmail_extra"><br><br><div class="gmail_quote">On Wed, Apr 16, 2014 at 5:36 PM, Atchley, Scott <span dir="ltr"><<a href="mailto:atchleyes@ornl.gov" target="_blank">atchleyes@ornl.gov</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div class="">On Apr 16, 2014, at 8:35 AM, Scott Atchley <<a href="mailto:atchleyes@ornl.gov">atchleyes@ornl.gov</a>> wrote:<br>
<br>
> Keep in mind that iperf is a streaming test. What MPI application are you running? IMB pingpong?<br>
><br>
> You should review you NIC vendor's tuning FAQ. Is interrupt coalescing turned on? Normally it is, but it increases small message latency.<br>
><br>
> As for bandwidth, have you tried a streaming MPI benchmark?<br>
<br>
</div>Or run Netperf's TCP_RR for each message size to see how it compares to pingpong.<br>
<div class="HOEnZb"><div class="h5"><br>
><br>
> Scott<br>
><br>
> On Apr 15, 2014, at 4:29 PM, Muhammad Ansar Javed <<a href="mailto:muhammad.ansar@seecs.edu.pk">muhammad.ansar@seecs.edu.pk</a>> wrote:<br>
><br>
>> Hi Antonio,<br>
>> Thanks for response.<br>
>> Here is output of iperf bandwidth test.<br>
>> .<br>
>> Client connecting to 10g2, TCP port 5001<br>
>> TCP window size: 1.41 MByte (default)<br>
>> ------------------------------------------------------------<br>
>> [ 5] local 192.168.1.33 port 34543 connected with 192.168.1.34 port 5001<br>
>> [ 4] local 192.168.1.33 port 5001 connected with 192.168.1.34 port 48531<br>
>> [ ID] Interval Transfer Bandwidth<br>
>> [ 5] 0.0-10.0 sec 11.2 GBytes 9.61 Gbits/sec<br>
>> [ 4] 0.0-10.0 sec 11.2 GBytes 9.59 Gbits/sec<br>
>><br>
>><br>
>><br>
>> On Tue, Apr 15, 2014 at 7:17 PM, "Antonio J. Peña" <<a href="mailto:apenya@mcs.anl.gov">apenya@mcs.anl.gov</a>> wrote:<br>
>><br>
>> Hi Ansar,<br>
>><br>
>> The bandwidth numbers seem a little low. I'd expect something around 7,000 Mbps. To discard issues with MPI, you can try generic benchmarking tools such as those described here><br>
>><br>
>> <a href="http://linuxaria.com/article/tool-command-line-bandwidth-linux?lang=en" target="_blank">http://linuxaria.com/article/tool-command-line-bandwidth-linux?lang=en</a><br>
>><br>
>> If you get similar numbers, you'll have to check your setup. Otherwise, let us know and we'll try to see if there's any performance issue with MPICH.<br>
>><br>
>> Best,<br>
>> Antonio<br>
>><br>
>><br>
>><br>
>> On 04/15/2014 06:35 AM, Muhammad Ansar Javed wrote:<br>
>>> Hi,<br>
>>> I am running benchmarks for MPICH v3.1 performance evaluation on 10G Ethernet connection between two hosts. The performance results are less than expected. Here is complete set of numbers for Latency and Bandwidth tests.<br>
>>><br>
>>> mpj@host3:~/code/benchmarks$ mpiexec -n 2 -f machines ./latency.out<br>
>>> Latency: MPICH-C<br>
>>> Size ______ Time (us)<br>
>>> 1 _________ 37.91<br>
>>> 2 _________ 37.76<br>
>>> 4 _________ 37.75<br>
>>> 8 _________ 39.02<br>
>>> 16 ________ 52.71<br>
>>> 32 ________ 39.08<br>
>>> 64 ________ 37.75<br>
>>> 128 _______ 37.77<br>
>>> 256 _______ 57.47<br>
>>> 512 _______ 37.86<br>
>>> 1024 ______ 37.76<br>
>>> 2048 ______ 37.88<br>
>>><br>
>>> mpj@host3:~/code/benchmarks$ mpiexec -n 2 -f machines ./bandwidth.out<br>
>>> Bandwidth: MPICH-C<br>
>>> Size(Bytes) Bandwidth (Mbps)<br>
>>> 2048 ______ 412.32<br>
>>> 4096 ______ 820.06<br>
>>> 8192 ______ 827.77<br>
>>> 16384 _____ 1644.36<br>
>>> 32768 _____ 2207.52<br>
>>> 65536 _____ 4368.76<br>
>>> 131072 ____ 2942.93<br>
>>> 262144 ____ 4281.17<br>
>>> 524288 ____ 4773.78<br>
>>> 1048576 ___ 5310.85<br>
>>> 2097152 ___ 5382.94<br>
>>> 4194304 ___ 5518.97<br>
>>> 8388608 ___ 5508.87<br>
>>> 16777216 __ 5498.93<br>
>>><br>
>>> My environments consists of two hosts having point-to-point (switch-less) 10Gbps Ethernet connection. Environment (OS, user, directory structure etc) on both hosts is exactly same. There is no NAS or shared file system between hosts. I have attached output of script<br>
>>> mpiexec -all.<br>
>>><br>
>>> Are these numbers okay? If not then please suggest performance improvement methodology...<br>
>>><br>
>>> Thanks<br>
>>><br>
>>> --<br>
>>> Regards<br>
>>><br>
>>> Ansar Javed<br>
>>> HPC Lab<br>
>>> SEECS NUST<br>
>>> Contact: +92 334 438 9394<br>
>>> Email: <a href="mailto:muhammad.ansar@seecs.edu.pk">muhammad.ansar@seecs.edu.pk</a><br>
>>><br>
>>><br>
>>> _______________________________________________<br>
>>> discuss mailing list<br>
>>> <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
>>><br>
>>> To manage subscription options or unsubscribe:<br>
>>><br>
>>> <a href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
>><br>
>><br>
>> --<br>
>> Antonio J. Peña<br>
>> Postdoctoral Appointee<br>
>> Mathematics and Computer Science Division<br>
>> Argonne National Laboratory<br>
>> 9700 South Cass Avenue, Bldg. 240, Of. 3148<br>
>> Argonne, IL 60439-4847<br>
>><br>
>> <a href="mailto:apenya@mcs.anl.gov">apenya@mcs.anl.gov</a><br>
>> <a href="http://www.mcs.anl.gov/~apenya" target="_blank">www.mcs.anl.gov/~apenya</a><br>
>><br>
>> _______________________________________________<br>
>> discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
>> To manage subscription options or unsubscribe:<br>
>> <a href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
>><br>
>><br>
>><br>
>> --<br>
>> Regards<br>
>><br>
>> Ansar Javed<br>
>> HPC Lab<br>
>> SEECS NUST<br>
>> Contact: <a href="tel:%2B92%20334%20438%209394" value="+923344389394">+92 334 438 9394</a><br>
>> Email: <a href="mailto:muhammad.ansar@seecs.edu.pk">muhammad.ansar@seecs.edu.pk</a><br>
>> _______________________________________________<br>
>> discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
>> To manage subscription options or unsubscribe:<br>
>> <a href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
><br>
<br>
_______________________________________________<br>
discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
</div></div></blockquote></div><br><br clear="all"><br>-- <br><div dir="ltr">Regards<br><br>Ansar Javed<br>HPC Lab<br>SEECS NUST <br>Contact: +92 334 438 9394<br>Email: <a href="mailto:muhammad.ansar@seecs.edu.pk" target="_blank">muhammad.ansar@seecs.edu.pk</a><br>
</div>
</div>