[mpich-discuss] Unexpected results

Ken Miller swonomo at gmail.com
Thu Apr 2 21:16:41 CDT 2015


Yes it works simply as mpiexec -n 4 hostname. Prints out Pi01 four times.

Have the 4 Pis connected to a 5 port switch, going a 24 port switch and then Asus RT66N roter.

MPICH installed is 3.1. Only downloaded and installed the one version.

Thanks.

Ken

Sent from my iPad

> On Apr 2, 2015, at 10:07 PM, Antonio J. Peña <apenya at mcs.anl.gov> wrote:
> 
> 
> That should be fine. Let's start from the basics: does it work without the "-f machinefile" parameter? if it does, it means that you have some connection problem (e.g.: do you have a firewall installed?); otherwise, there's something wrong in the installation (e.g.: multiple MPI implementations installed).
> 
> BTW, which MPICH version are you using?
> 
> 
>> On 04/02/2015 08:52 PM, Ken Miller wrote:
>> Antonio,
>> 
>> I can ssh without having to enter a password from:
>> Pi01 to Pi02
>> Pi01 to Pi03
>> Pi01 to Pi04
>> 
>> Also from each Pi back to Pi01
>> 
>> However, I'm asked for a password from Pi04 to Pi03 and Pi02, etc.
>> 
>> Does that matter since I am running the mpiexec command from Pi01 only?
>> 
>> Thanks.
>> 
>> Ken
>> 
>> Sent from my iPad
>> 
>> On Apr 2, 2015, at 9:29 PM, Antonio J. Peña <apenya at mcs.anl.gov> wrote:
>> 
>>> 
>>> Well, I meant to suggest to actually check if that had the expected effects, so rephrasing what I said: can you check if you can ssh without being asked for a password to all nodes from all nodes, using exactly what you have in the host file (i.e.: ip addresses)?
>>> 
>>> 
>>>> On 04/02/2015 08:20 PM, Ken Miller wrote:
>>>> Antonio,
>>>> 
>>>> Thanks for the quick response.
>>>> 
>>>> I did do the following on each of the other three Pis:
>>>> 
>>>> On Pi02:
>>>> ssh-keygen
>>>> cd .ssh
>>>> cd id_rsa.pub pi02
>>>> scp 192.168.2.121:/home/pi/.ssh/p01 .
>>>> cat pi01>>authorized_keys
>>>> 
>>>> and then I copied each from the other three Pis to Pi01 and cat those to the authorized_keys on Pi01
>>>> 
>>>> Is that what you are referring to?
>>>> 
>>>> If not, do you mind pointing me in the right direction?
>>>> 
>>>> Thanks in advance.
>>>> 
>>>> Ken
>>>> 
>>>>> On Apr 2, 2015, at 9:09 PM, Antonio J. Peña <apenya at mcs.anl.gov> wrote:
>>>>> 
>>>>> 
>>>>> Hi Ken,
>>>>> 
>>>>> Please check that you have password-less ssh connectivity between the nodes.
>>>>> 
>>>>> Best,
>>>>>   Antonio
>>>>> 
>>>>> 
>>>>>> On 04/02/2015 07:33 PM, Ken Miller wrote:
>>>>>> Hello,
>>>>>> 
>>>>>> I am hoping someone might be able to point out what I am doing wrong. I have setup a Raspberry Pi cluster by going through a couple tutorials. The last step is to use MPI to test the cluster. Everything was going great until I ran the following command:
>>>>>> 
>>>>>> mpiexec -f machinefile -n 4 hostname
>>>>>> 
>>>>>> My machinefile contains the following:
>>>>>> 
>>>>>> pi at Pi01 ~/mpi_test $ cat machinefile
>>>>>> 192.168.2.121
>>>>>> 192.168.2.122
>>>>>> 192.168.2.123
>>>>>> 192.168.2.124
>>>>>> 
>>>>>> But, when I run the mpiexec command, I get the following error:
>>>>>> 
>>>>>> pi at Pi01 ~/mpi_test $ mpiexec -f machinefile -n 4 hostname
>>>>>> Pi01
>>>>>> [mpiexec at Pi01] control_cb (/home/pi/mpich2/mpich-3.1/src/pm/hydra/pm/pmiserv/pmiserv_cb.c:200): assert (!closed) failed
>>>>>> [mpiexec at Pi01] HYDT_dmxu_poll_wait_for_event (/home/pi/mpich2/mpich-3.1/src/pm/hydra/tools/demux/demux_poll.c:76): callback returned error status
>>>>>> [mpiexec at Pi01] HYD_pmci_wait_for_completion (/home/pi/mpich2/mpich-3.1/src/pm/hydra/pm/pmiserv/pmiserv_pmci.c:198): error waiting for event
>>>>>> [mpiexec at Pi01] main (/home/pi/mpich2/mpich-3.1/src/pm/hydra/ui/mpich/mpiexec.c:336): process manager error waiting for completion
>>>>>> pi at Pi01 ~/mpi_test $ 
>>>>>> 
>>>>>> Clearly, I am missing something. Would appreciate any help in advance.
>>>>>> 
>>>>>> Thanks.
>>>>>> 
>>>>>> Ken
>>>>>> 
>>>>>> ps. I am connected to the other Pis evident by the following ping results. I can also log into each of the other 3 Pis.
>>>>>> 
>>>>>> ping 192.168.2.122
>>>>>> PING 192.168.2.122 (192.168.2.122) 56(84) bytes of data.
>>>>>> 64 bytes from 192.168.2.122: icmp_req=1 ttl=64 time=0.807 ms
>>>>>> 64 bytes from 192.168.2.122: icmp_req=2 ttl=64 time=0.626 ms
>>>>>> 64 bytes from 192.168.2.122: icmp_req=3 ttl=64 time=0.614 ms
>>>>>> 64 bytes from 192.168.2.122: icmp_req=4 ttl=64 time=0.605 ms
>>>>>> 64 bytes from 192.168.2.122: icmp_req=5 ttl=64 time=0.603 ms
>>>>>> ^C
>>>>>> --- 192.168.2.122 ping statistics ---
>>>>>> 5 packets transmitted, 5 received, 0% packet loss, time 4002ms
>>>>>> rtt min/avg/max/mdev = 0.603/0.651/0.807/0.078 ms
>>>>>> pi at Pi01 ~/mpi_test $ ping 192.168.2.123
>>>>>> PING 192.168.2.123 (192.168.2.123) 56(84) bytes of data.
>>>>>> 64 bytes from 192.168.2.123: icmp_req=1 ttl=64 time=0.794 ms
>>>>>> 64 bytes from 192.168.2.123: icmp_req=2 ttl=64 time=0.634 ms
>>>>>> 64 bytes from 192.168.2.123: icmp_req=3 ttl=64 time=0.628 ms
>>>>>> 64 bytes from 192.168.2.123: icmp_req=4 ttl=64 time=0.607 ms
>>>>>> ^C
>>>>>> --- 192.168.2.123 ping statistics ---
>>>>>> 4 packets transmitted, 4 received, 0% packet loss, time 3003ms
>>>>>> rtt min/avg/max/mdev = 0.607/0.665/0.794/0.081 ms
>>>>>> pi at Pi01 ~/mpi_test $ ping 192.168.2.124
>>>>>> PING 192.168.2.124 (192.168.2.124) 56(84) bytes of data.
>>>>>> 64 bytes from 192.168.2.124: icmp_req=1 ttl=64 time=0.787 ms
>>>>>> 64 bytes from 192.168.2.124: icmp_req=2 ttl=64 time=0.632 ms
>>>>>> 64 bytes from 192.168.2.124: icmp_req=3 ttl=64 time=0.612 ms
>>>>>> ^C
>>>>>> --- 192.168.2.124 ping statistics ---
>>>>>> 3 packets transmitted, 3 received, 0% packet loss, time 2002ms
>>>>>> rtt min/avg/max/mdev = 0.612/0.677/0.787/0.078 ms
>>>>>> pi at Pi01 ~/mpi_test $ 
>>>>>> 
>>>>>> 
>>>>>> _______________________________________________
>>>>>> discuss mailing list     discuss at mpich.org
>>>>>> To manage subscription options or unsubscribe:
>>>>>> https://lists.mpich.org/mailman/listinfo/discuss
>>>>> 
>>>>> 
>>>>> -- 
>>>>> Antonio J. Peña
>>>>> Postdoctoral Appointee
>>>>> Mathematics and Computer Science Division
>>>>> Argonne National Laboratory
>>>>> 9700 South Cass Avenue, Bldg. 240, Of. 3148
>>>>> Argonne, IL 60439-4847
>>>>> apenya at mcs.anl.gov
>>>>> www.mcs.anl.gov/~apenya
>>>>> _______________________________________________
>>>>> discuss mailing list     discuss at mpich.org
>>>>> To manage subscription options or unsubscribe:
>>>>> https://lists.mpich.org/mailman/listinfo/discuss
>>>> 
>>>> 
>>>> 
>>>> _______________________________________________
>>>> discuss mailing list     discuss at mpich.org
>>>> To manage subscription options or unsubscribe:
>>>> https://lists.mpich.org/mailman/listinfo/discuss
>>> 
>>> 
>>> -- 
>>> Antonio J. Peña
>>> Postdoctoral Appointee
>>> Mathematics and Computer Science Division
>>> Argonne National Laboratory
>>> 9700 South Cass Avenue, Bldg. 240, Of. 3148
>>> Argonne, IL 60439-4847
>>> apenya at mcs.anl.gov
>>> www.mcs.anl.gov/~apenya
>>> _______________________________________________
>>> discuss mailing list     discuss at mpich.org
>>> To manage subscription options or unsubscribe:
>>> https://lists.mpich.org/mailman/listinfo/discuss
>> 
>> 
>> _______________________________________________
>> discuss mailing list     discuss at mpich.org
>> To manage subscription options or unsubscribe:
>> https://lists.mpich.org/mailman/listinfo/discuss
> 
> 
> -- 
> Antonio J. Peña
> Postdoctoral Appointee
> Mathematics and Computer Science Division
> Argonne National Laboratory
> 9700 South Cass Avenue, Bldg. 240, Of. 3148
> Argonne, IL 60439-4847
> apenya at mcs.anl.gov
> www.mcs.anl.gov/~apenya
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mpich.org/pipermail/discuss/attachments/20150402/25f8cfbc/attachment.html>
-------------- next part --------------
_______________________________________________
discuss mailing list     discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss


More information about the discuss mailing list