[mpich-discuss] Unexpected results
Ken Miller
swonomo at gmail.com
Thu Apr 2 20:52:06 CDT 2015
Antonio,
I can ssh without having to enter a password from:
Pi01 to Pi02
Pi01 to Pi03
Pi01 to Pi04
Also from each Pi back to Pi01
However, I'm asked for a password from Pi04 to Pi03 and Pi02, etc.
Does that matter since I am running the mpiexec command from Pi01 only?
Thanks.
Ken
Sent from my iPad
> On Apr 2, 2015, at 9:29 PM, Antonio J. Peña <apenya at mcs.anl.gov> wrote:
>
>
> Well, I meant to suggest to actually check if that had the expected effects, so rephrasing what I said: can you check if you can ssh without being asked for a password to all nodes from all nodes, using exactly what you have in the host file (i.e.: ip addresses)?
>
>
>> On 04/02/2015 08:20 PM, Ken Miller wrote:
>> Antonio,
>>
>> Thanks for the quick response.
>>
>> I did do the following on each of the other three Pis:
>>
>> On Pi02:
>> ssh-keygen
>> cd .ssh
>> cd id_rsa.pub pi02
>> scp 192.168.2.121:/home/pi/.ssh/p01 .
>> cat pi01>>authorized_keys
>>
>> and then I copied each from the other three Pis to Pi01 and cat those to the authorized_keys on Pi01
>>
>> Is that what you are referring to?
>>
>> If not, do you mind pointing me in the right direction?
>>
>> Thanks in advance.
>>
>> Ken
>>
>>> On Apr 2, 2015, at 9:09 PM, Antonio J. Peña <apenya at mcs.anl.gov> wrote:
>>>
>>>
>>> Hi Ken,
>>>
>>> Please check that you have password-less ssh connectivity between the nodes.
>>>
>>> Best,
>>> Antonio
>>>
>>>
>>>> On 04/02/2015 07:33 PM, Ken Miller wrote:
>>>> Hello,
>>>>
>>>> I am hoping someone might be able to point out what I am doing wrong. I have setup a Raspberry Pi cluster by going through a couple tutorials. The last step is to use MPI to test the cluster. Everything was going great until I ran the following command:
>>>>
>>>> mpiexec -f machinefile -n 4 hostname
>>>>
>>>> My machinefile contains the following:
>>>>
>>>> pi at Pi01 ~/mpi_test $ cat machinefile
>>>> 192.168.2.121
>>>> 192.168.2.122
>>>> 192.168.2.123
>>>> 192.168.2.124
>>>>
>>>> But, when I run the mpiexec command, I get the following error:
>>>>
>>>> pi at Pi01 ~/mpi_test $ mpiexec -f machinefile -n 4 hostname
>>>> Pi01
>>>> [mpiexec at Pi01] control_cb (/home/pi/mpich2/mpich-3.1/src/pm/hydra/pm/pmiserv/pmiserv_cb.c:200): assert (!closed) failed
>>>> [mpiexec at Pi01] HYDT_dmxu_poll_wait_for_event (/home/pi/mpich2/mpich-3.1/src/pm/hydra/tools/demux/demux_poll.c:76): callback returned error status
>>>> [mpiexec at Pi01] HYD_pmci_wait_for_completion (/home/pi/mpich2/mpich-3.1/src/pm/hydra/pm/pmiserv/pmiserv_pmci.c:198): error waiting for event
>>>> [mpiexec at Pi01] main (/home/pi/mpich2/mpich-3.1/src/pm/hydra/ui/mpich/mpiexec.c:336): process manager error waiting for completion
>>>> pi at Pi01 ~/mpi_test $
>>>>
>>>> Clearly, I am missing something. Would appreciate any help in advance.
>>>>
>>>> Thanks.
>>>>
>>>> Ken
>>>>
>>>> ps. I am connected to the other Pis evident by the following ping results. I can also log into each of the other 3 Pis.
>>>>
>>>> ping 192.168.2.122
>>>> PING 192.168.2.122 (192.168.2.122) 56(84) bytes of data.
>>>> 64 bytes from 192.168.2.122: icmp_req=1 ttl=64 time=0.807 ms
>>>> 64 bytes from 192.168.2.122: icmp_req=2 ttl=64 time=0.626 ms
>>>> 64 bytes from 192.168.2.122: icmp_req=3 ttl=64 time=0.614 ms
>>>> 64 bytes from 192.168.2.122: icmp_req=4 ttl=64 time=0.605 ms
>>>> 64 bytes from 192.168.2.122: icmp_req=5 ttl=64 time=0.603 ms
>>>> ^C
>>>> --- 192.168.2.122 ping statistics ---
>>>> 5 packets transmitted, 5 received, 0% packet loss, time 4002ms
>>>> rtt min/avg/max/mdev = 0.603/0.651/0.807/0.078 ms
>>>> pi at Pi01 ~/mpi_test $ ping 192.168.2.123
>>>> PING 192.168.2.123 (192.168.2.123) 56(84) bytes of data.
>>>> 64 bytes from 192.168.2.123: icmp_req=1 ttl=64 time=0.794 ms
>>>> 64 bytes from 192.168.2.123: icmp_req=2 ttl=64 time=0.634 ms
>>>> 64 bytes from 192.168.2.123: icmp_req=3 ttl=64 time=0.628 ms
>>>> 64 bytes from 192.168.2.123: icmp_req=4 ttl=64 time=0.607 ms
>>>> ^C
>>>> --- 192.168.2.123 ping statistics ---
>>>> 4 packets transmitted, 4 received, 0% packet loss, time 3003ms
>>>> rtt min/avg/max/mdev = 0.607/0.665/0.794/0.081 ms
>>>> pi at Pi01 ~/mpi_test $ ping 192.168.2.124
>>>> PING 192.168.2.124 (192.168.2.124) 56(84) bytes of data.
>>>> 64 bytes from 192.168.2.124: icmp_req=1 ttl=64 time=0.787 ms
>>>> 64 bytes from 192.168.2.124: icmp_req=2 ttl=64 time=0.632 ms
>>>> 64 bytes from 192.168.2.124: icmp_req=3 ttl=64 time=0.612 ms
>>>> ^C
>>>> --- 192.168.2.124 ping statistics ---
>>>> 3 packets transmitted, 3 received, 0% packet loss, time 2002ms
>>>> rtt min/avg/max/mdev = 0.612/0.677/0.787/0.078 ms
>>>> pi at Pi01 ~/mpi_test $
>>>>
>>>>
>>>> _______________________________________________
>>>> discuss mailing list discuss at mpich.org
>>>> To manage subscription options or unsubscribe:
>>>> https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>
>>> --
>>> Antonio J. Peña
>>> Postdoctoral Appointee
>>> Mathematics and Computer Science Division
>>> Argonne National Laboratory
>>> 9700 South Cass Avenue, Bldg. 240, Of. 3148
>>> Argonne, IL 60439-4847
>>> apenya at mcs.anl.gov
>>> www.mcs.anl.gov/~apenya
>>> _______________________________________________
>>> discuss mailing list discuss at mpich.org
>>> To manage subscription options or unsubscribe:
>>> https://lists.mpich.org/mailman/listinfo/discuss
>>
>>
>>
>> _______________________________________________
>> discuss mailing list discuss at mpich.org
>> To manage subscription options or unsubscribe:
>> https://lists.mpich.org/mailman/listinfo/discuss
>
>
> --
> Antonio J. Peña
> Postdoctoral Appointee
> Mathematics and Computer Science Division
> Argonne National Laboratory
> 9700 South Cass Avenue, Bldg. 240, Of. 3148
> Argonne, IL 60439-4847
> apenya at mcs.anl.gov
> www.mcs.anl.gov/~apenya
> _______________________________________________
> discuss mailing list discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mpich.org/pipermail/discuss/attachments/20150402/9e625718/attachment.html>
-------------- next part --------------
_______________________________________________
discuss mailing list discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss
More information about the discuss
mailing list