<meta http-equiv="Content-Type" content="text/html; charset=utf-8"><div dir="ltr">changing the hostname and adding it along with IP address in /etc/hosts made it work.<div><br></div><div>Thanks </div><div>Doha</div></div><div class="gmail_extra"><br><div class="gmail_quote">On Mon, Jul 25, 2016 at 8:41 PM, Balaji, Pavan <span dir="ltr"><<a href="mailto:balaji@anl.gov" target="_blank">balaji@anl.gov</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><br>
This sometimes happens if the node cannot resolve its hostname itself. Can you try adding your hostname and IP to /etc/hosts?<br>
<span class="HOEnZb"><font color="#888888"><br>
- Pavan<br>
</font></span><div class="HOEnZb"><div class="h5"><br>
> On Jul 25, 2016, at 1:36 PM, Kenneth Raffenetti <<a href="mailto:raffenet@mcs.anl.gov">raffenet@mcs.anl.gov</a>> wrote:<br>
><br>
> Hi,<br>
><br>
> Normally this type of error would be the result of a firewall blocking communication, but your run is on a single node so that shouldn't be the case. I wonder, since it did work at one point, if there is some bad state on your system that might be cleared with a reboot?<br>
><br>
> It also looks like you are running on a Android system, which we do not have much experience with, and no good way to test ourselves so our ways to help may be limited.<br>
><br>
> Ken<br>
><br>
> On 07/24/2016 11:50 AM, Doha Ehab wrote:<br>
>> Hello<br>
>> I am using a cross compiled version of MPICH3, I was trying simple<br>
>> containing MPI_Send and MPI_Recv, it was working but suddenly I keep<br>
>> receiving this error messsage: can anyone point out what is wrong and<br>
>> how to fix it.<br>
>><br>
>> $ mpiexec -v -n 4 /data/parallelCode<br>
>> host: tab<br>
>><br>
>> ==================================================================================================<br>
>> mpiexec options:<br>
>> ----------------<br>
>> Base path: /system/xbin/<br>
>> Launcher: (null)<br>
>> Debug level: 1<br>
>> Enable X: -1<br>
>><br>
>> Global environment:<br>
>> -------------------<br>
>> _=/system/xbin/mpiexec<br>
>> PATH=/sbin:/vendor/bin:/system/sbin:/system/bin:/system/xbin<br>
>> LOOP_MOUNTPOINT=/mnt/obb<br>
>> ANDROID_ROOT=/system<br>
>> SHELL=/system/bin/sh<br>
>> ANDROID_DATA=/data<br>
>> ANDROID_ASSETS=/system/app<br>
>> TERM=vt100<br>
>> ANDROID_PROPERTY_WORKSPACE=8,0<br>
>> ANDROID_BOOTLOGO=1<br>
>> HOSTNAME=hwt1701<br>
>> LD_LIBRARY_PATH=/vendor/lib:/system/lib<br>
>><br>
>> BOOTCLASSPATH=/system/framework/core.jar:/system/framework/conscrypt.jar:/system/framework/okhttp.jar:/system/framework/core-junit.jar:/system/framework/bouncycastle.jar:/system/framework/ext.jar:/system/framework/framework.jar:/system/framework/framework2.jar:/system/framework/hwframework.jar:/system/framework/hwcustframework.jar:/system/framework/telephony-common.jar:/system/framework/voip-common.jar:/system/framework/mms-common.jar:/system/framework/android.policy.jar:/system/framework/services.jar:/system/framework/apache-xml.jar:/system/framework/webviewchromium.jar:/system/framework/hwEmui.jar:/system/framework/hwServices.jar:/system/framework/hwAndroid.policy.jar:/system/framework/hwTelephony-common.jar:/system/framework/hwpadext.jar<br>
>> EMULATED_STORAGE_SOURCE=/mnt/shell/emulated<br>
>> ANDROID_SOCKET_adbd=10<br>
>> EMULATED_STORAGE_TARGET=/storage/emulated<br>
>> ANDROID_STORAGE=/storage<br>
>> MKSH=/system/bin/sh<br>
>> EXTERNAL_STORAGE=/storage/emulated/legacy<br>
>> USBHOST_STORAGE=/storage/usbdisk<br>
>> RANDOM=11338<br>
>> ASEC_MOUNTPOINT=/mnt/asec<br>
>> SECONDARY_STORAGE=/storage/sdcard1<br>
>> USER=shell<br>
>> LEGACY_STORAGE=/storage/emulated/legacy<br>
>> HOME=/data<br>
>><br>
>> Hydra internal environment:<br>
>> ---------------------------<br>
>> GFORTRAN_UNBUFFERED_PRECONNECTED=y<br>
>><br>
>><br>
>> Proxy information:<br>
>> *********************<br>
>> [1] proxy: tab (1 cores)<br>
>> Exec list: /data/mmp100 (4 processes);<br>
>><br>
>><br>
>> ==================================================================================================<br>
>><br>
>> [mpiexec@tab] Timeout set to -1 (-1 means infinite)<br>
>> [mpiexec@tab] Got a control port string of tab:48661<br>
>><br>
>> Proxy launch args: /system/xbin/hydra_pmi_proxy --control-port tab:48661<br>
>> --debug --rmk user --launcher ssh --demux poll --pgid 0 --retries 10<br>
>> --usize -2 --proxy-id<br>
>><br>
>> Arguments being passed to proxy 0:<br>
>> --version 3.2 --iface-ip-env-name MPIR_CVAR_CH3_INTERFACE_HOSTNAME<br>
>> --hostname tab --global-core-map 0,1,1 --pmi-id-map 0,0<br>
>> --global-process-count 4 --auto-cleanup 1 --pmi-kvsname kvs_10003_0<br>
>> --pmi-process-mapping (vector,(0,1,1)) --ckpoint-num -1<br>
>> --global-inherited-env 26 '_=/system/xbin/mpiexec'<br>
>> 'PATH=/sbin:/vendor/bin:/system/sbin:/system/bin:/system/xbin'<br>
>> 'LOOP_MOUNTPOINT=/mnt/obb' 'ANDROID_ROOT=/system' 'SHELL=/system/bin/sh'<br>
>> 'ANDROID_DATA=/data' 'ANDROID_ASSETS=/system/app' 'TERM=vt100'<br>
>> 'ANDROID_PROPERTY_WORKSPACE=8,0' 'ANDROID_BOOTLOGO=1' 'HOSTNAME=hwt1701'<br>
>> 'LD_LIBRARY_PATH=/vendor/lib:/system/lib'<br>
>> 'BOOTCLASSPATH=/system/framework/core.jar:/system/framework/conscrypt.jar:/system/framework/okhttp.jar:/system/framework/core-junit.jar:/system/framework/bouncycastle.jar:/system/framework/ext.jar:/system/framework/framework.jar:/system/framework/framework2.jar:/system/framework/hwframework.jar:/system/framework/hwcustframework.jar:/system/framework/telephony-common.jar:/system/framework/voip-common.jar:/system/framework/mms-common.jar:/system/framework/android.policy.jar:/system/framework/services.jar:/system/framework/apache-xml.jar:/system/framework/webviewchromium.jar:/system/framework/hwEmui.jar:/system/framework/hwServices.jar:/system/framework/hwAndroid.policy.jar:/system/framework/hwTelephony-common.jar:/system/framework/hwpadext.jar'<br>
>> 'EMULATED_STORAGE_SOURCE=/mnt/shell/emulated' 'ANDROID_SOCKET_adbd=10'<br>
>> 'EMULATED_STORAGE_TARGET=/storage/emulated' 'ANDROID_STORAGE=/storage'<br>
>> 'MKSH=/system/bin/sh' 'EXTERNAL_STORAGE=/storage/emulated/legacy'<br>
>> 'USBHOST_STORAGE=/storage/usbdisk' 'RANDOM=11338'<br>
>> 'ASEC_MOUNTPOINT=/mnt/asec' 'SECONDARY_STORAGE=/storage/sdcard1'<br>
>> 'USER=shell' 'LEGACY_STORAGE=/storage/emulated/legacy' 'HOME=/data'<br>
>> --global-user-env 0 --global-system-env 1<br>
>> 'GFORTRAN_UNBUFFERED_PRECONNECTED=y' --proxy-core-count 1 --exec<br>
>> --exec-appnum 0 --exec-proc-count 4 --exec-local-env 0 --exec-wdir /<br>
>> --exec-args 1 /data/mmp100<br>
>><br>
>> [mpiexec@tab] Launch arguments: /system/xbin/hydra_pmi_proxy<br>
>> --control-port tab:48661 --debug --rmk user --launcher ssh --demux poll<br>
>> --pgid 0 --retries 10 --usize -2 --proxy-id 0<br>
>> [proxy:0:0@tab] got pmi command (from 0): init<br>
>> pmi_version=1 pmi_subversion=1<br>
>> [proxy:0:0@tab] PMI response: cmd=response_to_init pmi_version=1<br>
>> pmi_subversion=1 rc=0<br>
>> [proxy:0:0@tab] got pmi command (from 0): get_maxes<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=maxes kvsname_max=256 keylen_max=64<br>
>> vallen_max=1024<br>
>> [proxy:0:0@tab] got pmi command (from 6): init<br>
>> pmi_version=1 pmi_subversion=1<br>
>> [proxy:0:0@tab] PMI response: cmd=response_to_init pmi_version=1<br>
>> pmi_subversion=1 rc=0<br>
>> [proxy:0:0@tab] got pmi command (from 6): get_maxes<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=maxes kvsname_max=256 keylen_max=64<br>
>> vallen_max=1024<br>
>> [proxy:0:0@tab] got pmi command (from 9): init<br>
>> pmi_version=1 pmi_subversion=1<br>
>> [proxy:0:0@tab] PMI response: cmd=response_to_init pmi_version=1<br>
>> pmi_subversion=1 rc=0<br>
>> [proxy:0:0@tab] got pmi command (from 15): init<br>
>> pmi_version=1 pmi_subversion=1<br>
>> [proxy:0:0@tab] PMI response: cmd=response_to_init pmi_version=1<br>
>> pmi_subversion=1 rc=0<br>
>> [proxy:0:0@tab] got pmi command (from 0): get_appnum<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=appnum appnum=0<br>
>> [proxy:0:0@tab] got pmi command (from 9): get_maxes<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=maxes kvsname_max=256 keylen_max=64<br>
>> vallen_max=1024<br>
>> [proxy:0:0@tab] got pmi command (from 0): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 15): get_maxes<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=maxes kvsname_max=256 keylen_max=64<br>
>> vallen_max=1024<br>
>> [proxy:0:0@tab] got pmi command (from 0): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 9): get_appnum<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=appnum appnum=0<br>
>> [proxy:0:0@tab] got pmi command (from 0): get<br>
>> kvsname=kvs_10003_0 key=PMI_process_mapping<br>
>> [proxy:0:0@tab] PMI response: cmd=get_result rc=0 msg=success<br>
>> value=(vector,(0,1,1))<br>
>> [proxy:0:0@tab] got pmi command (from 15): get_appnum<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=appnum appnum=0<br>
>> [proxy:0:0@tab] got pmi command (from 6): get_appnum<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=appnum appnum=0<br>
>> [proxy:0:0@tab] got pmi command (from 9): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 15): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 6): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 6): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 9): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 0): put<br>
>> kvsname=kvs_10003_0 key=P0-businesscard<br>
>> value=port#49751$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] cached command:<br>
>> P0-businesscard=port#49751$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] PMI response: cmd=put_result rc=0 msg=success<br>
>> [proxy:0:0@tab] got pmi command (from 9): get<br>
>> kvsname=kvs_10003_0 key=PMI_process_mapping<br>
>> [proxy:0:0@tab] PMI response: cmd=get_result rc=0 msg=success<br>
>> value=(vector,(0,1,1))<br>
>> [proxy:0:0@tab] got pmi command (from 0): barrier_in<br>
>><br>
>> [proxy:0:0@tab] got pmi command (from 6): get<br>
>> kvsname=kvs_10003_0 key=PMI_process_mapping<br>
>> [proxy:0:0@tab] PMI response: cmd=get_result rc=0 msg=success<br>
>> value=(vector,(0,1,1))<br>
>> [proxy:0:0@tab] got pmi command (from 15): get_my_kvsname<br>
>><br>
>> [proxy:0:0@tab] PMI response: cmd=my_kvsname kvsname=kvs_10003_0<br>
>> [proxy:0:0@tab] got pmi command (from 9): put<br>
>> kvsname=kvs_10003_0 key=P2-businesscard<br>
>> value=port#60729$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] cached command:<br>
>> P2-businesscard=port#60729$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] PMI response: cmd=put_result rc=0 msg=success<br>
>> [proxy:0:0@tab] got pmi command (from 15): get<br>
>> kvsname=kvs_10003_0 key=PMI_process_mapping<br>
>> [proxy:0:0@tab] PMI response: cmd=get_result rc=0 msg=success<br>
>> value=(vector,(0,1,1))<br>
>> [proxy:0:0@tab] got pmi command (from 9): barrier_in<br>
>><br>
>> [proxy:0:0@tab] got pmi command (from 6): put<br>
>> kvsname=kvs_10003_0 key=P1-businesscard<br>
>> value=port#44344$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] cached command:<br>
>> P1-businesscard=port#44344$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] PMI response: cmd=put_result rc=0 msg=success<br>
>> [proxy:0:0@tab] got pmi command (from 15): put<br>
>> kvsname=kvs_10003_0 key=P3-businesscard<br>
>> value=port#51326$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] cached command:<br>
>> P3-businesscard=port#51326$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] PMI response: cmd=put_result rc=0 msg=success<br>
>> [proxy:0:0@tab] got pmi command (from 6): barrier_in<br>
>><br>
>> [proxy:0:0@tab] got pmi command (from 15): barrier_in<br>
>><br>
>> [proxy:0:0@tab] flushing 4 put command(s) out<br>
>> [mpiexec@tab] [pgid: 0] got PMI command: cmd=put<br>
>> P0-businesscard=port#49751$description#tab$ifname#192.168.1.4$<br>
>> P2-businesscard=port#60729$description#tab$ifname#192.168.1.4$<br>
>> P1-businesscard=port#44344$description#tab$ifname#192.168.1.4$<br>
>> P3-businesscard=port#51326$description#tab$ifname#192.168.1.4$<br>
>> [proxy:0:0@tab] forwarding command (cmd=put<br>
>> P0-businesscard=port#49751$description#tab$ifname#192.168.1.4$<br>
>> P2-businesscard=port#60729$description#tab$ifname#192.168.1.4$<br>
>> P1-businesscard=port#44344$description#tab$ifname#192.168.1.4$<br>
>> P3-businesscard=port#51326$description#tab$ifname#192.168.1.4$) upstream<br>
>> [proxy:0:0@tab] forwarding command (cmd=barrier_in) upstream<br>
>> [mpiexec@tab] [pgid: 0] got PMI command: cmd=barrier_in<br>
>> [mpiexec@tab] PMI response to fd 6 pid 15: cmd=keyval_cache<br>
>> P0-businesscard=port#49751$description#tab$ifname#192.168.1.4$<br>
>> P2-businesscard=port#60729$description#tab$ifname#192.168.1.4$<br>
>> P1-businesscard=port#44344$description#tab$ifname#192.168.1.4$<br>
>> P3-businesscard=port#51326$description#tab$ifname#192.168.1.4$<br>
>> [mpiexec@tab] PMI response to fd 6 pid 15: cmd=barrier_out<br>
>> [proxy:0:0@tab] PMI response: cmd=barrier_out<br>
>> [proxy:0:0@tab] PMI response: cmd=barrier_out<br>
>> [proxy:0:0@tab] PMI response: cmd=barrier_out<br>
>> [proxy:0:0@tab] PMI response: cmd=barrier_out<br>
>> [proxy:0:0@tab] got pmi command (from 0): get<br>
>> kvsname=kvs_10003_0 key=P1-businesscard<br>
>> [proxy:0:0@tab] PMI response: cmd=get_result rc=0 msg=success<br>
>> value=port#44344$description#tab$ifname#192.168.1.4$<br>
>> Fatal error in MPI_Send: Unknown error class, error stack:<br>
>> MPI_Send(174)...............................: MPI_Send(buf=0x15c56c,<br>
>> count=1, MPI_INT, dest=1, tag=1, MPI_COMM_WORLD) failed<br>
>> MPIDI_CH3i_Progress_wait(242)...............: an error occurred while<br>
>> handling an event returned by MPIDU_Sock_Wait()<br>
>> MPIDI_CH3I_Progress_handle_sock_event(697)..:<br>
>> MPIDI_CH3_Sockconn_handle_connect_event(597): [ch3:sock] failed to<br>
>> connnect to remote process<br>
>> MPIDU_Socki_handle_connect(808).............: connection failure<br>
>> (set=0,sock=1,errno=113:No route to host)<br>
>> [proxy:0:0@tab] got pmi command (from 0): abort<br>
>> exitcode=69331543<br>
>> [proxy:0:0@tab] we don't understand this command abort; forwarding upstream<br>
>> [mpiexec@tab] [pgid: 0] got PMI command: cmd=abort exitcode=69331543<br>
>><br>
>><br>
>><br>
>><br>
>> _______________________________________________<br>
>> discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
>> To manage subscription options or unsubscribe:<br>
>> <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
>><br>
> _______________________________________________<br>
> discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
> To manage subscription options or unsubscribe:<br>
> <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
<br>
_______________________________________________<br>
discuss mailing list <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailman/listinfo/discuss</a><br>
</div></div></blockquote></div><br></div>