<meta http-equiv="Content-Type" content="text/html; charset=utf-8"><div dir="ltr">Hi Min,<div> I have attached the two Config.log. and here is the code</div><div><br></div><div><div>#include <stdio.h></div><div>#include <mpi.h></div><div><br></div><div>int main (argc, argv)</div><div>     int argc;</div><div>     char *argv[];</div><div>{</div><div><br></div><div>int i=0;</div><div> MPI_Init (&argc, &argv);<span class="gmail-Apple-tab-span" style="white-space:pre"> </span>/* starts MPI */</div><div>// Find out rank, size</div><div>int world_rank;</div><div>MPI_Comm_rank(MPI_COMM_WORLD, &world_rank);</div><div>int world_size;</div><div>MPI_Comm_size(MPI_COMM_WORLD, &world_size);</div><div><br></div><div>int number;</div><div>if (world_rank == 0) {</div><div><br></div><div>    number = -1;</div><div>for( i=1; i < world_size; i++){</div><div><br></div><div>    MPI_Send(&number, 1, MPI_INT, i, 0, MPI_COMM_WORLD);</div><div>}</div><div>} </div><div>else  {</div><div>    MPI_Recv(&number, 1, MPI_INT, 0, 0, MPI_COMM_WORLD,</div><div>             MPI_STATUS_IGNORE);</div><div>    printf("Process %d received number %d from process 0\n",world_rank, number);</div><div>}</div><div>MPI_Finalize();</div><div><br></div><div>  return 0;</div><div>}</div></div><div><br></div><div>Regards,</div><div>Doha</div></div><div class="gmail_extra"><br><div class="gmail_quote">On Sun, Jan 22, 2017 at 10:47 PM, Min Si <span dir="ltr"><<a href="mailto:msi@anl.gov" target="_blank">msi@anl.gov</a>></span> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
  
    
  
  <div bgcolor="#FFFFFF" text="#000000">
    Hi Doha,<br>
    <br>
    Can you please send us the config.log file for each MPICH build and
    your helloworld source doe ? The config.log file should be under
    your MPICH build directory where you executed ./configure.<span class="HOEnZb"><font color="#888888"><br>
    <br>
    Min</font></span><div><div class="h5"><br>
    <div class="m_4759687804894361769moz-cite-prefix">On 1/21/17 4:53 AM, Doha Ehab wrote:<br>
    </div>
    <blockquote type="cite">
      
      <div dir="ltr">I have tried what you mentioned in the previous
        E-mail. 
        <div><br>
          <div>1- I have build MPICH for CPU node and ARM node.</div>
          <div>2- Uploaded the <span style="font-size:12.8px">binaries</span> on
            same path on the 2 nodes.</div>
          <div>3- Compiled helloWorld (it sends a number from process
            zero to all other processes ) for both nodes. Then tried <span style="font-size:12.8px">mpiexec -np 2 -f <hostfile
              with mic hostnames>./helloworld</span> </div>
          <div><br>
          </div>
          <div>I got this error</div>
          <div> Fatal error in MPI_Recv: Other MPI error, error stack:</div>
          <div>MPI_Recv(200).................<wbr>...............:
            MPI_Recv(buf=0xbe9460d0, count=1, MPI_INT, src=0, tag=0,
            MPI_COMM_WORLD, status=0x1) failed</div>
          <div>MPIDI_CH3i_Progress_wait(242).<wbr>...............: an error
            occurred while handling an event returned by
            MPIDU_Sock_Wait()</div>
          <div>MPIDI_CH3I_Progress_handle_<wbr>sock_event(554)...: </div>
          <div>MPIDI_CH3_Sockconn_handle_<wbr>connopen_event(899): unable to
            find the process group structure with id <></div>
          <div><br>
          </div>
          <div>Regards,</div>
          <div>Doha</div>
          <div><br>
          </div>
        </div>
      </div>
      <div class="gmail_extra"><br>
        <div class="gmail_quote">On Wed, Nov 16, 2016 at 6:38 PM, Min Si
          <span dir="ltr"><<a href="mailto:msi@anl.gov" target="_blank">msi@anl.gov</a>></span>
          wrote:<br>
          <blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">I guess
            you might need to put all the MPICH binaries (e.g.,
            hydra_pmi_proxy) to the same path on each node. I have
            executed MPICH on Intel MIC chips from the host CPU node
            where OS are different. The thing I did was:<br>
            1. build MPICH for both CPU node and MIC on the CPU node
            (you have done this step).<br>
            2. upload the MIC binaries to the same path on MIC chip as
            on the CPU node<br>
               For example:<br>
               - on CPU node : /tmp/mpich/install/bin holds the CPU
            version<br>
               - on MIC :          /tmp/mpich/install/bin holds the MIC
            version<br>
            3. compile helloworld.c with the MIC version mpicc<br>
            4. execute on CPU node: mpiexe -np 2 -f <hostfile with
            mic hostnames>./helloworld<br>
            <br>
            I think you should be able to follow step 2, but since your
            helloworld binary is also built with different OS, you might
            want to put it also into the same path on two nodes similar
            as we do for MPICH binaries.<span class="m_4759687804894361769HOEnZb"><font color="#888888"><br>
                <br>
                Min</font></span>
            <div class="m_4759687804894361769HOEnZb">
              <div class="m_4759687804894361769h5"><br>
                <br>
                On 11/16/16 8:29 AM, Kenneth Raffenetti wrote:<br>
                <blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
                  Have you disabled any and all firewalls on both nodes?
                  It sounds like they are unable to communicate in
                  initialization.<br>
                  <br>
                  Ken<br>
                  <br>
                  On 11/16/2016 07:34 AM, Doha Ehab wrote:<br>
                  <blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">
                    Yes, I built MPICH-3 on both systems and I tried the
                    code on each node<br>
                    separately and it worked, I tried each node with
                    other nodes that has<br>
                    the same operating system and it worked as well.<br>
                    When I try the code on the 2 nodes that have
                    different operating systems<br>
                    no result or error message appear.<br>
                    <br>
                    Regards<br>
                    Doha<br>
                    <br>
                    On Mon, Nov 14, 2016 at 6:25 PM, Kenneth Raffenetti<br>
                    <<a href="mailto:raffenet@mcs.anl.gov" target="_blank">raffenet@mcs.anl.gov</a>
                    <mailto:<a href="mailto:raffenet@mcs.anl.gov" target="_blank">raffenet@mcs.anl.gov</a>>>
                    wrote:<br>
                    <br>
                        It may be possible to run in such a setup, but
                    it would not be<br>
                        recommended. Did you build MPICH on both systems
                    you are trying to<br>
                        run on? What exactly happened when the code
                    didn't work?<br>
                    <br>
                        Ken<br>
                    <br>
                    <br>
                        On 11/13/2016 12:36 AM, Doha Ehab wrote:<br>
                    <br>
                            Hello,<br>
                             I tried to run a parallel (Hello World) C
                    code on a cluster<br>
                            that has 2<br>
                            nodes, the nodes have different operating
                    system so the code did not<br>
                            work and no results were printed.<br>
                             How to make such a cluster work? is there
                    is extra steps that<br>
                            should be<br>
                            done?<br>
                    <br>
                            Regards,<br>
                            Doha<br>
                    <br>
                    <br>
                            ______________________________<wbr>_________________<br>
                            discuss mailing list     <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
                            <mailto:<a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a>><br>
                            To manage subscription options or
                    unsubscribe:<br>
                            <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailma<wbr>n/listinfo/discuss</a><br>
                            <<a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailm<wbr>an/listinfo/discuss</a>><br>
                    <br>
                        ______________________________<wbr>_________________<br>
                        discuss mailing list     <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a>
                    <mailto:<a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a>><br>
                        To manage subscription options or unsubscribe:<br>
                        <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailma<wbr>n/listinfo/discuss</a><br>
                        <<a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailm<wbr>an/listinfo/discuss</a>><br>
                    <br>
                    <br>
                    <br>
                    <br>
                    ______________________________<wbr>_________________<br>
                    discuss mailing list     <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
                    To manage subscription options or unsubscribe:<br>
                    <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailma<wbr>n/listinfo/discuss</a><br>
                    <br>
                  </blockquote>
                  ______________________________<wbr>_________________<br>
                  discuss mailing list     <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
                  To manage subscription options or unsubscribe:<br>
                  <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailma<wbr>n/listinfo/discuss</a><br>
                </blockquote>
                <br>
                ______________________________<wbr>_________________<br>
                discuss mailing list     <a href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a><br>
                To manage subscription options or unsubscribe:<br>
                <a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/mailma<wbr>n/listinfo/discuss</a><br>
              </div>
            </div>
          </blockquote>
        </div>
        <br>
      </div>
      <br>
      <fieldset class="m_4759687804894361769mimeAttachmentHeader"></fieldset>
      <br>
      <pre>______________________________<wbr>_________________
discuss mailing list     <a class="m_4759687804894361769moz-txt-link-abbreviated" href="mailto:discuss@mpich.org" target="_blank">discuss@mpich.org</a>
To manage subscription options or unsubscribe:
<a class="m_4759687804894361769moz-txt-link-freetext" href="https://lists.mpich.org/mailman/listinfo/discuss" target="_blank">https://lists.mpich.org/<wbr>mailman/listinfo/discuss</a></pre>
    </blockquote>
    <br>
  </div></div></div>

<br>______________________________<wbr>_________________<br>
discuss mailing list     <a href="mailto:discuss@mpich.org">discuss@mpich.org</a><br>
To manage subscription options or unsubscribe:<br>
<a href="https://lists.mpich.org/mailman/listinfo/discuss" rel="noreferrer" target="_blank">https://lists.mpich.org/<wbr>mailman/listinfo/discuss</a><br></blockquote></div><br></div>