[mpich-discuss] success and failure report for mpich-3.0.2
Pavan Balaji
balaji at mcs.anl.gov
Mon Apr 29 21:14:18 CDT 2013
Adding discuss at mpich.org back to cc.
> [proxy:0:1 at linpc1] HYDU_create_process
> (../../../../mpich-3.0.2/src/pm/hydra/utils/launch/launch.c:74):
> execvp error on file date (No such file or directory)
This seems to indicate that "date" is not found on linpc1. Try giving
the full path to "date".
-- Pavan
On 04/29/2013 02:40 AM US Central Time, Siegmar Gross wrote:
> Hi
>
>> This is not an architectural problem. You need to have mpich installed
>> on all machines. It looks like some of your machines don't have it
>> installed.
>
> It is installed on all machines, but sometimes I have a problem if
> I use different operating systems on the same hardware architecture.
> I use "date" to exclude any errors in my program. I get an error,
> if I try to start my processes from a Solaris machine and everything
> works well, if I do the same thing from my Linux machine, so that
> it is an architectural problem in my opinion. I have unsubscribed
> from the list in the meantime, because I didn't get any response.
> Perhaps you can reply to my address directly, if you want to send
> a response. I get the following results.
>
>
> sunpc1 fd1026 105 which mpiexec
> /usr/local/mpich-3.0.2_64_cc/bin/mpiexec
> sunpc1 fd1026 106 ssh linpc1 which mpiexec
> /usr/local/mpich-3.0.2_64_cc/bin/mpiexec
> sunpc1 fd1026 107 mpiexec -np 2 -host sunpc0,sunpc1 date
> Mon Apr 29 09:26:34 CEST 2013
> Mon Apr 29 09:26:35 CEST 2013
> sunpc1 fd1026 108 mpiexec -np 2 -host linpc0,linpc1 date
> [proxy:0:1 at linpc1] HYDU_create_process (../../../../mpich-3.0.2/src/pm/hydra/utils/launch/launch.c:74): execvp
> error on file date (No such file or directory)
>
> ===================================================================================
> = BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> = EXIT CODE: 255
> = CLEANING UP REMAINING PROCESSES
> = YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> ===================================================================================
> [proxy:0:0 at linpc0] HYDU_create_process (../../../../mpich-3.0.2/src/pm/hydra/utils/launch/launch.c:74): execvp
> error on file date (No such file or directory)
>
> ===================================================================================
> = BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> = EXIT CODE: 255
> = CLEANING UP REMAINING PROCESSES
> = YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> ===================================================================================
> sunpc1 fd1026 109 ssh linpc1
> linpc1 fd1026 101 mpiexec -np 2 -host linpc0,linpc1 date
> Mon Apr 29 09:27:40 CEST 2013
> Mon Apr 29 09:27:41 CEST 2013
> linpc1 fd1026 102 mpiexec -np 2 -host sunpc0,sunpc1 date
> Mon Apr 29 09:27:49 CEST 2013
> Mon Apr 29 09:27:49 CEST 2013
> linpc1 fd1026 103 mpiexec -np 2 -host sunpc0,linpc0 date
> Mon Apr 29 09:30:55 CEST 2013
> Mon Apr 29 09:30:56 CEST 2013
> linpc1 fd1026 104 linpc1 fd1026 103
>
>
> As you can see below, I wasn't able to build the packacke on some
> machines with GNU C or in 32-bit versions. Thank you very much
> for any answer in advance.
>
> Kind regards
>
> Siegmar
>
>>
>> -- Pavan
>>
>> On 03/22/2013 10:22 AM US Central Time, Siegmar Gross wrote:
>>> Hi
>>>
>>> today I tried to install mpich-3.0.2. I could compile and install the
>>> package for the following architectures.
>>>
>>> openSuSE Linux 12.1, x86_64, Sun C 5.12, 64-bit
>>> Solaris 10, x86_64, Sun C 5.12, 64-bit
>>> Solaris 10, sparc, Sun C 5.12, 32-bit
>>> Solaris 10, sparc, Sun C 5.12, 64-bit
>>>
>>> As described in the RELEASE_NOTES a communication between little-endian
>>> and big-endian machines is not possible (broken pipe). Unfortunately
>>> you don't even allow different operating systems (it seems that you
>>> don't evaluate PATH to find a program).
>>>
>>> sunpc1 fd1026 103 mpiexec -np 2 -host sunpc0,sunpc1 init_finalize
>>> Hello!
>>> Hello!
>>>
>>> sunpc1 fd1026 104 mpiexec -np 2 -host sunpc0,linpc1 init_finalize
>>> [proxy:0:1 at linpc1] HYDU_create_process
>>> (../../../../mpich-3.0.2/src/pm/hydra/utils/launch/launch.c:74):
>>> execvp error on file init_finalize (No such file or directory)
>>>
>>> ========================================================================
>>> = BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>>> = EXIT CODE: 255
>>> = CLEANING UP REMAINING PROCESSES
>>> = YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>>> ========================================================================
>>> [proxy:0:0 at sunpc0] HYD_pmcd_pmip_control_cmd_cb
>>> (../../../../mpich-3.0.2/src/pm/hydra/pm/pmiserv/pmip_cb.c:886):
>>> assert (!closed) failed
>>> [proxy:0:0 at sunpc0] HYDT_dmxu_poll_wait_for_event
>>> (../../../../mpich-3.0.2/src/pm/hydra/tools/demux/demux_poll.c:77):
>>> callback returned error status
>>> [proxy:0:0 at sunpc0] main
>>> (../../../../mpich-3.0.2/src/pm/hydra/pm/pmiserv/pmip.c:206):
>>> demux engine error waiting for event
>>>
>>>
>>>
>>> The programs are available on both machines.
>>>
>>> sunpc1 fd1026 105 which init_finalize
>>> /home/fd1026/SunOS/x86_64/bin/init_finalize
>>> sunpc1 fd1026 106 ssh linpc1 which init_finalize
>>> /home/fd1026/Linux/x86_64/bin/init_finalize
>>> sunpc1 fd1026 107
>>>
>>>
>>>
>>>
>>> I wasn't able to build the package on the following architectures.
>>>
>>>
>>> openSuSE Linux 12.1, x86_64, Sun C 5.12, 32-bit
>>> openSuSE Linux 12.1, x86_64, GNU C 4.7.1, 32-bit
>>> ------------------------------------------------
>>>
>>> tyr mpich-3.0.2-Linux.x86_64.32_cc 158 tail log.make.Linux.x86_64.32_cc
>>> CC mplstr.lo
>>> CC mpltrmem.lo
>>> CC mplenv.lo
>>> CCLD libmpl.la
>>> /usr/lib/libxml2.so: could not read symbols: File in wrong format
>>> make[2]: *** [libmpl.la] Error 2
>>> make[2]: Leaving directory `.../src/mpl'
>>> make[1]: *** [all-recursive] Error 1
>>>
>>>
>>>
>>> openSuSE Linux 12.1, x86_64, GNU C 4.7.1, 64-bit
>>> ------------------------------------------------
>>>
>>> tyr mpich-3.0.2-Linux.x86_64.64_gcc 171 tail log.make.Linux.x86_64.64_gcc
>>> CC mpltrmem.lo
>>> CC mplenv.lo
>>> CCLD libmpl.la
>>> /usr/local/bin/ld: libmpl.so.1: No such file: No such file or directory
>>> collect2: error: ld returned 1 exit status
>>> make[2]: *** [libmpl.la] Error 1
>>> make[2]: Leaving directory `.../src/mpl'
>>>
>>>
>>>
>>> Solaris 10, x86_64, Sun C 5.12, 32-bit
>>> --------------------------------------
>>>
>>> hangs
>>>
>>>
>>>
>>> Solaris 10, x86_64, GNU C 4.7.1, 32-bit
>>> ---------------------------------------
>>>
>>> tyr mpich-3.0.2-SunOS.x86_64.32_gcc 193 tail -12 log.make.SunOS.x86_64.32_gcc
>>> CC src/mpi/coll/barrier_group.lo
>>> CC src/mpi/coll/helper_fns.lo
>>> Assembler: helper_fns.c
>>> "/var/tmp//ccIdXgrh.s", line 912 : Syntax error
>>> Near line: "mov $PREFETCHBLOCK/16, %eax"
>>> "/var/tmp//ccIdXgrh.s", line 920 : Syntax error
>>> Near line: "mov $PREFETCHBLOCK/8, %eax"
>>> make[2]: *** [src/mpi/coll/helper_fns.lo] Error 1
>>>
>>>
>>>
>>> Solaris 10, x86_64, GNU C 4.7.1, 64-bit
>>> Solaris 10, sparc, GNU C 4.7.1, 32-bit
>>> Solaris 10, sparc, GNU C 4.7.1, 64-bit
>>> ---------------------------------------
>>>
>>> tyr mpich-3.0.2-SunOS.x86_64.64_gcc 195 tail log.make.SunOS.x86_64.64_gcc
>>> CXX src/binding/cxx/initcxx.lo
>>> CXXLD lib/libmpichcxx.la
>>> ld: fatal: file libmpichcxx.so.10: open failed: No such file or directory
>>> ld: fatal: file processing errors. No output written to
>>> lib/.libs/libmpichcxx.so.10.0.2
>>> collect2: error: ld returned 1 exit status
>>> make[2]: *** [lib/libmpichcxx.la] Error 1
>>>
>>>
>>> Kind regards
>>>
>>> Siegmar
>>>
>>> _______________________________________________
>>> discuss mailing list discuss at mpich.org
>>> To manage subscription options or unsubscribe:
>>> https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>
>> --
>> Pavan Balaji
>> http://www.mcs.anl.gov/~balaji
>
--
Pavan Balaji
http://www.mcs.anl.gov/~balaji
More information about the discuss
mailing list