[mpich-discuss] cpi

Kenneth Raffenetti raffenet at mcs.anl.gov
Wed Mar 23 10:42:38 CDT 2016


cpi.c is a source file, not an executable.

On 03/23/2016 10:41 AM, יוסף אלון wrote:
> when i do this i reciev:
>
> cluster at elec-cluster-1 ~ $ *_mpiexec -n 5 -f machinefile
> ./mpich2/mpich-3.1/examples/cpi.c_*
> [proxy:0:0 at elec-cluster-1] HYDU_create_process
> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
> execvp error on file ./mpich2/mpich-3.1/examples/cpi.c (Permission denied)
>
> ===================================================================================
> =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> =   PID 2895 RUNNING AT 147.161.4.200
> =   EXIT CODE: 255
> =   CLEANING UP REMAINING PROCESSES
> =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> ===================================================================================
> [proxy:0:1 at elec-cluster-2] HYDU_create_process
> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
> execvp error on file ./mpich2/mpich-3.1/examples/cpi.c (Permission denied)
> [proxy:0:4 at elec-cluster-5] HYDU_create_process
> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
> ===================================================================================
> =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
> =   PID 7382 RUNNING AT 147.161.4.201
> =   EXIT CODE: 255
> =   CLEANING UP REMAINING PROCESSES
> =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
> ===================================================================================
>
> 2016-03-23 17:34 GMT+02:00 Kenneth Raffenetti <raffenet at mcs.anl.gov
> <mailto:raffenet at mcs.anl.gov>>:
>
>     The error in your first message was that mpiexec was unable to find
>     the file "./examples/cpi".
>
>     The error in your second message is that mpiexec was unable to find
>     your "machinefile".
>
>     Please make sure you are giving the correct paths to these files in
>     your mpiexec command.
>
>     Ken
>
>     On 03/23/2016 10:20 AM, יוסף אלון wrote:
>
>         i am not using a share network and i have it on the place in all
>         nodes.
>         and thus file are located in same place on every node.
>         i tried to run like this allso
>         תמונה מוטבעת 1
>
>         2016-03-23 17:06 GMT+02:00 Kenneth Raffenetti
>         <raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>
>         <mailto:raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>>>:
>
>
>              Are you executing your commands from a shared network
>         filesystem? If
>              not, have you copied your MPI installation and cpi binaries
>         into the
>              same location on all the machines in your cluster?
>
>              Ken
>
>              On 03/23/2016 09:47 AM, יוסף אלון wrote:
>
>
>                  hiii
>
>                  i am new here and i have a 18 node cluster that works
>         pretty
>                  good when i
>                  execute the command:
>
>                  *mpiexec -f machinefile -n 18 hostname
>                  * and the folowing output:
>                  elec-cluster-1
>                  elec-cluster-2
>                  elec-cluster-3
>                  elec-cluster-5
>                  elec-cluster-4
>                  elec-cluster-6
>                  elec-cluster-7
>                  elec-cluster-9
>                  elec-cluster-8
>                  elec-cluster-10
>                  elec-cluster-11
>                  elec-cluster-13
>                  elec-cluster-14
>                  elec-cluster-15
>                  elec-cluster-16
>                  elec-cluster-12
>                  elec-cluster-18
>                  elec-cluster-17
>
>                  when i execute the command:
>                  *mpiexec -n 5 -f machinefile ./examples/cpi*
>
>                     nothing is seems to work and i receive:
>
>                  [proxy:0:0 at elec-cluster-1] HYDU_create_process
>
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>                  execvp error on file ./examples/cpi (No such file or
>         directory)
>
>
>         ===================================================================================
>                  =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>                  =   PID 2806 RUNNING AT 147.161.4.200
>                  =   EXIT CODE: 255
>                  =   CLEANING UP REMAINING PROCESSES
>                  =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>
>         ===================================================================================
>                  [proxy:0:2 at elec-cluster-3] HYDU_create_process
>
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>                  execvp error on file ./examples/cpi (No such file or
>         directory)
>                  [proxy:0:3 at elec-cluster-4] HYDU_create_process
>
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>
>         ===================================================================================
>                  =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>                  =   PID 6718 RUNNING AT 147.161.4.202
>                  =   EXIT CODE: 255
>                  =   CLEANING UP REMAINING PROCESSES
>                  =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>
>         ===================================================================================
>
>                  i dont know what to do?
>                  another think is how to run and compile a c program?
>
>                  --
>                  בברכה, יוסף אלון
>                  050-4916740
>
>
>                  _______________________________________________
>                  discuss mailing list discuss at mpich.org
>         <mailto:discuss at mpich.org> <mailto:discuss at mpich.org
>         <mailto:discuss at mpich.org>>
>                  To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>              _______________________________________________
>              discuss mailing list discuss at mpich.org
>         <mailto:discuss at mpich.org> <mailto:discuss at mpich.org
>         <mailto:discuss at mpich.org>>
>              To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>
>
>
>         --
>         בברכה, יוסף אלון
>         050-4916740
>
>
>         _______________________________________________
>         discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
>         To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>     _______________________________________________
>     discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
>     To manage subscription options or unsubscribe:
>     https://lists.mpich.org/mailman/listinfo/discuss
>
>
>
>
> --
> בברכה, יוסף אלון
> 050-4916740
>
>
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
>
_______________________________________________
discuss mailing list     discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss


More information about the discuss mailing list