[mpich-discuss] cpi

Kenneth Raffenetti raffenet at mcs.anl.gov
Wed Mar 23 10:59:33 CDT 2016


The executable MPICH builds (and you want to run) is "cpi" in the 
examples. cpi.c is a C source code file and is not executable. I suggest 
you read more on how to build and execute C programs if you still have 
questions.

Ken

On 03/23/2016 10:43 AM, יוסף אלון wrote:
> what do you mean?
>
>
> 2016-03-23 17:42 GMT+02:00 Kenneth Raffenetti <raffenet at mcs.anl.gov
> <mailto:raffenet at mcs.anl.gov>>:
>
>     cpi.c is a source file, not an executable.
>
>     On 03/23/2016 10:41 AM, יוסף אלון wrote:
>
>         when i do this i reciev:
>
>         cluster at elec-cluster-1 ~ $ *_mpiexec -n 5 -f machinefile
>         ./mpich2/mpich-3.1/examples/cpi.c_*
>         [proxy:0:0 at elec-cluster-1] HYDU_create_process
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>         execvp error on file ./mpich2/mpich-3.1/examples/cpi.c
>         (Permission denied)
>
>         ===================================================================================
>         =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>         =   PID 2895 RUNNING AT 147.161.4.200
>         =   EXIT CODE: 255
>         =   CLEANING UP REMAINING PROCESSES
>         =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>         ===================================================================================
>         [proxy:0:1 at elec-cluster-2] HYDU_create_process
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>         execvp error on file ./mpich2/mpich-3.1/examples/cpi.c
>         (Permission denied)
>         [proxy:0:4 at elec-cluster-5] HYDU_create_process
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>         ===================================================================================
>         =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>         =   PID 7382 RUNNING AT 147.161.4.201
>         =   EXIT CODE: 255
>         =   CLEANING UP REMAINING PROCESSES
>         =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>         ===================================================================================
>
>         2016-03-23 17:34 GMT+02:00 Kenneth Raffenetti
>         <raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>
>         <mailto:raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>>>:
>
>              The error in your first message was that mpiexec was unable
>         to find
>              the file "./examples/cpi".
>
>              The error in your second message is that mpiexec was unable
>         to find
>              your "machinefile".
>
>              Please make sure you are giving the correct paths to these
>         files in
>              your mpiexec command.
>
>              Ken
>
>              On 03/23/2016 10:20 AM, יוסף אלון wrote:
>
>                  i am not using a share network and i have it on the
>         place in all
>                  nodes.
>                  and thus file are located in same place on every node.
>                  i tried to run like this allso
>                  תמונה מוטבעת 1
>
>                  2016-03-23 17:06 GMT+02:00 Kenneth Raffenetti
>                  <raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>
>         <mailto:raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>>
>                  <mailto:raffenet at mcs.anl.gov
>         <mailto:raffenet at mcs.anl.gov> <mailto:raffenet at mcs.anl.gov
>         <mailto:raffenet at mcs.anl.gov>>>>:
>
>
>
>                       Are you executing your commands from a shared network
>                  filesystem? If
>                       not, have you copied your MPI installation and cpi
>         binaries
>                  into the
>                       same location on all the machines in your cluster?
>
>                       Ken
>
>                       On 03/23/2016 09:47 AM, יוסף אלון wrote:
>
>
>                           hiii
>
>                           i am new here and i have a 18 node cluster
>         that works
>                  pretty
>                           good when i
>                           execute the command:
>
>                           *mpiexec -f machinefile -n 18 hostname
>                           * and the folowing output:
>                           elec-cluster-1
>                           elec-cluster-2
>                           elec-cluster-3
>                           elec-cluster-5
>                           elec-cluster-4
>                           elec-cluster-6
>                           elec-cluster-7
>                           elec-cluster-9
>                           elec-cluster-8
>                           elec-cluster-10
>                           elec-cluster-11
>                           elec-cluster-13
>                           elec-cluster-14
>                           elec-cluster-15
>                           elec-cluster-16
>                           elec-cluster-12
>                           elec-cluster-18
>                           elec-cluster-17
>
>                           when i execute the command:
>                           *mpiexec -n 5 -f machinefile ./examples/cpi*
>
>                              nothing is seems to work and i receive:
>
>                           [proxy:0:0 at elec-cluster-1] HYDU_create_process
>
>
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>                           execvp error on file ./examples/cpi (No such
>         file or
>                  directory)
>
>
>
>         ===================================================================================
>                           =   BAD TERMINATION OF ONE OF YOUR APPLICATION
>         PROCESSES
>                           =   PID 2806 RUNNING AT 147.161.4.200
>                           =   EXIT CODE: 255
>                           =   CLEANING UP REMAINING PROCESSES
>                           =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>
>
>         ===================================================================================
>                           [proxy:0:2 at elec-cluster-3] HYDU_create_process
>
>
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>                           execvp error on file ./examples/cpi (No such
>         file or
>                  directory)
>                           [proxy:0:3 at elec-cluster-4] HYDU_create_process
>
>
>         (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>
>
>         ===================================================================================
>                           =   BAD TERMINATION OF ONE OF YOUR APPLICATION
>         PROCESSES
>                           =   PID 6718 RUNNING AT 147.161.4.202
>                           =   EXIT CODE: 255
>                           =   CLEANING UP REMAINING PROCESSES
>                           =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>
>
>         ===================================================================================
>
>                           i dont know what to do?
>                           another think is how to run and compile a c
>         program?
>
>                           --
>                           בברכה, יוסף אלון
>                           050-4916740
>
>
>                           _______________________________________________
>                           discuss mailing list discuss at mpich.org
>         <mailto:discuss at mpich.org>
>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>
>         <mailto:discuss at mpich.org <mailto:discuss at mpich.org>
>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>>
>                           To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>                       _______________________________________________
>                       discuss mailing list discuss at mpich.org
>         <mailto:discuss at mpich.org>
>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>
>         <mailto:discuss at mpich.org <mailto:discuss at mpich.org>
>
>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>>
>                       To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>
>
>
>                  --
>                  בברכה, יוסף אלון
>                  050-4916740
>
>
>                  _______________________________________________
>                  discuss mailing list discuss at mpich.org
>         <mailto:discuss at mpich.org> <mailto:discuss at mpich.org
>         <mailto:discuss at mpich.org>>
>                  To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>              _______________________________________________
>              discuss mailing list discuss at mpich.org
>         <mailto:discuss at mpich.org> <mailto:discuss at mpich.org
>         <mailto:discuss at mpich.org>>
>              To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>
>
>
>         --
>         בברכה, יוסף אלון
>         050-4916740
>
>
>         _______________________________________________
>         discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
>         To manage subscription options or unsubscribe:
>         https://lists.mpich.org/mailman/listinfo/discuss
>
>     _______________________________________________
>     discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
>     To manage subscription options or unsubscribe:
>     https://lists.mpich.org/mailman/listinfo/discuss
>
>
>
>
> --
> בברכה, יוסף אלון
> 050-4916740
>
>
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
>
_______________________________________________
discuss mailing list     discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss


More information about the discuss mailing list