[mpich-discuss] cpi

sanjesh pant spant3474 at gmail.com
Wed Mar 23 11:09:24 CDT 2016


First compile it by using command as
mpicc -o cpi cpi.c
On 23 Mar 2016 21:33, "יוסף אלון" <yos104104 at gmail.com> wrote:

> but i do want to run a C source code all over the nodes?
> and allso when i run only the cpi witout .c nothing is hapning.
> i think you the only one who can help me.
> thank you
>
> 2016-03-23 17:59 GMT+02:00 Kenneth Raffenetti <raffenet at mcs.anl.gov>:
>
>> The executable MPICH builds (and you want to run) is "cpi" in the
>> examples. cpi.c is a C source code file and is not executable. I suggest
>> you read more on how to build and execute C programs if you still have
>> questions.
>>
>> Ken
>>
>> On 03/23/2016 10:43 AM, יוסף אלון wrote:
>>
>>> what do you mean?
>>>
>>>
>>> 2016-03-23 17:42 GMT+02:00 Kenneth Raffenetti <raffenet at mcs.anl.gov
>>> <mailto:raffenet at mcs.anl.gov>>:
>>>
>>>
>>>     cpi.c is a source file, not an executable.
>>>
>>>     On 03/23/2016 10:41 AM, יוסף אלון wrote:
>>>
>>>         when i do this i reciev:
>>>
>>>         cluster at elec-cluster-1 ~ $ *_mpiexec -n 5 -f machinefile
>>>         ./mpich2/mpich-3.1/examples/cpi.c_*
>>>         [proxy:0:0 at elec-cluster-1] HYDU_create_process
>>>
>>> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>>>         execvp error on file ./mpich2/mpich-3.1/examples/cpi.c
>>>         (Permission denied)
>>>
>>>
>>> ===================================================================================
>>>         =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>>>         =   PID 2895 RUNNING AT 147.161.4.200
>>>         =   EXIT CODE: 255
>>>         =   CLEANING UP REMAINING PROCESSES
>>>         =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>>>
>>> ===================================================================================
>>>         [proxy:0:1 at elec-cluster-2] HYDU_create_process
>>>
>>> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>>>         execvp error on file ./mpich2/mpich-3.1/examples/cpi.c
>>>         (Permission denied)
>>>         [proxy:0:4 at elec-cluster-5] HYDU_create_process
>>>
>>> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>>>
>>> ===================================================================================
>>>         =   BAD TERMINATION OF ONE OF YOUR APPLICATION PROCESSES
>>>         =   PID 7382 RUNNING AT 147.161.4.201
>>>         =   EXIT CODE: 255
>>>         =   CLEANING UP REMAINING PROCESSES
>>>         =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>>>
>>> ===================================================================================
>>>
>>>         2016-03-23 17:34 GMT+02:00 Kenneth Raffenetti
>>>         <raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>
>>>         <mailto:raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>>>:
>>>
>>>
>>>              The error in your first message was that mpiexec was unable
>>>         to find
>>>              the file "./examples/cpi".
>>>
>>>              The error in your second message is that mpiexec was unable
>>>         to find
>>>              your "machinefile".
>>>
>>>              Please make sure you are giving the correct paths to these
>>>         files in
>>>              your mpiexec command.
>>>
>>>              Ken
>>>
>>>              On 03/23/2016 10:20 AM, יוסף אלון wrote:
>>>
>>>                  i am not using a share network and i have it on the
>>>         place in all
>>>                  nodes.
>>>                  and thus file are located in same place on every node.
>>>                  i tried to run like this allso
>>>                  תמונה מוטבעת 1
>>>
>>>                  2016-03-23 17:06 GMT+02:00 Kenneth Raffenetti
>>>                  <raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>
>>>         <mailto:raffenet at mcs.anl.gov <mailto:raffenet at mcs.anl.gov>>
>>>                  <mailto:raffenet at mcs.anl.gov
>>>         <mailto:raffenet at mcs.anl.gov> <mailto:raffenet at mcs.anl.gov
>>>         <mailto:raffenet at mcs.anl.gov>>>>:
>>>
>>>
>>>
>>>                       Are you executing your commands from a shared
>>> network
>>>                  filesystem? If
>>>                       not, have you copied your MPI installation and cpi
>>>         binaries
>>>                  into the
>>>                       same location on all the machines in your cluster?
>>>
>>>                       Ken
>>>
>>>                       On 03/23/2016 09:47 AM, יוסף אלון wrote:
>>>
>>>
>>>                           hiii
>>>
>>>                           i am new here and i have a 18 node cluster
>>>         that works
>>>                  pretty
>>>                           good when i
>>>                           execute the command:
>>>
>>>                           *mpiexec -f machinefile -n 18 hostname
>>>                           * and the folowing output:
>>>                           elec-cluster-1
>>>                           elec-cluster-2
>>>                           elec-cluster-3
>>>                           elec-cluster-5
>>>                           elec-cluster-4
>>>                           elec-cluster-6
>>>                           elec-cluster-7
>>>                           elec-cluster-9
>>>                           elec-cluster-8
>>>                           elec-cluster-10
>>>                           elec-cluster-11
>>>                           elec-cluster-13
>>>                           elec-cluster-14
>>>                           elec-cluster-15
>>>                           elec-cluster-16
>>>                           elec-cluster-12
>>>                           elec-cluster-18
>>>                           elec-cluster-17
>>>
>>>                           when i execute the command:
>>>                           *mpiexec -n 5 -f machinefile ./examples/cpi*
>>>
>>>                              nothing is seems to work and i receive:
>>>
>>>                           [proxy:0:0 at elec-cluster-1] HYDU_create_process
>>>
>>>
>>>
>>> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>>>                           execvp error on file ./examples/cpi (No such
>>>         file or
>>>                  directory)
>>>
>>>
>>>
>>>
>>> ===================================================================================
>>>                           =   BAD TERMINATION OF ONE OF YOUR APPLICATION
>>>         PROCESSES
>>>                           =   PID 2806 RUNNING AT 147.161.4.200
>>>                           =   EXIT CODE: 255
>>>                           =   CLEANING UP REMAINING PROCESSES
>>>                           =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>>>
>>>
>>>
>>> ===================================================================================
>>>                           [proxy:0:2 at elec-cluster-3] HYDU_create_process
>>>
>>>
>>>
>>> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>>>                           execvp error on file ./examples/cpi (No such
>>>         file or
>>>                  directory)
>>>                           [proxy:0:3 at elec-cluster-4] HYDU_create_process
>>>
>>>
>>>
>>> (/home/cluster/mpich2/mpich-3.1/src/pm/hydra/utils/launch/launch.c:75):
>>>
>>>
>>>
>>> ===================================================================================
>>>                           =   BAD TERMINATION OF ONE OF YOUR APPLICATION
>>>         PROCESSES
>>>                           =   PID 6718 RUNNING AT 147.161.4.202
>>>                           =   EXIT CODE: 255
>>>                           =   CLEANING UP REMAINING PROCESSES
>>>                           =   YOU CAN IGNORE THE BELOW CLEANUP MESSAGES
>>>
>>>
>>>
>>> ===================================================================================
>>>
>>>                           i dont know what to do?
>>>                           another think is how to run and compile a c
>>>         program?
>>>
>>>                           --
>>>                           בברכה, יוסף אלון
>>>                           050-4916740
>>>
>>>
>>>                           _______________________________________________
>>>                           discuss mailing list discuss at mpich.org
>>>         <mailto:discuss at mpich.org>
>>>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>
>>>         <mailto:discuss at mpich.org <mailto:discuss at mpich.org>
>>>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>>
>>>                           To manage subscription options or unsubscribe:
>>>         https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>                       _______________________________________________
>>>                       discuss mailing list discuss at mpich.org
>>>         <mailto:discuss at mpich.org>
>>>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>
>>>         <mailto:discuss at mpich.org <mailto:discuss at mpich.org>
>>>
>>>                  <mailto:discuss at mpich.org <mailto:discuss at mpich.org>>>
>>>
>>>                       To manage subscription options or unsubscribe:
>>>         https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>
>>>
>>>
>>>                  --
>>>                  בברכה, יוסף אלון
>>>                  050-4916740
>>>
>>>
>>>                  _______________________________________________
>>>                  discuss mailing list discuss at mpich.org
>>>         <mailto:discuss at mpich.org> <mailto:discuss at mpich.org
>>>         <mailto:discuss at mpich.org>>
>>>                  To manage subscription options or unsubscribe:
>>>         https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>              _______________________________________________
>>>              discuss mailing list discuss at mpich.org
>>>         <mailto:discuss at mpich.org> <mailto:discuss at mpich.org
>>>         <mailto:discuss at mpich.org>>
>>>              To manage subscription options or unsubscribe:
>>>         https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>
>>>
>>>
>>>         --
>>>         בברכה, יוסף אלון
>>>         050-4916740
>>>
>>>
>>>         _______________________________________________
>>>         discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org
>>> >
>>>         To manage subscription options or unsubscribe:
>>>         https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>     _______________________________________________
>>>     discuss mailing list discuss at mpich.org <mailto:discuss at mpich.org>
>>>     To manage subscription options or unsubscribe:
>>>     https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>>
>>>
>>>
>>> --
>>> בברכה, יוסף אלון
>>> 050-4916740
>>>
>>>
>>> _______________________________________________
>>> discuss mailing list     discuss at mpich.org
>>> To manage subscription options or unsubscribe:
>>> https://lists.mpich.org/mailman/listinfo/discuss
>>>
>>> _______________________________________________
>> discuss mailing list     discuss at mpich.org
>> To manage subscription options or unsubscribe:
>> https://lists.mpich.org/mailman/listinfo/discuss
>>
>
>
>
> --
> בברכה, יוסף אלון
> 050-4916740
>
> _______________________________________________
> discuss mailing list     discuss at mpich.org
> To manage subscription options or unsubscribe:
> https://lists.mpich.org/mailman/listinfo/discuss
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mpich.org/pipermail/discuss/attachments/20160323/896767b2/attachment.html>
-------------- next part --------------
_______________________________________________
discuss mailing list     discuss at mpich.org
To manage subscription options or unsubscribe:
https://lists.mpich.org/mailman/listinfo/discuss


More information about the discuss mailing list