- Next message: Soumya,Teena, Ranjana,Liss,Navya: "Re: LAM: please suggest a solution."
- Previous message: Jeff Squyres: "Re: LAM: redirection"
- In reply to: Soumya,Teena, Ranjana,Liss,Navya: "LAM: please suggest a solution."
- Next in thread: Soumya,Teena, Ranjana,Liss,Navya: "Re: LAM: please suggest a solution."
- Reply: Soumya,Teena, Ranjana,Liss,Navya: "Re: LAM: please suggest a solution."
I *think* that you are seeing the canonical parallel output issue --
that output from different nodes shows up at seemingly random times
because, at least in part, all nodes are not exactly synchronized.
However, you should probably still see the "Hello world" messages
from the MPI processes on node-3 (note that you killed node-1, so you
won't see anything from node-1). These are silly suggestions, but
try them anyway:
- Output is not guaranteed to be flushed unless you trail the message
with a \n -- put a \n in your "After final" message.
- Put a "sleep(1);" before MPI_Finalize to give the output time to
flush over to mpirun before closing everything down.
I'm sure your application is running properly and that for some
reason the stdout is not getting flushed over to mpirun.
On Feb 24, 2006, at 4:26 AM, Soumya,Teena, Ranjana,Liss,Navya wrote:
> hi,
>
>
> hello.c program is as follows:
> #include<stdio.h>
> #include<mpi.h>
> int main(int argc,char* argv[]){
> int rank,size;
> MPI_Init(&argc,&argv);
> MPI_Comm_rank(MPI_COMM_WORLD,&rank);
> MPI_Comm_size(MPI_COMM_WORLD,&size);
>
> printf("Hello World %d of %d \n",rank,size);
> MPI_Finalize();
> printf("After final...");
> return 0;
> }
>
>
> Lam-mpi is configured on 4 systems.
> The command lamnodes gave:
> n0 master.mec.ac.in:1:origin,this_node
> n1 node-1.mec.ac.in:1:
> n2 node-2.mec.ac.in:1:
> n3 node-3.mec.ac.in:1:
>
> When we do mpirun -np 4 a.out after mpicc test.c -lmpi having
> compiled successfully we are getting the following output:
> Hello World 0 of 4
> Hello World 2 of 4
> After final...Afterfinal...
>
>
> Then we did mpirun -np 8 a.out and got output as follows
> Hello World 0 of 8
> Hello World 4 of 8
> After final...Hello World 2 of 8
> Hello World 6 of 8
> After final...After final...After final...
>
> Then we did lamshrink n1 .
> Then mpirun -np 8 a.out
> Hello World 3 of 8
> Hello World 1 of 8
> Hello World 4 of 8
> Hello World 7 of 8
> After final...After final...After final...After final...Hello World
> 0 of 8
> After final...Hello World 6 of 8
>
> It seems that the nodes node-1 and node-3 is not printing any
> message although communication between the nodes by means of
> message parsing is taking place and lam is properly installed in
> all 4 nodes.
>
> What could be the reason? Please suggest a solution.
>
> Thank u.
> --
> CE 2002-06
> _______________________________________________
> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
--
{+} Jeff Squyres
{+} The Open MPI Project
{+} http://www.open-mpi.org/
- Next message: Soumya,Teena, Ranjana,Liss,Navya: "Re: LAM: please suggest a solution."
- Previous message: Jeff Squyres: "Re: LAM: redirection"
- In reply to: Soumya,Teena, Ranjana,Liss,Navya: "LAM: please suggest a solution."
- Next in thread: Soumya,Teena, Ranjana,Liss,Navya: "Re: LAM: please suggest a solution."
- Reply: Soumya,Teena, Ranjana,Liss,Navya: "Re: LAM: please suggest a solution."
|