LAM/MPI logo

LAM/MPI General User's Mailing List Archives

  |   Home   |   Download   |   Documentation   |   FAQ   |   all just in this list

From: Elie Choueiri (elie.choueiri_at_[hidden])
Date: 2006-02-27 05:09:54


Would an MPI_Barrier before the MPI_Finalize make a difference?

On 2/27/06, Soumya,Teena, Ranjana,Liss,Navya <bytecode.compression_at_[hidden]>
wrote:
>
> Thank you for the prompt suggestions. But it did not help. The outputs
> are just the same as before except that the message comes on separate
> lines.
>
> On 2/26/06, Jeff Squyres <jsquyres_at_[hidden]> wrote:
> >
> > I *think* that you are seeing the canonical parallel output issue --
> > that output from different nodes shows up at seemingly random times
> > because, at least in part, all nodes are not exactly synchronized.
> >
> > However, you should probably still see the "Hello world" messages
> > from the MPI processes on node-3 (note that you killed node-1, so you
> > won't see anything from node-1). These are silly suggestions, but
> > try them anyway:
> >
> > - Output is not guaranteed to be flushed unless you trail the message
> > with a \n -- put a \n in your "After final" message.
> > - Put a "sleep(1);" before MPI_Finalize to give the output time to
> > flush over to mpirun before closing everything down.
> >
> > I'm sure your application is running properly and that for some
> > reason the stdout is not getting flushed over to mpirun.
> >
> >
> > On Feb 24, 2006, at 4:26 AM, Soumya,Teena, Ranjana,Liss,Navya wrote:
> >
> > > hi,
> > >
> > >
> > > hello.c program is as follows:
> > > #include<stdio.h>
> > > #include<mpi.h>
> > > int main(int argc,char* argv[]){
> > > int rank,size;
> > > MPI_Init(&argc,&argv);
> > > MPI_Comm_rank(MPI_COMM_WORLD,&rank);
> > > MPI_Comm_size(MPI_COMM_WORLD,&size);
> > >
> > > printf("Hello World %d of %d \n",rank,size);
> > > MPI_Finalize();
> > > printf("After final...");
> > > return 0;
> > > }
> > >
> > >
> > > Lam-mpi is configured on 4 systems.
> > > The command lamnodes gave:
> > > n0 master.mec.ac.in:1:origin,this_node
> > > n1 node-1.mec.ac.in:1:
> > > n2 node-2.mec.ac.in:1:
> > > n3 node-3.mec.ac.in:1:
> > >
> > > When we do mpirun -np 4 a.out after mpicc test.c -lmpi having
> > > compiled successfully we are getting the following output:
> > > Hello World 0 of 4
> > > Hello World 2 of 4
> > > After final...Afterfinal...
> > >
> > >
> > > Then we did mpirun -np 8 a.out and got output as follows
> > > Hello World 0 of 8
> > > Hello World 4 of 8
> > > After final...Hello World 2 of 8
> > > Hello World 6 of 8
> > > After final...After final...After final...
> > >
> > > Then we did lamshrink n1 .
> > > Then mpirun -np 8 a.out
> > > Hello World 3 of 8
> > > Hello World 1 of 8
> > > Hello World 4 of 8
> > > Hello World 7 of 8
> > > After final...After final...After final...After final...Hello World
> > > 0 of 8
> > > After final...Hello World 6 of 8
> > >
> > > It seems that the nodes node-1 and node-3 is not printing any
> > > message although communication between the nodes by means of
> > > message parsing is taking place and lam is properly installed in
> > > all 4 nodes.
> > >
> > > What could be the reason? Please suggest a solution.
> > >
> > > Thank u.
> > > --
> > > CE 2002-06
> > > _______________________________________________
> > > This list is archived at http://www.lam-mpi.org/MailArchives/lam/
> >
> >
> > --
> > {+} Jeff Squyres
> > {+} The Open MPI Project
> > {+} http://www.open-mpi.org/
> >
> >
> > _______________________________________________
> > This list is archived at http://www.lam-mpi.org/MailArchives/lam/
> >
>
>
>
> --
> CE 2002-06
> _______________________________________________
> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>
>