The short version is that Open MPI is the next generation beyond LAM/MPI. All of us LAM developers stopped working on LAM/MPI (literally) years ago and moved to Open MPI.
If your app uses the standard MPI API, then it should port quite easily (probably with no source code changes) to Open MPI.
On Apr 1, 2010, at 8:31 AM, Patricio Mac Adden wrote:
> It's quite complex, this application is for a college report. I'm not
> sure if I can use open mpi. What's the major differences between lam-mpi
> and open mpi?
>
> Thanks for your answer!!
>
> El jue, 01-04-2010 a las 05:33 -0500, Jeff Squyres (jsquyres) escribió:
> > Since you're just starting out, can you upgrade to open mpi?
> >
> > -jms
> > Sent from my PDA. No type good.
> >
> > ----- Original Message -----
> > From: lam-bounces_at_[hidden] <lam-bounces_at_[hidden]>
> > To: lam_at_[hidden] <lam_at_[hidden]>
> > Sent: Thu Apr 01 00:20:35 2010
> > Subject: LAM: Unix Errno: 14 Bad Address
> >
> > Hi, I'm new in this list.
> >
> > I'm developing a MPI application that uses math.h lib. My problem is
> > that when I compile my application, no error ocurrs, but when I run my
> > application with mpirun, I get this error:
> >
> > It seems that some error has occurred during MPI_INIT. This will
> > cause your process to abort. These kinds of errors are usually
> > system-related, such as running out of disk space, running out of
> > memory, or something more serious such as data not being passed
> > between processes properly. That is, you should not be seeing this
> > error message; if you are, something is likely Very Wrong with your
> > system. :-(
> >
> > Perhaps this Unix error message will help:
> >
> > Unix errno: 14
> > Bad address
> >
> > -----------------------------------------------------------------------------
> > -----------------------------------------------------------------------------
> > It seems that [at least] one of the processes that was started with
> > mpirun did not invoke MPI_INIT before quitting (it is possible that
> > more than one process did not invoke MPI_INIT -- mpirun was only
> > notified of the first one, which was on node n0).
> >
> > mpirun can *only* be used with MPI programs (i.e., programs that
> > invoke MPI_INIT and MPI_FINALIZE). You can use the "lamexec" program
> > to run non-MPI programs over the lambooted nodes.
> >
> > This is how I'm compiling the application:
> >
> > mpicc -Wall -g -O3 -c dist.c lib/*.c
> > mpicc -Wall -g -O3 -o dist *.o -lm
> >
> > Also, I'm running the application like this:
> >
> > mpirun -c 2 dist
> >
> > I'm using Ubuntu 9.10, lam 7.1.2, gcc 4.4.1
> >
> > I hope that someone can help me, I googled a lot but I've found no
> > solution to my problem.
> >
> > Thanks in advance.
> >
> > --
> > Patricio Mac Adden
> >
> > _______________________________________________
> > This list is archived at http://www.lam-mpi.org/MailArchives/lam/
> >
> >
> > _______________________________________________
> > This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>
>
> --
> Patricio Mac Adden
>
> _______________________________________________
> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>
--
Jeff Squyres
jsquyres_at_[hidden]
For corporate legal information go to:
http://www.cisco.com/web/about/doing_business/legal/cri/
|