Can you run a simple hello, world application (like the ones that ship
with LAM)? This looks like either a problem with your test code or a
misunderstanding with how you are supposed to run the executable -
unfortunately both of those issues are out of my area of expertise.
Brian
On Fri, 27 Mar 2009, Axel Bellivier wrote:
> All mpi run get this same message.
> Single runs don't have any problems.
>
> Here is more informations:
>
> My command: (lamd is running)
> mpirun -np 4 fds5_mpi_intel test1-Parallel.fds < /dev/null > log.run
> 2>&1 &
>
> My system:
> Dell R900 with 4 quad core Intel(R) Xeon(R) CPU X7350 @
> 2.93GHz
> Red Hat Enterprise Linux Server release 5.2 (Tikanga) (64 bit)
> Linux version 2.6.18-92.1.18.el5
> (brewbuilder_at_[hidden]) (gcc version 4.1.2 20071124
> (Red Hat 4.1.2-42)) #1 SMP Wed Nov 5 09:00:19 EST 2008
>
>
> LAM/MPI: 7.1.4
> Prefix: /usr/local
> Architecture: x86_64-unknown-linux-gnu
> Configured by: root
> Configured on: Mon Jan 12 14:31:06 CET 2009
> Configure host: srvcal-r900-1
> Memory manager: ptmalloc2
> C bindings: yes
> C++ bindings: yes
> Fortran bindings: yes
> C compiler: icc
> C++ compiler: icpc
> Fortran compiler: ifort
> Fortran symbols: underscore
> C profiling: yes
> C++ profiling: yes
> Fortran profiling: yes
> C++ exceptions: no
> Thread support: yes
> ROMIO support: yes
> IMPI support: no
> Debug support: no
> Purify clean: no
> SSI boot: globus (API v1.1, Module v0.6)
> SSI boot: rsh (API v1.1, Module v1.1)
> SSI boot: slurm (API v1.1, Module v1.0)
> SSI coll: lam_basic (API v1.1, Module v7.1)
> SSI coll: shmem (API v1.1, Module v1.0)
> SSI coll: smp (API v1.1, Module v1.2)
> SSI rpi: crtcp (API v1.1, Module v1.1)
> SSI rpi: lamd (API v1.0, Module v7.1)
> SSI rpi: sysv (API v1.0, Module v7.1)
> SSI rpi: tcp (API v1.0, Module v7.1)
> SSI rpi: usysv (API v1.0, Module v7.1)
> SSI cr: self (API v1.0, Module v1.0)
>
>
> 2009/3/26 Brian W. Barrett <brbarret_at_[hidden]>
> On Thu, 26 Mar 2009, Axel Bellivier wrote:
>
> i post because a realise that my mpi jobs was as
> longer as single ones !!
>
> The only thing strange i found on logs is:
>
> Process 0 of 0 is running on srvcal-r900-1
> Mesh 1 is assigned to Process 0
> Mesh 2 is assigned to Process 0
> Mesh 3 is assigned to Process 0
> Mesh 4 is assigned to Process 0
> Process 0 of 0 is running on srvcal-r900-1
>
>
> >From your output, it's hard to tell what's supposed to
> happen and what
>
> actually happened. If you run a simple hello world example, does
> it run properly? Is your log of many runs, or a single run? How
> did you start your parallel application?
>
> Brian
>
> --
> Brian Barrett
> LAM/MPI Developer
> Make today a LAM/MPI day!
> _______________________________________________
> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>
>
>
>
--
Brian Barrett
LAM/MPI Developer
Make today a LAM/MPI day!
|