LAM/MPI logo

LAM/MPI General User's Mailing List Archives

  |   Home   |   Download   |   Documentation   |   FAQ   |   all just in this list

From: Jeff Squyres (jsquyres_at_[hidden])
Date: 2008-05-19 15:59:22


lamrun is for non-MPI executables (e.g., /bin/uptime).

If Fedora installed both Open MPI and LAM, then it sounds like you
need to use the "alternatives" system to have it select LAM as your
preferred MPI.

On May 19, 2008, at 3:42 PM, Krzysztof Bandurski wrote:

> Thank you anyway, your hint was very helpful nonetheless :) I
> found an
> executable called "lamrun" in /usr/bin, and it looks like it's the
> right
> thing to use :)
>
> regards,
>
> kris
>
> McCalla, Mac wrote:
>> Sorry, I am not familiar with "yum" . My standard install process
>> for
>> LAM is download the tarball, extract it where I want it, configure
>> it
>> and make/make install to where $LAMHOME is pointing. then
>> $LAMHOME/bin
>> is in my $PATH environmental variable and I am set.
>>
>> HTH,
>>
>> mac
>>
>> -----Original Message-----
>> From: lam-bounces_at_[hidden] [mailto:lam-bounces_at_[hidden]] On
>> Behalf
>> Of Krzysztof Bandurski
>> Sent: Monday, May 19, 2008 2:16 PM
>> To: General LAM/MPI mailing list
>> Subject: Re: LAM: problem with mpirun - processes don't 'see' one
>> another
>>
>> Hi,
>>
>> Thanks a lot, it seems that you're right, mprun points to
>> /etc/alternatives/mpi-run, which in turn points to /usr/bin/
>> orterun, and
>> that, from what I see in the manual, is OpenMP's thing... I guess I
>> have
>> to find the correct lam's executable... any idea where that might be?
>>
>> kris
>>
>>
>> McCalla, Mac wrote:
>>
>>> Hi,
>>> This looks like a mixed LAM and OpenMPI environment. what does
>>>
>> a
>>
>>> "which mpirun" command show you?
>>>
>>> Cheers,
>>>
>>> Mac McCalla
>>>
>>> -----Original Message-----
>>> From: lam-bounces_at_[hidden] [mailto:lam-bounces_at_[hidden]] On
>>> Behalf Of Krzysztof Bandurski
>>> Sent: Monday, May 19, 2008 1:30 PM
>>> To: lam_at_[hidden]
>>> Subject: LAM: problem with mpirun - processes don't 'see' one
>>> another
>>>
>>> Hi All,
>>>
>>> I used lam before, but I upgraded my system and installed fedora 8
>>> from scratch. I have a dual-core athlon 64 on an nforce chipset. I
>>> wanted to install some mpi environment quickly to test my parallel
>>> programs on my machina at home before submitting them to the cluster
>>> that I use, so I just "yummed" lam to my machine. Lamboot seems to
>>> work fine, but I have a strange problem with mpirun/mpiexec.
>>>
>>> When I run a program using mpirun, e.g. like this:
>>>
>>> mpirun -np 4 testpopmpi_release <and then follow the command line
>>> arguments...>
>>>
>>> I do get 4 processes running, but each of them sees only itself in
>>> MPI_COMM_WORLD. When I run it with --display-map, I get something
>>> like
>>>
>>
>>
>>> this at the beginning of the output:
>>>
>>> [kris_at_nothing nnworkshop]$ mpirun --display-map -np 4
>>> testpopmpi_release -packley -d300 -T0f -v1 -Dcgpr -P256
>>> -Mdesa-best2bin [nothing:05733] Map for job: 1 Generated by mapping
>>>
>> mode: byslot
>>
>>> Starting vpid: 0 Vpid range: 4 Num app_contexts: 1
>>> Data for app_context: index 0 app: testpopmpi_release
>>> Num procs: 4
>>> Argv[0]: testpopmpi_release
>>> Argv[1]: -packley
>>> Argv[2]: -d300
>>> Argv[3]: -T0f
>>> Argv[4]: -v1
>>> Argv[5]: -Dcgpr
>>> Argv[6]: -P256
>>> Argv[7]: -Mdesa-best2bin
>>> Env[0]: OMPI_MCA_rmaps_base_display_map=1
>>> Env[1]:
>>>
>>>
>> OMPI_MCA_orte_precondition_transports
>> =444a2d3c430e64ba-6534b32b337c12e7
>>
>>> Env[2]: OMPI_MCA_rds=proxy
>>> Env[3]: OMPI_MCA_ras=proxy
>>> Env[4]: OMPI_MCA_rmaps=proxy
>>> Env[5]: OMPI_MCA_pls=proxy
>>> Env[6]: OMPI_MCA_rmgr=proxy
>>> Working dir: /home/kris/nnworkshop (user: 0)
>>> Num maps: 0
>>> Num elements in nodes list: 1
>>> Mapped node:
>>> Cell: 0 Nodename: nothing Launch id: -1
>>> Username: NULL
>>> Daemon name:
>>> Data type: ORTE_PROCESS_NAME Data Value:
>>>
>> NULL
>>
>>> Oversubscribed: True Num elements in procs list: 4
>>> Mapped proc:
>>> Proc Name:
>>> Data type: ORTE_PROCESS_NAME Data Value:
>>> [0,1,0]
>>> Proc Rank: 0 Proc PID: 0 App_context
>>> index: 0
>>>
>>> Mapped proc:
>>> Proc Name:
>>> Data type: ORTE_PROCESS_NAME Data Value:
>>> [0,1,1]
>>> Proc Rank: 1 Proc PID: 0 App_context
>>> index: 0
>>>
>>> Mapped proc:
>>> Proc Name:
>>> Data type: ORTE_PROCESS_NAME Data Value:
>>> [0,1,2]
>>> Proc Rank: 2 Proc PID: 0 App_context
>>> index: 0
>>>
>>> Mapped proc:
>>> Proc Name:
>>> Data type: ORTE_PROCESS_NAME Data Value:
>>> [0,1,3]
>>> Proc Rank: 3 Proc PID: 0 App_context
>>> index: 0
>>>
>>> and then follows the output of my program. As you can see, lam
>>> thinks
>>> that all the processes are in the same communicator (they all have
>>> different ranks), but when I call MPI_Comm_rank and MPI_Comm_size in
>>> my program, I always get rank == 0 and size == 1in each single
>>> process
>>>
>>
>>
>>> - needless to say, the processes can't communicate and I just have 4
>>> independent copies of my program running (and printin exactly the
>>> same
>>>
>>
>>
>>> output on the terminal....). Does anyone have any idea what might
>>> be
>>> going on? This is really driving me nuts, I will appreciate any
>>> hints.
>>>
>>> best regards,
>>>
>>> kris.
>>>
>>>
>>>
>>> _______________________________________________
>>> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>>>
>>> _______________________________________________
>>> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>>>
>>>
>>>
>>
>> _______________________________________________
>> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>>
>> _______________________________________________
>> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>>
>>
>
> _______________________________________________
> This list is archived at http://www.lam-mpi.org/MailArchives/lam/

-- 
Jeff Squyres
Cisco Systems