LAM/MPI logo

LAM/MPI General User's Mailing List Archives

  |   Home   |   Download   |   Documentation   |   FAQ   |   all just in this list

From: Srinivasa Prade Patri (Srinivasa.Patri_at_[hidden])
Date: 2005-03-29 15:36:31


Hi!
    I did look in to the web page and set the following in Bmake.inc

SYSINC =

INTFACE = -Df77IsF2C
/* when i ran xintface i got the following output
For this platform, set INTFACE = -Df77IsF2C */

TRANSCOMM = -DUseMpi2
/* when i ran xtc_CsameF77 i got the following output
[patri_at_kemc ~]$ mpirun -v -np 2 /home/BLACS/INSTALL/EXE/xtc_CsameF77
23986 /home/BLACS/INSTALL/EXE/xtc_CsameF77 running on n0 (o)
23987 /home/BLACS/INSTALL/EXE/xtc_CsameF77 running on n0 (o)
 If this routine does not complete successfully,
 Do _NOT_ set TRANSCOMM = -DCSameF77

 Do _NOT_ set TRANSCOMM = -DCSameF77 */

F77 = mpif77

CC = mpicc
    
    Thanking you.

Regards
Srinivasa Patri

-----Original Message-----
From: Jeff Squyres <jsquyres_at_[hidden]>
To: General LAM/MPI mailing list <lam_at_[hidden]>, sppatr2_at_[hidden]
Date: Tue, 29 Mar 2005 14:23:21 -0500
Subject: Re: LAM: Problem running the tester for BLACS

Have you see this page on how to build BLACS with LAM/MPI?

        http://www.lam-mpi.org/3rd-party/blacs.php3

It may be helpful in making the BLACS code do the Right Things with
regard to LAM/MPI.

On Mar 29, 2005, at 2:00 PM, Srinivasa Prade Patri wrote:

> Hi!
> First of all i would like thank all for your support. I got my
> previous problem solved and the test suite has complied succesfully.
> But when running the exe files iam getting the following errors listed
> below. What could be the problem. I once again like to thank all of
> you for your support.
>
> -----------------------------------------------------------------------
> --------------
> [patri_at_e01 ~]$ mpirun -v C /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0
> 1435 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n0 (o)
> 1196 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n1
> 1194 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n2
> 1194 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n3
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n4
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n5
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n6
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n7
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n8
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n9
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n10
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n11
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n12
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n13
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n14
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n15
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n16
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n17
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n18
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n19
> 1191 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n20
> 1192 /home/BLACS/TESTING/EXE/xFbtest_MPI-LINUX-0 running on n21
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=0, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 0,
> MPI_COMM_WORLD)
> Rank (0, MPI_COMM_WORLD): Call stack within LAM:
> Rank (0, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (0, MPI_COMM_WORLD): - main()
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=2, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=3, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=1, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 2,
> MPI_COMM_WORLD)
> Rank (2, MPI_COMM_WORLD): Call stack within LAM:
> Rank (2, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (2, MPI_COMM_WORLD): - main()
> MPI_Comm_group: invalid communicator: Invalid argument (rank 3,
> MPI_COMM_WORLD)
> Rank (3, MPI_COMM_WORLD): Call stack within LAM:
> Rank (3, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (3, MPI_COMM_WORLD): - main()
> MPI_Comm_group: invalid communicator: Invalid argument (rank 1,
> MPI_COMM_WORLD)
> Rank (1, MPI_COMM_WORLD): Call stack within LAM:
> Rank (1, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (1, MPI_COMM_WORLD): - main()
> -----------------------------------------------------------------------
> ------
> One of the processes started by mpirun has exited with a nonzero exit
> code. This typically indicates that the process finished in error.
> If your process did not finish in error, be sure to include a "return
> 0" or "exit(0)" in your C code before exiting the application.
>
> PID 1435 failed on node n0 (10.5.0.1) with exit status 1.
> -----------------------------------------------------------------------
> ------
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=18, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=5, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 18,
> MPI_COMM_WORLD)
> Rank (18, MPI_COMM_WORLD): Call stack within LAM:
> Rank (18, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (18, MPI_COMM_WORLD): - main()
> MPI_Comm_group: invalid communicator: Invalid argument (rank 5,
> MPI_COMM_WORLD)
> Rank (5, MPI_COMM_WORLD): Call stack within LAM:
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=14, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> Rank (5, MPI_COMM_WORLD): - MPI_Comm_group()
> MPI_Comm_group: invalid communicator: Invalid argument (rank 14,
> MPI_COMM_WORLD)
> Rank (14, MPI_COMM_WORLD): Call stack within LAM:
> Rank (5, MPI_COMM_WORLD): - main()
> Rank (14, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (14, MPI_COMM_WORLD): - main()
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=20, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 20,
> MPI_COMM_WORLD)
> Rank (20, MPI_COMM_WORLD): Call stack within LAM:
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=21, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> Rank (20, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (20, MPI_COMM_WORLD): - main()
> MPI_Comm_group: invalid communicator: Invalid argument (rank 21,
> MPI_COMM_WORLD)
> Rank (21, MPI_COMM_WORLD): Call stack within LAM:
> Rank (21, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (21, MPI_COMM_WORLD): - main()
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=16, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 16,
> MPI_COMM_WORLD)
> Rank (16, MPI_COMM_WORLD): Call stack within LAM:
> Rank (16, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (16, MPI_COMM_WORLD): - main()
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=17, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 17,
> MPI_COMM_WORLD)
> Rank (17, MPI_COMM_WORLD): Call stack within LAM:
> Rank (17, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (17, MPI_COMM_WORLD): - main()
> BLACS WARNING 'No need to set message ID range due to MPI
> communicator.'
> from {-1,-1}, pnum=19, Contxt=-1, on line 18 of file 'blacs_set_.c'.
>
> MPI_Comm_group: invalid communicator: Invalid argument (rank 19,
> MPI_COMM_WORLD)
> Rank (19, MPI_COMM_WORLD): Call stack within LAM:
> Rank (19, MPI_COMM_WORLD): - MPI_Comm_group()
> Rank (19, MPI_COMM_WORLD): - main()
> -----------------------------------------------------------------------
> --------------
>
> Regards
> Srinivasa Patri
>
>
> _______________________________________________
> This list is archived at http://www.lam-mpi.org/MailArchives/lam/
>

-- 
{+} Jeff Squyres
{+} jsquyres_at_[hidden]
{+} http://www.lam-mpi.org/