[MPICH] Getting BLACS/SCALAPACK running with mpich2

Rajeev Thakur thakur at mcs.anl.gov
Tue Nov 14 21:16:37 CST 2006


I don't know. These warnings are coming from BLACS. You can try contacting
the BLACS folks.

Rajeev
 

> -----Original Message-----
> From: owner-mpich-discuss at mcs.anl.gov 
> [mailto:owner-mpich-discuss at mcs.anl.gov] On Behalf Of Steve Kargl
> Sent: Tuesday, November 14, 2006 4:10 PM
> To: mpich-discuss at mcs.anl.gov
> Subject: [MPICH] Getting BLACS/SCALAPACK running with mpich2
> 
> I've installed MPICH2 on a small cluster running
> FreeBSD.  The Fortran compiler is gfortran-4.2.0.
> In setting up BLACS (to eventually use SCALAPACK),
> I've runs its test program
> 
> node10:kargl[212] mpiexec -n 4 ./xFbtest
> BLACS WARNING 'No need to set message ID range due to MPI 
> communicator.'
> from {-1,-1}, pnum=0, Contxt=-1, on line 18 of file 'blacs_set_.c'.
> 
> BLACS WARNING 'No need to set message ID range due to MPI 
> communicator.'
> from {-1,-1}, pnum=1, Contxt=-1, on line 18 of file 'blacs_set_.c'.
> 
> BLACS WARNING 'No need to set message ID range due to MPI 
> communicator.'
> from {-1,-1}, pnum=2, Contxt=-1, on line 18 of file 'blacs_set_.c'.
> 
> BLACS WARNING 'No need to set message ID range due to MPI 
> communicator.'
> from {-1,-1}, pnum=3, Contxt=-1, on line 18 of file 'blacs_set_.c'.
> 
> {0,2}, pnum=2, Contxt=0, killed other procs, exiting with error #-1.
> 
> rank 2 in job 17  node10.cimu.org_60092   caused collective 
> abort of all ranks
>   exit status of rank 2: killed by signal 9 
> [cli_2]: aborting job:
> application called MPI_Abort(MPI_COMM_WORLD, -1) - process 2
> 
> I've search the web for information and possible patches, but 
> came up empty.  Anyone have any comments?
> 
> PS: MPICH2 works on the cluster in that an application I wrote
> that uses MPICH2 gives the desired result.
> 
> -- 
> Steve
> 
> 




More information about the mpich-discuss mailing list