[petsc-users] Compiling 3.3 for Open-MPI on a SLURM system

Satish Balay balay at mcs.anl.gov
Wed Dec 19 16:03:56 CST 2012


Perhaps mpiexec is invoking srun internally? The details should be in configure.log

[petsc configure doesn't know about srun].

Satish

On Wed, 19 Dec 2012, Gunter, David O wrote:

> I am trying to compile PETSc on a SLURM-based system using GCC, openmpi-1.6.3.
> 
> Here's my configure line:
> 
> $ ./configure --prefix=/tmp/dog/petsc-3.3-p5 --with-mpiexec=mpiexec
> 
> configure bombs out on this test:
> 
> TESTING: configureMPITypes from config.packages.MPI(/usr/aprojects/hpctools/dog/petsc/petsc-3.3-p5/config/BuildSystem/config/packages/MPI.py:230)srun: error: slurm_send_recv_rc_msg_only_one: Connection timed out
> srun: error: slurm_send_recv_rc_msg_only_one: Connection timed out
> 
> It should not be trying to srun anything as we use mpiexec with Open-MPI.  Any ideas?
> 
> -david
> 
> --
> David Gunter
> HPC-3: Infrastructure Team
> Los Alamos National Laboratory
> 
> 
> 
> 
> 



More information about the petsc-users mailing list