MPI Initialisation Problem
Satish Balay
balay at mcs.anl.gov
Sat Jun 16 13:22:40 CDT 2007
On Sat, 16 Jun 2007, Tim Stitt wrote:
> Hi all,
>
> I am having some difficulty getting my parallel eigensolver to run over
> multiple processes.
>
> When I execute my parallel code on my distributed-memory machine ( with > 1
> processes) I keep getting the following runtime message:
>
> mpiexec: Warning: task 0 exited before completing MPI startup.
> mpiexec: Warning: task 1 exited oddly---report bug: status 0 done 0.
>
> MPI_Comm_Size returns 1 for #processes even though I mpiexec -n 2 or higher.
>
> The code still runs to completion but serially.......
can you reproduce this problem with a PETSc example? If so, send us
the complete log of your build [configure.log, make_log_PETSC_ARCH]
and the log for test examples [test_log_PETSC_ARCH] to
petsc-maint at mcs.anl.gov
Satish
>
> I have taken the same code and ran it on my shared-memory machine with no
> problems and all processes getting picked up so I know it is not a coding
> problem.....the sample ksp test codes that come with my petsc (2.3.3)
> distribution also exhibit this problem.
>
> So there seems to be a problem with the PETScInitialize routine with this
> particular architecture. I have tried both pathscale and pgi compilers with
> the same result. As far as I can see there are no noticeable warnings
> generated during the config and make phases. Others use MPI regularly on our
> cluster so I don't see how it could be a MPI library issue.
>
> Any thoughts on what could be happening gratefully received.
>
> Regards,
>
> Tim.
>
>
More information about the petsc-users
mailing list