[petsc-users] Communication time with threading enabled.
Shri
abhyshr at mcs.anl.gov
Thu Feb 14 10:32:02 CST 2013
This has something to do with OpenMPI. I cannot reproduce this issue with MPICH. Can you try switching to MPICH (--download-mpich).
Shri
On Feb 14, 2013, at 6:58 AM, Gonçalo Albuquerque wrote:
> Dear All,
>
> I'm experimenting with PETSc hybrid MPI/OpenMP capabilities and I have run a rather simple test case (2D magnetostatic) using PETSc compiled with both OpenMPI and thread support (both PThreads and OpenMP) on a Ubuntu 12.04 system. I cannot figure out the results obtained when comparing runs made using the same number of MPI processes (2) and specifying either no threads (-threadcomm_type nothread) or 1 OpenMP thread (-threadcomm_type openmp -threadcomm_nthreads 1). I attached the logs of both runs. It seems that the communication time has literally exploded. A grep over the logs gives:
>
> No threading:
> Average time for MPI_Barrier(): 1.38283e-06
> Average time for zero size MPI_Send(): 7.03335e-06
>
> 1 OpenMP thread:
> Average time for MPI_Barrier(): 0.00870218
> Average time for zero size MPI_Send(): 0.00614798
>
> The same things is occurring when running ksp ex5 (see attached logs).
>
> Any ideas as to what I'm missing?
>
> Many thanks in advance,
>
> Gonçalo
> <nothread.log><openmp_nthreads_1.log><ex5_nothread.log><ex5_openmp_nthreads_1.log>
More information about the petsc-users
mailing list