[petsc-users] How to use PETSc4py/SLEPc4py to solve an eigenvalue problem in parallel

Jed Brown jed at jedbrown.org
Wed Oct 29 22:52:48 CDT 2014


Mengqi Zhang <jollage at gmail.com> writes:
> It seems that the parallel running doesn't work, the elapsed time is
> roughly the same. 

http://www.mcs.anl.gov/petsc/documentation/faq.html#computers

> You see that in the process of assembling the matrix in the parallel
> case, the two cores are working according to
>
>     from  0  to  8192  on rank  0
>     from  8192  to  16384  on rank  1
>
> but in the SLEPc phase, only rank 0 is working (the print command only
> works for rank 0, but not for rank 1). 

That is just printing.  If every rank printed, you'd have a jumbled mess
of output.

Always send the output of -log_summary if you have questions about performance.

> I also tried to run the ex1.py in parallel and in serial, the
> computation times for a relative large matrix are also roughly the
> same for the two runnings, or in some cases, parallel running is even
> longer.

-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 818 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20141029/25beceae/attachment.pgp>


More information about the petsc-users mailing list