Thanks Satish for the reply. However, I also observed the same behavior with gamg and sor preconditioners and ksp_type bcgs as well as gmres. Could you tell which solver and preconditioners would behave same on whatever number of mpi I use?<div><br></div><div>Thanks, Mari<br><br><br>On Wednesday, April 17, 2019, Balay, Satish <<a href="mailto:balay@mcs.anl.gov">balay@mcs.anl.gov</a>> wrote:<br><blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Yes - the default preconditioner is block-jacobi - with one block on<br>
each processor.<br>
<br>
So when run on 1 proc vs 8 proc - the preconditioner is different<br>
(with 1block for bjacobi vs 8blocks for bjacobi)- hence difference in<br>
convergence.<br>
<br>
Satish<br>
<br>
On Wed, 17 Apr 2019, Marian Greg via petsc-users wrote:<br>
<br>
> Hi All,<br>
> <br>
> I am facing strange behavior of the ksp solvers with increasing number of<br>
> MPI. The solver is taking more and more iterations with increase in number<br>
> of MPIs. Is that a normal situation? I was expecting to get the same number<br>
> of iteration with whatever number of MPIs I use.<br>
> <br>
> E.g.<br>
> My matrix has about 2 million dofs<br>
> Solving with np 1 takes about 3500 iteration while solving with np 4 takes<br>
> 6500 iterations for the same convergence criteria.<br>
> <br>
> Thanks<br>
> Mari<br>
> <br>
<br>
</blockquote></div>