[petsc-users] Poor weak scaling when solvingsuccessivelinearsystems

Junchao Zhang jczhang at mcs.anl.gov
Wed May 30 07:53:00 CDT 2018


If you have an example code and can share that, I can test and further
profile it on our machine.

--Junchao Zhang

On Wed, May 30, 2018 at 3:27 AM, Michael Becker <
Michael.Becker at physik.uni-giessen.de> wrote:

> Barry: On its way. Could take a couple days again.
>
> Junchao: I unfortunately don't have access to a cluster with a faster
> network. This one has a mixed 4X QDR-FDR InfiniBand 2:1 blocking fat-tree
> network, which I realize causes parallel slowdown if the nodes are not
> connected to the same switch. Each node has 24 processors (2x12/socket) and
> four NUMA domains (two for each socket).
> The ranks are usually not distributed perfectly even, i.e. for 125
> processes, of the six required nodes, five would use 21 cores and one 20.
> Would using another CPU type make a difference communication-wise? I could
> switch to faster ones (on the same network), but I always assumed this
> would only improve performance of the stuff that is unrelated to
> communication.
>
> Michael
>
>
>
> The log files have something like "Average time for zero size MPI_Send():
> 1.84231e-05". It looks you ran on a cluster with a very slow network. A
> typical machine should give less than 1/10 of the latency you have. An easy
> way to try is just running the code on a machine with a faster network and
> see what happens.
>
> Also, how many cores & numa domains does a compute node have? I could not
> figure out how you distributed the 125 MPI ranks evenly.
>
> --Junchao Zhang
>
> On Tue, May 29, 2018 at 6:18 AM, Michael Becker <
> Michael.Becker at physik.uni-giessen.de> wrote:
>
>> Hello again,
>>
>> here are the updated log_view files for 125 and 1000 processors. I ran
>> both problems twice, the first time with all processors per node allocated
>> ("-1.txt"), the second with only half on twice the number of nodes
>> ("-2.txt").
>>
>> On May 24, 2018, at 12:24 AM, Michael Becker <Michael.Becker at physik.uni-giessen.de> <Michael.Becker at physik.uni-giessen.de> wrote:
>>
>> I noticed that for every individual KSP iteration, six vector objects are created and destroyed (with CG, more with e.g. GMRES).
>>
>>    Hmm, it is certainly not intended at vectors be created and destroyed within each KSPSolve() could you please point us to the code that makes you think they are being created and destroyed?   We create all the work vectors at KSPSetUp() and destroy them in KSPReset() not during the solve. Not that this would be a measurable distance.
>>
>>
>> I mean this, right in the log_view output:
>>
>> Memory usage is given in bytes:
>>
>> Object Type Creations Destructions Memory Descendants' Mem.
>> Reports information only for process 0.
>>
>> --- Event Stage 0: Main Stage
>>
>> ...
>>
>> --- Event Stage 1: First Solve
>>
>> ...
>>
>> --- Event Stage 2: Remaining Solves
>>
>> Vector 23904 23904 1295501184 0.
>>
>> I logged the exact number of KSP iterations over the 999 timesteps and
>> its exactly 23904/6 = 3984.
>>
>> Michael
>>
>>
>>
>> Am 24.05.2018 um 19:50 schrieb Smith, Barry F.:
>>
>>   Please send the log file for 1000 with cg as the solver.
>>
>>    You should make a bar chart of each event for the two cases to see which ones are taking more time and which are taking less (we cannot tell with the two logs you sent us since they are for different solvers.)
>>
>>
>>
>>
>> On May 24, 2018, at 12:24 AM, Michael Becker <Michael.Becker at physik.uni-giessen.de> <Michael.Becker at physik.uni-giessen.de> wrote:
>>
>> I noticed that for every individual KSP iteration, six vector objects are created and destroyed (with CG, more with e.g. GMRES).
>>
>>    Hmm, it is certainly not intended at vectors be created and destroyed within each KSPSolve() could you please point us to the code that makes you think they are being created and destroyed?   We create all the work vectors at KSPSetUp() and destroy them in KSPReset() not during the solve. Not that this would be a measurable distance.
>>
>>
>>
>>
>> This seems kind of wasteful, is this supposed to be like this? Is this even the reason for my problems? Apart from that, everything seems quite normal to me (but I'm not the expert here).
>>
>>
>> Thanks in advance.
>>
>> Michael
>>
>>
>>
>> <log_view_125procs.txt><log_view_1000procs.txt>
>>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20180530/1a1cb1d6/attachment-0001.html>


More information about the petsc-users mailing list