[petsc-users] log_summary time ratio and flops ratio

Matthew Knepley knepley at gmail.com
Wed Feb 10 08:31:23 CST 2016


On Wed, Feb 10, 2016 at 8:12 AM, Xiangdong <epscodes at gmail.com> wrote:

> On Mon, Feb 8, 2016 at 6:45 PM, Jed Brown <jed at jedbrown.org> wrote:
>
>> Xiangdong <epscodes at gmail.com> writes:
>>
>> > iii) since the time ratios of VecDot (2.5) and MatMult (1.5) are still
>> > high, I rerun the program with ipm module. The IPM summary is here:
>> >
>> https://drive.google.com/file/d/0BxEfb1tasJxhYXI0VkV0cjlLWUU/view?usp=sharing
>> .
>> > From this IPM reuslts, MPI_Allreduce takes 74% of MPI time. The
>> > communication by task figure (1st figure in p4) in above link showed
>> that
>> > it is not well-balanced. Is this related to the hardware and network
>> (which
>> > the users cannot control) or can I do something on my codes to improve?
>>
>> Here are a few functions that don't have any communication, but still
>> have significant load imbalance.
>>
>>   VecAXPY          1021815 1.0 2.2148e+01 2.1 1.89e+10 1.1 0.0e+00
>> 0.0e+00 0.0e+00  2  4  0  0  0   2  4  0  0  0 207057
>>   VecMAXPY          613089 1.0 1.3276e+01 2.2 2.27e+10 1.1 0.0e+00
>> 0.0e+00 0.0e+00  1  4  0  0  0   1  4  0  0  0 414499
>>   MatSOR            818390 1.0 1.9608e+02 1.5 2.00e+11 1.1 0.0e+00
>> 0.0e+00 0.0e+00 22 40  0  0  0  22 40  0  0  0 247472
>>
>>
> The result above is from a run with 256 cores (16 nodes * 16 cores/node).
> I did another run with 64 nodes * 4 cores/node. Now these functions are
> much better balanced ( a factor of 1.2-1.3, instead of 1.5-2.1).
>
> VecAXPY           987215 1.0 6.8469e+00 1.3 1.82e+10 1.1 0.0e+00 0.0e+00
> 0.0e+00  1  4  0  0  0   1  4  0  0  0 647096
> VecMAXPY          592329 1.0 6.0866e+00 1.3 2.19e+10 1.1 0.0e+00 0.0e+00
> 0.0e+00  1  4  0  0  0   1  4  0  0  0 873511
> MatSOR            790717 1.0 1.2933e+02 1.2 1.93e+11 1.1 0.0e+00 0.0e+00
> 0.0e+00 24 40  0  0  0  24 40  0  0  0 362525
>
> For the functions requires communication, the time ratio is about (1.4-1.6)
> VecDot            789772 1.0 8.4804e+01 1.4 1.46e+10 1.1 0.0e+00 0.0e+00
> 7.9e+05 14  3  0  0 40  14  3  0  0 40 41794
> VecNorm           597914 1.0 7.6259e+01 1.6 1.10e+10 1.1 0.0e+00 0.0e+00
> 6.0e+05 12  2  0  0 30  12  2  0  0 30 34996
>
> The full logsummary for this new run is here:
> https://googledrive.com/host/0BxEfb1tasJxhVkZ2NHJkSmF4LUU
>
> Can we say now the load imbalance is from the network communication,
> instead of memory bandwidth?
>

Actually now it looks even more like what Jed was saying. The 4 cores have
much more available bandwidth.

   Matt


> Thanks.
>
> Xiangdong
>
> You can and should improve load balance before stressing about network
>> costs.  This could be that the nodes aren't clean (running at different
>> speeds) or that the partition is not balancing data.
>>
>
>


-- 
What most experimenters take for granted before they begin their
experiments is infinitely more interesting than any results to which their
experiments lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20160210/1b6f03ea/attachment.html>


More information about the petsc-users mailing list