[petsc-users] log_summary time ratio and flops ratio
Jed Brown
jed at jedbrown.org
Mon Feb 8 17:45:29 CST 2016
Xiangdong <epscodes at gmail.com> writes:
> iii) since the time ratios of VecDot (2.5) and MatMult (1.5) are still
> high, I rerun the program with ipm module. The IPM summary is here:
> https://drive.google.com/file/d/0BxEfb1tasJxhYXI0VkV0cjlLWUU/view?usp=sharing.
> From this IPM reuslts, MPI_Allreduce takes 74% of MPI time. The
> communication by task figure (1st figure in p4) in above link showed that
> it is not well-balanced. Is this related to the hardware and network (which
> the users cannot control) or can I do something on my codes to improve?
Here are a few functions that don't have any communication, but still
have significant load imbalance.
VecAXPY 1021815 1.0 2.2148e+01 2.1 1.89e+10 1.1 0.0e+00 0.0e+00 0.0e+00 2 4 0 0 0 2 4 0 0 0 207057
VecMAXPY 613089 1.0 1.3276e+01 2.2 2.27e+10 1.1 0.0e+00 0.0e+00 0.0e+00 1 4 0 0 0 1 4 0 0 0 414499
MatSOR 818390 1.0 1.9608e+02 1.5 2.00e+11 1.1 0.0e+00 0.0e+00 0.0e+00 22 40 0 0 0 22 40 0 0 0 247472
You can and should improve load balance before stressing about network
costs. This could be that the nodes aren't clean (running at different
speeds) or that the partition is not balancing data.
-------------- next part --------------
A non-text attachment was scrubbed...
Name: signature.asc
Type: application/pgp-signature
Size: 818 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20160208/5efced4c/attachment.pgp>
More information about the petsc-users
mailing list