performance issue
Wei-Keng Liao
wkliao at northwestern.edu
Fri Aug 11 12:13:19 CDT 2023
OK. I will test it myself on Perlmutter.
Do you have a small test program to reproduce or is it still pioperf?
If pioperf, are the build instructions on Perlmutter the same?
Please let me know how you run on Perlmutter, i.e. no. process, nodes,
Lustre striping, problem size, etc.
Does "1 16 64" in your results mean 16 I/O tasks and 64 variables,
and only 16 MPI processes out of total ? processes call PnetCDF APIs?
Wei-keng
On Aug 11, 2023, at 9:35 AM, Jim Edwards <jedwards at ucar.edu> wrote:
I tried on perlmutter and am seeing the same issue only maybe even worse:
RESULT: write SUBSET 1 16 64 1261.0737058071 14.7176171500
RESULT: write SUBSET 1 16 64 90.3736534450 205.3695882870
On Fri, Aug 11, 2023 at 8:17 AM Jim Edwards <jedwards at ucar.edu<mailto:jedwards at ucar.edu>> wrote:
Hi Wei-Keng,
I released that the numbers in this table are all showing the slow performing file and the fast file
(the one without the scalar variable) are not represented - I will rerun and present these numbers again.
Here are corrected numbers for a few cases:
GPFS (/glade/work on derecho):
RESULT: write SUBSET 1 16 64 4570.2078677815 4.0610844270
RESULT: write SUBSET 1 16 64 4470.3231494386 4.1518251320
Lustre, default PFL's:
RESULT: write SUBSET 1 16 64 2808.6570137094 6.6081404420
RESULT: write SUBSET 1 16 64 1025.1671656858 18.1043644600
LUSTRE, no PFL's and very wide stripe:
RESULT: write SUBSET 1 16 64 4687.6852437580 3.9593102000
RESULT: write SUBSET 1 16 64 3001.4741125579 6.1836282120
On Thu, Aug 10, 2023 at 11:34 AM Jim Edwards <jedwards at ucar.edu<mailto:jedwards at ucar.edu>> wrote:
the stripe settings
lfs setstripe -c 96 -S 128M
logs/c96_S128M/
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/parallel-netcdf/attachments/20230811/2e904f03/attachment.html>
More information about the parallel-netcdf
mailing list