performance issue
Wei-Keng Liao
wkliao at northwestern.edu
Mon Aug 14 14:26:47 CDT 2023
I found results of using GPFS from one of your previous emails.
> GPFS (/glade/work on derecho):
> RESULT: write SUBSET 1 16 64 4570.2078677815 4.0610844270
> RESULT: write SUBSET 1 16 64 4470.3231494386 4.1518251320
I checked the ROMIO source codes. The codes for copying data look the same
to me, although they are not shared among different file system drivers.
Strange.
Wei-keng
> On Aug 14, 2023, at 2:11 PM, Wei-Keng Liao <wkliao at northwestern.edu> wrote:
>
> Did you run the same tests on a non-Lustre file system and see no difference?
> Can you show me the timings?
>
> Wei-keng
>
>> On Aug 14, 2023, at 11:54 AM, Jim Edwards <jedwards at ucar.edu> wrote:
>>
>> Hi Wei-Keng,
>>
>> Thanks for looking into this. Because the allocations in pioperformance.F90 are done on the compute nodes and
>> not the IO nodes I don't think that your suggestion would make any difference. I also wonder why this issue
>> appears to be so specific to the lustre file system - presumably the ROMIO functionality you speak of is general and not
>> specific to lustre? Anyway your analysis spurred me to try something else which seems to work: prior to calling
>> ncmpi_iput_varn in pio_darray_int.c I added a call to ncmpi_wait_all to make sure that any existing buffer was written.
>>
>> This seems to have fixed the problem and my writes are now
>> RESULT: write SUBSET 1 16 64 4787.6393342631 3.8766495770
>> RESULT: write SUBSET 1 16 64 4803.9296372205 3.8635037150
>>
>
More information about the parallel-netcdf
mailing list