[mpich2-dev] More ROMIO performance questions
Kevin Harms
harms at alcf.anl.gov
Mon Sep 14 17:18:02 CDT 2009
Bob,
maybe this is a dumb question/suggestion but is the write
contention created by locking? (i'm assuming this is GPFS?) If this
test case doesn't require actual locking; can they test with
bglockless and see what happens to the performance?
kevin harms
On Sep 14, 2009, at 4:57 PM, Bob Cernohous wrote:
>
> We have another i/o scenario with interesting performance issues.
>
> One again, it's large non-interleaved contiguous blocks being
> written/read (checkpointing software). We ran into the same
> problems with data sieving and romio_cb_write/read = enable as we
> discussed a couple weeks ago.
>
> We tried to tune it with hints for cb_block_size and get ok
> performance when we can avoid read/write data sieving.
>
> Trying romio_cb_write/read = automatic gets very poor
> performance.Similarly, pure non-collective writes get very poor
> performance. It seems like having too many writers/readers performs
> poorly on their configuration ... so
>
> They customized the testcase to coordinate/flow-control the non-
> collective i/o and they get great performance. They only have N
> simultaneous writers/readers active. They pass a token around and
> take turns. It's almost like having N aggregators but without the
> collective i/o overhead to pass the data around. Instead they pass
> a small token and take turns writing the large, non-interleaved
> contiguous data blocks.
>
> I'm not aware of anything in MPIIO or ROMIO that would do tihs?
> Has this been explored by the experts (meaning you guys)?
>
>
>
> Bob Cernohous: (T/L 553) 507-253-6093
>
> BobC at us.ibm.com
> IBM Rochester, Building 030-2(C335), Department 61L
> 3605 Hwy 52 North, Rochester, MN 55901-7829
>
> > Chaos reigns within.
> > Reflect, repent, and reboot.
> > Order shall return.
More information about the mpich2-dev
mailing list