[mpich2-dev] More ROMIO performance questions

Kevin Harms harms at alcf.anl.gov
Mon Sep 14 17:18:02 CDT 2009


Bob,

   maybe this is a dumb question/suggestion but is the write  
contention created by locking? (i'm assuming this is GPFS?) If this  
test case doesn't require actual locking; can they test with  
bglockless and see what happens to the performance?

kevin harms

On Sep 14, 2009, at 4:57 PM, Bob Cernohous wrote:

>
> We have another i/o scenario with interesting performance issues.
>
> One again, it's large non-interleaved contiguous blocks being  
> written/read (checkpointing software).  We ran into the same  
> problems with data sieving and romio_cb_write/read = enable as we  
> discussed a couple weeks ago.
>
> We tried to tune it with hints for cb_block_size and get ok  
> performance when we can avoid read/write data sieving.
>
> Trying romio_cb_write/read = automatic gets very poor  
> performance.Similarly, pure non-collective writes get very poor  
> performance.  It seems like having too many writers/readers performs  
> poorly on their configuration ... so
>
> They customized the testcase to coordinate/flow-control the non- 
> collective i/o and they get great performance.   They only have N  
> simultaneous writers/readers active.  They pass a token around and  
> take turns.  It's almost like having N aggregators but without the  
> collective i/o overhead to pass the data around.  Instead they pass  
> a small token and take turns writing the large, non-interleaved  
> contiguous data blocks.
>
> I'm not aware of anything in MPIIO or ROMIO that would do tihs?    
> Has this been explored by the experts (meaning you guys)?
>
>
>
> Bob Cernohous:  (T/L 553) 507-253-6093
>
> BobC at us.ibm.com
> IBM Rochester, Building 030-2(C335), Department 61L
> 3605 Hwy 52 North, Rochester,  MN 55901-7829
>
> > Chaos reigns within.
> > Reflect, repent, and reboot.
> > Order shall return.



More information about the mpich2-dev mailing list