[mpich-discuss] Performance issue when writing large files with MPI-IO/ROMIO and OrangeFS (PVFS)

Rob Latham robl at mcs.anl.gov
Thu Mar 15 16:16:21 CDT 2012


On Wed, Mar 14, 2012 at 06:18:34PM +0100, Louis-Claude Canon wrote:
> I am seeing significant variability and low performance when writing
> large files, whereas it is stable with small ones (the threshold
> seems to be around 100 MB) with one server and one client.

> IOR -a MPIIO -i 10 -o pvfs2:/mnt/pvfs2/iortest -t 10000000 -b 10000000 
># 10 MB
> Operation  Max (MiB)  Min (MiB)  Mean (MiB)   Std Dev
> write          92.06      88.43       91.61      1.06

> IOR -a MPIIO -i 10 -o pvfs2:/mnt/pvfs2/iortest -t 200000000 -b 200000000 
> # 200 MB
> write          94.67      35.72       68.07     26.17

Writes are funny operations.  They may or may not get cached.  What
might be happening is that once you've exceeded a certain size, the
PVFS servers will start exhausting cache and have to actually write to
disk.

Doesn't IOR report a bandwith that also includes open/close times?  If
so, I suspect you will get more consistent values that way.

> I am not sure if it is related to OrangeFS or ROMIO. When I use
> pvfs2-cp with a file of size 2GB, the bandwidth is correct and
> stable, which suggests that it comes from ROMIO:
> dd if=/dev/zero of=/tmp/test bs=1000000 count=2000
> pvfs2-cp -t /tmp/test /mnt/pvfs2/test
> Wrote 2000000000 bytes in 19.527506 seconds. 97.674973 MB/seconds
> 
> On the other hand when I enable TroveSyncData or when I put
> TroveMethod to directio with OrangeFS, the variability disappears.

This setting further makes me suspect the server-size VFS cache: both
those settings (in different ways) make the VFS cache irrelevant.

==rob

-- 
Rob Latham
Mathematics and Computer Science Division
Argonne National Lab, IL USA


More information about the mpich-discuss mailing list