[mpich-discuss] MPI_File_open() fails on local + NFS file system

Rajeev Thakur thakur at mcs.anl.gov
Thu Apr 21 16:31:38 CDT 2011


Try adding the prefix "nfs:" to file name passed to MPI_File_open.

Rajeev

On Apr 21, 2011, at 4:20 PM, Audet, Martin wrote:

> Hi MPICH_Developers,
> 
> We are unable to use MPI_File_open() on a cluster where the first node (master node) mount a local file system and export it via NFS to a few cluster nodes so that /home on both the master node and the compute node refers to the same directory.
> 
> When a job composed on one (or more) process on the master node and one (or more) process on a client node is started, MPI_File_open() to create a new file either make the program to abort (if the process of rank 0 is on the master node using the local file system) or to freeze (if the process of rank 0 is on a compute node acessing the file via NFS).
> 
> When the program freeze, an inspection with gdb shows that the process of rank 0 is stuck into a MPI_Bcast() called by MPI_File_open().
> 
> Note that this happens with many mpich2 versions from 1.0.7 to 1.4rc2.
> 
> So we would like to know if the configuration we use is local or not.
> 
> Thanks,
> 
> Martin Audet
> _______________________________________________
> mpich-discuss mailing list
> mpich-discuss at mcs.anl.gov
> https://lists.mcs.anl.gov/mailman/listinfo/mpich-discuss



More information about the mpich-discuss mailing list