[Darshan-users] Darshan Intel MPI, SLURM, Fortran
Harms, Kevin
harms at alcf.anl.gov
Fri Apr 28 10:07:15 CDT 2017
Alain,
Can you reproduce the crash by just running the binary on the login node? If so, you can run the test under GDB and see if you can get a backtrace. How are you setting the LD_PRELOAD variable? srun --export=LD_PRELOAD=<path> ?
kevin
-----Original Message-----
From: <darshan-users-bounces at lists.mcs.anl.gov> on behalf of Alain REFLOCH <alain.refloch at onera.fr>
Date: Friday, April 28, 2017 at 8:03 AM
To: "darshan-users at lists.mcs.anl.gov" <darshan-users at lists.mcs.anl.gov>
Subject: [Darshan-users] Darshan Intel MPI, SLURM, Fortran
I try my first test with Darshan (version darshan-3.1.4).
No problem for install
I use Intel compiler v 15 and Intel MPI, only a message at the compilation :
lib/darshan-mpiio.c(230): warning #147: declaration is incompatible with "int MPI_File_open(MPI_Comm={int}, const char *, int, MPI_Info={int}, MPI_File *)" (declared at line 192 of "/opt/software/common/intel/impi/5.1.3.258/include64/mpio.h")
Resolve by put option : -DHAVE_MPIIO_CONST
I
test with the file io-sample.c in the directory darshan-3.1.4/darshan-test
all is ok
For install I have do this :
./configure --prefix=$HOME/Darshan --with-mem-align=8 --with-log-path=$SCRATCHDIR/Darshan --with-jobid-env=SLURM_JOB_ID --with-zlib=/opt/software/occigen/libraries/zlib/1.2.8/intel/17.0/nompi CC='mpiicc -DHAVE_MPIIO_CONST'
I am on a Bull supercomputer with Haswell processor, with slum for the batch. I have use SLURM_JOB_ID or NONE
It's OK, I have a file in the repository ....../year/months/days.
After I try on my big program. All it's ok with no LD_PRELOAD, but with PRELOAD
I have No message but the job is kill by SLURM without message.
The main is in Fortran, call to mpi_init and finalize is in Fortran.
But in the code there is a part in C/C++ for the mesh file with MPI_IO (all parts
open, write, read, close for this file are in C/C++), I am interested by the Darshan informations
on this file (no on the ascii files open in Fortran). There is in the code a parallel partitionner
for domains decompostion (ParMetis), there is no recompilation of this part.
My question is : is Darshan ok for this configuration (main in Fortran, partial recompilation
of the MPI_IO part write in C)
bests regards
Alain
More information about the Darshan-users
mailing list