<html>
<head>
<meta http-equiv="Content-Type" content="text/html; charset=UTF-8">
</head>
<body>
<p>Hi Kevin,</p>
<p>thanks for the quick reply!<br>
</p>
<p>1. the IntelMPI version I have mentioned is based on MPICH 3.3. I
have attached the log (iimpi_error_log.txt).<br>
</p>
<p>2. I have attached the code as well (which just writes a couple
of lines to stdout) (hello_world.c).<br>
Compilation command: <i>mpicxx -fopenmp hello_world.c -o
hello_world.exe<br>
</i>Hint: With OpenMPI/4.1.0rc2 the configuration with APMPI and
execution works. The PDF report seems to be properly created.</p>
<p>3. If I interpret that correctly the data collected by APMPI are
not shown yet in the PDF report?<i><br>
</i>FYI: A couple of years ago I have written a python script that
extracts the data from the binary log file with darshan-parser to
get the raw data which you use to create the PDF report. I was
able to reproduce the statistics shown in the PDF and have added a
few more tables which helped us to get a deeper understanding of
applications IO at that time. Since I have not touched this script
for a long time it might not work anymore. That's why it also
interest for me to have a look at what pydarshan is offering.<br>
</p>
<p>Thomas<i><br>
</i></p>
<div class="moz-cite-prefix">Am 19.05.2021 um 16:46 schrieb Harms,
Kevin:<br>
</div>
<blockquote type="cite"
cite="mid:SA1PR09MB7709D0216B56278DA223BD278A2B9@SA1PR09MB7709.namprd09.prod.outlook.com">
<pre class="moz-quote-pre" wrap="">Thomas,
1. Not sure why the Intel MPI is tripping up on the configure check. I'm assuming it is MPI3 based. Can you send us the config.log output from that one? Maybe we can see why the check fails.
2. The partial log indicates the log is incorrect, so those parser errors are expected. I don't know why the finalize hangs. Was this a Fortran hello world example? I'm not familiar with ParaStationMPI but since it is based on MPICH, it should work. Can you send the test code and how you built it? We can try it on a system here.
3. Autoperf can't be disabled at runtime yet. We have a broader plan to add the ability to enable/disable modules during runtime, but not available yet. We have tested AutoPerf with CrayMPI, MPICH3.3 and OpenMPI. The systems we tested on were generic Linux laptop, Cray XC-40 and Nvidia DGX A100. As far as what can be done with APMPI data, we have some python analysis script based on pydarshan.
<a class="moz-txt-link-freetext" href="https://xgitlab.cels.anl.gov/AutoPerf/autoperf/-/blob/master/apmpi/util/apmpi-analysis.py">https://xgitlab.cels.anl.gov/AutoPerf/autoperf/-/blob/master/apmpi/util/apmpi-analysis.py</a>
The counters are also output by darshan-parser. We are still in the process of building more analysis based on this work.
kevin
________________________________________
From: Darshan-users <a class="moz-txt-link-rfc2396E" href="mailto:darshan-users-bounces@lists.mcs.anl.gov"><darshan-users-bounces@lists.mcs.anl.gov></a> on behalf of Thomas Breuer <a class="moz-txt-link-rfc2396E" href="mailto:t.breuer@fz-juelich.de"><t.breuer@fz-juelich.de></a>
Sent: Wednesday, May 19, 2021 6:52 AM
To: <a class="moz-txt-link-abbreviated" href="mailto:darshan-users@lists.mcs.anl.gov">darshan-users@lists.mcs.anl.gov</a>
Subject: [Darshan-users] darshan 3.3.0 issues
Dear Darshan Team,
I have installed the latest darshan version (3.3.0) for different MPIs on our HPC JUWELS (<a class="moz-txt-link-freetext" href="https://apps.fz-juelich.de/jsc/hps/juwels/configuration.html">https://apps.fz-juelich.de/jsc/hps/juwels/configuration.html</a>) and would like to report two issues:
1. Intel (19.1.3.304) Compiler with IntelMPI/2019.8.254:
- Configure Step fails for the new APMPI feature:
cd darshan-runtime; ./configure --prefix=/path/to/darshan-runtime/3.3.0-iimpi-2020-APMPI --with-mem-align=8 --with-log-path-by-env=DARSHAN_LOG_P
ATH --with-jobid-env=SLURM_JOBID CC=mpicc --enable-hdf5-mod=$EBROOTHDF5 --enable-apmpi-mod --enable-apmpi-coll-sync
- Error msg: configure: error: APMPI module requires MPI version 3+
- without the new APMPI Options the configure steps ends successfully:
cd darshan-runtime; ./configure --prefix=/p/software/juwels/stages/Devel-2020/software/darshan-runtime/3.3.0-iimpi-2020 --with-mem-align=8 --with-log-path-by-env=DARSHAN_LOG_PATH --with-jobid-env=SLURM_JOBID CC=mpicc --enable-hdf5-mod=$EBROOTHDF5
2. GCC/9.3.0 Compiler with ParaStationMPI/5.4.7-1 (based on MPICH 3.3.2) (<a class="moz-txt-link-freetext" href="https://github.com/ParaStation/psmpi/">https://github.com/ParaStation/psmpi/</a>):
- darshan-runtime configured with --enable-apmpi-mod --enable-apmpi-coll-sync
- For a simple helloworld code (MPI + OMP) the application seems to be hanging in the MPI_FINALIZE call.
- if I open the *.darshan_partial file with `darshan-parser`, then the following output is printed:
Error: incompatible darshan file.
Error: expected version 3.21, but got
Error: failed to read darshan log file header.
- There are no issues without APMPI.
3. Further questions:
- Is it possible to switch on/off APMPI during runtime?
- Are there any examples available that demonstrate the additional value that can be achieved by using the new AutoPerf feature?
- Can you confirm that APMPI works on none-Cray systems ?
Best regards,
Thomas
--
Thomas Breuer
Division Application Support Forschungszentrum Jülich GmbH
Jülich Supercomputing Centre (JSC) Wilhelm-Johnen-Straße
<a class="moz-txt-link-freetext" href="http://www.fz-juelich.de/ias/jsc">http://www.fz-juelich.de/ias/jsc</a> 52425 Jülich (Germany)
Phone: +49 2461 61-96742 (currently not available via phone)
Email: <a class="moz-txt-link-abbreviated" href="mailto:t.breuer@fz-juelich.de">t.breuer@fz-juelich.de</a><a class="moz-txt-link-rfc2396E" href="mailto:t.breuer@fz-juelich.de"><mailto:t.breuer@fz-juelich.de></a>
-------------------------------------------------------------------------------------
-------------------------------------------------------------------------------------
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Volker Rieke
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt, Prof. Dr. Frauke Melchior
-------------------------------------------------------------------------------------
-------------------------------------------------------------------------------------
</pre>
</blockquote>
<pre class="moz-signature" cols="72">--
Thomas Breuer
Division Application Support Forschungszentrum Jülich GmbH
Jülich Supercomputing Centre (JSC) Wilhelm-Johnen-Straße
<a class="moz-txt-link-freetext" href="http://www.fz-juelich.de/ias/jsc">http://www.fz-juelich.de/ias/jsc</a> 52425 Jülich (Germany)
Phone: +49 2461 61-96742 (currently not available via phone)
Email: <a class="moz-txt-link-abbreviated" href="mailto:t.breuer@fz-juelich.de">t.breuer@fz-juelich.de</a>
-------------------------------------------------------------------------------------
-------------------------------------------------------------------------------------
Forschungszentrum Juelich GmbH
52425 Juelich
Sitz der Gesellschaft: Juelich
Eingetragen im Handelsregister des Amtsgerichts Dueren Nr. HR B 3498
Vorsitzender des Aufsichtsrats: MinDir Volker Rieke
Geschaeftsfuehrung: Prof. Dr.-Ing. Wolfgang Marquardt (Vorsitzender),
Karsten Beneke (stellv. Vorsitzender), Prof. Dr.-Ing. Harald Bolt, Prof. Dr. Frauke Melchior
-------------------------------------------------------------------------------------
-------------------------------------------------------------------------------------</pre>
</body>
</html>