[petsc-users] parallel IO messages

Fande Kong fdkong.jd at gmail.com
Fri Nov 27 13:21:58 CST 2015


Hi Dave,

This not always happens. I am trying to get performance measurement so that
the PETSc is compiled with --with-debugging=no. I will try later.

Thanks,
Fande,

On Fri, Nov 27, 2015 at 12:08 PM, Dave May <dave.mayhem23 at gmail.com> wrote:

> There is little information in this stack trace.
> You would get more information if you use a debug build of petsc.
> e.g. configure with --with-debugging=yes
> It is recommended to always debug problems using a debug build of petsc
> and a debug build of your application.
>
> Thanks,
>   Dave
>
> On 27 November 2015 at 20:05, Fande Kong <fdkong.jd at gmail.com> wrote:
>
>> Hi all,
>>
>> I implemented a parallel IO based on the Vec and IS which uses HDF5. I am
>> testing this loader on a supercomputer. I occasionally (not always)
>> encounter the following errors (using 8192 cores):
>>
>> [7689]PETSC ERROR:
>> ------------------------------------------------------------------------
>> [7689]PETSC ERROR: Caught signal number 5 TRAP
>> [7689]PETSC ERROR: Try option -start_in_debugger or
>> -on_error_attach_debugger
>> [7689]PETSC ERROR: or see
>> http://www.mcs.anl.gov/petsc/documentation/faq.html#valgrind
>> [7689]PETSC ERROR: or try http://valgrind.org on GNU/linux and Apple Mac
>> OS X to find memory corruption errors
>> [7689]PETSC ERROR: configure using --with-debugging=yes, recompile, link,
>> and run
>> [7689]PETSC ERROR: to get more information on the crash.
>> [7689]PETSC ERROR: --------------------- Error Message
>> --------------------------------------------------------------
>> [7689]PETSC ERROR: Signal received
>> [7689]PETSC ERROR: See
>> http://www.mcs.anl.gov/petsc/documentation/faq.html for trouble shooting.
>> [7689]PETSC ERROR: Petsc Release Version 3.6.2, unknown
>> [7689]PETSC ERROR: ./fsi on a arch-linux2-cxx-opt named ys6103 by fandek
>> Fri Nov 27 11:26:30 2015
>> [7689]PETSC ERROR: Configure options --with-clanguage=cxx
>> --with-shared-libraries=1 --download-fblaslapack=1 --with-mpi=1
>> --download-parmetis=1 --download-metis=1 --with-netcdf=1
>> --download-exodusii=1
>> --with-hdf5-dir=/glade/apps/opt/hdf5-mpi/1.8.12/intel/12.1.5
>> --with-debugging=no --with-c2html=0 --with-64-bit-indices=1
>> [7689]PETSC ERROR: #1 User provided function() line 0 in  unknown file
>> Abort(59) on node 7689 (rank 7689 in comm 1140850688): application called
>> MPI_Abort(MPI_COMM_WORLD, 59) - process 7689
>> ERROR: 0031-300  Forcing all remote tasks to exit due to exit code 1 in
>> task 7689
>>
>> Make and configure logs are attached.
>>
>> Thanks,
>>
>> Fande Kong,
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20151127/78982dc7/attachment-0001.html>


More information about the petsc-users mailing list