How to loop Petsc in Fortran?

Letian Wang letian.wang at ghiocel-tech.com
Mon Apr 24 11:56:36 CDT 2006


Dear All:

 

Question 1):

 

For an optimization task, I need to loop Petsc (I'm using Petsc-2.3.0). But
I had problems to reinitialize Petsc after finalize, here is a simple
FORTRAN program to explain my problem:

 

      program petsc_test

#     include "include/finclude/petsc.h"

      call PetscInitialize(PETSC_NULL_CHARACTER, ierr)

      call PetscFinalize(ierr)

      print*,'ierr=',ierr

      call PetscInitialize(PETSC_NULL_CHARACTER, ierr)

      call PetscFinalize(ierr)

      end

 

When the program excuted to the second PetscInitial line, it shows error
message: "Error encountered before initializing MPICH". Can anyone help me
on this? Thanks.

 

Question 2):

 

Follow up my previous question, I also tried to Initiallize and Finalize
Petsc only once and perform the do-loop between Petscinitialize and
PetscFinalize. I used KSP CR solver with Prometheus PCs to solver large
linear equations. After several loops, the program was interrupted by
segmentation violation error. I suppose there was a memory leak somewhere.
The error message is like this: Any suggestion for this? Thanks

 

 

*********Doing job -- nosort0001

 

Task No.    1    Total CPU=       52.3    

 ---------------------------------------------------

 

 *********Doing job -- nosort0002

 

Task No.    2    Total CPU=       52.1    

 ---------------------------------------------------

 

 *********Doing job -- nosort0003

--------------------------------------------------------------------------

Petsc Release Version 2.3.0, Patch 44, April, 26, 2005

See docs/changes/index.html for recent updates.

See docs/faq.html for hints about trouble shooting.

See docs/index.html for manual pages.

-----------------------------------------------------------------------

../feap on a linux-gnu named GPTnode3.cl.ghiocel-tech.com by ltwang Mon Apr
24 15:25:04 2006

Libraries linked from /home/ltwang/Library/petsc-2.3.0/lib/linux-gnu

Configure run at Tue Mar 14 11:19:49 2006

Configure options --with-mpi-dir=/usr --with-debugging=0
--download-spooles=1 --download-f-blas-lapack=1 --download-parmetis=1
--download-prometheus=1 --with-shared=0

-----------------------------------------------------------------------

[1]PETSC ERROR: Caught signal number 11 SEGV: Segmentation Violation,
probably memory access out of range

[1]PETSC ERROR: Try option -start_in_debugger or -on_error_attach_debugger

[1]PETSC ERROR: configure using --with-debugging=yes, recompile, link, and
run 

[1]PETSC ERROR: to get more information on the crash.

[1]PETSC ERROR: User provided function() line 0 in unknown directory unknown
file

[1]PETSC ERROR: Signal received!

[1]PETSC ERROR:  !

[cli_1]: aborting job:

application called MPI_Abort(MPI_COMM_WORLD, 59) - process 1

[cli_0]: aborting job:

Fatal error in MPI_Allgather: Other MPI error, error stack:

MPI_Allgather(949)........................: MPI_Allgather(sbuf=0xbffeea14,
scount=1, MPI_INT, rbuf=0x8bf0a0c, rcount=1, MPI_INT, comm=0x84000000)
failed

MPIR_Allgather(180).......................: 

MPIC_Sendrecv(161)........................: 

MPIC_Wait(321)............................: 

MPIDI_CH3_Progress_wait(199)..............: an error occurred while handling
an event returned by MPIDU_Sock_Wait()

MPIDI_CH3I_Progress_handle_sock_event(422): 

MPIDU_Socki_handle_read(649)..............: connection failure
(set=0,sock=2,errno=104:(strerror() not found))

rank 1 in job 477  GPTMaster_53830   caused collective abort of all ranks

  exit status of rank 1: return code 59 

 

 

 

Letian 

 

 

 

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20060424/d5957cda/attachment.htm>


More information about the petsc-users mailing list