a problem with error of "An error occurred in MPI_Comm_rank after MPI was finalized"

Yin Feng yfeng1 at tigers.lsu.edu
Sun Oct 11 22:07:02 CDT 2009


yes, exactly

On Sun, Oct 11, 2009 at 10:04 PM, Yaakoub El Khamra <yelkhamra at gmail.com>wrote:

> This is for Mayank's project isn't it
>
> Regards
> Yaakoub El Khamra
>
>
>
>
> On Sun, Oct 11, 2009 at 9:59 PM, Yin Feng <yfeng1 at tigers.lsu.edu> wrote:
> > The completed code is:
> >
> > void PETSc(double mat[],double vec[],double sol[],long N,double tol) {
> >
> > Vec            x,b;
> >
> > Mat            A;
> >
> > KSP            ksp;
> >
> > PC             pc;
> >
> > PetscInt       i,j,col[N];
> >
> > PetscScalar    value[N];
> >
> > PetscScalar    val;
> >
> > PetscInitialize(PETSC_NULL,PETSC_NULL,PETSC_NULL,PETSC_NULL);
> >
> > VecCreate(PETSC_COMM_WORLD,&x);
> >
> > VecSetSizes(x,PETSC_DECIDE,N);
> >
> > VecSetFromOptions(x);
> >
> > VecDuplicate(x,&b);
> >
> > MatCreate(PETSC_COMM_WORLD,&A);
> >
> > MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,N,N);
> >
> > MatSetFromOptions(A);
> >
> > for (i=0; i<N; i++) {
> >
> > for (j=0;j<N;j++) {
> >
> > value[j]=mat[i*N+j];
> >
> > col[j]=j;
> >
> > }
> >
> > MatSetValues(A,1,&i,N,col,value,INSERT_VALUES);
> >
> > }
> >
> > MatAssemblyBegin(A,MAT_FINAL_ASSEMBLY);
> >
> > MatAssemblyEnd(A,MAT_FINAL_ASSEMBLY);
> >
> >
> >
> > for (i=0; i<N; i++) {
> >
> > val=vec[i];
> >
> > VecSetValues(b,1,&i,&val,INSERT_VALUES);
> >
> > val=sol[i];
> >
> > VecSetValues(x,1,&i,&val,INSERT_VALUES);
> >
> > }
> >
> > VecAssemblyBegin(b);
> >
> > VecAssemblyEnd(b);
> >
> > KSPCreate(PETSC_COMM_WORLD,&ksp);
> >
> > KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN);
> >
> > KSPGetPC(ksp,&pc);
> >
> >         KSPSetType(ksp, KSPGMRES);
> >
> >         PCSetType(pc,  PCBJACOBI);
> >
> >  KSPSetTolerances(ksp,tol,PETSC_DEFAULT,PETSC_DEFAULT,PETSC_DEFAULT);
> >
> >   KSPSetFromOptions(ksp);
> >
> >   KSPSolve(ksp,b,x);
> >
> >   for (i=0;i<N;i++) {
> >
> >       VecGetValues(x,1,&i,&sol[i]);
> >
> >   }
> >
> >
> >
> >   VecDestroy(x);
> >
> >   VecDestroy(b);
> >
> >   MatDestroy(A);
> >
> >   KSPDestroy(ksp);
> >
> >   PetscFinalize();
> >
> > }
> >
> > Once this function is call in a loop, it reports error.
> > On Sun, Oct 11, 2009 at 9:38 PM, Yin Feng <yfeng1 at tigers.lsu.edu> wrote:
> >>
> >> I put PETSc solver in one function and use another function to call
> that.
> >> This problem only appears when I put the function with PETSc solver in a
> >> loop,
> >> it works well at first step, and reports error "An error occurred in
> >> MPI_Comm_rank after MPI was finalized"
> >> at second time. The program is designed to support only one processor
> >> like:
> >>
> >> Vec            x,b;
> >>
> >> Mat            A;
> >>
> >> KSP            ksp;
> >>
> >> PC             pc;
> >>
> >> PetscInt       i,j,col[N];
> >>
> >> PetscScalar    value[N];
> >>
> >> PetscScalar    val;
> >>
> >> PetscInitialize(PETSC_NULL,PETSC_NULL,PETSC_NULL,PETSC_NULL);
> >>
> >> VecCreate(PETSC_COMM_WORLD,&x);
> >>
> >> VecSetSizes(x,PETSC_DECIDE,N);
> >>
> >> VecSetFromOptions(x);
> >>
> >> VecDuplicate(x,&b);
> >>
> >> MatCreate(PETSC_COMM_WORLD,&A);
> >>
> >> MatSetSizes(A,PETSC_DECIDE,PETSC_DECIDE,N,N);
> >>
> >> MatSetFromOptions(A);
> >>
> >>         ........
> >>
> >>         ........
> >>
> >> KSPCreate(PETSC_COMM_WORLD,&ksp);
> >>
> >> KSPSetOperators(ksp,A,A,DIFFERENT_NONZERO_PATTERN);
> >>
> >> KSPGetPC(ksp,&pc);
> >>
> >>         ................
> >>         ...............
> >>
> >>
> >>  KSPSetTolerances(ksp,tol,PETSC_DEFAULT,PETSC_DEFAULT,PETSC_DEFAULT);
> >>
> >>         KSPSetFromOptions(ksp);
> >>
> >>         KSPSolve(ksp,b,x);
> >>
> >>         VecDestroy(x);
> >>
> >>         VecDestroy(b);
> >>
> >>         MatDestroy(A);
> >>
> >>         KSPDestroy(ksp);
> >>
> >>         PetscFinalize();
> >>
> >> Any one has ideal about this?
> >>
> >> The detailed error description is:
> >>
> >> An error occurred in MPI_Comm_rank
> >>
> >> *** after MPI was finalized
> >>
> >> *** MPI_ERRORS_ARE_FATAL (goodbye)
> >>
> >> Abort before MPI_INIT completed successfully; not able to guarantee that
> >> all other processes were killed!
> >>
> >> Thank you so much in advance!
> >
> >
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20091011/6269c394/attachment.htm>


More information about the petsc-users mailing list