[petsc-users] Collective abort of all ranks
Matthew Knepley
knepley at gmail.com
Fri Jul 9 05:43:23 CDT 2010
On Fri, Jul 9, 2010 at 12:42 AM, Немања Илић (Nemanja Ilic) <
nemanja.ilic.81 at gmail.com> wrote:
> Hello,
>
> I am developing a PETSc application and I encountered the following
> problem:
> My parallel program does everything ok, finishes the meaningful job, and
> near the end it breaks with the following line:
>
> "rank 0 in job 17 wg_49925 caused collective abort of all ranks
> exit status of rank 0: killed by signal 9"
>
> It happends regardless of number of processes in the job.
>
> Here is the ending of the function 'calculate' which does the work:
>
> "....
> // write result to file
> PetscViewer viewer;
> ierr = PetscViewerASCIIOpen(PETSC_COMM_WORLD, output_file_name,
> &viewer); CHKERRQ(ierr);
> ierr = PetscViewerSetFormat(viewer, PETSC_VIEWER_ASCII_COMMON);
> CHKERRQ(ierr);
> ierr = VecView(bp, viewer); CHKERRQ(ierr);
>
>
> // clean-up
> ierr = MatDestroy(Ap); CHKERRQ(ierr);
> ierr = VecDestroy(bp); CHKERRQ(ierr);
> ierr = VecDestroy(xp); CHKERRQ(ierr);
>
> free(col);
> free(row);
> free(row_values);
>
Did you malloc() or PetscMalloc() these? You cannot free() stuff you get
from PetscMalloc().
Matt
> return b;"
>
> The file 'viewer' is outputed correctly.
> When 'calculate' finishes the finalization function gets called:
> "ierr = PetscFinalize(); CHKERRQ(ierr);"
>
> Does anyone have any idea to why my program breaks?
>
> Thank you in advance,
> Best regards,
> Nemanja Ilic
>
> PS
> Sorry if this is not the right place for my question :)
>
--
What most experimenters take for granted before they begin their experiments
is infinitely more interesting than any results to which their experiments
lead.
-- Norbert Wiener
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20100709/9302147d/attachment-0001.htm>
More information about the petsc-users
mailing list