[petsc-users] strange error by changing the number of cells.

Yann JOBIC yann.jobic at univ-amu.fr
Tue Dec 18 07:45:54 CST 2018


Le 18/12/2018 à 14:26, Matthew Knepley a écrit :
> On Tue, Dec 18, 2018 at 5:23 AM Yann JOBIC via petsc-users 
> <petsc-users at mcs.anl.gov <mailto:petsc-users at mcs.anl.gov>> wrote:
>
>     Le 18/12/2018 à 03:01, Matthew Knepley a écrit :
>>     On Mon, Dec 17, 2018 at 4:24 PM Yann Jobic via petsc-users
>>     <petsc-users at mcs.anl.gov <mailto:petsc-users at mcs.anl.gov>> wrote:
>>
>>         Dear petsc users,
>>
>>         I'm using petsc FEM framework, with p4est. I'm getting a
>>         strange error,
>>         when changing the number of point in DMPlexCreateBoxMesh()
>>         (2D, 20x20 is
>>         ok, 22x22 is not) :
>>
>>         [0]PETSC ERROR: Nonconforming object sizes
>>         [0]PETSC ERROR: The section point closure size 20 != dual
>>         space dimension 18
>>
>>
>>     Okay, here is what the nomenclature means. The closure size is
>>     the size I get from the Section
>>     when I take the closure of the given point. So for example, if we
>>     have quads and Q2 elements,
>>     that would have a closure size of
>>
>>       3 x 3 = 9
>>
>>     per component. So a 2-component Q2 element would have closure
>>     size 18. The dual space
>>     dimension is the number of dual space basis vectors. If we use a
>>     Lagrange element, there
>>     would be 1 dof per vertex, 1 per edge and 1 per cell for
>>     quadratic fields, so on a quad that is
>>     9 dofs, and for 2 component fields that is 18. This is why the
>>     two numbers are designed to match.
>>     Could you give
>>
>>       -petscds_view
>>
>>     so I can see the element you are using, and also
>>     PetscSectionView() on the DM section (or use
>>     -petscsection_view although that catches all sections, so it hard
>>     to find the right one).
>>
>>       Thanks,
>     The problem arise in parallel, and for a 6x6 grid size. I'm doing
>     something wrong, but i don't know where. It's working fine in
>     older petsc version.
>     You will find in attachment the log files.
>
>     The code for the section is the following :
>         ierr = DMGetSection(dmAux,&section);CHKERRQ(ierr);
>         PetscSectionView(section, PETSC_VIEWER_STDOUT_WORLD);
>
>     Thanks for the help !
>
>
> On 1 proc, you have 120 cells, but on 2 procs you have 126? Can you 
> also send the output with -dm_view?
> I can't see what is going on with the parallelization. Are you 
> specfiying some overlap in the partition?
I don't have an overlap. My DM is created by :

     ierr = DMPlexCreateBoxMesh(comm, dim, user->simplex, user->cells, 
DM_BOUNDARY_NONE, DM_BOUNDARY_NONE, DM_BOUNDARY_NONE, interpolate, 
dm);CHKERRQ(ierr);

I then distribute the DM :

     ierr = DMPlexDistribute(*dm, 0, NULL, &distributedMesh);CHKERRQ(ierr);

And finally i convert to DMP4EST the resulting DM.
I copy this mesh to create the auxiliary one, for the auxiliary velocity 
field.

Is that right ?

Thanks,

Yann

>
>   Thanks,
>
>     Matt
>
>     Yann
>
>>
>>         Matt
>>
>>         [0]PETSC ERROR: See
>>         http://www.mcs.anl.gov/petsc/documentation/faq.html
>>         for trouble shooting.
>>         [0]PETSC ERROR: Petsc Release Version 3.10.2, Oct, 09, 2018
>>         [0]PETSC ERROR: ./gauss on a  named crabe by yann Mon Dec 17
>>         22:17:38 2018
>>         [0]PETSC ERROR: Configure options
>>         --prefix=/local/lib/petsc/3.10/p2/gcc/openmpi_gcc_all
>>         --with-single-library=0 --with-debugging=1
>>         --download-scalapack=1
>>         --download-metis=1 --download-parmetis=1 --download-ptscotch=1
>>         --download-mumps=1 --download-hypre=1 --download-superlu=1
>>         --download-superlu_dist=1 --download-fblaslapack=1
>>         --download-metis=1
>>         --download-ml=1 --download-p4est=1 --download-netcdf=1
>>         --download-pragmatic=1 --with-cxx-dialect=C++11
>>         --download-eigen=1
>>         --download-parms=1 --download-triangle=1 --download-hdf5=1
>>         --with-zlib=1
>>         --download-szlib=1 --download-suitesparse=1
>>         --with-shared-libraries=0
>>         PETSC_ARCH=openmpi_gcc_all
>>         [0]PETSC ERROR: #1 DMProjectLocal_Generic_Plex() line 400 in
>>         /local/yann/petsc-3.10.2/src/dm/impls/plex/plexproject.c
>>         [0]PETSC ERROR: #2 DMProjectFunctionLocal_Plex() line 530 in
>>         /local/yann/petsc-3.10.2/src/dm/impls/plex/plexproject.c
>>         [0]PETSC ERROR: #3 DMProjectFunctionLocal() line 6320 in
>>         /local/yann/petsc-3.10.2/src/dm/interface/dm.c
>>         [0]PETSC ERROR: #4 DMProjectFunctionLocal_p4est() line 4393 in
>>         /local/yann/petsc-3.10.2/src/dm/impls/forest/p4est/pforest.c
>>         [0]PETSC ERROR: #5 DMProjectFunctionLocal() line 6320 in
>>         /local/yann/petsc-3.10.2/src/dm/interface/dm.c
>>         [0]PETSC ERROR: #6 SetupVelocity() line 332 in
>>         /local/yann/fe-utils/petsc/3.10/gauss.c
>>         [0]PETSC ERROR: #7 SetupVelBC() line 359 in
>>         /local/yann/fe-utils/petsc/3.10/gauss.c
>>         [0]PETSC ERROR: #8 main() line 652 in
>>         /local/yann/fe-utils/petsc/3.10/gauss.c
>>
>>         I'm using petscspace_degree == 2 for the primary field, and
>>         also for the
>>         auxiliary one (velocity). The error occurs when i project the
>>         velocity
>>         function into the auxiliary field.
>>
>>         Do you see where the problem could come from ?
>>
>>         Best regards,
>>
>>         Yann
>>
>
>
>
> -- 
> What most experimenters take for granted before they begin their 
> experiments is infinitely more interesting than any results to which 
> their experiments lead.
> -- Norbert Wiener
>
> https://www.cse.buffalo.edu/~knepley/ 
> <http://www.cse.buffalo.edu/%7Eknepley/>


-- 
___________________________

Yann JOBIC
HPC engineer
Polytech Marseille DME
IUSTI-CNRS UMR 6595
Technopôle de Château Gombert
5 rue Enrico Fermi
13453 Marseille cedex 13
Tel : (33) 4 91 10 69 39
   ou  (33) 4 91 10 69 43
Fax : (33) 4 91 10 69 69

-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20181218/cfdc147b/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: oneproc.log
Type: text/x-log
Size: 6886 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20181218/cfdc147b/attachment-0002.bin>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: twoproc.log
Type: text/x-log
Size: 12672 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20181218/cfdc147b/attachment-0003.bin>


More information about the petsc-users mailing list