[petsc-users] dmplex extruded layers

Bhargav Subramanya bhargav.subramanya at kaust.edu.sa
Wed Nov 24 11:27:09 CST 2021


Dear Matt and Mark,

Thank you very much for your reply. Your inputs are very useful to me.

On Mon, Nov 22, 2021 at 9:38 PM Matthew Knepley <knepley at gmail.com> wrote:

> On Mon, Nov 22, 2021 at 12:10 PM Bhargav Subramanya <
> bhargav.subramanya at kaust.edu.sa> wrote:
>
>> Dear All,
>>
>> I have a prismatic mesh generated by extruding a base icosahedron
>> spherical surface mesh. The mesh is distributed on 2 processes after
>> extrusion. The dof for my problem are defined only on the simplices of the
>> spherical layers.
>>
>> 1. For a given spherical layer, I want to get points, which are basically
>> simplices, lying on that layer only. DMPlexGetHeightStratum returns the
>> points on all the spherical Layers. I can probably use DMPlexFindVertices
>> (based on the radius of the spherical layer) and
>> DMPlexGetSimplexOrBoxCells. Could you suggest if there is a better way to
>> retrieve the points (which are simplices) on a given spherical layer in the
>> extruded mesh?
>>
>
> DMPlexGetHeightStratum() refers to height in the Hasse Diagram, which is a
> DAG, not in the mesh. For example, height 0 points are the cells, height 1
> are the faces, etc.
>
> I believe the default numbering for extrusion (in the main branch), is
> that all vertices produced from a given vertex be in order. This would mean
> that if v were the vertex point number, then
>
>   (v - vStart) % Nl == l
>
> where Nl is the number of layers and l is the layer of that vertex. It is
> also the same for triangles. So if you want to segregate each shell, after
> extrusion, make a label that gives triangles this marker, and then use
> DMPlexLabelComplete(). Then after refinement you will still have the shells
> labeled correctly.
>
> I would be happy to help you make an example that does this. It seems cool.
>

----- Sure, I am interested to make an example of it. And, I really
appreciate your help.

>
>
>> 2. I am able to visualize the entire mesh using dm_view. Is there a way
>> to get the mesh file for the local dm from a specific process?
>>
>
> You can use -dm_partition_view which outputs a field with the process
> number. Then use Clip inside Paraview and clip to the process number you
> want,
> or just view that field so each process has a different color.
>
>
>> 3. One of my previous emails might have got lost as I mistakenly attached
>> a large mesh file and sent it. So I am repeating the question here.
>> DMPlexExtrude gives the following error after distributing the base 2D
>> spherical mesh. Both refinement or/and extrusion after dm distribution does
>> not work for me. In fact, I tried with src/dm/impls/plex/tutorials/ex10.c
>> also. Although the mesh is distributed after the extrusion in ex10.c (which
>> is working fine for me), I tried to distribute before extrusion, and I get
>> the following error. Could you please suggest where I might be making a
>> mistake?
>>
>
> So you want to distribute the mesh before extruding. For that small
> example (using the main branch), I run
>
>   mpiexec -n 3 ./meshtest -dm_plex_shape sphere -dm_refine_pre 2
> -dm_distribute -dm_partition_view -dm_view hdf5:mesh.h5 -dm_extrude 3
>

----- I pulled the code from the main branch and configured my petsc again.
I am now finally able to reproduce the mesh that you have attached. I am
able to do parallel mesh refinement, followed by extrusion, and use the
following code to check that. However, I still have one problem. Attached
are the extruded mesh files.

1. Mesh doesn't seem to be refined properly near interfaces between the
processes. I then used -init_dm_distribute_overlap of 1 and that fixed the
problem. I hope this is the way to do that.

2. The parallel mesh refinement doesn't seem to take care of the spherical
geometry (shown in the attached mesh file). Could you please suggest how to
fix it?

command line options:
-init_dm_plex_shape sphere -init_dm_plex_sphere_radius 1.0
-init_dm_refine_pre 2 -init_dm_distribute -dm_refine 2 -dm_partition_view
-dm_view hdf5:mesh.h5 -dm_extrude 3

 ierr = DMCreate(PETSC_COMM_WORLD, &dm);CHKERRQ(ierr);
ierr = DMSetType(dm, DMPLEX);CHKERRQ(ierr);
ierr = PetscObjectSetOptionsPrefix((PetscObject) dm, "init_");CHKERRQ(ierr);
ierr = DMSetFromOptions(dm);CHKERRQ(ierr);

ierr = DMPlexIsDistributed(dm, &distributed);
ierr = PetscSynchronizedPrintf(PETSC_COMM_WORLD, "%d \n", distributed);
CHKERRQ(ierr);
ierr = PetscSynchronizedFlush(PETSC_COMM_WORLD,PETSC_STDOUT);

ierr = PetscObjectSetOptionsPrefix((PetscObject) dm, NULL);CHKERRQ(ierr);
ierr = DMSetFromOptions(dm);CHKERRQ(ierr);
ierr = DMViewFromOptions(dm, NULL, "-dm_view");CHKERRQ(ierr);


> and I get the attached picture.
>
>   Thanks,
>
>     Matt
>
>
>> [0]PETSC ERROR: --------------------- Error Message
>> --------------------------------------------------------------
>> [0]PETSC ERROR: Object is in wrong state
>> [0]PETSC ERROR: This DMPlex is distributed but its PointSF has no graph
>> set
>> [0]PETSC ERROR: See https://petsc.org/release/faq/
>> <https://urldefense.com/v3/__https://petsc.org/release/faq/__;!!Nmw4Hv0!hb2EaePjZqAA4Ns2bFzuHqtjvdZrySmfJcO-7YL5Twmqdh_n0Ym4CneeBI7x1Z5UiaOW2dk_COg$>
>> for trouble shooting.
>> [0]PETSC ERROR: Petsc Release Version 3.16.1, unknown
>> [0]PETSC ERROR: ./cavity_flow.out on a arch-darwin-c-debug named kl-21859
>> by subrambm Mon Nov 22 19:47:14 2021
>> [0]PETSC ERROR: Configure options --with-cc=gcc --with-cxx=g++
>> --with-fc=gfortran --download-mpich --download-fblaslapack
>> --download-superlu_dist --download-hypre --download-fiat
>> --download-generator --download-triangle --download-tetgen --download-chaco
>> --download-make -download-boost --download-cmake --download-ml
>> --download-mumps=https://bitbucket.org/petsc/pkg-mumps.git
>> <https://urldefense.com/v3/__https://bitbucket.org/petsc/pkg-mumps.git__;!!Nmw4Hv0!hb2EaePjZqAA4Ns2bFzuHqtjvdZrySmfJcO-7YL5Twmqdh_n0Ym4CneeBI7x1Z5UiaOW414fjd0$>
>> --download-mumps-commit=v5.4.1-p1 --download-scalapack --download-ptscotch
>> --download-hdf5 --force
>> [0]PETSC ERROR: #1 DMPlexCheckPointSF() at
>> /Users/subrambm/petsc/src/dm/impls/plex/plex.c:8554
>> [0]PETSC ERROR: #2 DMPlexOrientInterface_Internal() at
>> /Users/subrambm/petsc/src/dm/impls/plex/plexinterpolate.c:595
>> [0]PETSC ERROR: #3 DMPlexInterpolate() at
>> /Users/subrambm/petsc/src/dm/impls/plex/plexinterpolate.c:1357
>> [0]PETSC ERROR: #4 DMPlexExtrude() at
>> /Users/subrambm/petsc/src/dm/impls/plex/plexcreate.c:1543
>> [0]PETSC ERROR: #5 CreateMesh() at ex10.c:161
>> [0]PETSC ERROR: #6 main() at ex10.c:180
>> [0]PETSC ERROR: PETSc Option Table entries:
>> [0]PETSC ERROR: -dm_plex_extrude_layers 3
>> [0]PETSC ERROR: -dm_view vtk:mesh.vtk
>> [0]PETSC ERROR: -init_dm_plex_dim 2
>> [0]PETSC ERROR: -petscpartitioner_type simple
>> [0]PETSC ERROR: -srf_dm_distribute
>> [0]PETSC ERROR: -srf_dm_refine 0
>> [0]PETSC ERROR: ----------------End of Error Message -------send entire
>> error message to petsc-maint at mcs.anl.gov----------
>>
>> Thanks,
>> Bhargav
>>
>>
>> ------------------------------
>> This message and its contents, including attachments are intended solely
>> for the original recipient. If you are not the intended recipient or have
>> received this message in error, please notify me immediately and delete
>> this message from your computer system. Any unauthorized use or
>> distribution is prohibited. Please consider the environment before printing
>> this email.
>
>
>
> --
> What most experimenters take for granted before they begin their
> experiments is infinitely more interesting than any results to which their
> experiments lead.
> -- Norbert Wiener
>
> https://www.cse.buffalo.edu/~knepley/
> <https://urldefense.com/v3/__http://www.cse.buffalo.edu/*knepley/__;fg!!Nmw4Hv0!hb2EaePjZqAA4Ns2bFzuHqtjvdZrySmfJcO-7YL5Twmqdh_n0Ym4CneeBI7x1Z5UiaOWdEg9at8$>
>

-- 

This message and its contents, including attachments are intended solely 
for the original recipient. If you are not the intended recipient or have 
received this message in error, please notify me immediately and delete 
this message from your computer system. Any unauthorized use or 
distribution is prohibited. Please consider the environment before printing 
this email.
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20211124/fa139865/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: mesh_overlap_0.png
Type: image/png
Size: 342980 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20211124/fa139865/attachment-0002.png>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: mesh_overlap_1.png
Type: image/png
Size: 331015 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20211124/fa139865/attachment-0003.png>


More information about the petsc-users mailing list