<html><head><style type="text/css"><!-- DIV {margin:0px;} --></style></head><body><div style="font-family:times new roman, new york, times, serif;font-size:12pt"><div>Hi,<br><br>I have observed that the memory usage of the petsc mesh is much higher than my previous code, if both were to be run serially.<br>For example, for a simple cubic box with 750,000 tetrahedral elements, my old code takes about 200MB for the whole array, including all the mappings required for later use such as the inverse connectivity table. For the same mesh, my PETSc code takes about 4GB for the mesh alone. <br><br>The same can be found in the provided examples. I made a few changes to the navierStokes code to output the virtual memory usage and got<br><span style="font-style: italic;">./navierStokes -dim 3 -generate -structured 0 -refinement_limit 1e-6</span><br style="font-style: italic;"><span style="font-style: italic;">
109,283 elements, 139,030 edges , 21,523 vertexes </span><br style="font-style: italic;"><span style="font-style: italic;">
[0]:after mesh created:mem=574.46 MB<br></span>This is consistent with my Petsc code. <br><br>I understand that for the mesh to scale in parallel, extra information needs to be stored. But the current cost seems too expensive. I am wondering if there is a way to cut the memory usage for the mesh.<br>Thank you very much.<br><br>Shi <br><span style="font-style: italic;"></span></div></div><br>__________________________________________________<br>Do You Yahoo!?<br>Tired of spam? Yahoo! Mail has the best spam protection around <br>http://mail.yahoo.com </body></html>