[petsc-users] Memory requirements in SUPERLU_DIST
Barry Smith
bsmith at mcs.anl.gov
Tue Sep 23 20:59:25 CDT 2014
This is something you better ask Sherri about. She’s the one who wrote and understands SuperLU_DIST
Barry
On Sep 23, 2014, at 7:00 PM, Zin Lin <zinlin.zinlin at gmail.com> wrote:
> Hi
> I am solving a frequency domain Maxwell problem for a dielectric structure of size 90x90x50, (the total matrix size is (90x90x50x6)^2 which includes the three vector components as well as real and imaginary parts.)
> I am using SUPERLU_DIST for the direct solver with the following options
>
> parsymbfact = 1, (parallel symbolic factorization)
> permcol = PARMETIS, (parallel METIS)
> permrow = NATURAL (natural ordering).
>
> First, I tried to use 4096 cores with 2GB / core memory which totals to about 8 TB of memory.
> I get the following error:
>
> Using ParMETIS for parallel ordering.
> Structual symmetry is:100%
> Current memory used: 1400271832 bytes
> Maximum memory used: 1575752120 bytes
> ***Memory allocation failed for SetupCoarseGraph: adjncy. Requested size: 148242928 bytes
>
> So it seems to be an insufficient memory allocation problem (which apparently happens at the METIS analysis phase?).
>
> Then, I tried to use 64 large-memory cores which have a total of 2 TB memory (so larger memory per each core), it seems to work fine (though the solver takes about 900 sec ).
> What I don't understand is why memory per core matters rather than the total memory? If the work space is distributed across the processors, shouldn't it work as long as I choose a sufficient number of smaller-memory cores? What kind of role does the memory per core play in the algorithm in contrast to the total memory over all the cores?
>
> The issue is I would rather use a large number of small-memory cores than any number of the large-memory cores. The latter are two times more expensive in terms of service units (I am running on STAMPEDE at TACC) and not many cores are available either.
>
> Any idea would be appreciated.
>
> Zin
>
> --
> Zin Lin
>
More information about the petsc-users
mailing list