[petsc-dev] Understanding Some Parallel Results with PETSc
Dave Nystrom
dnystrom1 at comcast.net
Thu Feb 23 23:41:33 CST 2012
I could also send you my mpi/numactl command lines for gpu and cpu when I am
back in the office.
Dave Nystrom writes:
> Jed Brown writes:
> > On Thu, Feb 23, 2012 at 18:53, Nystrom, William D <wdn at lanl.gov> wrote:
> >
> > > Rerunning the CPU case with numactl results in a 25x speedup and
> > > log_summary
> > > results that look reasonable to me now.
> > >
> >
> > What command are you using for this? We usually use the affinity options to
> > mpiexec instead of using numactl/taskset manually.
>
> I was using openmpi-1.5.4 as installed by the system admins on our testbed
> cluster. I talked to a couple of our openmpi developers and they indicated
> that the affinity stuff was broken in that version but should be fixed when
> 1.5.5 and 1.6 come out - which should be within the next month.
>
> I also tried mvapich2-1.7 built with slurm and tried using the affinity stuff
> with srun. That also did not seem to work. But I should probably revisit
> that and try to make sure that I really understand how to use srun.
>
> I was pretty surprised that getting the numa stuff right made such a huge
> difference. I'm also wondering if getting the affinity right will make much
> of a difference for the gpu case.
>
> > Did you also set a specific memory policy?
>
> I'm not sure what you mean by the above question but I'm kind of new to all
> this numa stuff.
>
> > Which Linux kernel is this?
>
> The OS was the latest beta of TOSS2. If I remember, I can check next time I
> am in my office. It is probably RHEL6.
More information about the petsc-dev
mailing list