NetBSD port

Satish Balay balay at mcs.anl.gov
Wed Dec 16 16:56:19 CST 2009


Ok - the code runs locally fine - but not on  'SunGridEngine'

> % qsub -pe kmbmpi 4 my_mpirun_job.sh

Wrt SGE - what does it require from MPI. Is it MPI agnostic - or does
it need a perticular MPI to be used?

What if you use PETSc installed with --download-mpich instead? How
will it know how to schedule these MPI jobs?

Or does it require a peritcular MPI - installed in a perticular way -
on all the nodes on the grid?

BTW: what do you have for 'ldd ex19'?

Satish


On Thu, 17 Dec 2009, Kevin.Buckley at ecs.vuw.ac.nz wrote:

> > Lets ignore 'Sun Grid Engine environment' initially and just figureout
> > your PETSc install.
> >
> > - What MPI is it built with? Send us the output for the compile of ex19
> >
> > - you claim 'make test' worked fine - i.e this example ran fine
> > paralley.  can you confrim this with manual run?
> >
> > [if thats the case - then PETSc would be working correctly with the
> > MPI specified]
> >
> >
> >>From the info below -- the example crashes happen only in 'Sun Grid
> > Engine environment' What is that? And why should binaries compiled
> > with this default 'MPI' - work in that grid enviornment - without
> > recompiling with a different 'sun-grid-mpi' ?
> >
> >
> > Satish
> 
> Someone else using the PISM software, over in Alaska as it happens,
> which sits on top of PETSc here, has seen similar errors so I am
> thinking that it may not be just my environment, which I doubt
> matches theirs.
> 
> For your consideration though:
> 
> ===========
> 
> My PETSc was built against OpenMPI 1.4
> 
> ===========
> 
> Compilation of the example in question shows:
> 
> $ export PETSC_DIR=/vol/grid/pkg/petsc-3.0.0-p7
> 
> $gmake ex19
> mpicc -o ex19.o -c -Wall -Wwrite-strings -Wno-strict-aliasing -g3
> -I/vol/grid/pkg/petsc-3.0.0-p7/include
> -I/vol/grid/pkg/petsc-3.0.0-p7/include -I/usr/pkg/include
> -D__SDIR__="src/snes/examples/tutorials/" ex19.c
> mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -g3  -o ex19  ex19.o
> -Wl,-rpath,/vol/grid/pkg/petsc-3.0.0-p7/lib
> -L/vol/grid/pkg/petsc-3.0.0-p7/lib -lpetscsnes -lpetscksp -lpetscdm
> -lpetscmat -lpetscvec -lpetsc   -L/usr/pkg/lib -lX11 -llapack -lblas
> -L/usr/pkg/lib -lmpi -lopen-rte -lopen-pal -lutil -lpthread -lgcc_eh
> -Wl,-rpath,/usr/pkg/lib -lmpi_f77 -lf95 -lm -lm
> -L/usr/pkg/lib/gcc-lib/i386--netbsdelf/4.0.3 -L/lib -lm -lm -lmpi_cxx
> -lstdc++ -lgcc_s -lmpi_cxx -lstdc++ -lgcc_s -lmpi -lopen-rte -lopen-pal
> -lutil -lpthread -lgcc_eh
> /bin/rm -f ex19.o
> 
> $./ex19
> lid velocity = 0.0204082, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> lid velocity = 0.0204082, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> 
> ==========
> 
> Running a parallel invocation local to one machine
> 
> $ mpirun -n 2 ./ex19 -dmmg_nlevels 4
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> 
> $ mpirun -n 4 ./ex19 -dmmg_nlevels 4
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> 
> however when submitting within the SGE environment, we see a
> similar story to that seen with the PISM package
> 
> % qsub -pe kmbmpi 2 my_mpirun_job.sh
> % cat my_mpirun_job.sh.o425710
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1
> Number of Newton iterations = 2
> 
> 
> % qsub -pe kmbmpi 4 my_mpirun_job.sh
> 
> A swathe of PETSc errors.
> 
> ==========
> 
> 




More information about the petsc-dev mailing list