NetBSD port
Kevin.Buckley at ecs.vuw.ac.nz
Kevin.Buckley at ecs.vuw.ac.nz
Wed Dec 16 16:38:20 CST 2009
> Lets ignore 'Sun Grid Engine environment' initially and just figureout
> your PETSc install.
>
> - What MPI is it built with? Send us the output for the compile of ex19
>
> - you claim 'make test' worked fine - i.e this example ran fine
> paralley. can you confrim this with manual run?
>
> [if thats the case - then PETSc would be working correctly with the
> MPI specified]
>
>
>>From the info below -- the example crashes happen only in 'Sun Grid
> Engine environment' What is that? And why should binaries compiled
> with this default 'MPI' - work in that grid enviornment - without
> recompiling with a different 'sun-grid-mpi' ?
>
>
> Satish
Someone else using the PISM software, over in Alaska as it happens,
which sits on top of PETSc here, has seen similar errors so I am
thinking that it may not be just my environment, which I doubt
matches theirs.
For your consideration though:
===========
My PETSc was built against OpenMPI 1.4
===========
Compilation of the example in question shows:
$ export PETSC_DIR=/vol/grid/pkg/petsc-3.0.0-p7
$gmake ex19
mpicc -o ex19.o -c -Wall -Wwrite-strings -Wno-strict-aliasing -g3
-I/vol/grid/pkg/petsc-3.0.0-p7/include
-I/vol/grid/pkg/petsc-3.0.0-p7/include -I/usr/pkg/include
-D__SDIR__="src/snes/examples/tutorials/" ex19.c
mpicc -Wall -Wwrite-strings -Wno-strict-aliasing -g3 -o ex19 ex19.o
-Wl,-rpath,/vol/grid/pkg/petsc-3.0.0-p7/lib
-L/vol/grid/pkg/petsc-3.0.0-p7/lib -lpetscsnes -lpetscksp -lpetscdm
-lpetscmat -lpetscvec -lpetsc -L/usr/pkg/lib -lX11 -llapack -lblas
-L/usr/pkg/lib -lmpi -lopen-rte -lopen-pal -lutil -lpthread -lgcc_eh
-Wl,-rpath,/usr/pkg/lib -lmpi_f77 -lf95 -lm -lm
-L/usr/pkg/lib/gcc-lib/i386--netbsdelf/4.0.3 -L/lib -lm -lm -lmpi_cxx
-lstdc++ -lgcc_s -lmpi_cxx -lstdc++ -lgcc_s -lmpi -lopen-rte -lopen-pal
-lutil -lpthread -lgcc_eh
/bin/rm -f ex19.o
$./ex19
lid velocity = 0.0204082, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
lid velocity = 0.0204082, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
==========
Running a parallel invocation local to one machine
$ mpirun -n 2 ./ex19 -dmmg_nlevels 4
lid velocity = 0.0016, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
lid velocity = 0.0016, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
$ mpirun -n 4 ./ex19 -dmmg_nlevels 4
lid velocity = 0.0016, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
lid velocity = 0.0016, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
however when submitting within the SGE environment, we see a
similar story to that seen with the PISM package
% qsub -pe kmbmpi 2 my_mpirun_job.sh
% cat my_mpirun_job.sh.o425710
lid velocity = 0.0016, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
lid velocity = 0.0016, prandtl # = 1, grashof # = 1
Number of Newton iterations = 2
% qsub -pe kmbmpi 4 my_mpirun_job.sh
A swathe of PETSc errors.
==========
--
Kevin M. Buckley Room: CO327
School of Engineering and Phone: +64 4 463 5971
Computer Science
Victoria University of Wellington
New Zealand
More information about the petsc-dev
mailing list