<div dir="ltr">Hi Barry,<div><br></div><div>I indeed deleted some of the OpenMPI related warning messages, it should be irrelevant. Here is the full output:</div><div><p class="">======================<br>[kaisong@n0009 petsc-3.5.0]$ sh <a href="http://petsc-3.5.0-gcc.sh">petsc-3.5.0-gcc.sh</a><br>
Running test examples to verify correct installation<br>Using PETSC_DIR=/clusterfs/voltaire/home/software/modules/petsc/3.5.0 and PETSC_ARCH=arch-linux2-c-debug<br>Possible error running C/C++ src/snes/examples/tutorials/ex19 with 1 MPI process<br>
See <a href="http://www.mcs.anl.gov/petsc/documentation/faq.html">http://www.mcs.anl.gov/petsc/documentation/faq.html</a><br>--------------------------------------------------------------------------<br>WARNING: There is at least non-excluded one OpenFabrics device found,<br>
but there are no active ports detected (or Open MPI was unable to use<br>them). This is most certainly not what you wanted. Check your<br>cables, subnet manager configuration, etc. The openib BTL will be<br>ignored for this job.<br>
<br> Local host: n0009.scs00<br>--------------------------------------------------------------------------<br>lid velocity = 0.0016, prandtl # = 1, grashof # = 1<br>Number of SNES iterations = 2<br>Possible error running C/C++ src/snes/examples/tutorials/ex19 with 2 MPI processes<br>
See <a href="http://www.mcs.anl.gov/petsc/documentation/faq.html">http://www.mcs.anl.gov/petsc/documentation/faq.html</a><br>--------------------------------------------------------------------------<br>WARNING: There is at least non-excluded one OpenFabrics device found,<br>
but there are no active ports detected (or Open MPI was unable to use<br>them). This is most certainly not what you wanted. Check your<br>cables, subnet manager configuration, etc. The openib BTL will be<br>ignored for this job.<br>
<br> Local host: n0009.scs00<br>--------------------------------------------------------------------------<br>lid velocity = 0.0016, prandtl # = 1, grashof # = 1<br>Number of SNES iterations = 2<br>[n0009.scs00:07638] 1 more process has sent help message help-mpi-btl-openib.txt / no active ports found<br>
[n0009.scs00:07638] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages<br>Possible error running Fortran example src/snes/examples/tutorials/ex5f with 1 MPI process<br>See <a href="http://www.mcs.anl.gov/petsc/documentation/faq.html">http://www.mcs.anl.gov/petsc/documentation/faq.html</a><br>
--------------------------------------------------------------------------<br>WARNING: There is at least non-excluded one OpenFabrics device found,<br>but there are no active ports detected (or Open MPI was unable to use<br>
them). This is most certainly not what you wanted. Check your<br>cables, subnet manager configuration, etc. The openib BTL will be<br>ignored for this job.<br><br> Local host: n0009.scs00<br>--------------------------------------------------------------------------<br>
Number of SNES iterations = 4<br>Completed test examples<br>=========================================<br>Now to evaluate the computer systems you plan use - do:<br>make PETSC_DIR=/clusterfs/voltaire/home/software/modules/petsc/3.5.0 PETSC_ARCH=arch-linux2-c-debug streams NPMAX=<number of MPI processes you intend to use><br>
======================</p></div><div>I also followed your suggestion and built ex19 separately, and made sure using the same compiler that built petsc. Here is the complete output:<br></div><div>======================</div>
[kaisong@n0009 tutorials]$ mpiexec -n 2 ./ex19<br>--------------------------------------------------------------------------<br>WARNING: There is at least non-excluded one OpenFabrics device found,<br>but there are no active ports detected (or Open MPI was unable to use<br>
them). This is most certainly not what you wanted. Check your<br>cables, subnet manager configuration, etc. The openib BTL will be<br>ignored for this job.<br><br> Local host: n0009.scs00<br>--------------------------------------------------------------------------<br>
lid velocity = 0.0625, prandtl # = 1, grashof # = 1<br>Number of SNES iterations = 2<br>[n0009.scs00:09124] 1 more process has sent help message help-mpi-btl-openib.txt / no active ports found<br>[n0009.scs00:09124] Set MCA parameter "orte_base_help_aggregate" to 0 to see all help / error messages<div>
======================</div><div><br></div><div>Thanks,</div><div><br></div><div>Kai</div><div><br></div></div><div class="gmail_extra"><br><br><div class="gmail_quote">On Fri, Jul 25, 2014 at 1:32 PM, Barry Smith <span dir="ltr"><<a href="mailto:bsmith@mcs.anl.gov" target="_blank">bsmith@mcs.anl.gov</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex"><div class=""><br>
On Jul 25, 2014, at 12:23 PM, Kai Song <<a href="mailto:ksong@lbl.gov">ksong@lbl.gov</a>> wrote:<br>
<br>
> Hi All,<br>
><br>
> Thanks for all the suggestions! I took Satish's advice, removed the source, and did a fresh built.<br>
><br>
> It got through the test compiling, but the tests failed as follow:<br>
> ==================<br>
> Running test examples to verify correct installation<br>
> Using PETSC_DIR=/clusterfs/voltaire/home/software/modules/petsc/3.5.0 and PETSC_ARCH=arch-linux2-c-debug<br>
> Possible error running C/C++ src/snes/examples/tutorials/ex19 with 1 MPI process<br>
> See <a href="http://www.mcs.anl.gov/petsc/documentation/faq.html" target="_blank">http://www.mcs.anl.gov/petsc/documentation/faq.html</a><br>
> lid velocity = 0.0016, prandtl # = 1, grashof # = 1<br>
> Number of SNES iterations = 2<br>
> Possible error running C/C++ src/snes/examples/tutorials/ex19 with 2 MPI processes<br>
> See <a href="http://www.mcs.anl.gov/petsc/documentation/faq.html" target="_blank">http://www.mcs.anl.gov/petsc/documentation/faq.html</a><br>
</div>> …<br>
<br>
Surely it did not print … here? What was the complete message?<br>
<br>
Usually this happens due to an incompatible mpiexec being used because it appears earlier in the path.<br>
<br>
What happens if you do<br>
<br>
cd src/snes/examples/tutorials<br>
make ex19<br>
mpiexec -n 2 ./ex19<br>
<br>
for mpiexec use the one associated with the MPI that you built PETSc with<br>
<span class="HOEnZb"><font color="#888888"><br>
Barry<br>
</font></span><div class="HOEnZb"><div class="h5"><br>
> ==================<br>
><br>
> I tried to go into src/snes/examples/tutorials/ex19, but didn't find any logs. Is this error for the test Okay?<br>
><br>
> Thanks,<br>
><br>
> Kai<br>
><br>
><br>
><br>
> On Fri, Jul 25, 2014 at 11:06 AM, Satish Balay <<a href="mailto:balay@mcs.anl.gov">balay@mcs.anl.gov</a>> wrote:<br>
> Its probably best to do a fresh build with the latest petsc-3.5.1 tarball and see<br>
> if the problem is reproduceable [with --with-shared-libraries=0]<br>
><br>
> Satish<br>
><br>
> On Fri, 25 Jul 2014, Jed Brown wrote:<br>
><br>
> > Matthew Knepley <<a href="mailto:knepley@gmail.com">knepley@gmail.com</a>> writes:<br>
> ><br>
> > > On Thu, Jul 24, 2014 at 5:33 PM, Kai Song <<a href="mailto:ksong@lbl.gov">ksong@lbl.gov</a>> wrote:<br>
> > ><br>
> > >> Hi Satish,<br>
> > >><br>
> > >> Thanks for the quick response. I attached the make.log in this email, and<br>
> > >> a quick glance, I didn't see any obvious errors.<br>
> > >><br>
> > ><br>
> > > This never built the MUMPS directory, even though you have PETSC_HAVE_MUMPS<br>
> > > defined. I can see two<br>
> > > possible reasons for this:<br>
> > ><br>
> > > 1) You distribution as missing files<br>
> > ><br>
> > > 2) You have two copies of PETSc fighting each other here (old PETSC_DIR?)<br>
> > ><br>
> > > Can you make sure you have the file mumps.c in<br>
> > ><br>
> > > src/mat/impls/aij/mpi/mumps<br>
> > ><br>
> > > Jed, how do we check what make is doing here?<br>
> ><br>
> > mumps.c should be listed in $PETSC_ARCH/conf/files after the code<br>
> > generator runs. Possibly a file system problem or tinkering caused it<br>
> > to appear up-to-date after reconfiguring? (Just speculating because we<br>
> > haven't seen this problem before.)<br>
> ><br>
><br>
><br>
><br>
><br>
> --<br>
> Kai Song<br>
> <<a href="mailto:ksong@lbl.gov">ksong@lbl.gov</a>> <a href="tel:1.510.495.2180" value="+15104952180">1.510.495.2180</a><br>
> 1 Cyclotron Rd. Berkeley, CA94720, MS-50B 3209<br>
> High Performance Computing Services (HPCS)<br>
> Lawrence Berkeley National Laboratory - <a href="http://scs.lbl.gov" target="_blank">http://scs.lbl.gov</a><br>
<br>
</div></div></blockquote></div><br><br clear="all"><div><br></div>-- <br>Kai Song<br><<a href="mailto:ksong@lbl.gov" target="_blank">ksong@lbl.gov</a>> 1.510.495.2180<br>1 Cyclotron Rd. Berkeley, CA94720, MS-50B 3209<br>
High Performance Computing Services (HPCS) <br>Lawrence Berkeley National Laboratory - <a href="http://scs.lbl.gov" target="_blank">http://scs.lbl.gov</a>
</div>