[petsc-users] Installation question

Pham Pham pvsang002 at gmail.com
Wed Apr 19 10:37:37 CDT 2017


Hi,

I just installed petsc-3.7.5 into my university cluster. When evaluating
the computer system, PETSc reports "It appears you have 1 node(s)", I donot
understand this, since the system is a multinodes system. Could you please
explain this to me?

Thank you very much.

S.

Output:
=========================================
Now to evaluate the computer systems you plan use - do:
make PETSC_DIR=/home/svu/mpepvs/petsc/petsc-3.7.5
PETSC_ARCH=arch-linux-cxx-opt streams
[mpepvs at atlas7-c10 petsc-3.7.5]$ make
PETSC_DIR=/home/svu/mpepvs/petsc/petsc-3.7.5 PETSC_ARCH=arch-linux-cxx-opt
streams
cd src/benchmarks/streams; /usr/bin/gmake  --no-print-directory
PETSC_DIR=/home/svu/mpepvs/petsc/petsc-3.7.5 PETSC_ARCH=arch-linux-cxx-opt
streams
/home/svu/mpepvs/petsc/petsc-3.7.5/arch-linux-cxx-opt/bin/mpicxx -o
MPIVersion.o -c -Wall -Wwrite-strings -Wno-strict-aliasing
-Wno-unknown-pragmas -fvisibility=hidden -g -O
-I/home/svu/mpepvs/petsc/petsc-3.7.5/include
-I/home/svu/mpepvs/petsc/petsc-3.7.5/arch-linux-cxx-opt/include
`pwd`/MPIVersion.c
Running streams with
'/home/svu/mpepvs/petsc/petsc-3.7.5/arch-linux-cxx-opt/bin/mpiexec ' using
'NPMAX=12'
Number of MPI processes 1 Processor names  atlas7-c10
Triad:         9137.5025   Rate (MB/s)
Number of MPI processes 2 Processor names  atlas7-c10 atlas7-c10
Triad:         9707.2815   Rate (MB/s)
Number of MPI processes 3 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
Triad:        13559.5275   Rate (MB/s)
Number of MPI processes 4 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10
Triad:        14193.0597   Rate (MB/s)
Number of MPI processes 5 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10
Triad:        14492.9234   Rate (MB/s)
Number of MPI processes 6 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10
Triad:        15476.5912   Rate (MB/s)
Number of MPI processes 7 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
Triad:        15148.7388   Rate (MB/s)
Number of MPI processes 8 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
Triad:        15799.1290   Rate (MB/s)
Number of MPI processes 9 Processor names  atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
Triad:        15671.3104   Rate (MB/s)
Number of MPI processes 10 Processor names  atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10
Triad:        15601.4754   Rate (MB/s)
Number of MPI processes 11 Processor names  atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10
Triad:        15434.5790   Rate (MB/s)
Number of MPI processes 12 Processor names  atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
atlas7-c10 atlas7-c10 atlas7-c10 atlas7-c10
Triad:        15134.1263   Rate (MB/s)
------------------------------------------------
np  speedup
1 1.0
2 1.06
3 1.48
4 1.55
5 1.59
6 1.69
7 1.66
8 1.73
9 1.72
10 1.71
11 1.69
12 1.66
Estimation of possible speedup of MPI programs based on Streams benchmark.
It appears you have 1 node(s)
Unable to plot speedup to a file
Unable to open matplotlib to plot speedup
[mpepvs at atlas7-c10 petsc-3.7.5]$
[mpepvs at atlas7-c10 petsc-3.7.5]$
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20170419/a0c56a94/attachment-0001.html>
-------------- next part --------------
A non-text attachment was scrubbed...
Name: configure.log
Type: text/x-log
Size: 6194578 bytes
Desc: not available
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20170419/a0c56a94/attachment-0001.bin>


More information about the petsc-users mailing list