[petsc-users] Streams scaling.log

Manuel Valera mvalera-w at sdsu.edu
Wed Apr 25 14:12:56 CDT 2018


Hi and thanks for the quick answer,

Yes it looks i am using MPICH for my configure instead of using the system
installation of OpenMPI, in the past i had better experience using MPICH
but maybe this will be a conflict, should i reconfigure using the system
MPI installation?

I solved the problem in a different way by login into the nodes i wanted to
use and doing the make streams tests there, but i get the following:

np  speedup
1 1.0
2 1.51
3 2.17
4 2.66
5 2.87
6 3.06
7 3.44
8 3.84
9 3.81
10 3.17
11 3.69
12 3.81
13 3.26
14 3.51
15 3.61
16 3.81
17 3.8
18 3.64
19 3.48
20 4.01

So very modest scaling, this is about the same i get with my application,
how can i make it work faster? i am already using --map-by and
--machinefile arguments for mpirun, maybe this is also a conflict with the
different MPI installations?

Thanks,



On Wed, Apr 25, 2018 at 11:51 AM, Karl Rupp <rupp at iue.tuwien.ac.at> wrote:

> Hi Manuel,
>
> this looks like the wrong MPI gets used. You should see an increasing
> number of processes, e.g.
>
> Number of MPI processes 1 Processor names  node37
> Triad:         6052.3571   Rate (MB/s)
> Number of MPI processes 2 Processor names  node37 node37
> Triad:        9138.9376   Rate (MB/s)
> Number of MPI processes 3 Processor names  node37 node37 node37
> Triad:        11077.5905   Rate (MB/s)
> Number of MPI processes 4 Processor names  node37 node37 node37 node37
> Triad:        12055.9123   Rate (MB/s)
>
> Best regards,
> Karli
>
>
>
>
> On 04/25/2018 08:26 PM, Manuel Valera wrote:
>
>> Hi,
>>
>> I'm running scaling tests on my system to check why my scaling is so
>> poor, and after following the MPIVersion guidelines my scaling.log output
>> looks like this:
>>
>> Number of MPI processes 1 Processor names  node37
>> Triad:        12856.9252   Rate (MB/s)
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Triad:         9138.3320   Rate (MB/s)
>> Triad:         9945.0006   Rate (MB/s)
>> Triad:        10480.8471   Rate (MB/s)
>> Triad:        12055.4846   Rate (MB/s)
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Triad:         7394.1014   Rate (MB/s)
>> Triad:         5528.9757   Rate (MB/s)
>> Triad:         6052.7506   Rate (MB/s)
>> Triad:         6188.5710   Rate (MB/s)
>> Triad:         6944.4515   Rate (MB/s)
>> Triad:         7407.1594   Rate (MB/s)
>> Triad:         9508.1984   Rate (MB/s)
>> Triad:        10699.7551   Rate (MB/s)
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Number of MPI processes 1 Processor names  node37
>> Triad:         6682.3749   Rate (MB/s)
>> Triad:         6825.3243   Rate (MB/s)
>> Triad:         7217.8178   Rate (MB/s)
>> Triad:         7525.1025   Rate (MB/s)
>> Triad:         7882.1781   Rate (MB/s)
>> Triad:         8071.1430   Rate (MB/s)
>> Triad:        10341.9424   Rate (MB/s)
>> Triad:        10418.4740   Rate (MB/s)
>>
>>
>> Is this normal? i feel is different from what i get from an usual streams
>> test, how can i get it to work properly?
>>
>> Thanks,
>>
>>
>>
>>
>>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.mcs.anl.gov/pipermail/petsc-users/attachments/20180425/b7867391/attachment-0001.html>


More information about the petsc-users mailing list